DSA-C03 試験問題を無料オンラインアクセス
試験コード: | DSA-C03 |
試験名称: | SnowPro Advanced: Data Scientist Certification Exam |
認定資格: | Snowflake |
無料問題数: | 289 |
更新日: | 2025-09-05 |
Which of the following statements about Z-tests and T-tests are generally true? Select all that apply.
You've trained a model using Snowflake ML and want to deploy it for real-time predictions using a Snowflake UDF. To ensure minimal latency, you need to optimize the UDF's performance. Which of the following strategies and considerations are most important when creating and deploying a UDF for model inference in Snowflake to minimize latency, especially when the model is large (e.g., > 100MB)?
Select all that apply.
You are building a churn prediction model for a telecommunications company using Snowflake and Snowpark ML. You have trained a Gradient Boosting Machine (GBM) model and want to understand the feature importance to identify key drivers of churn. You've used SHAP (SHapley Additive exPlanations) values to explain individual predictions. Given a customer with a high churn risk, you observe that the 'monthly_charges' feature has a significantly large negative SHAP value for that specific prediction. Which of the following statements best interprets this observation in the context of feature impact?
You are analyzing customer transaction data in Snowflake to identify fraudulent activities. The 'TRANSACTION AMOUNT' column exhibits a right-skewed distribution. Which of the following Snowflake queries is MOST effective in identifying outliers based on the Interquartile Range (IQR) method, specifically targeting unusually large transaction amounts? Assume IQR is already calculated as variable and QI as and Q3 as in snowflake session.
A marketing team uses Snowflake to store customer purchase data'. They want to segment customers based on their spending habits using a derived feature called The 'PURCHASES' table has columns 'customer id' (IN T), 'purchase_date' (DATE), and 'purchase_amount' (NUMBER). The team needs a way to handle situations where a customer might have missing months (no purchases in a particular month). They want to impute a 0 spend for those months before calculating the average. Which approach provides the most accurate and robust calculation, especially when considering users with sparse purchase history?