Databricks Certified Machine Learning Professional Exam Dumps & Practice Test Questions

Question 1:

In a real-world machine learning scenario, data often changes as time progresses. This can affect the stability of the relationship between input features and the target output.

Which option correctly defines what is meant by concept drift in this context?

A. A shift in the distribution of one or more input features over time.
B. A shift in the distribution of the target variable only.
C. A change in the relationship between input variables and the target variable over time.
D. A change in the distribution of model predictions over time, independent of data.
E. None of the above describe concept drift accurately.

Correct Answer: C

Explanation:

Concept drift occurs when the underlying relationship between input variables and the target variable changes over time. In machine learning, models are built under the assumption that historical data patterns will remain consistent in the future. Specifically, they assume that the joint distribution P(X, Y)—where X represents features and Y the target—stays stable. But in many real-world applications, this assumption breaks down.

For instance, consider a fraud detection model trained on transaction data. As fraudsters adopt new strategies, the patterns that previously indicated fraudulent behavior may no longer apply. This means that the model’s understanding of how input features relate to the target (fraud or not) becomes outdated. This evolution in the feature-target relationship is what defines concept drift.

Concept drift differs from simple data distribution changes. A shift in feature distribution (known as covariate shift) or in the target distribution (called prior probability shift) does not necessarily mean that the actual relationship between X and Y has changed. Concept drift specifically addresses the case when P(Y|X) changes—even if P(X) or P(Y) remain unchanged.

There are different types of drift to be aware of:

  • Sudden drift: The change happens abruptly.

  • Gradual drift: The transition occurs over time.

  • Incremental drift: Small consistent changes accumulate.

  • Recurring drift: Past relationships reappear cyclically.

If not detected and addressed, concept drift can lead to significant declines in model accuracy and reliability. Countermeasures include model retraining, using sliding windows for training data, or employing online learning strategies. Drift detection algorithms like DDM (Drift Detection Method) and ADWIN (Adaptive Windowing) are commonly used.

In summary, concept drift is best defined as a change in the functional relationship between inputs and outputs, and recognizing it is key for maintaining long-term model performance.

Question 2:

A machine learning engineer suspects that a specific category within a categorical feature is experiencing more missing values in recent data compared to older records.

Which statistical method should the engineer use to test whether this shift in missing values is significant?

A. Kolmogorov-Smirnov (KS) Test
B. One-way Chi-squared Test
C. Two-way Chi-squared Test
D. Jensen-Shannon Distance
E. None of these

Correct Answer: B

Explanation:

To analyze whether missing values in a categorical variable are increasing over time, the One-way Chi-squared Test (also known as the Chi-squared Goodness-of-Fit Test) is the most suitable tool. This test determines whether the distribution of observed frequencies for a single categorical variable significantly deviates from an expected distribution.

In this scenario, the engineer can segment the data into time periods (e.g., older vs. recent data) and compute the frequency of missing values for the category in question. The One-way Chi-squared Test is then applied to compare these observed frequencies against what would be expected if there were no change over time. The null hypothesis asserts that the frequency of missing values remains consistent, while the alternative hypothesis suggests a statistically significant shift.

Here's why other options are less appropriate:

  • Kolmogorov-Smirnov (KS) Test is designed for continuous distributions, not categorical ones, and especially not for analyzing frequencies of missing values.

  • Two-way Chi-squared Test is generally used to examine the relationship between two categorical variables (e.g., gender and preference), not to test distribution consistency over time.

  • Jensen-Shannon Distance measures the similarity between two probability distributions but doesn’t provide a statistical hypothesis test for categorical frequency shifts.

  • "None of these" is incorrect because the One-way Chi-squared Test fits the scenario well.

Monitoring the prevalence of missing data is critical in production ML pipelines. Unchecked increases in missing values can degrade model performance, bias predictions, and complicate feature engineering. By statistically validating changes using the One-way Chi-squared Test, data scientists can take proactive steps, such as modifying imputation strategies or flagging upstream data quality issues.

Therefore, Option B is the most appropriate and statistically valid choice for the engineer’s task.

Question 3:

A machine learning engineer wants to implement a basic and affordable method for tracking changes in numeric features over time that could impact model performance.

Which of the following techniques is best suited for detecting drift in numerical input features using a simple and cost-effective approach?

A. Jensen-Shannon test
B. Summary statistics trends
C. Chi-squared test
D. None of these methods are suitable for monitoring feature drift
E. Kolmogorov-Smirnov (KS) test

Correct Answer: B

Explanation:

Monitoring data drift—especially in numeric features—is essential for ensuring that machine learning models continue to perform accurately in real-world conditions. Feature drift, a type of data drift, occurs when the statistical properties of input features change over time, even if the relationship between inputs and outputs remains stable.

One of the most efficient and cost-effective strategies for tracking numeric feature drift is the use of summary statistics trends. This approach involves calculating and plotting basic statistics such as mean, standard deviation, median, minimum, and maximum over defined time intervals. By monitoring these values regularly, you can quickly detect significant changes or patterns that indicate drift. This method is easy to implement, computationally inexpensive, and highly interpretable, making it a practical solution for many production environments.

Here’s why the other options are less suitable:

  • A. Jensen-Shannon test is used to compare two probability distributions and is particularly useful with categorical data or discrete probability distributions. It’s more complex and resource-intensive than necessary for basic drift detection.

  • C. Chi-squared test is designed for evaluating categorical variables by comparing expected versus observed frequencies. It’s not effective for continuous numeric data.

  • D. None of these methods are suitable is incorrect because summary statistics trends (Option B) clearly offer a practical method.

  • E. Kolmogorov-Smirnov (KS) test is a statistical method to compare two distributions and can be applied to numeric data. However, it involves more complex implementation and computation than summary statistics, making it less ideal for quick and scalable drift monitoring.

In conclusion, summary statistics trends provide a straightforward, transparent, and low-overhead way to identify numeric feature drift and are often used as a first line of defense in production monitoring systems.

Question 4:

A predictive model was trained to estimate ice cream sales using expected temperature and sunshine duration as input features. The model performed well when input temperatures stayed within a certain range. However, recent forecasts predict temperatures dropping below the training range.

Which type of data drift best explains this situation?

A. Label Drift
B. None of These
C. Concept Drift
D. Prediction Drift
E. Feature Drift

Correct Answer: E

Explanation:

When real-world data starts to diverge from the patterns observed in the training data, machine learning models may become less effective. In this scenario, the model was trained using temperature values within a specific historical range. Now, with predicted temperatures falling below that known range, the distribution of the input feature (temperature) has changed. This situation is best described as Feature Drift.

Feature Drift occurs when the distribution of input variables changes over time. Even if the relationship between the input and the output remains intact, changes in the input data alone can degrade model accuracy, especially when inputs fall outside the domain the model was trained on. In this case, the model might struggle to make accurate predictions because it hasn’t “seen” temperatures this low before.

Let’s examine the other types of drift for clarity:

  • A. Label Drift refers to changes in the target variable’s distribution. For example, if ice cream sales decline for unrelated reasons, like supply issues or economic shifts, that would be label drift. But that’s not the issue here.

  • C. Concept Drift occurs when the relationship between features and the label changes over time. If, for instance, ice cream sales no longer depend on temperature as they did previously, that would be concept drift. In this case, the relationship is likely intact—it’s the input data that’s changing.

  • D. Prediction Drift is not a formally recognized type of drift. It may refer informally to shifts in predicted values but doesn’t help categorize the root cause of model degradation.

  • B. None of These is clearly incorrect because the described scenario fits the definition of feature drift.

To summarize, Feature Drift is the most accurate classification here. Monitoring input distributions and ensuring the model adapts to changes like these—through retraining or feature engineering—is crucial to maintaining model performance over time.

Question 5:

A data scientist wants to remove the column star_rating from a Delta table stored at a specific file path. To do this, they first need to load the data and then drop the star_rating column.

Which of the following code snippets correctly performs this operation?

A. spark.read.format("delta").load(path).drop("star_rating")
B. spark.read.format("delta").table(path).drop("star_rating")
C. Delta tables cannot be modified.
D. spark.read.table(path).drop("star_rating")
E. spark.sql("SELECT * EXCEPT star_rating FROM path")

Correct Answer: A

Explanation:

When working with Delta Lake tables, it’s common to load the data for processing and then apply transformations such as dropping unwanted columns. In this case, the goal is to remove the star_rating column from the Delta table located at a specific path.

Option A is the correct choice because spark.read.format("delta").load(path) loads the Delta table data from the exact file system location specified by path. Calling .drop("star_rating") on the resulting DataFrame removes the star_rating column in the data held in memory. This does not directly alter the Delta table on disk but produces a DataFrame excluding that column. To make the change permanent, the modified DataFrame must be written back to the Delta table.

Option B is incorrect because .table(path) expects a table name registered in the Spark catalog, not a file path. Since the question mentions a file path, .load(path) is the proper method.

Option C is false — Delta tables are designed to support modifications, including updates, deletes, and schema changes.

Option D misuses .table(path) which, like option B, is for table names, not paths, so it won’t work as intended here.

Option E attempts a SQL query with SELECT * EXCEPT star_rating FROM path. This is invalid because SQL queries refer to table names, not file paths, and the EXCEPT clause isn’t a valid SQL syntax for dropping columns in this context.

In summary, option A correctly loads the Delta table from a file path and drops the star_rating column in the in-memory DataFrame. To persist this change, further steps to write the data back to the Delta table would be necessary.

Question 6:

Which operation from the Feature Store Client (fs) is used to load a Feature Store table’s data into a Spark DataFrame?

A. fs.create_table
B. fs.write_table
C. fs.get_table
D. There is no way to do this with fs
E. fs.read_table

Correct Answer: E

Explanation:

In feature engineering and machine learning workflows, Feature Stores serve as centralized repositories that store precomputed features to be reused across models and pipelines. When working within such systems, a typical need is to retrieve the feature data in a form that can be processed and analyzed efficiently, such as a Spark DataFrame.

Among the operations offered by the Feature Store Client (fs), fs.read_table is the method specifically designed for this purpose. It loads the data associated with a named Feature Store table and returns it as a Spark DataFrame, enabling distributed processing and integration with Spark-based ML pipelines.

Let’s examine why the other options are incorrect:

  • A. fs.create_table: This operation is used for creating new Feature Store tables, not for reading data.

  • B. fs.write_table: This method writes data (usually a DataFrame) into an existing Feature Store table. It’s for data ingestion, not retrieval.

  • C. fs.get_table: This typically returns metadata or schema details about the Feature Store table, but not the actual feature data as a DataFrame.

  • D. The statement that no such operation exists in fs is false since fs.read_table explicitly performs data loading.

Thus, fs.read_table is the correct operation to fetch the actual feature data from the Feature Store as a Spark DataFrame, facilitating downstream tasks like model training and validation within the Spark ecosystem.

Question 7:

A machine learning engineer is setting up a system to monitor concept drift in a deployed model. The planned steps include:

  1. Deploying the model and generating predictions.

  2. Collecting the actual observed label values.

  3. Running a statistical test to detect any changes over time.

Which action should be performed as Step #3?

A. Collect the actual feature values.
B. Measure the prediction latency.
C. Retrain the model.
D. None of the above.
E. Calculate the evaluation metric using the predicted and actual labels.

Correct Answer: E

Explanation:

Concept drift occurs when the underlying data distribution changes over time, causing a previously trained machine learning model to degrade in accuracy. Detecting such drift promptly is critical for maintaining reliable predictions in production environments.

In the scenario, the engineer’s goal is to identify whether the model’s performance is deteriorating due to changes in the data. The initial two steps are straightforward: first, deploy the model to generate predictions on new data, and second, collect the corresponding true outcomes or labels.

After these values are available, the logical next step is to quantify how well the model’s predictions match the actual observations. This is done by calculating evaluation metrics such as accuracy, precision, recall, F1-score, or RMSE, depending on the task. These metrics provide an objective measure of model performance over the recent data.

Step #4 involves conducting a statistical test, which compares model performance metrics or distributional properties over time to detect significant shifts indicative of concept drift. For example, a decline in accuracy or a statistical change detected by tests like the Kolmogorov-Smirnov test would signal drift.

Measuring prediction latency (option B) or gathering feature values again (option A) are not immediate steps for drift detection, and retraining (option C) should only occur after drift has been confirmed. Hence, calculating evaluation metrics is the critical intermediate step to objectively assess the model’s current accuracy before conducting statistical tests.

Thus, Step #3 correctly involves computing the evaluation metric to quantify prediction quality, making E the best answer.

Question 8:

When detecting numeric feature drift in machine learning models, what is a primary advantage of using the Jensen-Shannon (JS) distance instead of the Kolmogorov-Smirnov (KS) test?

A. All of the above reasons
B. JS distance is neither normalized nor smoothed
C. None of the above reasons
D. JS distance is more robust with large datasets
E. JS distance does not require manual threshold or cutoff settings

Correct Answer: E

Explanation:

Feature drift describes changes in the statistical properties of input features over time, which can negatively impact the accuracy of deployed machine learning models. Detecting such drift early is essential for triggering model maintenance or retraining.

Two common statistical approaches for detecting numeric feature drift are the Jensen-Shannon (JS) distance and the Kolmogorov-Smirnov (KS) test. Each has distinct characteristics and use cases.

The Jensen-Shannon distance measures similarity between two probability distributions and has several advantages for drift detection:

  • It produces values within a fixed, interpretable range (0 to 1), which makes results easier to understand.

  • It inherently smooths distributions, making it less sensitive to noise or small sample sizes.

  • Importantly, it does not require manual threshold selection. This means practitioners don’t have to define arbitrary cutoff points to decide when drift has occurred, which reduces subjectivity and the need for extensive domain expertise or trial and error.

In contrast, the Kolmogorov-Smirnov test compares empirical distributions to assess if they differ significantly. Although widely used, it requires setting a significance level or threshold to flag drift, which can be difficult and dataset-dependent. The KS test is also sensitive to sample size and can produce overly sensitive results with very large datasets.

While JS distance can be robust with large datasets, its key strength lies in its automatic, threshold-free interpretation. This makes it a preferred tool in many monitoring systems where simplicity, reliability, and reduced manual tuning are important.

Therefore, the primary advantage of JS distance over the KS test in numeric feature drift detection is that it eliminates the need for manual threshold or cutoff settings, making option E the correct choice.

Question 9:

You have trained a gradient-boosted decision tree model using the Databricks AutoML feature on a large dataset for a multiclass classification task. After the model is trained, you want to optimize the model further by tuning hyperparameters. 

Which approach is best suited for hyperparameter tuning within Databricks, and why?

A. Use MLflow to manually adjust hyperparameters and rerun experiments.
B. Use Databricks AutoML's built-in hyperparameter tuning capabilities.
C. Export the data and model to an external platform for tuning.
D. Use only grid search without parallelizing the workload.

Correct Answer: B

Explanation:

The Databricks Certified Machine Learning Professional exam tests your ability to efficiently train, optimize, and deploy machine learning models using Databricks’ advanced tools and features.

When working with large datasets and complex models such as gradient-boosted decision trees, hyperparameter tuning is a critical step to enhance performance. Databricks AutoML is designed to simplify this process by automatically selecting the best algorithms and tuning hyperparameters within the Databricks environment.

Option B is the best choice because Databricks AutoML integrates hyperparameter optimization directly, often using efficient techniques such as Bayesian optimization or random search to explore the hyperparameter space intelligently. This is more efficient than manual tuning or simple grid search, especially for large datasets.

Option A suggests manual tuning via MLflow. While MLflow tracks experiments, manually adjusting hyperparameters and retraining models is time-consuming and inefficient for complex models.

Option C involves exporting models/data, which adds overhead, risk of inconsistency, and loss of integration benefits within the Databricks ecosystem.

Option D mentions grid search without parallelization. Grid search is often expensive and slow, especially without distributed parallelism, making it impractical on large datasets.

Using AutoML's tuning leverages the platform's distributed computing power and automation to quickly and effectively identify optimal hyperparameters, reducing time to production.

Mastering these concepts is essential for the Databricks Certified Machine Learning Professional exam, which emphasizes practical skills in building and optimizing ML workflows on the Databricks platform.

Question 10:

In a production ML pipeline on Databricks, a model is deployed and serves predictions via REST API endpoints. After some time, you notice prediction performance degrading due to concept drift in incoming data. 

Which approach best helps detect and address concept drift within Databricks?

A. Retrain the model on a fixed historical dataset periodically.
B. Use Databricks Feature Store monitoring and implement drift detection using batch data comparisons.
C. Ignore the drift and rely on manual inspection monthly.
D. Replace the model every week without monitoring.

Correct Answer: B

Explanation:

One key responsibility covered in the Databricks Certified Machine Learning Professional exam is managing ML models in production, including monitoring model performance and detecting concept drift.

Concept drift occurs when the statistical properties of the input data change over time, causing a degradation in model accuracy. Detecting drift early is crucial for maintaining reliable predictions.

Option B is the correct approach. Databricks provides a Feature Store that supports monitoring feature distributions and automated detection of data drift by comparing recent input batches against historical data. This can be integrated into the ML pipeline to trigger alerts or retraining workflows, enabling proactive model maintenance.

Option A suggests retraining on a static historical dataset, which does not address drift in current data effectively because it ignores the evolving data distribution.

Option C delays detection and relies on manual intervention, which is inefficient and prone to delays, risking poor prediction quality.

Option D blindly replaces the model without monitoring, which wastes resources and may not improve performance.

Thus, using the Databricks Feature Store monitoring tools combined with automated drift detection allows teams to maintain model accuracy dynamically, which is a core best practice for production ML systems. This is a critical topic for the professional-level Databricks ML exam, as it demonstrates knowledge of operationalizing ML workflows responsibly.


SPECIAL OFFER: GET 10% OFF

ExamCollection Premium

ExamCollection Premium Files

Pass your Exam with ExamCollection's PREMIUM files!

  • ExamCollection Certified Safe Files
  • Guaranteed to have ACTUAL Exam Questions
  • Up-to-Date Exam Study Material - Verified by Experts
  • Instant Downloads
Enter Your Email Address to Receive Your 10% Off Discount Code
A Confirmation Link will be sent to this email address to verify your login
We value your privacy. We will not rent or sell your email address

SPECIAL OFFER: GET 10% OFF

Use Discount Code:

MIN10OFF

A confirmation link was sent to your e-mail.
Please check your mailbox for a message from support@examcollection.com and follow the directions.

Next

Download Free Demo of VCE Exam Simulator

Experience Avanset VCE Exam Simulator for yourself.

Simply submit your e-mail address below to get started with our interactive software demo of your free trial.

Free Demo Limits: In the demo version you will be able to access only first 5 questions from exam.