New Year Special Limited Time Flat 70% Discount offer - Ends in 0d 00h 00m 00s - Coupon code: 70spcl

Databricks Databricks-Machine-Learning-Professional Databricks Certified Machine Learning Professional Exam Practice Test

Page: 1 / 6
Total 60 questions

Databricks Certified Machine Learning Professional Questions and Answers

Question 1

A machine learning engineer is converting a Hyperopt-based hyperparameter tuning process from manual MLflow logging to MLflow Autologging. They are trying to determine how to manage nested Hyperopt runs with MLflow Autologging.

Which of the following approaches will create a single parent run for the process and a child run for each unique combination of hyperparameter values when using Hyperopt and MLflow Autologging?

Options:

A.

Startinq amanual parent run before callingfmin

B.

Ensuring that a built-in model flavor is used for the model logging

C.

Starting a manual child run within the objective function

D.

There is no way to accomplish nested runs with MLflow Autoloqqinq and Hyperopt

E.

MLflow Autoloqqinq will automatically accomplish this task with Hyperopt

Question 2

A machine learning engineer has developed a model and registered it using the FeatureStoreClient fs. The model has model URI model_uri. The engineer now needs to perform batch inference on customer-level Spark DataFrame spark_df, but it is missing a few of the static features that were used when training the model. The customer_id column is the primary key of spark_df and the training set used when training and logging the model.

Which of the following code blocks can be used to compute predictions for spark_df when the missing feature values can be found in the Feature Store by searching for features by customer_id?

Options:

A.

df = fs.get_missing_features(spark_df, model_uri)

fs.score_model(model_uri, df)

B.

fs.score_model(model_uri, spark_df)

C.

df = fs.get_missing_features(spark_df, model_uri)

fs.score_batch(model_uri, df)

df = fs.get_missing_features(spark_df)

D.

fs.score_batch(model_uri, df)

E.

fs.score_batch(model_uri, spark_df)

Question 3

A data scientist is utilizing MLflow to track their machine learning experiments. After completing a series of runs for the experiment with experiment ID exp_id, the data scientist wants to programmatically work with the experiment run data in a Spark DataFrame. They have an active MLflow Client client and an active Spark session spark.

Which of the following lines of code can be used to obtain run-level results for exp_id in a Spark DataFrame?

Options:

A.

client.list_run_infos(exp_id)

B.

spark.read.format("delta").load(exp_id)

C.

There is no way to programmatically return row-level results from an MLflow Experiment.

D.

mlflow.search_runs(exp_id)

E.

spark.read.format("mlflow-experiment").load(exp_id)

Question 4

A data scientist has computed updated feature values for all primary key values stored in the Feature Store table features. In addition, feature values for some new primary key values have also been computed. The updated feature values are stored in the DataFrame features_df. They want to replace all data in features with the newly computed data.

Which of the following code blocks can they use to perform this task using the Feature Store Client fs?

A)

Question # 4

B)

Question # 4

C)

Question # 4

D)

Question # 4

E)

Question # 4

Options:

A.

Option A

B.

Option B

C.

Option C

D.

Option D

E.

Option E

Question 5

A machine learning engineer is in the process of implementing a concept drift monitoring solution. They are planning to use the following steps:

1. Deploy a model to production and compute predicted values

2. Obtain the observed (actual) label values

3. _____

4. Run a statistical test to determine if there are changes over time

Which of the following should be completed as Step #3?

Options:

A.

Obtain the observed values (actual) feature values

B.

Measure the latency of the prediction time

C.

Retrain the model

D.

None of these should be completed as Step #3

E.

Compute the evaluation metric using the observed and predicted values

Question 6

A machine learning engineer wants to programmatically create a new Databricks Job whose schedule depends on the result of some automated tests in a machine learning pipeline.

Which of the following Databricks tools can be used to programmatically create the Job?

Options:

A.

MLflow APIs

B.

AutoML APIs

C.

MLflow Client

D.

Jobs cannot be created programmatically

E.

Databricks REST APIs

Question 7

Which of the following statements describes streaming with Spark as a model deployment strategy?

Options:

A.

The inference of batch processed records as soon as a trigger is hit

B.

The inference of all types of records in real-time

C.

The inference of batch processed records as soon as a Spark job is run

D.

The inference of incrementally processed records as soon as trigger is hit

E.

The inference of incrementally processed records as soon as a Spark job is run

Question 8

A data scientist has developed and logged a scikit-learn random forest model model, and then they ended their Spark session and terminated their cluster. After starting a new cluster, they want to review the feature_importances_ of the original model object.

Which of the following lines of code can be used to restore the model object so that feature_importances_ is available?

Options:

A.

mlflow.load_model(model_uri)

B.

client.list_artifacts(run_id)["feature-importances.csv"]

C.

mlflow.sklearn.load_model(model_uri)

D.

This can only be viewed in the MLflow Experiments UI

E.

client.pyfunc.load_model(model_uri)

Question 9

A machine learning engineering team wants to build a continuous pipeline for data preparation of a machine learning application. The team would like the data to be fully processed and made ready for inference in a series of equal-sized batches.

Which of the following tools can be used to provide this type of continuous processing?

Options:

A.

Spark UDFs

B.

[Structured Streaming

C.

MLflow

D Delta Lake

D.

AutoML

Question 10

A machine learning engineer has registered a sklearn model in the MLflow Model Registry using the sklearn model flavor with UI model_uri.

Which of the following operations can be used to load the model as an sklearn object for batch deployment?

Options:

A.

mlflow.spark.load_model(model_uri)

B.

mlflow.pyfunc.read_model(model_uri)

C.

mlflow.sklearn.read_model(model_uri)

D.

mlflow.pyfunc.load_model(model_uri)

E.

mlflow.sklearn.load_model(model_uri)

Question 11

Which of the following is a reason for using Jensen-Shannon (JS) distance over a Kolmogorov-Smirnov (KS) test for numeric feature drift detection?

Options:

A.

All of these reasons

B.

JS is not normalized or smoothed

C.

None of these reasons

D.

JS is more robust when working with large datasets

E.

JS does not require any manual threshold or cutoff determinations

Question 12

Which of the following lists all of the model stages are available in the MLflow Model Registry?

Options:

A.

Development. Staging. Production

B.

None. Staging. Production

C.

Staging. Production. Archived

D.

None. Staging. Production. Archived

E.

Development. Staging. Production. Archived

Question 13

Which of the following describes the concept of MLflow Model flavors?

Options:

A.

A convention that deployment tools can use to wrap preprocessing logic into a Model

B.

A convention that MLflow Model Registry can use to version models

C.

A convention that MLflow Experiments can use to organize their Runs by project

D.

A convention that deployment tools can use to understand the model

E.

A convention that MLflow Model Registrycan use to organize its Models by project

Question 14

A machine learning engineer has created a webhook with the following code block:

Question # 14

Which of the following code blocks will trigger this webhook to run the associate job?

A)

Question # 14

B)

Question # 14

C)

Question # 14

D)

Question # 14

E)

Question # 14

Options:

A.

Option A

B.

Option B

C.

Option C

D.

Option D

E.

Option E

Question 15

Which of the following is a benefit of logging a model signature with an MLflow model?

Options:

A.

The model will have a unique identifier in the MLflow experiment

B.

The schema of input data can be validated when serving models

C.

The model can be deployed using real-time serving tools

D.

The model will be secured by the user that developed it

E.

The schema of input data will be converted to match the signature

Question 16

Which of the following describes the purpose of the context parameter in the predict method of Python models for MLflow?

Options:

A.

The context parameter allows the user to specify which version of the registered MLflowModel should be used based on the given application's current scenario

B.

The context parameter allows the user to document the performance of a model after it has been deployed

C.

The context parameter allows the user to include relevant details of the business case to allow downstream users to understand the purpose of the model

D.

The context parameter allows the user to provide the model with completely custom if-else logic for the given application's current scenario

E.

The context parameter allows the user to provide the model access to objects like preprocessing models or custom configuration files

Question 17

Which of the following is a simple, low-cost method of monitoring numeric feature drift?

Options:

A.

Jensen-Shannon test

B.

Summary statistics trends

C.

Chi-squared test

D.

None of these can be used to monitor feature drift

E.

Kolmogorov-Smirnov (KS) test

Question 18

A machine learning engineer wants to view all of the active MLflow Model Registry Webhooks for a specific model.

They are using the following code block:

Question # 18

Which of the following changes does the machine learning engineer need to make to this code block so it will successfully accomplish the task?

Options:

A.

There are no necessary changes

B.

Replace list with view in the endpoint URL

C.

Replace POST with GET in the call to http request

D.

Replace list with webhooks in the endpoint URL

E.

Replace POST with PUT in the call to http request

Page: 1 / 6
Total 60 questions