Skip to main content

Databricks Certified Machine Learning Associate

The Databricks Certified Machine Learning Associate validates the skills required to implement machine learning workflows using Databricks and MLflow. It covers data preparation, model training, and the tracking of experimental results to ensure reproducible ML outcomes. Professionals with the symbol DTB_MLA are recognized for their ability to build and manage basic machine learning pipelines in a collaborative environment.



---------- Question 1
A data scientist is analyzing a dataset containing financial transaction amounts for a fraud detection model. The distribution of transaction amounts is highly skewed, with a long tail of very large values, indicating the presence of significant outliers. A direct application of linear models or models sensitive to feature scales might be negatively impacted by these characteristics. To mitigate the influence of these extreme values and normalize the distribution for better model performance, what are the most appropriate data transformation techniques to consider?
  1. Removing all transactions above the 95th percentile and then applying min-max scaling to the remaining data.
  2. Applying a log transformation to the transaction amounts and then removing outliers based on the interquartile range (IQR).
  3. Using a standard scaler for the transaction amounts and then imputing any missing values with the mean.
  4. Applying a log transformation to the transaction amounts and then assessing the transformed data for further outlier treatment if necessary.

---------- Question 2
A medical research team is developing a machine learning model to detect a rare disease using patient data. The dataset is severely imbalanced, with only 1% of patients actually having the disease (positive class). The primary objective of this model is to maximize the identification of true positive cases (patients with the disease) to ensure early intervention, while keeping the number of false positives (healthy patients incorrectly diagnosed) at a manageable level to avoid unnecessary anxiety and costly follow-up procedures. Which data balancing technique and evaluation metric would be most appropriate for this specific scenario?
  1. Use undersampling of the majority class, and prioritize accuracy as the main evaluation metric.
  2. Use oversampling techniques like SMOTE on the minority class, and prioritize the F1-score as the main evaluation metric.
  3. Adjust class weights in the model training algorithm, and prioritize recall (sensitivity) while monitoring precision.
  4. Remove all samples from the majority class to balance the dataset, and prioritize Log Loss as the main evaluation metric.

---------- Question 3
A data science team has successfully deployed a champion model for customer churn prediction in a production environment. Subsequently, a new challenger model has been developed which demonstrates a marginal improvement in AUC on offline validation data. Before completely replacing the current champion, the team intends to conduct A/B testing in production to validate the challengers real-world performance. Both the champion and challenger models have been meticulously registered in the Unity Catalog Model Registry. To facilitate agile switching, version control, and streamlined management during the A/B test and the eventual promotion of the challenger, which MLflow feature within Unity Catalog should the team leverage?
  1. Manually update the model serving endpoint configuration to point to the new challenger model artifact path.
  2. Archive the current champion model version and promote the challenger model to production stage.
  3. Utilize MLflow model aliases to assign a champion alias to the current production model and a challenger alias to the new model, enabling easy programmatic switching.
  4. Create a new model entry in the Unity Catalog Model Registry for the challenger model and manage its serving independently.

---------- Question 4
An e-commerce platform wants to deploy a new product recommendation model. They need to provide immediate, personalized recommendations to users browsing the website, which requires very low latency and high availability. Furthermore, they want to test the performance of a new challenger model against the existing champion model on a small, controlled segment of live user traffic before fully rolling it out to all users. This approach enables data-driven decision making for model updates and ensures minimal disruption. Which model deployment strategy and Databricks serving capability should be employed to achieve both real-time personalized recommendations and A/B testing of model versions effectively?
  1. Deploy both models for batch inference overnight and update recommendations periodically for all users.
  2. Deploy both models as streaming inference jobs using Delta Live Tables and route user requests through this continuous pipeline.
  3. Deploy the champion model to a Databricks Model Serving endpoint for real-time inference, and configure a separate endpoint for the challenger model, directing a small percentage of live traffic to it for A/B testing.
  4. Integrate the challenger model directly into the application backend and manually switch between models based on user identification numbers.

---------- Question 5
A data science team has developed a new high-performing churn prediction model and wants to integrate it into a production system that requires robust versioning, staged deployment, and clear lineage tracking across different environments. They need to promote the model from a development stage to a staging environment for testing and then eventually to production. Furthermore, they anticipate needing to run A/B tests between current and challenger models. Which Databricks MLflow capability, when combined with Unity Catalog, provides the most comprehensive and governed approach for managing this model lifecycle, including promotion and A/B testing with aliases?
  1. Using the workspace-level MLflow Model Registry to register model versions and manually move them between workspace-specific stages.
  2. Employing the MLflow Client API to log models as artifacts in an MLflow Run and then manually deploying them to production endpoints.
  3. Leveraging the Unity Catalog Model Registry to register model versions, manage aliases, and transition models across environments.
  4. Packaging the model as a custom Docker image and deploying it directly to a Kubernetes cluster for versioning and staging.

---------- Question 6
A data science team is implementing a robust MLOps strategy for their predictive maintenance models. They require a system where every model training run is fully reproducible and traceable, capturing all parameters, metrics, and the model artifact itself. This information must be easily queryable for auditing, comparison, and deployment purposes. Which MLflow Client API function calls are essential to manually log the learning rate, the achieved RMSE metric, and the final trained scikit-learn model artifact within a single MLflow Run, adhering to these MLOps best practices?
  1. mlflow.start_run(), mlflow.log_metric(key='rmse', value=model_rmse), mlflow.log_param(key='learning_rate', value=0.01), mlflow.sklearn.save_model(sk_model=trained_model, path='model')
  2. mlflow.active_run(), mlflow.log_params({'learning_rate': 0.01}), mlflow.log_metrics({'rmse': model_rmse}), mlflow.log_artifact(local_path='trained_model.pkl')
  3. mlflow.start_run(), mlflow.log_param('learning_rate', 0.01), mlflow.log_metric('rmse', model_rmse), mlflow.sklearn.log_model(sk_model=trained_model, artifact_path='model')
  4. mlflow.run_id(), mlflow.set_tag('learning_rate', 0.01), mlflow.set_tag('rmse', model_rmse), mlflow.upload_file(local_path='trained_model.pkl', artifact_path='model')

---------- Question 7
A manufacturing company uses a vast network of IoT sensors to monitor the health of critical machinery. They want to build a system that continuously processes the incoming sensor data stream, predicts potential equipment failures in near real-time, and makes these predictions immediately available for operational dashboards and alerting systems. The solution needs to be robust, scalable, and capable of handling data quality issues and schema evolution automatically. How can Databricks and specifically Delta Live Tables DLT facilitate this streaming inference pipeline?
  1. Deploy a real-time model serving endpoint that directly consumes the raw sensor data stream and performs inference, then writes results to a batch table.
  2. Use Delta Live Tables DLT to define a streaming pipeline that ingests raw sensor data, performs necessary feature engineering and data quality checks, applies a pre-trained machine learning model for inference, and continuously updates a Delta table with predictions for consumption by dashboards.
  3. Perform daily batch inference jobs on the accumulated sensor data in a Delta table, and then manually refresh the dashboards with the updated predictions.
  4. Utilize a traditional Spark Streaming job with custom Python scripts for data processing and model inference, storing results in a non-Delta table.

---------- Question 8
A bank is building a machine learning model to detect fraudulent transactions. The dataset is highly imbalanced, with a tiny fraction of transactions (less than 1%) being fraudulent. The business goal is critical: to minimize false negatives (i.e., failing to detect actual fraud) even if it means a slight increase in false positives, because missing fraud is far more costly than false alarms. Which approach would be most effective for addressing the data imbalance and selecting an appropriate evaluation metric for this specific problem?
  1. Use SMOTE to oversample the minority class and evaluate the model primarily using accuracy.
  2. Undersample the majority class and focus on the R-squared metric for model evaluation.
  3. Employ techniques like class weighting during model training or oversampling/undersampling, and prioritize evaluation metrics such as Recall, F1-score, or Area Under the Receiver Operating Characteristic Curve (ROC/AUC).
  4. Ignore the imbalance, train a standard algorithm like Linear Regression, and evaluate using Mean Absolute Error (MAE).

---------- Question 9
A global e-commerce company manages various machine learning projects across different geographical regions and Databricks workspaces. They are developing a unified fraud detection system that requires consistent features like customer purchase history and device fingerprints, accessible by multiple data science teams. They plan to use Databricks AutoML for rapid model development. Which strategy for creating and managing feature store tables would best support this organizational structure and why?
  1. Create separate feature store tables within each regional workspace, as this provides workspace-specific isolation and easier management for individual teams.
  2. Utilize Unity Catalog to create and manage feature store tables at the account level, enabling centralized feature discovery, sharing, and governance across all workspaces.
  3. Store features directly within Delta Lake tables in each project's workspace, relying on manual data synchronization between workspaces for feature consistency.
  4. Use an external feature store solution entirely separate from Databricks and integrate it via custom APIs, sacrificing the native integration benefits with AutoML and Unity Catalog.
  5. Create feature store tables as local CSV files within individual notebooks to allow for quick iteration, then manually merge them for model training.

---------- Question 10
A data scientist is analyzing a large dataset of customer transaction amounts for an e-commerce platform. The dataset exhibits a highly right-skewed distribution with a long tail, indicating the presence of a few extremely high transaction values that could significantly impact downstream model training for customer segmentation. The goal is to remove these extreme outliers to prevent them from unduly influencing the model, without discarding too much valid data from the main distribution. Which method for outlier detection and removal would generally be more appropriate in this specific scenario, and why?
  1. Using the standard deviation method to remove data points falling beyond 3 standard deviations from the mean, as it is simple and widely understood.
  2. Applying the Interquartile Range (IQR) method to remove data points outside 1.5 times the IQR from the first and third quartiles, as it is more robust to skewed distributions.
  3. Directly applying a log transformation to the transaction amounts without removing outliers, as this transformation naturally reduces the impact of large values.
  4. Removing the top 1% of transaction values directly, assuming these are always outliers and will not impact the data distribution in any other way.


Are they useful?
Click here to get 288 more questions to pass this certification at the first try! Explanation for each answer is included!

Follow the below LINKEDIN channel to stay updated about 89+ exams!

Comments

Popular posts from this blog

Microsoft Certified: Azure Fundamentals (AZ-900)

The Microsoft Certified: Azure Fundamentals (AZ-900) is the essential starting point for anyone looking to validate their foundational knowledge of cloud services and how those services are provided with Microsoft Azure. It is designed for both technical and non-technical professionals ---------- Question 1 A new junior administrator has joined your IT team and needs to manage virtual machines for a specific development project within your Azure subscription. This project has its own dedicated resource group called dev-project-rg. The administrator should be able to start, stop, and reboot virtual machines, but should not be able to delete them or modify network configurations, and crucially, should not have access to virtual machines or resources in other projects or subscription-level settings. Which Azure identity and access management concept, along with its appropriate scope, should be used to grant these specific permissions? Microsoft Entra ID Conditional Access, applied at...

Google Associate Cloud Engineer

The Google Associate Cloud Engineer (ACE) certification validates the fundamental skills needed to deploy applications, monitor operations, and manage enterprise solutions on the Google Cloud Platform (GCP). It is considered the "gatekeeper" certification, proving a candidate's ability to perform practical cloud engineering tasks rather than just understanding theoretical architecture.  ---------- Question 1 Your team is developing a serverless application using Cloud Functions that needs to process data from Cloud Storage. When a new object is uploaded to a specific Cloud Storage bucket, the Cloud Function should automatically trigger and process the data. How can you achieve this? Use Cloud Pub/Sub as a message broker between Cloud Storage and Cloud Functions. Directly access Cloud Storage from the Cloud Function using the Cloud Storage Client Library. Use Cloud Scheduler to periodically check for new objects in the bucket. Configure Cloud Storage to directly ca...

CompTIA Cybersecurity Analyst (CySA+)

CompTIA Cybersecurity Analyst (CySA+) focuses on incident detection, prevention, and response through continuous security monitoring. It validates a professional's expertise in vulnerability management and the use of threat intelligence to strengthen organizational security. Achieving the symbol COMP_CYSA marks an individual as a proficient security analyst capable of mitigating modern cyber threats. ---------- Question 1 A security analyst is reviewing logs in the SIEM and identifies a series of unusual PowerShell executions on a critical application server. The logs show the use of the -EncodedCommand flag followed by a long Base64 string. Upon decoding, the script appears to be performing memory injection into a legitimate system process. Which of the following is the most likely indicator of malicious activity being observed, and what should be the analysts immediate technical response using scripting or tools? The activity indicates a fileless malware attack attempting to ...