Google Associate Data Practitioner Practice Exams
Last updated on Apr 01,2025- Exam Code: Associate Data Practitioner
- Exam Name: Google Cloud Associate Data Practitioner ( ADP Exam )
- Certification Provider: Google
- Latest update: Apr 01,2025
Your team is building several data pipelines that contain a collection of complex tasks and dependencies that you want to execute on a schedule, in a specific order. The tasks and dependencies consist of files in Cloud Storage, Apache Spark jobs, and data in BigQuery. You need to design a system that can schedule and automate these data processing tasks using a fully managed approach.
What should you do?
- A . Use Cloud Scheduler to schedule the jobs to run.
- B . Use Cloud Tasks to schedule and run the jobs asynchronously.
- C . Create directed acyclic graphs (DAGs) in Cloud Composer. Use the appropriate operators to connect to Cloud Storage, Spark, and BigQuery.
- D . Create directed acyclic graphs (DAGs) in Apache Airflow deployed on Google Kubernetes Engine.
Use the appropriate operators to connect to Cloud Storage, Spark, and BigQuery.
You used BigQuery ML to build a customer purchase propensity model six months ago. You want to compare the current serving data with the historical serving data to determine whether you need to retrain the model.
What should you do?
- A . Compare the two different models.
- B . Evaluate the data skewness.
- C . Evaluate data drift.
- D . Compare the confusion matrix.
You work for a global financial services company that trades stocks 24/7. You have a Cloud SGL for PostgreSQL user database. You need to identify a solution that ensures that the database is continuously operational, minimizes downtime, and will not lose any data in the event of a zonal outage.
What should you do?
- A . Continuously back up the Cloud SGL instance to Cloud Storage. Create a Compute Engine instance with Postgre SCL in a different region. Restore the backup in the Compute Engine instance if a failure occurs.
- B . Create a read replica in another region. Promote the replica to primary if a failure occurs.
- C . Configure and create a high-availability Cloud SQL instance with the primary instance in zone A and a secondary instance in any zone other than zone A.
- D . Create a read replica in the same region but in a different zone.
You recently inherited a task for managing Dataflow streaming pipelines in your organization and noticed that proper access had not been provisioned to you. You need to request a Google-provided IAM role so you can restart the pipelines. You need to follow the principle of least privilege.
What should you do?
- A . Request the Dataflow Developer role.
- B . Request the Dataflow Viewer role.
- C . Request the Dataflow Worker role.
- D . Request the Dataflow Admin role.
You want to process and load a daily sales CSV file stored in Cloud Storage into BigQuery for downstream reporting. You need to quickly build a scalable data pipeline that transforms the data while providing insights into data quality issues.
What should you do?
- A . Create a batch pipeline in Cloud Data Fusion by using a Cloud Storage source and a BigQuery sink.
- B . Load the CSV file as a table in BigQuery, and use scheduled queries to run SQL transformation scripts.
- C . Load the CSV file as a table in BigQuery. Create a batch pipeline in Cloud Data Fusion by using a BigQuery source and sink.
- D . Create a batch pipeline in Dataflow by using the Cloud Storage CSV file to BigQuery batch template.
You need to create a weekly aggregated sales report based on a large volume of data. You want to use Python to design an efficient process for generating this report.
What should you do?
- A . Create a Cloud Run function that uses NumPy. Use Cloud Scheduler to schedule the function to run once a week.
- B . Create a Colab Enterprise notebook and use the bigframes.pandas library. Schedule the notebook to execute once a week.
- C . Create a Cloud Data Fusion and Wrangler flow. Schedule the flow to run once a week.
- D . Create a Dataflow directed acyclic graph (DAG) coded in Python. Use Cloud Scheduler to schedule the code to run once a week.
Your team uses the Google Ads platform to visualize metrics. You want to export the data to BigQuery to get more granular insights. You need to execute a one-time transfer of historical data and automatically update data daily. You want a solution that is low-code, serverless, and requires minimal maintenance.
What should you do?
- A . Export the historical data to BigQuery by using BigQuery Data Transfer Service. Use Cloud Composer for daily automation.
- B . Export the historical data to Cloud Storage by using Storage Transfer Service. Use Pub/Sub to trigger a Dataflow template that loads data for daily automation.
- C . Export the historical data as a CSV file. Import the file into BigQuery for analysis. Use Cloud Composer for daily automation.
- D . Export the historical data to BigQuery by using BigQuery Data Transfer Service. Use BigQuery Data Transfer Service for daily automation.
Your team uses the Google Ads platform to visualize metrics. You want to export the data to BigQuery to get more granular insights. You need to execute a one-time transfer of historical data and automatically update data daily. You want a solution that is low-code, serverless, and requires minimal maintenance.
What should you do?
- A . Export the historical data to BigQuery by using BigQuery Data Transfer Service. Use Cloud Composer for daily automation.
- B . Export the historical data to Cloud Storage by using Storage Transfer Service. Use Pub/Sub to trigger a Dataflow template that loads data for daily automation.
- C . Export the historical data as a CSV file. Import the file into BigQuery for analysis. Use Cloud Composer for daily automation.
- D . Export the historical data to BigQuery by using BigQuery Data Transfer Service. Use BigQuery Data Transfer Service for daily automation.
Your team needs to analyze large datasets stored in BigQuery to identify trends in user behavior. The analysis will involve complex statistical calculations, Python packages, and visualizations. You need to recommend a managed collaborative environment to develop and share the analysis.
What should you recommend?
- A . Create a Colab Enterprise notebook and connect the notebook to BigQuery. Share the notebook with your team. Analyze the data and generate visualizations in Colab Enterprise.
- B . Create a statistical model by using BigQuery ML. Share the query with your team. Analyze the data and generate visualizations in Looker Studio.
- C . Create a Looker Studio dashboard and connect the dashboard to BigQuery. Share the dashboard with your team. Analyze the data and generate visualizations in Looker Studio.
- D . Connect Google Sheets to BigQuery by using Connected Sheets. Share the Google Sheet with your team. Analyze the data and generate visualizations in Gooqle Sheets.
You are working on a data pipeline that will validate and clean incoming data before loading it into BigQuery for real-time analysis. You want to ensure that the data validation and cleaning is performed efficiently and can handle high volumes of data.
What should you do?
- A . Write custom scripts in Python to validate and clean the data outside of Google Cloud. Load the cleaned data into BigQuery.
- B . Use Cloud Run functions to trigger data validation and cleaning routines when new data arrives in Cloud Storage.
- C . Use Dataflow to create a streaming pipeline that includes validation and transformation steps.
- D . Load the raw data into BigQuery using Cloud Storage as a staging area, and use SQL queries in BigQuery to validate and clean the data.