Easter Sale - 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: dm70dm

Associate-Data-Practitioner Google Cloud Associate Data Practitioner (ADP Exam) Questions and Answers

Questions 4

Your organization has a petabyte of application logs stored as Parquet files in Cloud Storage. You need to quickly perform a one-time SQL-based analysis of the files and join them to data that already resides in BigQuery. What should you do?

Options:

A.

Create a Dataproc cluster, and write a PySpark job to join the data from BigQuery to the files in Cloud Storage.

B.

Launch a Cloud Data Fusion environment, use plugins to connect to BigQuery and Cloud Storage, and use the SQL join operation to analyze the data.

C.

Create external tables over the files in Cloud Storage, and perform SQL joins to tables in BigQuery to analyze the data.

D.

Use the bq load command to load the Parquet files into BigQuery, and perform SQL joins to analyze the data.

Buy Now
Questions 5

Your retail company collects customer data from various sources:

Associate-Data-Practitioner Question 5Online transactions: Stored in a MySQL database

Associate-Data-Practitioner Question 5Customer feedback: Stored as text files on a company server

Associate-Data-Practitioner Question 5Social media activity: Streamed in real-time from social media platforms

You are designing a data pipeline to extract this data. Which Google Cloud storage system(s) should you select for further analysis and ML model training?

Options:

A.

1. Online transactions: Cloud Storage

2. Customer feedback: Cloud Storage

3. Social media activity: Cloud Storage

B.

1. Online transactions: BigQuery

2. Customer feedback: Cloud Storage

3. Social media activity: BigQuery

C.

1. Online transactions: Bigtable

2. Customer feedback: Cloud Storage

3. Social media activity: CloudSQL for MySQL

D.

1. Online transactions: Cloud SQL for MySQL

2. Customer feedback: BigQuery

3. Social media activity: Cloud Storage

Buy Now
Questions 6

Your data science team needs to collaboratively analyze a 25 TB BigQuery dataset to support the development of a machine learning model. You want to use Colab Enterprise notebooks while ensuring efficient data access and minimizing cost. What should you do?

Options:

A.

Export the BigQuery dataset to Google Drive. Load the dataset into the Colab Enterprise notebook using Pandas.

B.

Use BigQuery magic commands within a Colab Enterprise notebook to query and analyze the data.

C.

Create a Dataproc cluster connected to a Colab Enterprise notebook, and use Spark to process the data in BigQuery.

D.

Copy the BigQuery dataset to the local storage of the Colab Enterprise runtime, and analyze the data using Pandas.

Buy Now
Questions 7

Your organization has several datasets in BigQuery. The datasets need to be shared with your external partners so that they can run SQL queries without needing to copy the data to their own projects. You have organized each partner’s data in its own BigQuery dataset. Each partner should be able to access only their data. You want to share the data while following Google-recommended practices. What should you do?

Options:

A.

Use Analytics Hub to create a listing on a private data exchange for each partner dataset. Allow each partner to subscribe to their respective listings.

B.

Create a Dataflow job that reads from each BigQuery dataset and pushes the data into a dedicated Pub/Sub topic for each partner. Grant each partner the pubsub. subscriber IAM role.

C.

Export the BigQuery data to a Cloud Storage bucket. Grant the partners the storage.objectUser IAM role on the bucket.

D.

Grant the partners the bigquery.user IAM role on the BigQuery project.

Buy Now
Questions 8

Another team in your organization is requesting access to a BigQuery dataset. You need to share the dataset with the team while minimizing the risk of unauthorized copying of data. You also want tocreate a reusable framework in case you need to share this data with other teams in the future. What should you do?

Options:

A.

Create authorized views in the team’s Google Cloud project that is only accessible by the team.

B.

Create a private exchange using Analytics Hub with data egress restriction, and grant access to the team members.

C.

Enable domain restricted sharing on the project. Grant the team members the BigQuery Data Viewer IAM role on the dataset.

D.

Export the dataset to a Cloud Storage bucket in the team’s Google Cloud project that is only accessible by the team.

Buy Now
Questions 9

Your retail organization stores sensitive application usage data in Cloud Storage. You need to encrypt the data without the operational overhead of managing encryption keys. What should you do?

Options:

A.

Use Google-managed encryption keys (GMEK).

B.

Use customer-managed encryption keys (CMEK).

C.

Use customer-supplied encryption keys (CSEK).

D.

Use customer-supplied encryption keys (CSEK) for the sensitive data and customer-managed encryption keys (CMEK) for the less sensitive data.

Buy Now
Questions 10

Your organization sends IoT event data to a Pub/Sub topic. Subscriber applications read and perform transformations on the messages before storing them in the data warehouse. During particularly busy times when more data is being written to the topic, you notice that the subscriber applications are not acknowledging messages within the deadline. You need to modify your pipeline to handle these activity spikes and continue to process the messages. What should you do?

Options:

A.

Retry messages until they are acknowledged.

B Implement flow control on the subscribers

B.

Forward unacknowledged messages to a dead-letter topic.

C.

Seek back to the last acknowledged message.

Buy Now
Questions 11

You work for an online retail company. Your company collects customer purchase data in CSV files and pushes them to Cloud Storage every 10 minutes. The data needs to be transformed and loaded into BigQuery for analysis. The transformation involves cleaning the data, removing duplicates, and enriching it with product information from a separate table in BigQuery. You need to implement a low-overhead solution that initiates data processing as soon as the files are loaded into Cloud Storage. What should you do?

Options:

A.

Use Cloud Composer sensors to detect files loading in Cloud Storage. Create a Dataproc cluster, and use a Composer task to execute a job on the cluster to process and load the data into BigQuery.

B.

Schedule a direct acyclic graph (DAG) in Cloud Composer to run hourly to batch load the data from Cloud Storage to BigQuery, and process the data in BigQuery using SQL.

C.

Use Dataflow to implement a streaming pipeline using anOBJECT_FINALIZEnotification from Pub/Sub to read the data from Cloud Storage, perform the transformations, and write the data to BigQuery.

D.

Create a Cloud Data Fusion job to process and load the data from Cloud Storage into BigQuery. Create anOBJECT_FINALIZE notification in Pub/Sub, and trigger a Cloud Run function to start the Cloud Data Fusion job as soon as new files are loaded.

Buy Now
Questions 12

You created a customer support application that sends several forms of data to Google Cloud. Your application is sending:

1. Audio files from phone interactions with support agents that will be accessed during trainings.

2. CSV files of users’ personally identifiable information (Pll) that will be analyzed with SQL.

3. A large volume of small document files that will power other applications.

You need to select the appropriate tool for each data type given the required use case, while following Google-recommended practices. Which should you choose?

Options:

A.

1. Cloud Storage

2. CloudSQL for PostgreSQL

3. Bigtable

B.

1. Filestore

2. Cloud SQL for PostgreSQL

3. Datastore

C.

1. Cloud Storage

2. BigQuery

3. Firestore

D.

1. Filestore

2. Bigtable

3. BigQuery

Buy Now
Questions 13

You manage a BigQuery table that is used for critical end-of-month reports. The table is updated weekly with new sales data. You want to prevent data loss and reporting issues if the table is accidentally deleted. What should you do?

Options:

A.

Configure the time travel duration on the table to be exactly seven days. On deletion, re-create the deleted table solely from the time travel data.

B.

Schedule the creation of a new snapshot of the table once a week. On deletion, re-create the deleted table using the snapshot and time travel data.

C.

Create a clone of the table. On deletion, re-create the deleted table by copying the content of the clone.

D.

Create a view of the table. On deletion, re-create the deleted table from the view and time travel data.

Buy Now
Questions 14

Your organization is building a new application on Google Cloud. Several data files will need to be stored in Cloud Storage. Your organization has approved only two specific cloud regions where these data files can reside. You need to determine a Cloud Storage bucket strategy that includes automated high availability. What should you do?

Options:

A.

Create a dual-region bucket, and upload the files to this bucket.

B.

Create a single-region bucket in each of the two regions, and use the gcloud storage command to replicate the data across the buckets in both regions.

C.

Create a multi-region bucket, and upload the files to this bucket.

D.

Create a single-region bucket in each of the two regions, and use Storage Transfer Service to replicate the data across the buckets in both regions.

Buy Now
Questions 15

You have a Dataproc cluster that performs batch processing on data stored in Cloud Storage. You need to schedule a daily Spark job to generate a report that will be emailed to stakeholders. You need a fully-managed solution that is easy to implement and minimizes complexity. What should you do?

Options:

A.

Use Cloud Composer to orchestrate the Spark job and email the report.

B.

Use Dataproc workflow templates to define and schedule the Spark job, and to email the report.

C.

Use Cloud Run functions to trigger the Spark job and email the report.

D.

Use Cloud Scheduler to trigger the Spark job. and use Cloud Run functions to email the report.

Buy Now
Questions 16

You are migrating data from a legacy on-premises MySQL database to Google Cloud. The database contains various tables with different data types and sizes, including large tables with millions of rowsand transactional data. You need to migrate this data while maintaining data integrity, and minimizing downtime and cost. What should you do?

Options:

A.

Set up a Cloud Composer environment to orchestrate a custom data pipeline. Use a Python script to extract data from the MySQL database and load it to MySQL on Compute Engine.

B.

Export the MySQL database to CSV files, transfer the files to Cloud Storage by using Storage Transfer Service, and load the files into a Cloud SQL for MySQL instance.

C.

Use Database Migration Service to replicate the MySQL database to a Cloud SQL for MySQL instance.

D.

Use Cloud Data Fusion to migrate the MySQL database to MySQL on Compute Engine.

Buy Now
Questions 17

You are a database administrator managing sales transaction data by region stored in a BigQuery table. You need to ensure that each sales representative can only see the transactions in their region. What should you do?

Options:

A.

Add a policy tag in BigQuery.

B.

Create a row-level access policy.

C.

Create a data masking rule.

D.

Grant the appropriate 1AM permissions on the dataset.

Buy Now
Questions 18

Your organization uses Dataflow pipelines to process real-time financial transactions. You discover that one of your Dataflow jobs has failed. You need to troubleshoot the issue as quickly as possible. What should you do?

Options:

A.

Set up a Cloud Monitoring dashboard to track key Dataflow metrics, such as data throughput, error rates, and resource utilization.

B.

Create a custom script to periodically poll the Dataflow API for job status updates, and send email alerts if any errors are identified.

C.

Navigate to the Dataflow Jobs page in the Google Cloud console. Use the job logs and worker logs to identify the error.

D.

Use the gcloud CLI tool to retrieve job metrics and logs, and analyze them for errors and performance bottlenecks.

Buy Now
Questions 19

You need to create a weekly aggregated sales report based on a large volume of data. You want to use Python to design an efficient process for generating this report. What should you do?

Options:

A.

Create a Cloud Run function that uses NumPy. Use Cloud Scheduler to schedule the function to run once a week.

B.

Create a Colab Enterprise notebook and use the bigframes.pandas library. Schedule the notebook to execute once a week.

C.

Create a Cloud Data Fusion and Wrangler flow. Schedule the flow to run once a week.

D.

Create a Dataflow directed acyclic graph (DAG) coded in Python. Use Cloud Scheduler to schedule the code to run once a week.

Buy Now
Questions 20

You need to create a data pipeline that streams event information from applications in multiple Google Cloud regions into BigQuery for near real-time analysis. The data requires transformation before loading. You want to create the pipeline using a visual interface. What should you do?

Options:

A.

Push event information to a Pub/Sub topic. Create a Dataflow job using the Dataflow job builder.

B.

Push event information to a Pub/Sub topic. Create a Cloud Run function to subscribe to the Pub/Sub topic, apply transformations, and insert the data into BigQuery.

C.

Push event information to a Pub/Sub topic. Create a BigQuery subscription in Pub/Sub.

D.

Push event information to Cloud Storage, and create an external table in BigQuery. Create a BigQuery scheduled job that executes once each day to apply transformations.

Buy Now
Questions 21

Your team needs to analyze large datasets stored in BigQuery to identify trends in user behavior. The analysis will involve complex statistical calculations, Python packages, and visualizations. You need to recommend a managed collaborative environment to develop and share the analysis. What should you recommend?

Options:

A.

Create a Colab Enterprise notebook and connect the notebook to BigQuery. Share the notebook with your team. Analyze the data and generate visualizations in Colab Enterprise.

B.

Create a statistical model by using BigQuery ML. Share the query with your team. Analyze the data and generate visualizations in Looker Studio.

C.

Create a Looker Studio dashboard and connect the dashboard to BigQuery. Share the dashboard with your team. Analyze the data and generate visualizations in Looker Studio.

D.

Connect Google Sheets to BigQuery by using Connected Sheets. Share the Google Sheet with your team. Analyze the data and generate visualizations in Gooqle Sheets.

Buy Now
Questions 22

Your organization needs to implement near real-time analytics for thousands of events arriving each second in Pub/Sub. The incoming messages require transformations. You need to configure a pipeline that processes, transforms, and loads the data into BigQuery while minimizing development time. What should you do?

Options:

A.

Use a Google-provided Dataflow template to process the Pub/Sub messages, perform transformations, and write the results to BigQuery.

B.

Create a Cloud Data Fusion instance and configure Pub/Sub as a source. Use Data Fusion to process the Pub/Sub messages, perform transformations, and write the results to BigQuery.

C.

Load the data from Pub/Sub into Cloud Storage using a Cloud Storage subscription. Create a Dataproc cluster, use PySpark to perform transformations in Cloud Storage, and write the results to BigQuery.

D.

Use Cloud Run functions to process the Pub/Sub messages, perform transformations, and write the results to BigQuery.

Buy Now
Questions 23

Your organization’s business analysts require near real-time access to streaming data. However, they are reporting that their dashboard queries are loading slowly. After investigating BigQuery query performance, you discover the slow dashboard queries perform several joins and aggregations.

You need to improve the dashboard loading time and ensure that the dashboard data is as up-to-date as possible. What should you do?

Options:

A.

Disable BiqQuery query result caching.

B.

Modify the schema to use parameterized data types.

C.

Create a scheduled query to calculate and store intermediate results.

D.

Create materialized views.

Buy Now
Questions 24

You are a data analyst at your organization. You have been given a BigQuery dataset that includes customer information. The dataset contains inconsistencies and errors, such as missing values, duplicates, and formatting issues. You need to effectively and quickly clean the data. What should you do?

Options:

A.

Develop a Dataflow pipeline to read the data from BigQuery, perform data quality rules and transformations, and write the cleaned data back to BigQuery.

B.

Use Cloud Data Fusion to create a data pipeline to read the data from BigQuery, perform data quality transformations, and write the clean data back to BigQuery.

C.

Export the data from BigQuery to CSV files. Resolve the errors using a spreadsheet editor, and re-import the cleaned data into BigQuery.

D.

Use BigQuery's built-in functions to perform data quality transformations.

Buy Now
Questions 25

Your company’s ecommerce website collects product reviews from customers. The reviews are loaded as CSV files daily to a Cloud Storage bucket. The reviews are in multiple languages and need to be translated to Spanish. You need to configure a pipeline that is serverless, efficient, and requires minimal maintenance. What should you do?

Options:

A.

Load the data into BigQuery using Dataproc. Use Apache Spark to translate the reviews by invoking the Cloud Translation API. Set BigQuery as the sink.U

B.

Use a Dataflow templates pipeline to translate the reviews using the Cloud Translation API. Set BigQuery as the sink.

C.

Load the data into BigQuery using a Cloud Run function. Use the BigQuery ML create model statement to train a translation model. Use the model to translate the product reviews within BigQuery.

D.

Load the data into BigQuery using a Cloud Run function. Create a BigQuery remote function that invokes the Cloud Translation API. Use a scheduled query to translate new reviews.

Buy Now
Questions 26

You work for a home insurance company. You are frequently asked to create and save risk reports with charts for specific areas using a publicly available storm event dataset. You want to be able to quickly create and re-run risk reports when new data becomes available. What should you do?

Options:

A.

Export the storm event dataset as a CSV file. Import the file to Google Sheets, and use cell data in the worksheets to create charts.

B.

Copy the storm event dataset into your BigQuery project. Use BigQuery Studio to query and visualize the data in Looker Studio.

C.

Reference and query the storm event dataset using SQL in BigQuery Studio. Export the results to Google Sheets, and use cell data in the worksheets to create charts.

D.

Reference and query the storm event dataset using SQL in a Colab Enterprise notebook. Display the table results and document with Markdown, and use Matplotlib to create charts.

Buy Now
Exam Name: Google Cloud Associate Data Practitioner (ADP Exam)
Last Update: Apr 16, 2025
Questions: 106

PDF + Testing Engine

$49.5  $164.99

Testing Engine

$37.5  $124.99
buy now Associate-Data-Practitioner testing engine

PDF (Q&A)

$31.5  $104.99
buy now Associate-Data-Practitioner pdf
dumpsmate guaranteed to pass
24/7 Customer Support

DumpsMate's team of experts is always available to respond your queries on exam preparation. Get professional answers on any topic of the certification syllabus. Our experts will thoroughly satisfy you.

Site Secure

mcafee secure

TESTED 20 Apr 2025