Google GCP-PDE Certification Exam Sample Questions

GCP-PDE Braindumps, GCP-PDE Exam Dumps, GCP-PDE Examcollection, GCP-PDE Questions PDF, GCP-PDE Sample Questions, Professional Data Engineer Dumps, Professional Data Engineer Official Cert Guide PDF, Professional Data Engineer VCEWe have prepared Google Professional Data Engineer (GCP-PDE) certification sample questions to make you aware of actual exam properties. This sample question set provides you with information about the Professional Data Engineer exam pattern, question formate, a difficulty level of questions and time required to answer each question. To get familiar with Google Cloud Platform - Professional Data Engineer (GCP-PDE) exam, we suggest you try our Sample Google GCP-PDE Certification Practice Exam in simulated Google certification exam environment.

To test your knowledge and understanding of concepts with real-time scenario based Google GCP-PDE questions, we strongly recommend you to prepare and practice with Premium Google Professional Data Engineer Certification Practice Exam. The premium Google Professional Data Engineer certification practice exam helps you identify topics in which you are well prepared and topics in which you may need further training to achieving great score in actual Google Cloud Platform - Professional Data Engineer (GCP-PDE) exam.

Google GCP-PDE Sample Questions:

01. You have 250,000 devices which produce a JSON device status event every 10 seconds. You want to capture this event data for outlier time series analysis. What should you do?
a) Ship the data into BigQuery. Develop a custom application that uses the BigQuery API to query the dataset and displays device outlier data based on your business requirements.
b) Ship the data into BigQuery. Use the BigQuery console to query the dataset and display device outlier data based on your business requirements.
c) Ship the data into Cloud Bigtable. Use the Cloud Bigtable cbt tool to display device outlier data based on your business requirements.
d) Ship the data into Cloud Bigtable. Install and use the HBase shell for Cloud Bigtable to query the table for device outlier data based on your business requirements.
 
02. You are designing storage for CSV files and using an I/O-intensive custom Apache Spark transform as part of deploying a data pipeline on Google Cloud. You intend to use ANSI SQL to run queries for your analysts.
How should you transform the input data?
a) Use BigQuery for storage. Use Dataflow to run the transformations.
b) Use BigQuery for storage. Use Dataproc to run the transformations.
c) Use Cloud Storage for storage. Use Dataflow to run the transformations.
d) Use Cloud Storage for storage. Use Dataproc to run the transformations.
 
03. Your company is loading comma-separated values (CSV) files into BigQuery. The data is fully imported successfully; however, the imported data is not matching byte-to-byte to the source file.
What is the most likely cause of this problem?
a) The CSV data loaded in BigQuery is not flagged as CSV.
b) The CSV data had invalid rows that were skipped on import.
c) The CSV data has not gone through an ETL phase before loading into BigQuery.
d) The CSV data loaded in BigQuery is not using BigQuery’s default encoding.
 
04. You are using Pub/Sub to stream inventory updates from many point-of-sale (POS) terminals into BigQuery.
Each update event has the following information: product identifier "prodSku", change increment "quantityDelta", POS identification "termId", and "messageId" which is created for each push attempt from the terminal.
During a network outage, you discovered that duplicated messages were sent, causing the inventory system to over-count the changes. You determine that the terminal application has design problems and may send the same event more than once during push retries.
You want to ensure that the inventory update is accurate. What should you do?
a) Add another attribute orderId to the message payload to mark the unique check-out order across all terminals. Make sure that messages whose "orderId" and "prodSku" values match corresponding rows in the BigQuery table are discarded.
b) Inspect the "messageId" of each message. Make sure that any messages whose "messageId" values match corresponding rows in the BigQuery table are discarded.
c) Instead of specifying a change increment for "quantityDelta", always use the derived inventory value after the increment has been applied. Name the new attribute "adjustedQuantity".
d) Inspect the "publishTime" of each message. Make sure that messages whose "publishTime" values match rows in the BigQuery table are discarded.
 
05. You are building storage for files for a data pipeline on Google Cloud. You want to support JSON files. The schema of these files will occasionally change.
Your analyst teams will use running aggregate ANSI SQL queries on this data. What should you do?
a) Use BigQuery for storage. Provide format files for data load. Update the format files as needed.
b) Use BigQuery for storage. Select "Automatically detect" in the Schema section.
c) Use Cloud Storage for storage. Link data as temporary tables in BigQuery and turn on the "Automatically detect" option in the Schema section of BigQuery.
d) Use Cloud Storage for storage. Link data as permanent tables in BigQuery and turn on the "Automatically detect" option in the Schema section of BigQuery.
 
06. You need to stream time-series data in Avro format, and then write this to both BigQuery and Cloud Bigtable simultaneously using Dataflow. You want to achieve minimal end-to-end latency.
Your business requirements state this needs to be completed as quickly as possible. What should you do?
a) Create a pipeline and use ParDo transform.
b) Create a pipeline that groups the data into a PCollection and uses the Combine transform.
c) Create a pipeline that groups data using a PCollection, and then use Avro I/O transform to write to Cloud Storage. After the data is written, load the data from Cloud Storage into BigQuery and Bigtable.
d) Create a pipeline that groups data using a PCollection and then uses Bigtable and BigQueryIO transforms.
 
07. You are working on a project with two compliance requirements. The first requirement states that your developers should be able to see the Google Cloud billing charges for only their own projects.
The second requirement states that your finance team members can set budgets and view the current charges for all projects in the organization.
The finance team should not be able to view the project contents. You want to set permissions. What should you do?
a) Add the finance team members to the Billing Administrator role for each of the billing accounts that they need to manage. Add the developers to the Viewer role for the Project.
b) Add the finance team members to the default IAM Owner role. Add the developers to a custom role that allows them to see their own spend only.
c) Add the developers and finance managers to the Viewer role for the Project.
d) Add the finance team to the Viewer role for the Project. Add the developers to the Security Reviewer role for each of the billing accounts.
 
08. You want to publish system metrics to Google Cloud from a large number of on-prem hypervisors and VMs for analysis and creation of dashboards.
You have an existing custom monitoring agent deployed to all the hypervisors and your on-prem metrics system is unable to handle the load. You want to design a system that can collect and store metrics at scale. You don't want to manage your own time series database.
Metrics from all agents should be written to the same table but agents must not have permission to modify or read data written by other agents. What should you do?
a) Modify the monitoring agent to write protobuf messages directly to BigTable.
b) Modify the monitoring agent to publish protobuf messages to Pub/Sub. Use a Dataproc cluster or Dataflow job to consume messages from Pub/Sub and write to BigTable.
c) Modify the monitoring agent to write protobuf messages to HBase deployed on Compute Engine VM Instances
d) Modify the monitoring agent to write protobuf messages to Pub/Sub. Use a Dataproc cluster or Dataflow job to consume messages from Pub/Sub and write to Cassandra deployed on Compute Engine VM Instances.
 
09. Your company is streaming real-time sensor data from their factory floor into Bigtable and they have noticed extremely poor performance.
How should the row key be redesigned to improve Bigtable performance on queries that populate real-time dashboards?
a) Use a row key of the form <timestamp>.
b) Use a row key of the form <sensorid>.
c) Use a row key of the form <timestamp>#<sensorid>.
d) Use a row key of the form <sensorid>#<timestamp>.
 
10. You are designing a relational data repository on Google Cloud to grow as needed. The data will be transactionally consistent and added from any location in the world.
You want to monitor and adjust node count for input traffic, which can spike unpredictably. What should you do?
a) Use Cloud Spanner for storage. Monitor CPU utilization and increase node count if more than 70% utilized for your time span.
b) Use Cloud Spanner for storage. Monitor storage usage and increase node count if more than 70% utilized.
c) Use Cloud Bigtable for storage. Monitor data stored and increase node count if more than 70% utilized.
d) Use Cloud Bigtable for storage. Monitor CPU utilization and increase node count if more than 70% utilized for your time span.

Answers:

Question: 01
Answer: c
Question: 02
Answer: b
Question: 03
Answer: d
Question: 04
Answer: a
Question: 05
Answer: b
Question: 06
Answer: d
Question: 07
Answer: a
Question: 08
Answer: b
Question: 09
Answer: d
Question: 10
Answer: a

Note: Please update us by writing an email on feedback@vmexam.com for any error in Google Cloud Platform - Professional Data Engineer (GCP-PDE) certification exam sample questions

Your rating: None Rating: 4.9 / 5 (82 votes)