Deal of The Day! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Google Exam Professional Data Engineer Topic 3 Question 85 Discussion

Actual exam question for Google's Professional Data Engineer exam
Question #: 85
Topic #: 3
[All Professional Data Engineer Questions]

You operate a database that stores stock trades and an application that retrieves average stock price for a given company over an adjustable window of time. The data is stored in Cloud Bigtable where the datetime of the stock trade is the beginning of the row key. Your application has thousands of concurrent users, and you notice that performance is starting to degrade as more stocks are added. What should you do to improve the performance of your application?

Show Suggested Answer Hide Answer
Suggested Answer: A

Data Fusion's advantages:

Visual interface: Offers a user-friendly interface for designing data pipelines without extensive coding, making it accessible to a wider range of users.

Built-in transformations: Includes a wide range of pre-built transformations to handle common data quality issues, such as:

Data type conversions

Data cleansing (e.g., removing invalid characters, correcting formatting)

Data validation (e.g., checking for missing values, enforcing constraints)

Data enrichment (e.g., adding derived fields, joining with other datasets)

Custom transformations: Allows for custom transformations using SQL or Java code for more complex cleaning tasks.

Scalability: Can handle large datasets efficiently, making it suitable for processing CSV files with potential data quality issues.

Integration with BigQuery: Integrates seamlessly with BigQuery, allowing for direct loading of transformed data.


Contribute your Thoughts:

Kassandra
14 days ago
C) Using BigQuery for storing the stock trades and updating the application seems like a reasonable approach. It could help handle the increased data volume and concurrency.
upvoted 0 times
...
Elke
17 days ago
I see both points, but I think option C) changing the data pipeline to use BigQuery might be the most efficient solution in the long run.
upvoted 0 times
...
Adelina
18 days ago
B) A random number per second? That's just crazy! How are we supposed to query that efficiently? Definitely not the way to go.
upvoted 0 times
...
Ammie
19 days ago
A) Changing the row key syntax to start with the stock symbol sounds like a good idea. That way, we can group all the trades for a single stock together and improve query performance.
upvoted 0 times
...
Jaime
25 days ago
I disagree, I believe option D) using Cloud Dataflow to write summary of each day's stock trades to an Avro file is the best approach.
upvoted 0 times
...
Tiera
30 days ago
I think we should go with option A) Change the row key syntax to begin with the stock symbol.
upvoted 0 times
...

Save Cancel
az-700  pass4success  az-104  200-301  200-201  cissp  350-401  350-201  350-501  350-601  350-801  350-901  az-720  az-305  pl-300  

Warning: Cannot modify header information - headers already sent by (output started at /pass.php:70) in /pass.php on line 77