Google Professional Data Engineer Certificate EXAMTOPIC DUMPS Q96-Q100
Q 96.
You want to analyze hundreds of thousands of social media posts daily at the lowest cost and with the fewest steps. You have the following requirements:
- You will batch-load the posts once per day and run them through the
Cloud Natural Language API
.- You will extract topics and sentiment from the posts.
- You must store the raw posts for archiving and reprocessing.
- You will create dashboards to be shared with people both inside and outside your organization.
You need to store both the data extracted from the API to perform analysis as well as the raw social media posts for historical archiving. What should you do?
- ❌ A. Store the social media posts and the data extracted from the API in BigQuery.
- ❌ B. Store the social media posts and the data extracted from the API in Cloud SQL.
- ⭕ C. Store the raw social media posts in Cloud Storage, and write the data extracted from the API into BigQuery.
- ❌ D. Feed to social media posts into the API
directly from the source,and write the extracted data from the API into BigQuery.
→ Must store the raw posts
Integrated repository for analytics and ML
- Media content storage
Cloud Storage
- Analysis / Dashboards
BigQuery
Q 97.
You store historic data in
Cloud Storage
. You need to perform analytics on the historic data. You want to use a solution to detect invalid data entries and perform data transformations that will not require programming or knowledge of SQL. What should you do?
- ❌ A. Use
Cloud Dataflow with Beamto detect errors and perform transformations. - ⭕ B. Use
Cloud Dataprep
with recipes to detect errors and perform transformations. - ❌ C. Use
Cloud Dataproc with a Hadoop jobto detect errors and perform transformations. - ❌ D. Use
federated tables in BigQuerywith queries to detect errors and perform transformations.
Cloud Dataprep
ETL processes on GCP for low/no-code solutions : Data Fusion & Cloud Dataprep
Dataprep
- More refined and cost-effective but limited in capability
- Third-party application : Managed _Trifacta for preparing and analysing quality and transforming the input data
- Service for visually exploring, cleaning, and preparing data for analysis.
- Can transform data of any size stored in CSV, JSON, or relational table formats
'Certificate - DS > Data engineer' 카테고리의 다른 글
[PDE CERTIFICATE - EXAMTOPIC] DUMPS Q121-Q125 (0) | 2022.02.17 |
---|---|
[PDE CERTIFICATE - EXAMTOPIC] DUMPS Q111-Q115 (0) | 2022.02.17 |
[PDE CERTIFICATE - EXAMTOPIC] DUMPS Q91-Q95 (0) | 2022.02.17 |
Professional Data Engineer 샘플 문제 정리 (0) | 2022.02.16 |
[PDE CERTIFICATE - EXAMTOPIC] DUMPS Q76-Q80 (0) | 2022.02.16 |