Your company is running their first dynamic campaign, serving different offers by analyzing real-time data during the holiday season. The data scientists are collecting terabytes of data that rapidly grows every hour during their 30-day campaign. They are using Google Cloud Dataflow to preprocess the data and collect the feature (signals) data that is needed for the machine learning model in Google Cloud Bigtable. The team is observing suboptimal performance with reads and writes of their initial load of 10 TB of data. They want to improve this performance while minimizing cost. What should they do?
A) Redefine the schema by evenly distributing reads and writes across the row space of the table.
B) The performance issue should be resolved over time as the site of the BigDate cluster is increased.
C) Redesign the schema to use a single row key to identify values that need to be updated frequently in the cluster.
D) Redesign the schema to use row keys based on numeric IDs that increase sequentially per user viewing the offers.
Correct Answer:
Verified
Q10: Your company is migrating their 30-node Apache
Q11: You have spent a few days loading
Q12: Flowlogistic Case Study Company Overview Flowlogistic is
Q13: Your company is performing data preprocessing for
Q14: You create an important report for your
Q16: Your weather app queries a database every
Q17: MJTelco Case Study Company Overview MJTelco is
Q18: Your company built a TensorFlow neutral-network model
Q19: You are building a model to make
Q20: You are creating a model to predict
Unlock this Answer For Free Now!
View this answer and more for free by performing one of the following actions
Scan the QR code to install the App and get 2 free unlocks
Unlock quizzes for free by uploading documents