You need to create a data pipeline that copies time-series transaction data so that it can be queried from within BigQuery by your data science team for analysis.
Every hour, thousands of transactions are updated with a new status. The size of the initial dataset is 1.5 PB, and it will grow by 3 TB per day. The data is heavily structured, and your data science team will build machine learning models based on this data. You want to maximize performance and usability for your data science team. Which two strategies should you adopt? (Choose two.)
rickywck
Highly Voted 3 years, 9 months ago[Removed]
Highly Voted 3 years, 9 months agoawssp12345
2 years, 6 months agomidgoo
Most Recent 9 months, 3 weeks agovaga1
7 months agoWillemHendr
7 months, 1 week agoodacir
1 year, 1 month agozellck
1 year, 1 month agoAzureDP900
1 year agoNicolasN
1 year, 2 months agojkhong
1 year agoNicolasN
1 year, 2 months agoDerickTW
1 year, 3 months agodevaid
1 year, 2 months agoMaxNRG
2 years agomedeis_jar
2 years agodoninakula
2 years, 1 month agosumanshu
2 years, 6 months agoJeysolomon
2 years, 6 months agoChelseajcole
2 years, 2 months agoretep007
2 years, 3 months agodaghayeghi
2 years, 10 months agoHithesh
2 years, 10 months agosumanshu
2 years, 6 months agoraf2121
2 years, 5 months agohdmi_switch
2 years, 5 months agodaghayeghi
2 years, 10 months agokarthik89
2 years, 10 months agodaghayeghi
2 years, 10 months agoNams_139
3 years, 1 month agofedericohi
3 years, 1 month ago