You need to create a data pipeline that copies time-series transaction data so that it can be queried from within BigQuery by your data science team for analysis.
Every hour, thousands of transactions are updated with a new status. The size of the initial dataset is 1.5 PB, and it will grow by 3 TB per day. The data is heavily structured, and your data science team will build machine learning models based on this data. You want to maximize performance and usability for your data science team. Which two strategies should you adopt? (Choose two.)
rickywck
Highly Voted 4 years, 1 month ago[Removed]
Highly Voted 4 years, 1 month agoawssp12345
2 years, 9 months agomidgoo
Most Recent 1 year, 1 month agovaga1
10 months, 3 weeks agoWillemHendr
10 months, 3 weeks agoodacir
1 year, 4 months agozellck
1 year, 4 months agoAzureDP900
1 year, 3 months agoNicolasN
1 year, 5 months agojkhong
1 year, 4 months agoNicolasN
1 year, 5 months agoDerickTW
1 year, 7 months agodevaid
1 year, 6 months agoMaxNRG
2 years, 3 months agomedeis_jar
2 years, 3 months agodoninakula
2 years, 5 months agosumanshu
2 years, 9 months agoJeysolomon
2 years, 10 months agoChelseajcole
2 years, 6 months agoretep007
2 years, 7 months agodaghayeghi
3 years, 1 month agoHithesh
3 years, 1 month agosumanshu
2 years, 9 months agoraf2121
2 years, 9 months agohdmi_switch
2 years, 9 months agodaghayeghi
3 years, 2 months agokarthik89
3 years, 2 months agodaghayeghi
3 years, 1 month agoNams_139
3 years, 5 months agofedericohi
3 years, 5 months ago