A production workload incrementally applies updates from an external Change Data Capture feed to a Delta Lake table as an always-on Structured Stream job. When data was initially migrated for this table, OPTIMIZE was executed and most data files were resized to 1 GB. Auto Optimize and Auto Compaction were both turned on for the streaming production job. Recent review of data files shows that most data files are under 64 MB, although each partition in the table contains at least 1 GB of data and the total table size is over 10 TB.
Which of the following likely explains these smaller file sizes?
cotardo2077
Highly Voted 1 year, 9 months agoVitality
1 week, 3 days agomeatpoof
4 months, 3 weeks agoVitality
Most Recent 1 week, 3 days agoKadELbied
1 month, 1 week agoJoG1221
2 months agoJoG1221
2 months agoJoG1221
2 months agokishanu
2 months, 1 week agoAlHerd
2 months, 3 weeks agoEZZALDIN
2 months, 3 weeks agoTedet
3 months, 3 weeks agoTedet
3 months, 3 weeks agorollno1
4 months agoMelik3
10 months, 2 weeks agoimatheushenrique
1 year agoRiktRikt007
1 year, 4 months agoPrashantTiwari
1 year, 4 months agoAziLa
1 year, 4 months agoJay_98_11
1 year, 5 months ago