Hendry Suwanda
01/27/2023, 2:49 PMdescribe detail
the total of files has been reduced.
but, when I run the readStream job, the job still read the old files (scan more than 5000 files), then I try to run the vacuum(0)
, but the readStream still read the file that has been deleted by vacuum command.
But, when I try to run the select
command, there is no error.
Is there any way to invalidate the delta table? why the readStream still read the file that has been deleted?
note: I use delta version 1.0.0
ThanksGerhard Brueckl
01/27/2023, 7:26 PMHendry Suwanda
01/28/2023, 1:19 AMGerhard Brueckl
01/30/2023, 8:05 AMvacuum
on the 5th and delete all changes before, and restart the stream, it will complain that the files between 2nd and 5th are missing.
you could use ``ignoreMissingFiles`` but obviously the changes between 2nd and 5th will be missing from your stream then
https://learn.microsoft.com/en-us/azure/databricks/kb/delta/file-transaction-log-not-foundDominique Brezinski
01/30/2023, 5:08 PMvijay
01/30/2023, 10:10 PMDominique Brezinski
01/30/2023, 11:12 PM