You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Given that we never merge existing delta layers, if a page is updated multiple times throughout a long time, the read cost of reconstructing the page would be high. Currently it would be reading 6 files on average and 20 files p99 to reconstruct a page.
This would be resolved as a long-term task as in #4359
Environment
Logs, links
The text was updated successfully, but these errors were encountered:
I read through the compaction epic and seems like the compaction isn’t done yet and we are trying to reconstruct the page by reading all the small files.
I have few questions
Do you have an index (sparse index -> key : byte offset) as metadata with a fixed size beginning of the file ? More like pointing to a block within a file. Future optimisation : that block can be compressed to reduce I/O bandwidth.
Might be possible the read amplification are just false positives where you go through all the files just to find nothing. Maybe a bloom filter might help in this case.
Steps to reproduce
Expected result
Actual result
Given that we never merge existing delta layers, if a page is updated multiple times throughout a long time, the read cost of reconstructing the page would be high. Currently it would be reading 6 files on average and 20 files p99 to reconstruct a page.
This would be resolved as a long-term task as in #4359
Environment
Logs, links
The text was updated successfully, but these errors were encountered: