Skip to content
This repository has been archived by the owner on Feb 12, 2020. It is now read-only.

Minimize number of random record file reads during Curly compaction #225

Open
thomas-kielbus opened this issue Dec 8, 2011 · 1 comment

Comments

@thomas-kielbus
Copy link
Contributor

While we perform the key file merge sort, we do random reads of records to be written from the record files being merged. It seems like there is a lot of potential to optimize this process. We could merge sort a fixed number of keys (e.g. 1024) and then only read once from each record file and get all the necessary records in memory before writing them to the compacted partition.

@thomas-kielbus
Copy link
Contributor Author

It is important to note that there is no guarantee that the record file will be sorted. However, it will usually be the case and we should take advantage of that.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant