Skip to content

Conversation

@ZJONSSON
Copy link
Contributor

Splits each column into pages as determined by pageSize. Encoding into buffer is done as soon as we have enough rows to encode a page, even though the rowBuffer is far from being full. This ideally should reduce memory usage since the encoded pages should be substantially smaller than the shred data. Also pages facilitate scrolling fast through a column, skipping pages of no interest, and also help in locating individual records without having to read the whole column.

* bitpacking should work for any length of data, not just multiple of 8 (last packed is padded if less than 8)

* Improve runs estimation - only start a new run if we are at a mod 8 === 0, otherwise use bitpacking
This moves data into encoded buffer as soon as possible, reducing memory requirements for the whole rowGroup
into ParquetReader so they can be called without arguments.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant