You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
One idea to increase roaringbitmap compression ratio is create a global contiguous transaction index, and use the roaringbitmap to mark the transaction index instead of sparse, rarely contiguous consensus timestamp. In doing so, there is a much higher change to encode the bits into run-length containers instead of the other two types which don't save as much space.
Also we want to use 32-bit roaringbitmap to reduce both write and read amplification. This however requires the table to have three columns, entity_id, high_value, and bitmap, where the high value is the high 32-bit of the transaction index.
The design would change how the rows are created, merged, and optimized during ingestion, as well as how to query the data based on the rest api query filters since essentially the bitmaps are not split into chunks.
The goal of the ticket is to run PoCs, given transactions with production env pattern, test and identify the potential bottlenecks, and optimize if possible.
Solution
check the description
Alternatives
No response
The text was updated successfully, but these errors were encountered:
xin-hedera
changed the title
Poc: test read / write performance of 32-bit roaringbitmap grouped by entity_id and high 32-bit of transaction index
Poc: test read / write performance of chucked 32-bit roaringbitmap
May 10, 2024
steven-sheehy
changed the title
Poc: test read / write performance of chucked 32-bit roaringbitmap
Poc: test read / write performance of chunked 32-bit roaringbitmap
May 10, 2024
Problem
One idea to increase roaringbitmap compression ratio is create a global contiguous transaction index, and use the roaringbitmap to mark the transaction index instead of sparse, rarely contiguous consensus timestamp. In doing so, there is a much higher change to encode the bits into run-length containers instead of the other two types which don't save as much space.
Also we want to use 32-bit roaringbitmap to reduce both write and read amplification. This however requires the table to have three columns,
entity_id
,high_value
, andbitmap
, where the high value is the high 32-bit of the transaction index.The design would change how the rows are created, merged, and optimized during ingestion, as well as how to query the data based on the rest api query filters since essentially the bitmaps are not split into chunks.
The goal of the ticket is to run PoCs, given transactions with production env pattern, test and identify the potential bottlenecks, and optimize if possible.
Solution
check the description
Alternatives
No response
The text was updated successfully, but these errors were encountered: