1BRC in SQL with Databend Cloud(<10.4s, ~2.2GB/s from S3) #230
BohuTANG
started this conversation in
Show and tell
Replies: 3 comments 1 reply
-
Beta Was this translation helpful? Give feedback.
0 replies
-
Beta Was this translation helpful? Give feedback.
0 replies
-
how about join multiple table with billion rows ? have you test it? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
SQL and read the 1RBC data from AWS S3.
Databend Cloud is a cost-effective cloud warehouse(alternative to Snowflake).
We delve into a practical case where a medium-sized data warehouse, equipped with 2 nodes (each node powered by 16vCPUs, in AWS us-east-2), tackles the challenge of analyzing 1BRC file, which is in parquet within AWS S3( 5.1GB, us-east-2), generated follow the steps similar to Robin Moffatt for generating the raw data and exporting it to a Parquet file.
This is the query(You can run and verified on Databend Cloud):
Reference:
[The One Billion Rows Challenge with Snowflake and Databend]: https://www.databend.com/blog/2024-01-05-1brows/
Beta Was this translation helpful? Give feedback.
All reactions