Parquet – NVIDIA Technical Blog News and tutorials for developers, data scientists, and IT admins 2025-05-29T19:05:22Z http://www.open-lab.net/blog/feed/ Matt Ahrens <![CDATA[Accelerating Apache Parquet Scans on Apache Spark with GPUs]]> http://www.open-lab.net/blog/?p=98350 2025-04-22T23:57:50Z 2025-04-03T16:18:03Z As data sizes have grown in enterprises across industries, Apache Parquet has become a prominent format for storing data. Apache Parquet is a columnar storage...]]> As data sizes have grown in enterprises across industries, Apache Parquet has become a prominent format for storing data. Apache Parquet is a columnar storage...Decorative image.

As data sizes have grown in enterprises across industries, Apache Parquet has become a prominent format for storing data. Apache Parquet is a columnar storage format designed for efficient data processing at scale. By organizing data by columns rather than rows, Parquet enables high-performance querying and analysis, as it can read only the necessary columns for a query instead of scanning entire��

Source

]]>
3
Gregory Kimball <![CDATA[Encoding and Compression Guide for Parquet String Data Using RAPIDS]]> http://www.open-lab.net/blog/?p=85090 2024-08-08T18:48:49Z 2024-07-17T16:00:00Z Parquet writers provide encoding and compression options that are turned off by default. Enabling these options may provide better lossless compression for your...]]> Parquet writers provide encoding and compression options that are turned off by default. Enabling these options may provide better lossless compression for your...

Parquet writers provide encoding and compression options that are turned off by default. Enabling these options may provide better lossless compression for your data, but understanding which options to use for your specific use case is critical to making sure they perform as intended. In this post, we explore which encoding and compression options work best for your string data.

Source

]]>
1
���˳���97caoporen����