Skip to content
Avatar

Highlights

  • Arctic Code Vault Contributor
  • Pro

Organizations

@apache @tokio-rs
andygrove/README.md

Hi there 👋

I'm Andy Grove, a software engineer specializing in distributed systems and query engines. I am the author of the book How Query Engines Work.

Work

I work for NVIDIA where I contribute to the RAPIDS Accelerator for Apache Spark, which is an open source plugin that provides GPU-acceleration for ETL workloads, leveraging the RAPIDS cuDF GPU DataFrame library.

Open Source

  • I am a PMC member of Apache Arrow, where I donated the initial Rust implementation and later donated the DataFusion query engine.
  • I believe that Rust and Apache Arrow together provide a great foundation for distributed systems and I created the Ballista Distributed Compute platform to demonstrate the possibilities.

Hobbies

I occasionally work on hobby projects involving digital electronics, microcontrollers, woodworking, and similar skills. Some of my projects have been featured on popular "maker" web sites.

Social Media

Pinned

  1. Apache Arrow is a cross-language development platform for in-memory data. It specifies a standardized language-independent columnar memory format for flat and hierarchical data, organized for effic…

    C++ 6.9k 1.7k

  2. Distributed compute platform implemented in Rust, using Apache Arrow memory model.

    Rust 1.5k 82

  3. Spark RAPIDS plugin - accelerate Apache Spark with GPUs

    Scala 229 65

  4. Extensible SQL Lexer and Parser for Rust

    Rust 534 107

  5. cuDF - GPU DataFrame Library

    C++ 3.6k 479

  6. Rust DataBase Connectivity (RDBC) :: Common Rust API for database drivers

    Rust 309 15

1,932 contributions in the last year

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec Jan Mon Wed Fri
Activity overview
Contributed to ballista-compute/ballista, NVIDIA/spark-rapids, apache/arrow and 5 other repositories

Contribution activity

January 2021

Created 1 repository

Created a pull request in NVIDIA/spark-rapids that received 3 comments

Add --use-decimals flag to TPC-DS ConvertFiles

This adds the ability to specify to use decimal types when converting TPC-DS data to Parquet when using spark-submit with the ConvertFiles utility,

+16 −13 3 comments
Reviewed 27 pull requests in 4 repositories
ballista-compute/ballista 12 pull requests
apache/arrow 12 pull requests
NVIDIA/spark-rapids 2 pull requests
ballista-compute/sqlparser-rs 1 pull request

Created an issue in ballista-compute/ballista that received 4 comments

Finish implementing new serde module

There is a partially implemented serde module for serializing and deserializing logical and physical plans. I could use some help in completing thi…

4 comments

Seeing something unexpected? Take a look at the GitHub profile guide.

You can’t perform that action at this time.