Import Data

This section is an overview of the various options for importing data into Pinot.

There are multiple options for importing data into Pinot. These guides are ready-made examples that show you step-by-step instructions for importing records into Pinot, supported by our plugin architecture.

These guides are meant to get you up and running with imported data as quick as possible. Pinot supports multiple file input formats without needing to change anything other than the file name. Each example imports a ready-made dataset so you can see how things work without needing to bring your own dataset.

Pinot Batch Ingestion

SparkHadoop

Pinot Stream Ingestion

This guide will show you how to import data using stream ingestion from Apache Kafka topics.

Apache Kafka

This guide will show you how to import data using stream ingestion with upsert.

Stream Ingestion with Upsert

This guide will show you how to import data using stream ingestion with deduplication.

Stream Ingestion with Dedup

Pinot File Systems

By default, Pinot does not come with a storage layer, so all the data sent, won't be stored in case of system crash. In order to persistently store the generated segments, you will need to change controller and server configs to add a deep storage. Checkout File systems for all the info and related configs.

These guides will show you how to import data as well as persist it in the file systems.

Amazon S3Azure Data Lake StorageGoogle Cloud StorageHDFS

Pinot Input Formats

These guides will show you how to import data from a Pinot supported input format.

Input formats

This guide will show you how to handle the complex type in the ingested data, such as map and array.

Complex Type (Array, Map) Handling

Last updated