Import Data
This section is an overview of the various options for importing data into Pinot.
There are multiple options for importing data into Pinot. These guides are ready-made examples that show you step-by-step instructions for importing records into Pinot, supported by our plugin architecture.
These guides are meant to get you up and running with imported data as quick as possible. Pinot supports multiple file input formats without needing to change anything other than the file name. Each example imports a ready-made dataset so you can see how things work without needing to bring your own dataset.
Pinot Batch Ingestion
SparkHadoopPinot Stream Ingestion
This guide will show you how to import data using stream ingestion from Apache Kafka topics.
Apache KafkaThis guide will show you how to import data using stream ingestion with upsert.
Stream Ingestion with UpsertThis guide will show you how to import data using stream ingestion with deduplication.
Stream Ingestion with DedupPinot File Systems
By default, Pinot does not come with a storage layer, so all the data sent, won't be stored in case of system crash. In order to persistently store the generated segments, you will need to change controller and server configs to add a deep storage. Checkout File systems for all the info and related configs.
These guides will show you how to import data as well as persist it in the file systems.
Amazon S3Azure Data Lake StorageGoogle Cloud StorageHDFSPinot Input Formats
These guides will show you how to import data from a Pinot supported input format.
Input formatsThis guide will show you how to handle the complex type in the ingested data, such as map and array.
Complex Type (Array, Map) HandlingLast updated