CREATE PIPELINE

Effortlessly set up streaming ingest feeds from Apache Kafka, Amazon S3,
and HDFS using a single CREATE PIPELINE command
extract
Extract
Pull data directly from Apache Kafka, Amazon S3, Azure Blob, or HDFS with no additional middleware required
transform
Transform
Map and enrich data with user defined or Apache Spark transformations for real-time scoring, cleaning and de-duplication
load
Load
Guarantee message delivery and eliminate duplicate or incomplete stream data for accurate reporting and analysis

Optimized for Streaming

rapid-parallel
Rapid Parallel Loading
Load multiple data feeds into a single database using scalable parallel ingestion
joins
Live De-Duplication
Eliminate duplicate records at the time of ingestion for real-time data cleansing
simplified-architecture
Simplified Architecture
Reduce or eliminate costly middleware tools and processing with direct ingest from message brokers
edit-pencil
Build Your Own
Add custom connectivity using an extensible plug-in framework
exactlyonce
Exactly Once Semantics
Ensure accurate delivery of every message for reporting and analysis of enterprise critical data
monitor
Built-in Management
Connect, add transformations and monitor performance using intuitive web UI

Ready to get started?

Experience the performance of The Real-Time Distributed SQL Database for your data today

Integrated Architecture

Efficiently load data into database tables using parallel ingestion from individual Apache Kafka or Amazon S3 brokers.
Pipelines Architecture Diagram