Say goodbye to indexing challenges, data scalability issues, and high operational costs with StarTree Cloud, the leading Elasticsearch alternative for OLAP.
Elasticsearch is a great search engine, but it wasn’t designed to be an analytics database. StarTree Cloud, the ultimate Elasticsearch alternative for real-time analytics, powered by Apache Pinot is the leading platform providing unmatched performance and ease of use.
Maintain performant aggregations against petabytes of data, with latencies measured in milliseconds.
Leverage multiple indexing options, including the star-tree index for fast and efficient query results.
Get results in seconds, not minutes with tiered storage designed for fast analytics.
Unlock your data for internal & external users by supporting extremely high concurrency queries (100,000+ QPS).
Save significantly on infrastructure spend with more efficient memory & column storage.
Built for fast, real-time user-facing online analytics processing (OLAP).
StarTree Cloud, the leading Elasticsearch alternative for OLAP, was purposefully designed for real-time analytics and reporting. It is well-suited for applications that require instant insights into large datasets without needing additional customizations, configurations, or plugins for complex queries. StarTree Cloud can handle high-throughput, low-latency queries, making it ideal for a variety of use cases across social media and collaboration platforms, delivery and ridesharing services, retail and telecommunications companies, financial services, and more.
StarTree Cloud also has additional advantages over Apache Pinot, including StarTree Data Manager, which makes for easy no-code ingestion of data from event streaming systems like Apache Kafka®, live Change Data Capture (CDC) from transactional databases, as well as batch-oriented systems, object stores and a wide variety of data formats.
"We use Apache Pinot as a core system to empower mission-critical use cases."
Uber fully replaced Elasticsearch with Apache Pinot for its time-sensitive real-time analytics. By migrating, they saved more than $2M on infrastructure costs per year and reduced their overall data size by 120TB. The team also saw a 50% reduction in database cores and reduced page load time from 14 seconds to less than 5 seconds. With Apache Pinot, Uber can do real-time upserts and get query results from 1.5PB of data with <100ms P99 latencies.
"Once we saw the raw numbers we decided Elasticsearch should no longer be considered for our further analyses."
Cisco WebEx moved from Elasticsearch to Apache Pinot after seeing how it outperformed their existing infrastructure. Apache Pinot brought query times of 10-30 seconds in Elasticsearch down to sub-second speeds. Apache Pinot was also found to be 4× faster than Clickhouse in most cases in Cisco WebEx’ head-to-head benchmarking comparison.
Cisco WebEx found that Apache Pinot provided 5x to 150x lower latencies than Elasticsearch
Cisco Webex obtained subsecond latencies with Apache Pinot in most tested cases, whereas Elasticsearch timed out (>30 seconds) in 67% of cases
Cisco Webex shrank their cluster by >500 nodes moving from Elasticsearch to Apache Pinot
Uniqode (formerly Beaconstac) found that Apache Pinot provided 10x faster queries overall, and 20x lower P95 latencies than Elasticsearch
Uniqode shrank their cluster footprint by 50%, reducing cloud infrastructure spending
Uber was able to shrink the number of cores used by 50% switching to Apache Pinot vs. Elasticsearch, saving >$2 million in infrastructure costs
Compare the features of Apache Pinot to Elasticsearch and you’ll see Pinot offers far more flexible indexing and ingestion capabilities to perform real-time analytics.
Data Structure | ||
Column Store for Efficient Analytics |
||
SQL | ||
Multi-Stage Query-Time JOINs |
||
Conformance to ANSI SQL |
||
Indexing Strategies | ||
Inverted Index |
||
Sorted Index |
||
Range Index |
||
JSON Index |
||
Geospatial Index |
||
Star-Tree Index |
||
Bloom Filter |
||
Text Index |
||
Timestamp Index |
||
Sparse Index |
||
Ingestion | ||
Upserts (Full-row and partial row) |
||
Change Data Capture (CDC) |
||
Out-of-Order Handling |
||
Real-Time Deduplication |
||
Event Streaming Integration | ||
Apache Kafka |
||
Apache Kinesis |
||
Apache Pulsar |
||
Google PubSub |
||
Data Warehouse Connectors (Batch Ingestion) | ||
Snowflake |
||
Delta Lake |
||
Google Big Query |
||
Object Store Support (Batch Ingestion) | ||
Amazon S3 |
Via Amazon Lambda |
|
Google Cloud Storage (GCS) |
||
Azure Data Lake Storage (ADLS) Gen2 |
||
Hadoop Distributed File System (HDFS) |
||
Batch Ingestion File Formats | ||
Avro |
As Logstash Events |
|
CSV |
||
JSON |
||
ORC |
||
Parquet |
||
Protocol Buffers (Profobuf) |
As Logstash Events |
|
Thrift |
||
TSV |
||
Data Analytics Integration | ||
Apache Spark 3 |
||
Tiered Storage | ||
Multi-Volume Tiering |
||
Compute Node Separation |
Via Mulitple Tenants |
|
Cloud Object Storage* |
Via "Frozen Tier" |
StarTree Cloud offers tiered storage in a way that far exceeds the performance of Elasticsearch. StarTree allows you to run your application’s fastest data on locally attached NVMe storage. Or you can also use block storage for greater resiliency. Plus you can use cost-effective distributed object stores such as Amazon S3, Google Cloud Storage, or Azure Data Lake Storage. Performance on these objects stores will be far faster than Elasticsearch’s “Frozen Tier,” which will produce query results in scales measured by minutes, not seconds.
Start a free trial or meet with our experts. Discover how easy it is to get started migrating your workloads to StarTree Cloud.