Ceph and Apache Hudi Integration
Powerful performance with an easy integration, powered by Telegraf, the open source data connector built by InfluxData.
5B+
Telegraf downloads
#1
Time series database
Source: DB Engines
1B+
Downloads of InfluxDB
2,800+
Contributors
Table of Contents
Powerful Performance, Limitless Scale
Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.
See Ways to Get Started
Input and output integration overview
<p>The Ceph plugin for Telegraf helps in gathering performance metrics from both MON and OSD nodes in a Ceph storage cluster for effective monitoring and management.</p>
<p>Writes metrics to Parquet files via Telegraf’s Parquet output plugin, preparing them for ingestion into Apache Hudi’s lakehouse architecture.</p>
Integration details
Ceph
<p>The Ceph Storage Telegraf plugin is designed to collect performance metrics from Monitor (MON) and Object Storage Daemon (OSD) nodes within a Ceph storage cluster. Ceph, a highly scalable storage system, integrates its metrics collection through this plugin, facilitating easy monitoring of its components. With the introduction of this plugin in the 13.x Mimic release, users can effectively gather detailed insights into the performance and health of their Ceph infrastructure. It functions by scanning configured socket directories for specific Ceph service socket files, executing commands via the Ceph administrative interface, and parsing the returned JSON data for metrics. The metrics are organized based on top-level keys, allowing for efficient monitoring and analysis of cluster performance. This plugin provides valuable capabilities for managing and maintaining the performance of a Ceph cluster by allowing administrators to understand system behavior and identify potential issues proactively.</p>
Apache Hudi
<p>This configuration leverages Telegraf’s Parquet plugin to serialize metrics into columnar Parquet files suitable for downstream ingestion by Apache Hudi. The plugin writes metrics grouped by metric name into files in a specified directory, buffering writes for efficiency and optionally rotating files on timers. It considers schema compatibility—metrics with incompatible schemas are dropped—ensuring consistency. Apache Hudi can then consume these Parquet files via tools like DeltaStreamer or Spark jobs, enabling transactional ingestion, time-travel queries, and upserts on your time series data.</p>
Configuration
Ceph
Apache Hudi
Input and output integration examples
Ceph
<ol> <li> <p><strong>Dynamic Monitoring Dashboard</strong>: Utilize the Ceph plugin to create a real-time monitoring dashboard that visually represents the performance metrics of your Ceph cluster. By integrating these metrics into a centralized dashboard, system administrators can gain immediate insights into the health of the storage infrastructure, which aids in quickly identifying and addressing potential issues before they escalate.</p> </li> <li> <p><strong>Automated Alerting System</strong>: Implement the Ceph plugin in conjunction with an alerting solution to automatically notify administrators of performance degradation or operational issues within the Ceph cluster. By defining thresholds for key metrics, organizations can ensure prompt response actions, thereby improving overall system reliability and performance.</p> </li> <li> <p><strong>Performance Benchmarking</strong>: Use the metrics collected by this plugin to conduct performance benchmarking tests across different configurations or hardware setups of your Ceph storage cluster. This process can assist organizations in identifying optimal configurations that enhance performance and resource utilization, promoting a more efficient storage environment.</p> </li> <li> <p><strong>Capacity Planning and Forecasting</strong>: Integrate the metrics gathered from the Ceph storage plugin into broader data analytics and reporting tools to facilitate capacity planning. By analyzing historical metrics, organizations can forecast future utilization trends, enabling informed decisions about scaling storage resources effectively.</p> </li> </ol>
Apache Hudi
<ol> <li> <p><strong>Transactional Lakehouse Metrics</strong>: Buffer and write Web service metrics as Parquet files for DeltaStreamer to ingest into Hudi, enabling upserts, ACID compliance, and time-travel on historical performance data.</p> </li> <li> <p><strong>Edge Device Batch Analytics</strong>: Telegraf running on IoT gateways writes metrics to Parquet locally, where periodic Spark jobs ingest them into Hudi for long-term analytics and traceability.</p> </li> <li> <p><strong>Schema-Enforced Abnormal Metric Handling</strong>: Use Parquet plugin’s strict schema-dropping behavior to prevent malformed or unexpected metric changes. Hudi ingestion then guarantees consistent schema and data quality in downstream datasets.</p> </li> <li> <p><strong>Data Platform Integration</strong>: Store Telegraf metrics as Parquet files in an S3/ADLS landing zone. Hudi’s Spark-based ingestion pipeline then loads them into a unified, queryable lakehouse with business events and logs.</p> </li> </ol>
Feedback
Thank you for being part of our community! If you have any general feedback or found any bugs on these pages, we welcome and encourage your input. Please submit your feedback in the InfluxDB community Slack.
Powerful Performance, Limitless Scale
Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.
See Ways to Get Started
Related Integrations
Related Integrations
HTTP and InfluxDB Integration
The HTTP plugin collects metrics from one or more HTTP(S) endpoints. It supports various authentication methods and configuration options for data formats.
View IntegrationKafka and InfluxDB Integration
This plugin reads messages from Kafka and allows the creation of metrics based on those messages. It supports various configurations including different Kafka settings and message processing options.
View IntegrationKinesis and InfluxDB Integration
The Kinesis plugin allows for reading metrics from AWS Kinesis streams. It supports multiple input data formats and offers checkpointing features with DynamoDB for reliable message processing.
View Integration