IPMI Sensor and Parquet Integration
Powerful performance with an easy integration, powered by Telegraf, the open source data connector built by InfluxData.
5B+
Telegraf downloads
#1
Time series database
Source: DB Engines
1B+
Downloads of InfluxDB
2,800+
Contributors
Table of Contents
Powerful Performance, Limitless Scale
Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.
See Ways to Get Started
Input and output integration overview
<p>The IPMI Sensor Plugin facilitates the collection of server health metrics directly from hardware via the IPMI protocol, querying sensor data from either local or remote systems.</p>
<p>This plugin writes metrics to parquet files, utilizing a schema based on the metrics grouped by name. It supports file rotation and buffered writing for optimal performance.</p>
Integration details
IPMI Sensor
<p>The IPMI Sensor plugin is designed to gather bare metal metrics via the command line utility <code>ipmitool</code>, which interfaces with the Intelligent Platform Management Interface (IPMI). This protocol provides management and monitoring capabilities for hardware components in server systems, allowing for the retrieval of critical system health metrics such as temperature, fan speeds, and power supply status from both local and remote servers. When configured without specified servers, the plugin defaults to querying the local machine’s sensor statistics using the <code>ipmitool sdr</code> command. In scenarios covering remote hosts, authentication is supported through username and password using the command format <code>ipmitool -I lan -H SERVER -U USERID -P PASSW0RD sdr</code>. This flexibility allows users to monitor systems effectively across various environments. The plugin also supports multiple sensor types, including chassis power status and DCMI power readings, catering to administrators needing real-time insight into server operations.</p>
Parquet
<p>The Parquet output plugin for Telegraf writes metrics to parquet files, which are columnar storage formats optimized for analytics. By default, this plugin groups metrics by their name, writing them to a single file. If a metric’s schema does not align with existing schemas, those metrics are dropped. The plugin generates an Apache Arrow schema based on all grouped metrics, ensuring that the schema reflects the union of all fields and tags. It operates in a buffered manner, meaning it temporarily holds metrics in memory before writing them to disk for efficiency. Parquet files require proper closure to ensure readability, and this is crucial when using the plugin, as improper closure can lead to unreadable files. Additionally, the plugin supports file rotation after specific time intervals, preventing overwrites of existing files and schema conflicts when a file with the same name already exists.</p>
Configuration
IPMI Sensor
Parquet
Input and output integration examples
IPMI Sensor
<ol> <li> <p><strong>Centralized Monitoring Dashboard</strong>: Utilize the IPMI Sensor plugin to gather metrics from multiple servers and compile them into a centralized monitoring dashboard. This enables real-time visibility into server health across data centers. Administrators can track metrics like temperature and power usage, helping them make data-driven decisions about resource allocation, potential failures, and maintenance schedules.</p> </li> <li> <p><strong>Automated Power Alerts</strong>: Incorporate the plugin into an alerting system that monitors chassis power status and triggers alerts when anomalies are detected. For instance, if the power status indicates a failure or if watt values exceed expected thresholds, automated notifications can be sent to operations teams, ensuring prompt attention to hardware issues.</p> </li> <li> <p><strong>Energy Consumption Analysis</strong>: Leverage the DCMI power readings collected via the plugin to analyze energy consumption patterns of hardware over time. By integrating these readings with analytics platforms, organizations can identify opportunities to reduce power usage, optimize efficiency, and potentially decrease operational costs in large server farms or cloud infrastructures.</p> </li> <li> <p><strong>Health Check Automation</strong>: Schedule regular health checks by using the IPMI Sensor Plugin to collect data from a fleet of servers. This data can be logged and compared against historical performance metrics to identify trends, outliers, or signs of impending hardware failure, allowing IT teams to take proactive measures and reduce downtime.</p> </li> </ol>
Parquet
<ol> <li> <p><strong>Data Lake Ingestion</strong>: Utilize the Parquet plugin to store metrics from various sources into a data lake. By writing metrics in parquet format, you establish a standardized and efficient way to manage time-series data, enabling faster querying capabilities and seamless integration with analytics tools like Apache Spark or AWS Athena. This setup can significantly improve data retrieval times and analysis workflows.</p> </li> <li> <p><strong>Long-term Storage of Metrics</strong>: Implement the Parquet plugin in a monitoring setup where metrics are collected over time from multiple applications. This allows for long-term storage of performance data in a compact format, making it cost-effective to store vast amounts of historical data while preserving the ability for quick retrieval and analysis later on. By archiving metrics in parquet files, organizations can maintain compliance and create detailed reports from historical performance trends.</p> </li> <li> <p><strong>Analytics and Reporting</strong>: After writing metrics to parquet files, leverage tools like Apache Arrow or PyArrow to perform complex analytical queries directly on the files without needing to load all the data into memory. This can enhance reporting capabilities, allowing teams to generate insights and visualization from large datasets efficiently, thereby improving decision-making processes based on accurate, up-to-date performance metrics.</p> </li> <li> <p><strong>Integrating with Data Warehouses</strong>: Use the Parquet plugin as part of a data integration pipeline that feeds into a modern data warehouse. By converting metrics to parquet format, the data can be easily ingested by systems like Snowflake or Google BigQuery, enabling powerful analytics and business intelligence capabilities that drive actionable insights from the collected metrics.</p> </li> </ol>
Feedback
Thank you for being part of our community! If you have any general feedback or found any bugs on these pages, we welcome and encourage your input. Please submit your feedback in the InfluxDB community Slack.
Powerful Performance, Limitless Scale
Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.
See Ways to Get Started
Related Integrations
Related Integrations
HTTP and InfluxDB Integration
The HTTP plugin collects metrics from one or more HTTP(S) endpoints. It supports various authentication methods and configuration options for data formats.
View IntegrationKafka and InfluxDB Integration
This plugin reads messages from Kafka and allows the creation of metrics based on those messages. It supports various configurations including different Kafka settings and message processing options.
View IntegrationKinesis and InfluxDB Integration
The Kinesis plugin allows for reading metrics from AWS Kinesis streams. It supports multiple input data formats and offers checkpointing features with DynamoDB for reliable message processing.
View Integration