-
Notifications
You must be signed in to change notification settings - Fork 3.2k
Description
Component(s)
exporter/awss3
Is your feature request related to a problem? Please describe.
awss3exporter supports JSON and OTLP-proto formats, which are functional for telemetry storage but inefficient for large-scale analytic tools. Folks who might want to analyze data with tools like Athena, Spark, or Glue are forced to manage this conversion out of band (separate process) or deal with less efficient query times.
Describe the solution you'd like
Add Parquet encoding support to awss3exporter, allowing data to be written in a columnar format with optional Hive-style partitioning.
Something like:
exporters:
awss3:
s3uploader:
region: 'eu-central-1'
s3_bucket: 'databucket'
s3_prefix: 'metric'
s3_partition_format: '%Y/%m/%d/%H/%M'
encoding: parquetI'd imagine we'd define separate Parquet schemas per signal type.
Describe alternatives you've considered
Looked at other similar requests (resulting in parquet files in S3) such as #44124
Export to file as JSON and convert to Parquet with tools like duckdb/custom scripts and upload
- Requires separate service to manage data pipeline
- Requires something to manage truncation of otel written json file
Additional context
No response
Tip
React with 👍 to help prioritize this issue. Please use comments to provide useful context, avoiding +1 or me too, to help us triage it. Learn more here.