Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Hey, Ofek from Datadog here!

I recently implemented our ClickHouse integration [1], so if any of you would like to try it out we would appreciate feedback. I really enjoyed learning about this database, and it has excellent docs :)

Oh fun fact, speaking of docs, this was the first integration of ours that we scrape docs for as part of the test suite. So when a new built-in metric is added it will fail our CI until we support it [2]. We just did this again for Apache Airflow [3].

[1]: https://github.com/DataDog/integrations-core/pull/4957

[2]: https://github.com/DataDog/integrations-core/pull/5233

[3]: https://github.com/DataDog/integrations-core/pull/5311



So at datadog, after aggregation with spark and storage into parquet, what is used for serving queries of all the datadog aggregated telemetry data (logs, apm and infra telemetry) to the consumers?

(interestingly, we have a nearly identical data ingestion/ETL stack running on spot instances and saving to parquet/s3)


Hi there! We actually talked quite a bit about that in a recent podcast https://www.dataengineeringpodcast.com/datadog-timeseries-da...




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: