DealForge autonomously sources, scores, and writes investment memos on venture deals. Stop manually hunting.

1,180+ deals tracked  ·  22 AI investment memos  ·  Updated daily

← Back to leaderboard

500k+ events/sec transformations for ClickHouse ingestion

Show HN: 500k+ events/sec transformations for ClickHouse ingestion

62 AI Score
Show_hn devtools Added Apr 8, 2026

Details

Sector
devtools
Total Funding
$0
Last Round
$0

About

Hi HN! We are Ashish and Armend, founders of GlassFlow.<p>Over the last year, we worked with teams running high-throughput pipelines into self-hosted ClickHouse. Mostly for observability and real-time analytics.<p>A question that came repeatedly was: What happens when throughput grows?<p>Usually, things work fine at 10k events&#x2F;sec, but we started seeing backpressure and errors at &gt;100k.<p>When the throughput per pipeline stops scaling, then adding more CPU&#x2F;memory doesn’t help because often parts of the pipeline are not parallelized or are bottlenecked by state handling.<p>At this point, engineers usually scale by adding more pipeline instances.<p>That works but comes with some trade-offs: - You have to split the workload (e.g., multiple pipelines reading from the same source) - Transformation logic gets duplicated across pipelines - Stateful logic becomes harder to manage and keep consistent - Debugging and changes get more difficult because the data flow is fragmented<p>Another challenge arises when working with high-cardinality keys like user IDs, session IDs, or request IDs, and when you need to handle longer time windows (24h or more). The state grows quickly and many systems rely on in-memory state, which makes it expensive and harder to recover from failures.<p>We wanted to solve this problem and rebuild our approach at GlassFlow.<p>Instead of scaling by adding more pipelines, we scale within a single pipeline by using replicas. Each replica consumes, processes, and writes independently, and the workload is distributed across them.<p>In the benchmarks we’re sharing, this scales to 500k+ events&#x2F;sec while still running stateful transformations and writing into ClickHouse.<p>A few things we think are interesting: - Scaling is close to linear as you add replicas - Works with stateful transformations (not just stateless ingestion) - State is backed by a file-based KV store instead of relying purely on memory - The ClickHouse sink is optimized for batching to avoid small inserts - The product is built with Go<p>Full write-up + benchmarks: <a href="https:&#x2F;&#x2F;www.glassflow.dev&#x2F;blog&#x2F;glassflow-now-scales-to-500k-events-per-sec" rel="nofollow">https:&#x2F;&#x2F;www.glassflow.dev&#x2F;blog&#x2F;glassflow-now-scales-to-500k-...</a><p>Repo: <a href="https:&#x2F;&#x2F;github.com&#x2F;glassflow&#x2F;clickhouse-etl" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;glassflow&#x2F;clickhouse-etl</a><p>Happy to answer questions about the design or trade-offs.

AI Score Reasoning

GlassFlow addresses a high-value technical bottleneck in the rapidly growing ClickHouse ecosystem, specifically regarding stateful scaling. While the product shows strong technical depth and addresses a clear pain point for high-throughput data engineering, it remains in an early 'Show HN' stage with significant competition from established stream processing incumbents.

Source

Show_hn — View original →