The moment you hit “reply” on Hacker News, your comment joins a dataset that’s not just readable by humans, it’s systematically encoded into 384-dimensional mathematical space, indexed for instant semantic retrieval, and packaged for commercial deployment. ClickHouse’s release of 28.74 million Hacker News comments as vector embeddings represents both a technical breakthrough and a philosophical turning point for online communities.
This dataset isn’t just another benchmark, it’s a complete archive of community knowledge transformed into searchable numerical representations. The 55GB Parquet file contains every story, comment, and poll option from HN’s history, each converted into vector embeddings using the SentenceTransformers all-MiniLM-L6-v2 model. What makes this genuinely interesting isn’t the raw capability, but what happens when you pair massive-scale semantic search with the collective intelligence of one of the internet’s most opinionated technical communities.
The Technical Reality: Building Vector Search at Internet Scale
ClickHouse’s implementation demonstrates what “scale” actually means in vector database terms. The table structure reveals the complexity beneath the surface:
CREATE TABLE hackernews
(
`id` Int32,
`doc_id` Int32,
`text` String,
`vector` Array(Float32),
`node_info` Tuple(start Nullable(UInt64), end Nullable(UInt64)),
`metadata` String,
`type` Enum8('story' = 1, 'comment' = 2, 'poll' = 3, 'pollopt' = 4, 'job' = 5),
`by` LowCardinality(String),
`time` DateTime,
`title` String,
`post_score` Int32,
`dead` UInt8,
`deleted` UInt8,
`length` UInt32
)
ENGINE = MergeTree
ORDER BY id;
The indexing strategy uses cosine similarity with HNSW (hierarchical navigable small world) parameters optimized for 28+ million vectors:
ALTER TABLE hackernews ADD INDEX vector_index vector
TYPE vector_similarity('hnsw', 'cosineDistance', 384, 'bf16', 64, 512);
This isn’t academic, building this index “could take a few minutes/hour for the full 28.74 million dataset”, demonstrating the real-world infrastructure demands of industrial-scale semantic search. The dataset becomes immediately usable for applications like semantic Q&A systems, where you can find “similar sentences” across decades of technical discussion or build summarization tools that distill collective wisdom on specific topics.
The Embedding Elephant in the Room
One of the first technical debates sparked by this release concerns the choice of embedding model. The all-MiniLM-L6-v2 model used here generates 384-dimensional vectors with a 512-token context window, essentially the “default” choice in many tutorials. But as critics quickly pointed out in Hacker News discussions, this model is showing its age.
Developer sentiment suggests that all-MiniLM-L6-v2 represents “the open-weights embedding model used in all the tutorials” from vector search’s infancy, but newer alternatives offer significant improvements. Models like Google’s EmbeddingGemma-300M (with 2k context), bge-base-en-v1.5, and nomic-embed-text-v1.5 now provide better performance and context handling.

The pragmatic choice for ClickHouse makes sense for a demonstration dataset, the smaller model size (70MB vs. 300-700MB for newer models) makes client-side processing feasible and reduces computational overhead. But it highlights the rapid evolution in embedding technology where yesterday’s standard becomes today’s legacy.
The Legal Gray Area That Nobody Talks About
Buried in the technical enthusiasm lies a simmering ethical debate. As one commenter noted, Y Combinator’s terms explicitly state: “Commercial Use: Unless otherwise expressly authorized herein or in the Site, you agree not to display, distribute, license, perform, publish, reproduce, duplicate, copy, create derivative works from, modify, sell, resell, exploit, transfer or upload for any commercial purposes, any portion of the Site.”
Yet here we have a commercial database company distributing a derivative work, vector embeddings of HN comments, as a product demonstration. The dataset itself may not directly monetize HN content, but it certainly serves ClickHouse’s commercial interests by showcasing their vector capabilities.
The user reactions reveal the cognitive dissonance many feel: “I know it’s unrelated but does anyone knows a good paper comparing vector searches vs ‘normal’ full text search?” one commenter asked, neatly sidestepping the elephant in the room that their own contributions are now training data.
Beyond Search: The Summarization Use Case
Where this gets genuinely useful is in applications like the provided summarization demo. The Python script demonstrates a practical RAG (Retrieval-Augmented Generation) workflow:
- Generate embeddings for search queries using the same
all-MiniLM-L6-v2model - Retrieve semantically similar HN discussions via ClickHouse vector search
- Feed context to OpenAI’s GPT-3.5-turbo for intelligent summarization
When searching for “ClickHouse performance experiences”, the system returned comprehensive summaries comparing ClickHouse against TimescaleDB, Apache Spark, AWS Redshift, and QuestDB, knowledge distilled from years of community discussion that no single person could synthesize manually.
This demonstrates the real power: turning collective intelligence into actionable insights. But it also raises questions about attribution and compensation. As one commenter noted, “Without our contributions, readership, and comments, their ability to hire and recruit founders is diminished.”
The Infrastructure Scaling Challenge
Building semantic search at this scale involves significant engineering tradeoffs. The ClickHouse documentation notes that the first load of the vector index into memory “could take a few seconds/minutes”, a non-trivial consideration for production systems handling real-time queries.
For comparison, independent projects like hn.fiodorov.es demonstrate smaller-scale implementations using different embedding models and PostgreSQL with pgvector. The creator reported that “daily updates I do on my m4 mac air: takes about 5 minutes to process roughly 10k fresh comments”, highlighting the computational cost of maintaining freshness in these systems.
Where This Leaves Us: Technical Excellence vs. Community Rights
The ClickHouse dataset represents a technical achievement, demonstrating industrial-scale vector search capabilities with real-world data. But it also embodies the central tension of our AI moment: the gap between technological possibility and community consent.
As one commenter articulated the violation felt: “When I first joined it was unconceivable that someone could just take everything and build a trivially queryable conversational model around everything I’ve posted just like that.” The feeling that “this might affect my involvement here going forward” echoes broader concerns about data sovereignty in the age of AI.
The dataset reveals two competing realities: vector search can unlock incredible value from collective knowledge, but it does so by treating community contributions as raw material for commercial applications.
The Uncomfortable Truth About Scale
What makes this dataset compelling isn’t just its size, it’s what it represents about the future of knowledge management. We’re transitioning from document-based retrieval to concept-based discovery, where “find me discussions about OLAP cubes” returns semantically similar content regardless of keyword matching.
But as these systems scale, they create new challenges. The 28 million comments in this dataset represent years of community investment, technical expertise, thoughtful debate, personal experiences, now reduced to numerical representations serving corporate interests.
The technical implementation is impressive. The ethical implications? Those remain very much unsettled. As vector databases continue to evolve, we’re forced to confront whether the communities generating the data should have more say in how it’s used, or whether technical capability alone justifies commercial exploitation.
In the end, the ClickHouse Hacker News dataset serves as both a technical tutorial and a cautionary tale: our collective intelligence has never been more valuable, and we’ve never had less control over how it gets used. The vectors may be numerical, but the implications are profoundly human.




