<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: vasa_</title><link>https://news.ycombinator.com/user?id=vasa_</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sat, 02 May 2026 20:46:03 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=vasa_" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[Show HN: I got Claude Code to run in Binary]]></title><description><![CDATA[
<p>Successfully restructured the entire leaked source code of Anthropic's Claude Code using OAI's Codex, rewritten in Binary without copyright issues . You won't get it more performant than this</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47616741">https://news.ycombinator.com/item?id=47616741</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 02 Apr 2026 16:37:11 +0000</pubDate><link>https://github.com/topoteretes/cognee-claude-code-binary</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=47616741</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47616741</guid></item><item><title><![CDATA[New comment by vasa_ in "Show HN: PageIndex – Vectorless RAG"]]></title><description><![CDATA[
<p>We've done this for a while with cognee, where we have graph completition retrieval that does that + many other things like weighting, self improving feedback and more
<a href="https://github.com/topoteretes/cognee" rel="nofollow">https://github.com/topoteretes/cognee</a></p>
]]></description><pubDate>Sat, 30 Aug 2025 16:33:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=45075936</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=45075936</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45075936</guid></item><item><title><![CDATA[New comment by vasa_ in "We need better ways to evaluate how AI memory systems perform"]]></title><description><![CDATA[
<p>The usual benchmarks for language models—Exact Match, F1, and even multi-hop QA datasets—weren’t designed to measure what matters most about persistent AI memory: connecting concepts across time, documents, and contexts.<p>We just completed our most extensive internal evaluation of cognee to date, using HotPotQA as a baseline. While the results showed strong gains, they also reinforced a growing realization: we need better ways to evaluate how AI memory systems actually perform.<p>We ran Cognee through 45 evaluation cycles on 24 questions from HotPotQA, using ChatGPT 4o for the analysis. Each part of the evaluation process is affected by the inherent variance in GPT’s output: cognification, answer generation, and answer evaluation. We especially noticed significant variance across different metrics on small runs, which is why we chose the repeated, end-to-end approach.<p>We compared results using the same questions and setup with:<p>Mem0
Lightrag
Graphiti<p>While they are standard in QA, EM and F1 scores reward surface-level overlap and miss the core value proposition of AI memory systems. For example, a syntactically perfect answer can be factually wrong, and a fuzzy-but-correct response can be penalized for missing the reference phrasing.<p>LLMs are inconsistent, that is another issue.<p>Even HotPotQA assumes all relevant information sits neatly in two paragraphs. That’s not how memory works. Real-world AI memory systems need to link information across documents, conversations, and knowledge domains that traditional QA benchmarks just can’t capture.<p>Consider the difference:<p>Traditional QA:<p>“What year was the company that acquired X founded?”<p>Memory Challenge:<p>“How do the concerns raised in last month’s security review relate to the authentication changes discussed in the architecture meeting three weeks ago?”<p>Only one of these tests long-term knowledge, reasoning across sources, and organizational memory—care to guess which one?<p>We are working on a new dataset and benchmarks to measure memory, and would love feedback!</p>
]]></description><pubDate>Fri, 08 Aug 2025 18:09:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=44839900</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=44839900</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44839900</guid></item><item><title><![CDATA[We need better ways to evaluate how AI memory systems perform]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.cognee.ai/blog/deep-dives/ai-memory-evals-0825">https://www.cognee.ai/blog/deep-dives/ai-memory-evals-0825</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=44839899">https://news.ycombinator.com/item?id=44839899</a></p>
<p>Points: 1</p>
<p># Comments: 1</p>
]]></description><pubDate>Fri, 08 Aug 2025 18:09:08 +0000</pubDate><link>https://www.cognee.ai/blog/deep-dives/ai-memory-evals-0825</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=44839899</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44839899</guid></item><item><title><![CDATA[New comment by vasa_ in "Ask HN: Who is hiring? (July 2025)"]]></title><description><![CDATA[
<p>cognee - memory tool for AI apps and Agents | Remote and Berlin | Full time | 100k EUR<p>At cognee we are working on building graph/vector based memory on top of vector and graph stores. Our pipelines had 116k runs last month and we are projecting north of a million in a few months. We need help on the infra side.<p>Check our OSS tool here: <a href="https://github.com/topoteretes/cognee">https://github.com/topoteretes/cognee</a><p>Open Roles: Platform Engineer - <a href="https://apply.workable.com/topoteretes-ug-haftungsbeschrankt/j/1D48B939E2/" rel="nofollow">https://apply.workable.com/topoteretes-ug-haftungsbeschrankt...</a></p>
]]></description><pubDate>Mon, 07 Jul 2025 16:58:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=44492301</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=44492301</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44492301</guid></item><item><title><![CDATA[New comment by vasa_ in "Show HN: Cognee – Open-Source AI Memory Layer That Remembers Context"]]></title><description><![CDATA[
<p>Hi, founder of cognee here<p>We have temporal resolution mechanisms we are building, and framework is generalizable enough to build any custom logic. We have a few ideas and some things will be posted there soon.<p>cognee has notion of nodesets which work similar to tags: <a href="https://docs.cognee.ai/core-concepts/node-sets" rel="nofollow">https://docs.cognee.ai/core-concepts/node-sets</a><p>And also we have graphs per user now available. So, user permisions + graph filtering</p>
]]></description><pubDate>Tue, 03 Jun 2025 17:21:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=44172358</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=44172358</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44172358</guid></item><item><title><![CDATA[Show HN: Cognee – Open-Source AI Memory Layer That Remembers Context]]></title><description><![CDATA[
<p>Hey HN! We're Vasilije, Laszlo and Lazar, the authors of a new paper and part of <a href="https://www.cognee.ai" rel="nofollow">https://www.cognee.ai</a>. cognee let’s you build memory layers for AI applications and agents, allowing them to personalize results, connect various data sources, and add custom rules. This enables AI apps to deliver increasingly accurate responses, we reached almost 90% on standard industry benchmarks as you can see here <a href="https://github.com/topoteretes/cognee/tree/main/evals">https://github.com/topoteretes/cognee/tree/main/evals</a> and our paper can be accessed at: <a href="https://arxiv.org/abs/2505.24478" rel="nofollow">https://arxiv.org/abs/2505.24478</a> and collab is in the repository<p>LLMs don't remember context well—they can’t ingest your data and keep it in memory. This limitation leads to lacking interactions, a lack of accuracy, and the inability to connect your data sources cheaply because developers must include long, unmanaged context in every prompt.<p>When we were building RAGs we saw that we can’t find the data we need and that there are too many knobs to turn in RAG frameworks. We had to tweak many parameters and also could not specify the rules we wanted the data to follow. No ontologies, no rulesets, no state or good data engineering practices and a lot of manual work. That’s why cognee<p>cognee builds memory that combines graph, vector, and relational stores. Here is how it works:<p>Adding data: When you use cognee with your AI App, it can take in any message, string, S3 bucket or even a relational database and automatically ingest it<p>Managing information: cognee sorts this information into semantic graph: - It extracts entities and connections between things. - It embeds the data in the vector store, It enriches data with custom ontologies you provide, that help ground the graphs and make them more reliable. - The overall information is stored in many layers of a graph and vector store that allows for finding similar information later using a variety of types of searches.<p>Retrieving data: When given an input query, cognee searches for and retrieves related stored information by leveraging a combination of graph traversal techniques, vector similarity, and COT techniques. It can use the internal benchmarking system to make sure that your pipelines are returning only accurate data when you need it.<p>---<p>cognee introduces a self-improving group of memory layers that cover various topics, data sources and can be customized. This reduces the need to build everything from scratch, and you can use our primitives to get started and move more quickly. On the other hand, you can do everything yourself, from start to end. We’ve designed the system to be modular and extensible.<p>We’ve open-sourced cognee —specifically the framework and various vector and graph database adapters, as well as our default data pipeline, cognify—under the Apache 2.0 license. This includes the ability to add, cognify, and retrieve data within your AI applications and also extend it with custom components that are just pure Python.<p>However, many keep features that are optimized for production use, as a part of our paid platform. We release these functionalities too, including in the last few weeks permission management + distributed pipelines(dev). These are a part of our open-source package and are available to those who in production environments need things like rate limiting and credential management. We will release a paid offering for developers to get easy API access to our platform.<p>Our automation tooling allows us to optimize the pipelines and find the best combination of parameters that answer questions our stakeholders have.<p>We’d love to hear what you think! Please feel free to try our demo, check out the code, read the research paper, and share thoughts or suggestions with us. Your feedback will help shape where we take cognee from here!</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=44169594">https://news.ycombinator.com/item?id=44169594</a></p>
<p>Points: 9</p>
<p># Comments: 2</p>
]]></description><pubDate>Tue, 03 Jun 2025 13:05:15 +0000</pubDate><link>https://github.com/topoteretes/cognee</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=44169594</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44169594</guid></item><item><title><![CDATA[Show HN: Cognee – Turn RAG and GraphRAG into custom dynamic semantic memory]]></title><description><![CDATA[
<p>Hey there HN! We’re Vasilije, Boris, and Laszlo, and we’re excited to introduce cognee, an open-source Python library that approaches building evolving semantic memory using knowledge graphs + data pipelines<p>Before we built cognee, Vasilije(B Economics and Clinical Psychology) worked at a few unicorns (Omio, Zalando, Taxfix), while Boris managed large-scale applications in production at Pera and StuDocu. Laszlo joined after getting his PhD in Graph Theory at the University of Szeged.<p>Using LLMs to connect to large datasets (RAG) has been popularized and has shown great promise. Unfortunately, this approach doesn’t live up to the hype.<p>Let’s assume we want to load a large repository from GitHub to a vector store.
Connectingfiles in larger systems with RAG would fail because a fixed RAG limit is too constraining in longer dependency chains. While we need results that are aware of the context of the whole repository, RAG’s similarity-based retrieval does not capture the full context of interdependent files spread across the repository.<p>This approach allows cognee to retrieve all relevant and correct context at inference time. For example, if `function A` in one file calls `function B` in another file, which calls `function C` in a third file, all code and summaries that further explain their position and purpose in that chain are served as context. As a result, the system has complete visibility into how different code parts work together within the repo.<p>Last year, Microsoft took a leap published GraphRAG - i.e. RAG with Knowledge Graphs. We think it is the right direction.
Our initial ideas were similar to this paper and this got some attention on Twitter (<a href="https://x.com/tricalt/status/1722216426709365024" rel="nofollow">https://x.com/tricalt/status/1722216426709365024</a>)<p>Over time we understood we needed tooling to create dynamically evolving groups of graphs, cross-connected and evaluated together.
Our tool is named after a process called cognification. We prefer the definition that Vakalo (1978) uses to explain that cognify represents "building a fitting (mental) picture"<p>We believe that agents of tomorrow will require a correct dynamic “mental picture” or context to operate in a rapidly evolving landscape.<p>To address this, we built ECL pipelines, where we do the following:
- Extract data from various sources using dlt and existing frameworks
- Cognify - create a graph/vector representation of the data
- Load - store the data in the vector (in this case our partner FalkorDB), graph, and relational stores<p>We can also continuously feed the graph with new information, and when testing this approach we found that on HotpotQA, with human labeling, we achieved 87% answer accuracy (<a href="https://docs.cognee.ai/evaluations" rel="nofollow">https://docs.cognee.ai/evaluations</a>).<p>To show how the approach works we did an integration with continue.dev and built a codegraph<p>Here is how codegraph was implemented: 
We're explicitly including repository structure details and integrating custom dependency graph versions. Think of it as a more insightful way to understand your codebase's architecture.
By transforming dependency graphs into knowledge graphs, we're creating a quick, graph-based version of tools like tree-sitter. This means faster and more accurate code analysis.
We worked on modeling causal relationships within code and enriching them with LLMs. This helps you understand how different parts of your code influence each other.
We created graph skeletons in memory which allows us to perform various operations on graphs and power custom retrievers.<p>If you want to integrate cognee into your systems or have a look at codegraph, our GitHub repository is (<a href="https://github.com/topoteretes/cognee">https://github.com/topoteretes/cognee</a>)<p>Thank you for reading! We’re definitely early and welcome your ideas and experiences as it relates to agents, graphs, evals, and human+LLM memory.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43031915">https://news.ycombinator.com/item?id=43031915</a></p>
<p>Points: 6</p>
<p># Comments: 1</p>
]]></description><pubDate>Thu, 13 Feb 2025 01:59:26 +0000</pubDate><link>https://github.com/topoteretes/cognee</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=43031915</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43031915</guid></item><item><title><![CDATA[New comment by vasa_ in "Show HN: FastGraphRAG – Better RAG using good old PageRank"]]></title><description><![CDATA[
<p>Neat, we are doing something similar with cognee, but are letting users define graph ingestion, generation, and retrieval themselves instead of making assumptions: <a href="https://github.com/topoteretes/cognee">https://github.com/topoteretes/cognee</a></p>
]]></description><pubDate>Mon, 18 Nov 2024 19:21:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=42175895</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=42175895</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42175895</guid></item><item><title><![CDATA[Show HN: Auto-optimizing deterministic LLM outputs using knowledge graphs]]></title><description><![CDATA[
<p>Hi,
We are building an open-source framework for loading and structuring LLM context to create accurate and explainable LLM answers using knowledge graphs and vector stores.<p>We built the tool with four main concepts in mind:<p>1. Loader -> uses dlt in the backend to load and structure the data<p>2. Cognify step -> creates a graph with summaries, labels and factoids that are interconnected across the documents and stored as a representation in the vector store<p>3. Optimizer -> Uses DSPy to optimize LLM queries, and we plan to extend it to most of the knobs we can turn, like chunking etc.<p>4. Search -> allows for searching using search types supported in graph stores (ex. Neo4j) or hybrid, BM25, or other search types available in vector stores.<p>We are quite early with the product but we would love to hear feedback on what we can improve.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=40117217">https://news.ycombinator.com/item?id=40117217</a></p>
<p>Points: 7</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 22 Apr 2024 18:25:59 +0000</pubDate><link>https://github.com/topoteretes/cognee</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=40117217</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40117217</guid></item><item><title><![CDATA[New comment by vasa_ in "Show HN: Dlt – Python library to automate the creation of datasets"]]></title><description><![CDATA[
<p>I worked with dlt guys on exactly that. Using OpenAI functions to generate a schema for the data based on the raw data structure. You can check that work here: <a href="https://github.com/topoteretes/PromethAI-Memory">https://github.com/topoteretes/PromethAI-Memory</a>
It's in the level 1 folder</p>
]]></description><pubDate>Wed, 01 Nov 2023 17:40:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=38101882</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=38101882</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38101882</guid></item><item><title><![CDATA[Show HN: PromethAI – Personalized AI assistant app]]></title><description><![CDATA[
<p>Hi! We created UI/UX flow on top of OpenAI API + Langchain and gave connected agents access to tools. The outcome is a mobile app that can decompose your thoughts, remember them, help with ideas and automate actions via Zapier integration. Would love to hear the thoughts and advice on how to make it better! Repos: <a href="https://github.com/topoteretes/">https://github.com/topoteretes/</a></p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=36480598">https://news.ycombinator.com/item?id=36480598</a></p>
<p>Points: 3</p>
<p># Comments: 1</p>
]]></description><pubDate>Mon, 26 Jun 2023 15:08:01 +0000</pubDate><link>https://www.prometh.ai</link><dc:creator>vasa_</dc:creator><comments>https://news.ycombinator.com/item?id=36480598</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36480598</guid></item></channel></rss>