Ask HN: How to measure how much data one can effectively process or understand?
Is there a scale of how much data one can effectively process, something similar to the "Kardashev scale for data"? What would be a name for such a thing? During Memgraph's Community Call (https://youtu.be/ygr8yvIouZk?t=1307), the point is that AgenticRuntimes + GraphRAG moves you up on the "Kardashev scale for data" because you suddenly can get much more insight from any dataset, and everyone can use it (a large corporation does not control it). I found something similar under https://ift.tt/eMVqf8v, but the definition/example looks very narrow. 1 comments on Hacker News.
Is there a scale of how much data one can effectively process, something similar to the "Kardashev scale for data"? What would be a name for such a thing? During Memgraph's Community Call (https://youtu.be/ygr8yvIouZk?t=1307), the point is that AgenticRuntimes + GraphRAG moves you up on the "Kardashev scale for data" because you suddenly can get much more insight from any dataset, and everyone can use it (a large corporation does not control it). I found something similar under https://ift.tt/eMVqf8v, but the definition/example looks very narrow.
Is there a scale of how much data one can effectively process, something similar to the "Kardashev scale for data"? What would be a name for such a thing? During Memgraph's Community Call (https://youtu.be/ygr8yvIouZk?t=1307), the point is that AgenticRuntimes + GraphRAG moves you up on the "Kardashev scale for data" because you suddenly can get much more insight from any dataset, and everyone can use it (a large corporation does not control it). I found something similar under https://ift.tt/eMVqf8v, but the definition/example looks very narrow. 1 comments on Hacker News.
Is there a scale of how much data one can effectively process, something similar to the "Kardashev scale for data"? What would be a name for such a thing? During Memgraph's Community Call (https://youtu.be/ygr8yvIouZk?t=1307), the point is that AgenticRuntimes + GraphRAG moves you up on the "Kardashev scale for data" because you suddenly can get much more insight from any dataset, and everyone can use it (a large corporation does not control it). I found something similar under https://ift.tt/eMVqf8v, but the definition/example looks very narrow.
Hacker News story: Ask HN: How to measure how much data one can effectively process or understand?
Reviewed by Tha Kur
on
February 20, 2026
Rating:
No comments: