Is there a scale of how much data one can effectively process, something similar to the "Kardashev scale for data"? What would be a name for such a thing? During Memgraph's Community Call (https://youtu.be/ygr8yvIouZk?t=1307), the point is that AgenticRuntimes + GraphRAG moves you up on the "Kardashev scale for data" because you suddenly can get much more insight from any dataset, and everyone can use it (a large corporation does not control it). I found something similar under https://adamdrake.com/from-enterprise-decentralization-to-tokenization-and-beyond.html#productize, but the definition/example looks very narrow.
mbuda|11 days ago
allinonetools_|11 days ago
mikewarot|11 days ago
Fir example, a correct grand unified theory isn't useful if you don't know the physics to understand it.
rgavuliak|11 days ago
kellkell|11 days ago
[deleted]
mbuda|11 days ago
Agree with the measures; follow-up question: what's the insight definition? I think exposing some of those measures would help people better understand what the analysis covered, in other words, how much data was actually analyzed. Maybe an additional measure is some kind of breadth (I guess it could be derived from the throughput).
"Informational leverage" reminded me of "retrieval leverage" because yeah, the scale of data didn't change, the ability to extract insights did :D
Natfan|11 days ago