(no title)
KhaledAlam | 10 months ago
It simulates chunking using the same logic as GraphRAG’s actual pipeline, pulls live model pricing from a hosted JSON, and includes output token projections. The user is also prompted whether to proceed with full indexing after seeing the estimates.
This is particularly useful when working with large corpora or limited OpenAI quotas.
Blog post (with technical deep dive and lessons learned): https://blog.khaledalam.net/how-i-added-token-llm-cost-estim...
No comments yet.