Show HN: Replace "hub" by "ingest" in GitHub URLs for a prompt-friendly extract
185 points| cyclotruc | 1 year ago |gitingest.com
It turns any public Github repository into a text extract that you can give to your favourite LLM easily.
Today I added this url trick to make it even easier to use!
How I use it myself: - Quickly generate a README.md boilerplate for a project - Ask LLMs questions about an undocumented codebase
It is still very much work in progress and I plan to add many more options (file size limits, exclude patterns..) and a public API
I hope this tool can help you Your feedback is very valuable to help me prioritize And contributions are welcome!
wwoessi|1 year ago
I've made https://uithub.com 2 months ago. Its speciality is the fact that seeing a repo's raw extract is a matter of changing 'g' to 'u'. It also works for subdirectories, so if you just want the docs of Upstash QStash, for example, just go to https://uithub.com/upstash/docs/tree/main/qstash
Great to see this keeps being worthwhile!
Arcuru|1 year ago
helsinki|1 year ago
prophesi|1 year ago
[0] https://github.com/cyclotruc/gitingest
bryant|1 year ago
[1] https://gitingest.com/cyclotruc/gitingest
Mockapapella|1 year ago
Fokamul|1 year ago
evmunro|1 year ago
I made a similar CLI tool[0] with the added feature that you can pass `--outline` and it'll omit function bodies (while leaving their signatures). I've found it works really well for giving a high-level overview of huge repos.
You can then progressively expand specific functions as the LLM needs to see their implementation, without bloating up your context window.
[0] https://github.com/everestmz/llmcat
lukejagg|1 year ago
shawnz|1 year ago
Jet_Xu|1 year ago
A few observations from building large-scale repo analysis systems:
1. Simple text extraction often misses critical context about code dependencies and architectural decisions 2. Repository structure varies significantly across languages and frameworks - what works for Python might fail for complex C++ projects 3. Caching strategies become crucial when dealing with enterprise-scale monorepos
The real challenge is building a universal knowledge graph that captures both explicit (code, dependencies) and implicit (architectural patterns, evolution history) relationships. We've found that combining static analysis with selective LLM augmentation provides better context than pure extraction approaches.
Curious about others' experiences with handling cross-repository knowledge transfer, especially in polyrepo environments?
ComputerGuru|1 year ago
Ctrl-a + ctrl-c would remain fast.
vallode|1 year ago
wwoessi|1 year ago
- for browsers it shows html - for curl is gets raw text
nfilzi|1 year ago
What'd you say are the differences with using sth like Cursor, which has access to your codebase already?
cyclotruc|1 year ago
fastball|1 year ago
cyclotruc|1 year ago
smcleod|1 year ago
nonethewiser|1 year ago
Cedricgc|1 year ago
I actually use txtar with a custom CLI to quickly copy multiple files to my clipboard and paste it into an LLM chat. I try not to get too far from the chat paradigm so I can stay flexible with which LLM provider I use
maleldil|1 year ago
It's quite useful, with some filtering options (hidden files, gitignore, extensions) and support for Claude-style tags.
wonderfuly|1 year ago
anamexis|1 year ago
cyclotruc|1 year ago
modelorona|1 year ago
Do you have any plans to expand it?
cyclotruc|1 year ago
bosky101|1 year ago
unknown|1 year ago
[deleted]
Exuma|1 year ago
xnx|1 year ago
lolinder|1 year ago
What you can do with something like this is store it in a database and then query it for relevant chunks, which you then feed to the LLM as needed.
matt3210|1 year ago
hereme888|1 year ago
spencerchubb|1 year ago
barbazoo|1 year ago
seventytwo|1 year ago
gardenhedge|1 year ago
dim13|1 year ago
752963e64|1 year ago
[deleted]
moralestapia|1 year ago
I also really like this idea in general of APIs being domains, eventually making the web a giant supercomputer.
Edit: There is literally nothing wrong with this comment but feel free to keep downvoting, only 5,600 clicks to go!