(no title)
michaelgiba | 4 months ago
1. Curate a continuously extended inventory of prebuilt llamafiles for models as they are released 2. Create both flexible builds (with dynamic backend loading for cpu and cuda) and slim minimalist builds 3. Upstreaming as much as they can into llama.cpp and partner with the project
michaelgiba|4 months ago
njbrake|4 months ago