It sounds like you don’t need immediate llm responses and can batch process your data nightly? Have you considered running a local llm? May not need to pay for api calls. Today’s local models are quite good. I started off with cpu and even that was fine for my pipelines.
kreetx|1 month ago
queenkjuul|1 month ago
ok_orco|1 month ago
LTL_FTC|1 month ago
My old threadripper pro was seeing about 15tps, which was quite acceptable for the background tasks I was running.
ydu1a2fovb|1 month ago
LTL_FTC|1 month ago
R_D_Olivaw|1 month ago