top | item 37790273

(no title)

keonix | 2 years ago

Mistral 7B ~ 8 GiB

StableLM 3B ~4 GiB

You could go even lower with smaller quantization if necessary. I personally wouldn't use anything smaller than 7B and Mistral already pushing it in coherence. Overall it depends on your use case, not everyone needs smart models, or large context that sometimes takes half of required memory.

Codellama is also surprisingly good even for non-coding tasks

discuss

order

No comments yet.