top | item 35829800

MosaicML MPT-7B: A Commercially-Usable LLaMa-Quality Model

119 points| ml_hardware | 2 years ago |mosaicml.com

11 comments

order

jpdus|2 years ago

I wonder why this is getting so few traction here.

These models seems to beat all other available open-source models easily and the Blogpost is extremely well written, with very good documentation and fine-tuning instructions.

Well done MosaicML, I am excited what comes next and will definitely test out you platform!

deepsquirrelnet|2 years ago

I'm perplexed as well. Here's a model with commercial use licensing that is competitive (better in half of the major benchmarks) with llama 7B, and has been tuned in several variants and has 2048 token width inputs.

This is BY FAR the best model of its size that is usable by businesses. I plan to start testing it out soon.

vsroy|2 years ago

This has a context window of 65K for the storywriter version.

ftxbro|2 years ago

How can I run some inference with this model locally? Do I have to make a huggingface account?

fswd|2 years ago

if you can't figure it out send me an email and I can help you figure it out