Congrats! I went to your demo and asked for words that end in agi. This is what I got:
--
agi, agi, agi, agi, agi, agi, agi
These are some of the words that end in agi. You can also use the word agi in a sentence. For example, "I am going to the grocery store to get some agi."
It's a fair criticism, and ChatGPT does better, but this isn't a great test of model quality. All LLMS that rely on tokenization struggle with being introspective on language. Try asking chatGPT to count how many e's are in a sentence, or to list all words that start with "to" and end wide "de".
I haven't heard anyone describe the phenomenon clearly, but I expect it is a challenge with reasoning over both intent of the prompt and specific token IDs.
yeah as usual these model can barely sustain a conversation and fall apart the moment actual instructions are given. typical prompt they fail to udnerstand:
"what is pistacchio? explain the question, not the answer."
all these toy llm: "pistacchio is..."
gpt is the only one that consistently understand these instructions: "The question "what is pistachio?" is asking for an explanation or description of the food item..."
this makes these llm basically useless for obtaining anything but hallucinated data.
I’m super excited to announce Lamini, the LLM engine that gives every developer the superpowers that took the world from GPT-3 to ChatGPT!
I’ve seen a lot of developers get stuck after prompt-tuning for a couple days or after fine-tuning an LLM and it just gets worse—there’s no good way to debug it. I have a PhD in AI from Stanford, and don’t think anyone should need one to build an LLM as good as ChatGPT. A world full of LLMs as different & diverse as people would be even more creative, productive, and inspiring.
That’s why I’m building Lamini, the LLM engine for developers to rapidly customize models from amazing foundation models from a ton of institutions: OpenAI, EleutherAI, Cerebras, Databricks, HuggingFace, Meta, and more.
Here’s what Lamini does for you:
Your LLM outperforms general-purpose models on your specific use case
You own the model, weights and all, not us (if foundation model allows it, of course!)
Your data helps the LLM, and build you an AI moat
Any developer can do it today in just a few lines of code
Commercial-use-friendly with a CC-BY license
We’re also releasing several tools on Github:
Today, you can try out our hosted data generator for training your own LLMs, weights and all, without spinning up any GPUs, in just a few lines of code from the Lamini library. https://github.com/lamini-ai/lamini/
Sign up for early access to the training module that took the generated data and trained it into this LLM, including enterprise features like virtual private cloud (VPC) deployments. https://lamini.ai/contact
Im confused, what are you actually offering? Does my fine tuning data get shared with your platform’? Does the model get fine tuned on your end or my own system? Do you host the model?
I've been playing a bit with stacking transformer adapters to add knowledge to models and so far it has met my needs. It doesn't have the same illusion of intelligence, but so far it's just as good as a multitasking intern, so I am still having fun with it. I wonder if this is basically doing the same thing.
Interesting. Do you know if this can be done with Sentence Transformers, too? Picking a good performing one from HF. Then training an adapter for the domain (unsupervised). Then adding another one using actual training triplets (base, similar, non-similar)?
GPT at this point is more than an LLM, it is a baseline layer of logic using the underlying transformer technology. This will be challenging to replicate without the same size of data sets
primordialsoup|2 years ago
--
agi, agi, agi, agi, agi, agi, agi
These are some of the words that end in agi. You can also use the word agi in a sentence. For example, "I am going to the grocery store to get some agi."
These are some of words that end in agi.
These are some words that end in agi.
maximize, maximize, maximize, maximize, maximize, maximize, maximize, maximize
These are some words that ends in agi
--
So I think this needs more work to get to "as good as ChatGPT". But having said that, congrats on the landing
brookst|2 years ago
I haven't heard anyone describe the phenomenon clearly, but I expect it is a challenge with reasoning over both intent of the prompt and specific token IDs.
avereveard|2 years ago
"what is pistacchio? explain the question, not the answer."
all these toy llm: "pistacchio is..."
gpt is the only one that consistently understand these instructions: "The question "what is pistachio?" is asking for an explanation or description of the food item..."
this makes these llm basically useless for obtaining anything but hallucinated data.
TwoFactor|2 years ago
---GPT-3.5---
Here are some words that end in "agi":
Strategy
Swarajya
Arthroplasty
Sialagogue
Podagric
Gynecology
Physiognomy
Ophthalmology
Esophagitis
Otalgia
--- GPT-4 ---
Here are some words that end in "agi":
Swaggy
Raggi
Magi
Gagi
Stagi
Please note that some of these words may not be commonly used or may be specific to certain dialects or regions.
sharonzhou|2 years ago
I’m super excited to announce Lamini, the LLM engine that gives every developer the superpowers that took the world from GPT-3 to ChatGPT!
I’ve seen a lot of developers get stuck after prompt-tuning for a couple days or after fine-tuning an LLM and it just gets worse—there’s no good way to debug it. I have a PhD in AI from Stanford, and don’t think anyone should need one to build an LLM as good as ChatGPT. A world full of LLMs as different & diverse as people would be even more creative, productive, and inspiring.
That’s why I’m building Lamini, the LLM engine for developers to rapidly customize models from amazing foundation models from a ton of institutions: OpenAI, EleutherAI, Cerebras, Databricks, HuggingFace, Meta, and more.
Here’s our blog announcing us and a few special open-source features! https://lamini.ai/blog/introducing-lamini
Here’s what Lamini does for you: Your LLM outperforms general-purpose models on your specific use case You own the model, weights and all, not us (if foundation model allows it, of course!) Your data helps the LLM, and build you an AI moat Any developer can do it today in just a few lines of code Commercial-use-friendly with a CC-BY license
We’re also releasing several tools on Github: Today, you can try out our hosted data generator for training your own LLMs, weights and all, without spinning up any GPUs, in just a few lines of code from the Lamini library. https://github.com/lamini-ai/lamini/
You can play with an open-source LLM, trained on generated data using Lamini. https://huggingface.co/spaces/lamini/instruct-playground
Sign up for early access to the training module that took the generated data and trained it into this LLM, including enterprise features like virtual private cloud (VPC) deployments. https://lamini.ai/contact
jeffybefffy519|2 years ago
jasonjmcghee|2 years ago
ec109685|2 years ago
So much click bait in the LLM space.
mkl|2 years ago
unknown|2 years ago
[deleted]
iguana|2 years ago
mensetmanusman|2 years ago
furyofantares|2 years ago
I don't really care to click on something I know is obviously lying to me.
batch12|2 years ago
leobg|2 years ago
eschluntz|2 years ago
gdiamos|2 years ago
Paid LLM hosting. 50% cheaper than OpenAI, pay per compute needed to run & create the LLM. Export the weights anytime you want.
Enterprise VPC deployments.
atulika612|2 years ago
mise_en_place|2 years ago
cultofmetatron|2 years ago
acapybara|2 years ago
digitcatphd|2 years ago
unknown|2 years ago
[deleted]
gdiamos|2 years ago
What's the difference between this and other data pipelines like Alpaca?
verdverm|2 years ago
8thcross|2 years ago