(no title)
whytai | 2 years ago
categories of startups that will be affected by these launches:
- vectorDB startups -> don't need embeddings anymore
- file processing startups -> don't need to process files anymore
- fine tuning startups -> can fine tune directly from the platform now, with GPT4 fine tuning coming
- cost reduction startups -> they literally lowered prices and increased rate limits
- structuring startups -> json mode and GPT4 turbo with better output matching
- vertical ai agent startups -> GPT marketplace
- anthropic/claude -> now GPT-turbo has 128k context window!
That being said, Sam Altman is an incredible founder for being able to have this close a watch on the market. Pretty much any "ai tooling" startup that was created in the past year was affected by this announcement.
For those asking: vectorDB, chunking, retrieval, and RAG are all implemented in a new stateful AI for you! No need to do it yourself anymore. [2] Exciting times to be a developer!
[1] https://youtu.be/smHw9kEwcgM
[2] https://openai.com/blog/new-models-and-developer-products-an...
morkalork|2 years ago
dragonwriter|2 years ago
But also one that their terms of service, which are designed to exclude the markets that they can't or won't touch, don't make it impractical for you to service with their tools.
Cali_cramoisie|2 years ago
ushakov|2 years ago
Why? Because they lack specificity. We're domain experts, we know how to prompt it correctly to get the best results for a given domain. The moat is having model do one task extremely well rather than do 100 things "alright"
Rastonbury|2 years ago
renewiltord|2 years ago
ren_engineer|2 years ago
What's stopping OpenAI from cranking up the inference pricing once they choke out the competition? That combined with the expanded context length makes it seem like they are trying to lead developers towards just throwing everything into context without much thought, which could be painful down the road
klabb3|2 years ago
I mean.. the lock in risks have been known with every new technology since forever now, and not just the risk but the actual costs are very real. People still buy HP printers with InkDRM and companies willingly write petabytes of data into AWS that they can’t even afford to egress at current prices.
To be clear, I despise this business practice more than most, but those of us who care are screaming into the void. People are surprisingly eager to walk into a leaking boat, as long as thousands of others are as well.
keithwhor|2 years ago
They can then either act as a distributor and take a marketplace fee or go full Amazon and start competing in their own marketplace.
stuckkeys|2 years ago
vikramkr|2 years ago
larodi|2 years ago
There's plenty more to innovate, really, saying OpenAI killed startups it's like saying that PHP/Wordpress/NameIt killed small shops doing static HTML. or IBM killing the... typewriter companies. Well, as I said - they could've known better. Competition is not always to blame.
karmasimida|2 years ago
The sad thing is, GPT-4 is its own league in the whole LLM game, whatever those other startups are selling, it isn't competing with OpenAI.
teaearlgraycold|2 years ago
You know you’re doing the wrong thing if you dread the OpenAI keynotes. Pick a niche, stop riding on OpenAI’s coat tails.
riku_iki|2 years ago
they don't provide embedings, but storage and query engines for embeddings, so still very relevant
> - file processing startups -> don't need to process files anymore
curious what is that exactly?..
> - vertical ai agent startups -> GPT marketplace
sure, those startups will be selling their agents on marketplace
dragonwriter|2 years ago
But you don't need any of the chain of: extract data, calculate embeddings, store data indexed by embeddings, detect need to retrieve data by embeddings and stuff it into LLM context along with your prompt if you use OpenAI's Assistants API, which, in addition to letting you store your own prompts and manage associated threads, also lets you upload data for it to extract, store, and use for RAG on the level of either a defined Assistant or a particular conversation (Thread.)
visarga|2 years ago
make3|2 years ago
blibble|2 years ago
I suspect that video is going to end up more notorious, it's even funnier given it's the VCs themselves
thisgoesnowhere|2 years ago
Those were valid concerns at the time and the market for non technical file storage like they were building was non existant.
Perfectly rational to be skeptical and Drew answered all his questions with well thought out responses.
arcanemachiner|2 years ago
EDIT: I guess it's this:
https://news.ycombinator.com/item?id=8863#9224
lazzlazzlazz|2 years ago
dragonwriter|2 years ago
Finbarr|2 years ago
Yadayadaaaa|2 years ago
If it would be only me, no one would buy azure or aws but just gcp.
bluecrab|2 years ago
m3kw9|2 years ago
Der_Einzige|2 years ago
throwaway-jim|2 years ago
yawnxyz|2 years ago
unknown|2 years ago
[deleted]
andrewjl|2 years ago
felixding|2 years ago
mvkel|2 years ago
bilsbie|2 years ago
monkeydust|2 years ago
colordrops|2 years ago
sharemywin|2 years ago
The model then decides when to retrieve content based on the user Messages. The Assistants API automatically chooses between two retrieval techniques:
it either passes the file content in the prompt for short documents, or performs a vector search for longer documents Retrieval currently optimizes for quality by adding all relevant content to the context of model calls. We plan to introduce other retrieval strategies to enable developers to choose a different tradeoff between retrieval quality and model usage cost.
karmasimida|2 years ago
There is a cost argument to make still, embedding-based approach will be cheaper and faster, but worse result than full text.
That being said, I don't see how those embedding startups compete with OpenAI, no one will be able to offer better embedding than OpenAI itself. It is hardly a convincing business.
The elephant in the room is the open source models aren't able to match up to OpenAI models, and it is qualitative, not quantitive.
lazzlazzlazz|2 years ago
emadabdulrahim|2 years ago
unknown|2 years ago
[deleted]
treprinum|2 years ago
baq|2 years ago
seydor|2 years ago
echelon|2 years ago
cityzen|2 years ago
atleastoptimal|2 years ago
unknown|2 years ago
[deleted]