mistymountains
|
1 year ago
Again, the problem is custom kernels in CUDA. It’s not straightforward for many applications (LLMs are probably the most straightforward).
mistymountains
|
1 year ago
|
on: Venting Doesn't Reduce Anger, but Something Else Does, Study Says
It doesn’t sound like having a house will magically make you feel better. Plenty of people are just as exposed to markets as you yet respond differently. I suggest exercise, nutrition, nature, and extended travel if those are not already a part of your life.
mistymountains
|
1 year ago
|
on: AMD's MI300X Outperforms Nvidia's H100 for LLM Inference
These kinds of comments make me think few people have actually tried. My experience has been 1 work day of getting things set up to work the same as before for training and testing (PyTorch).
mistymountains
|
1 year ago
|
on: AMD's MI300X Outperforms Nvidia's H100 for LLM Inference
Unless you develop in CUDA, you can easily train code (e.g. PyTorch) written for training on Nvidia hardware on AMD hardware. You can even keep the .cuda() calls.
mistymountains
|
1 year ago
|
on: AMD's MI300X Outperforms Nvidia's H100 for LLM Inference
I’m a AI Scientist and train a lot of models. Personally I think AMD is undervalued relative to Nvidia. No, chips aren’t as fast as Nvidia’s latest and yes, there are some hoops to get things working. But for most workloads in most industries (ignoring for the moment that AI is likely a poor use of capital), it will be much more cost effective and achieve about the same results.
mistymountains
|
1 year ago
|
on: Tesla cuts even more workers as unsold inventory stacks up
Why would you want (versus need) a Tesla? It’s no longer an aspirational product, it’s an appliance.
mistymountains
|
2 years ago
For all you know he had a hard adjustment to the college workload. Maybe his high school was not serious and nobody really challenged him. You all could have interrogated why he may have struggled and shared your strategies for success, lifting him up rather than beating him down. That would make you true “brothers”.
mistymountains
|
2 years ago
My (biotech, mostly remote) company does this. They may not love it but they realize they have to hire from SF, Boston, NYC etc to get the best ML talent and people expect market salaries / don’t want to up and move if they don’t have to.
mistymountains
|
2 years ago
The issue is that actually, outside of technical paths, businesses do like to hire business grads as it allows them to do even less training and they usually don’t care if someone is well read or opinionated (possibly prefer the opposite).
mistymountains
|
2 years ago
|
on: The GPT Store
Or, just maybe, AGI is a mirage with the bulk of its current utility as a marketing tool for much more realistic, if ultimately mundane, applications. OpenAI, of course, knows this.
mistymountains
|
2 years ago
You won’t teach yourself by running papers through Claude, and you won’t need to if you went from first principles rather than rushing.
mistymountains
|
2 years ago
You gave your 9 year old a smartphone?
mistymountains
|
2 years ago
It’s annoying how much things have shifted now that you can’t really own a performance car without worrying if someone will mess with it.
mistymountains
|
2 years ago
Yeah we all need to do better and trust China more! Lol.
mistymountains
|
2 years ago
|
on: A Texas semiconductor boom on the horizon?
Is this your diagnosis? That things are more corporate in terms of outcome? Genuinely curious, as a Bay Area native, why UT/Texas can’t begin to compete given how many issues face SF currently.
mistymountains
|
2 years ago
Cool it with the italics.
mistymountains
|
2 years ago
|
on: X/Twitter has updated its terms of service to let it use posts for AI training
Comment datasets are valuable for conversational AI, it’s the same reason Reddit locked down the API I imagine.
mistymountains
|
2 years ago
|
on: DMV tells Cruise to reduce its driverless vehicle fleet in SF by 50%
It’s probably the best from their training standpoint. The city is quite small relative to others and their only hope is to essentially memorize San Francisco in the neural networks. IMO you could not take a cruise and drop it into any other city and have it work.
mistymountains
|
2 years ago
|
on: DMV tells Cruise to reduce its driverless vehicle fleet in SF by 50%
They’re too deep, their networks are overfit to San Francisco at this point. Making it work in other cities would require the insane training hours to basically memorize that city.
mistymountains
|
2 years ago
|
on: Ask HN: Is GPT 4's quality lately worst than GPT 3.5?
I’ve seen this stated a ton and it’s not really true. Once trained, the model (except for decoding) is deterministic, and you can enforce determinism fairly easily. ChatGPT is not deterministic at the chat window but that’s not inherent to the model.