gandalfgeek | 2 months ago | on: Nvidia just paid $20B for a company that missed its revenue target by 75%
gandalfgeek's comments
gandalfgeek | 3 months ago | on: The Undermining of the CDC
This was obviously false during the pandemic when these “health” agencies did what the White House wanted, from the actual “science” to the messaging.
gandalfgeek | 6 months ago | on: Warp Code: the fastest way from prompt to production
Can we please standardize this and just have one markdown file that all the agents can use?
gandalfgeek | 6 months ago | on: MIT Study Finds AI Use Reprograms the Brain, Leading to Cognitive Decline
Is it safe to say that LLMs are, in essence, making us "dumber"? No! Please do not use the words like “stupid”, “dumb”, “brain rot”, "harm", "damage", "brain damage", "passivity", "trimming" , "collapse" and so on. It does a huge disservice to this work, as we did not use this vocabulary in the paper, especially if you are a journalist reporting on it.
[1]: https://www.media.mit.edu/projects/your-brain-on-chatgpt/ove...
gandalfgeek | 6 months ago | on: MIT Study Finds AI Use Reprograms the Brain, Leading to Cognitive Decline
gandalfgeek | 10 months ago | on: Wasting Inferences with Aider
gandalfgeek | 10 months ago | on: Wasting Inferences with Aider
gandalfgeek | 11 months ago | on: Heavy chatbot usage is correlated with loneliness and reduced socialization
gandalfgeek | 1 year ago | on: Stay Gold, America
If the last election was any indication then more than half the country explicitly rejected many of them.
gandalfgeek | 1 year ago | on: AI Predictions for 2025, from Gary Marcus
Of course there are plenty of problems with the current state of AI and LLMs, but to have such a preconceived pessimistic outlook that can't even acknowledge their massive and quick adoption and usefulness in multiple domains seems not intellectually honest.
gandalfgeek | 1 year ago | on: Fogus: Things and Stuff of 2024
gandalfgeek | 1 year ago | on: How Google spent 15 years creating a culture of concealment
Pretty much every public company, at least every bigtech company, follows the same conventions -- don't say incriminating things in chat, trainings for "communicate with care" (definitely don't say "we will kill the competition!!" in email or chat), automatic retention policy etc etc.
No need to single out Google.
gandalfgeek | 1 year ago | on: Scientific American's departing editor and the politicization of science
That's not how science works. Religions are "followed". Science is based on questioning and skepticism and falsifiability.
gandalfgeek | 1 year ago | on: Alan Kay on Messaging (1998)
"Call by meaning" sounds exactly like LLMs with tool-calling. The LLM is the component that has "common-sense understanding" of which tool to invoke when, based purely on natural language understanding of each tool's description and signature.
gandalfgeek | 1 year ago | on: The Friendship that made Google huge (2018)
It was really special to see how this pair basically laid out the foundations of large-scale distributed computing. Protobufs, huge parts of the search stack, GFS, MapReduce, BigTable... the list goes on.
They are the only two people at Google at level 11 (senior fellow) on a scale that goes from 3 (fresh grad) to 10 (fellow).
gandalfgeek | 1 year ago | on: DJI ban passes the House and moves on to the Senate
Or maybe they all mass-migrate to Anduril solutions?
gandalfgeek | 1 year ago | on: What we've learned from a year of building with LLMs
"Teaching" the LLM an entirely new language (like a DSL) might actually need fine-tuning, but you can probably build a pretty decent first-cut of your system with n-shot prompts, then fine-tune to get the accuracy higher.
gandalfgeek | 1 year ago | on: What we've learned from a year of building with LLMs
- "Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?" https://arxiv.org/abs//2405.05904
- "Fine-Tuning or Retrieval? Comparing Knowledge Injection in LLMs" https://arxiv.org/abs/2312.05934
gandalfgeek | 1 year ago | on: Ask HN: Is RAG the Future of LLMs?
The latest connotation of RAG includes mixing in real-time data from tools or RPC calls. E.g. getting data specific to the user issuing the query (their orders, history etc) and adding that to the context.
So will very large context windows (1M tokens!) "kill RAG"?
- at the simple end of the app complexity spectrum: when you're spinning up a prototype or your "corpus" is not very large, yes-- you can skip the complexity of RAG and just dump everything into the window.
- but there are always more complex use-cases that will want to shape the answer by limiting what they put into the context window.
- cost-- filling up a significant fraction of a 1M window is expensive, both in terms of money and latency. So at scale, you'll want to filter out and RAG relevant info rather than indiscriminately dump everything into the window.
gandalfgeek | 1 year ago | on: Groq CEO: 'We No Longer Sell Hardware'
No doubt fast SRAM helps, but from a computation pov imho its that they've statically planned computation and eliminated all locks.
Short explainer here: https://www.youtube.com/watch?v=H77tV1KcWIE (Based on their paper).
Not following the core argument here. Author seems to be comparing valuation in funding rounds to revenue projections. Revenue projection was revised downward, valuation was not.
Good point about not running the proprietary models, but that doesn't preclude strategic fit with Nvidia.