myownpetard | 2 months ago | on: NVIDIA frenemy relation with OpenAI and Oracle
myownpetard's comments
myownpetard | 1 year ago | on: MCP vs. API Explained
> Regardless, again: if the AI is so smart, and it somehow needs something akin to MCP as input (which seems silly), then we can use the AI to take, as input, the human readable documentation -- which is what we claim these AIs can read and understand -- and just have it output something akin to MCP.
This example is like telling someone who just wants to check their email to build an IMAP client. It's an unnecessary and expensive distraction from whatever goal they are actually trying to accomplish.
As others have said, models are now being trained on MCP interactions. It's analogous to having shared UI/UX patterns across different webapps. The result is we humans don't have to think as hard to understand how to use a new tool because of the familiar visual and interaction patterns. As the design book title says, 'don't make me think.'
myownpetard | 1 year ago | on: AlphaQubit: AI to identify errors in Quantum Computers
myownpetard | 1 year ago | on: The deep learning boom caught almost everyone by surprise
Our DNA specifies the model structure and hyperparameters for our brains. Then it is compiled and instantiated into a physical medium, our bodies, and our connectome is trained.
If you want to make a comparison about the quantity of information contained in different components of an artificial and a biological system, then it only makes sense if you compare apples to apples. DNA:Code :: Connectome:Weights
myownpetard | 1 year ago | on: The deep learning boom caught almost everyone by surprise
Then we compile and run that source code and our individual lived experience is the training data for the instantiation of that architecture, e.g. our brain.
It's two different but interrelated training/optimization processes.
myownpetard | 1 year ago | on: The deep learning boom caught almost everyone by surprise
myownpetard | 1 year ago | on: Training Language Models to Self-Correct via Reinforcement Learning
I take this to mean during weight updates, e.g. training.
> "runtime train of thought"
I take runtime here to mean inference, not during RL. What does runtime mean to you?
Previous approaches [0] successfully used inference time chain of thought to improve model responses. That has nothing to do with RL though.
The grandparent is wrong about the paper. They are doing chain of thought responses during training and doing RL on that to update the weights, not just during inference/runtime.
myownpetard | 1 year ago | on: Training Language Models to Self-Correct via Reinforcement Learning
One is talking about an improvement made by making control flow changes during inference (no weights updates).
The other is talking about using reinforcement learning to do weight updates during training to promote a particular type response.
OpenAI had previously used reinforcement learning with human feedback (RLHF), which essentially relies on manual human scoring as its reward function, which is inherently slow and limited.
o1 and this paper talk about using techniques to create a useful reward function to use in RL that doesn't rely on human feedback.
myownpetard | 1 year ago | on: Ask HN: What is the most useless project you have worked on?
myownpetard | 2 years ago | on: More product, fewer product managers
It also sounds like you have a PM on your team, but they actually have the title of SWE or Eng. Mgr. They probably spend > 50% of their time on the above listed responsibilities rather than engineering. Hopefully they don't get docked in their performance reviews for essentially performing the duties of a PM rather than those of an engineer.
myownpetard | 2 years ago | on: The business of extracting knowledge from academic publications
Literally the first line in the comment that started this thread.
myownpetard | 2 years ago | on: The business of extracting knowledge from academic publications
People are being dismissive of your comments because to say that proteins are niche in the context of pharma is like saying advertising is niche in the context of Meta and Google.
myownpetard | 2 years ago | on: John Riccitiello steps down as CEO of Unity
Which as we know puts you down an irreversible ideological path of... something?
myownpetard | 2 years ago | on: John Riccitiello steps down as CEO of Unity
myownpetard | 2 years ago | on: I’m not a programmer, and I used AI to build my first bot
We could also talk about how the word 'executes' implies some kind of agency which computers lack. It's like saying a rock just executes the laws of physics when it rolls down a hill...
myownpetard | 2 years ago | on: I’m not a programmer, and I used AI to build my first bot
Well in the case of Python, the interpreter is indeed interpreting.
myownpetard | 2 years ago | on: An Old Conjecture Falls, Making Spheres a Lot More Complicated
myownpetard | 2 years ago | on: Show HN: Learn a language quickly by practising speaking with AI
One followup: does it take into account grammatical errors, incorrect conjugations and that kind of thing?
myownpetard | 2 years ago | on: Show HN: Learn a language quickly by practising speaking with AI
myownpetard | 2 years ago | on: In 17th century, Leibniz dreamed of a machine that could calculate ideas (2019)
This downplays the importance of the set of systems for which his proof holds and makes it sound like it applies to some obscure branch of mathematics.
It applies to a huge set of important systems, not least of which is any system that is sufficiently expressive as to uniquely identify the natural numbers.
- "Our coin hit $100M daily volume, get on this rocketship before it's too late!"
- "Our exchange does $1B annually, so you know we're trustworthy!"
- "Hey investors, look at the massive demand for our GPUs (driven by the company we invested $100B)!"