> I find that AI substantially boosts materials discovery, leading to an increase in patent filing and a rise in downstream product innovation. However, the technology is effective only when paired with sufficiently skilled scientists.
I can see the point here. Today I was exploring the possibility of some new algorithm. I asked Claude to generate some part which is well know (but there are not a lot of examples on the internet) and it hallucinated some function. In spite of being bad, it was sufficiently close to the solution that I could myself "rehallucinate it" from my side, and turn it into a creative solution. Of course, the hallucination would have been useless if I was not already an expert in the field.
I came to the same conclusion a while back. LLMs are very useful when user expertise level is medium to high, and task complexity is low to medium. Why ? because it those scenarios, the user can use the LLM as a tool for brainstorming on drawing the first sketch before improving it. Human in the loop is the key and will stay key for the forceable future no matter what the autonomous AI agent gurus are saying.
https://www.lycee.ai/blog/mistral-ai-strategy-openai
If a model is right 99.99% of the time (which nobody has come close to), we still need something that understands what it's doing enough to observe and catch that 0.01% where it's wrong.
Because wrong at that level is often dangerously wrong.
This is explored (in an earlier context) in the 1983 paper "Ironies of Automation".
I wonder if the next generation of experts will be held back by use of AI tools. Having learned things “the hard way” without AI tools may allow better judgement of these semi-reliable outputs. A younger generation growing up in this era would not yet have that experience and may be more accepting of AI generated results.
I find proofreading the code gen ai less satisfying than writing it myself though it does depend on the nature of the function. Migrating mindless mapping type functions to autocomplete is nice
Yes, I have experienced it, too. I was building a web crawler using Replit as an agent. I could have done that in 2 hours without LLM help but I wanted to see how the LLM would do it. I gave it a set of of instructions but the LLM could not execute on it. It later choose an alternative path but that also did not yield. I then gave an exact list of steps. Results were slightly better but not what I was expecting. Overall, it's good to get something going but you still have to hold hands. It is not the best but also not the worst experience.
Yeah I had similar experience where I ask why a bug was happening but it gave me some thing that looked wrong, but upon closer inspection it pointed to a vague general direction where I haven’t thought of and i solved my bug with its help. The caveat is you still need to know your shit to decipher/recognize it.
“Survey evidence
reveals that these gains come at a cost, however, as 82% of scientists report reduced
satisfaction with their work due to decreased creativity and skill underutilization.”
What an interesting finding and not what I was expecting. Is this an issue with the UX/tooling? Could we alleviate this with an interface that still incorporates the joy of problem solving.
I haven’t seen any research that Copilot and similar tools for programmers have a similar reduction in satisfaction. Likely with how much the tools feel like an extension of traditional auto complete, and you still spend a lot of time “programming”. You haven’t abandoned your core skill.
Related: I often find myself disabling copilot when I have a fun problem I want the satisfaction of solving myself.
I feel if people are finding programming as creative and interesting with AI as without there is a chance they actually prefer product management?
Half statement, half question… I have personally stopped using AI assistance in programming as I felt it was making my mind lazy, and I stopped learning.
> Related: I often find myself disabling copilot when I have a fun problem I want the satisfaction of solving myself.
The way things seem to be going, I'd be worried management will find a way to monitor and try cut out this "security risk" in the coming months and years.
"The tool automates a majority of “idea
generation” tasks, reallocating scientists to the new task of evaluating model-suggested candidate
compounds. In the absence of AI, researchers devote nearly half their time to conceptualizing
potential materials. This falls to less than 16% after the tool’s introduction. Meanwhile, time spent
assessing candidate materials increases by 74%"
So the AI is in charge, and mostly needs a bunch of lab assistants.
"Machines should think. People should work." - not a joke any more.
It’s interesting to see how this research emphasizes the continued need for human expertise, even in the era of advanced AI. It highlights that while AI can significantly boost productivity, the value of human judgment and domain knowledge remains crucial.
Even Warren McCulloch and Walter Pitts were the two who originally modeled neurons with OR statements, realized it wasn't sufficient for a full replacement.
Biological neurons have many features like active dendritic compartmentalization that perceptrons cannot duplicate.
They are different with different advantages and limitations.
We have also known about the specification and frame problems for a long time also.
Note that part of the reason for the split between the symbolic camp and statistical camp in the 90s was due to more practical models being possible with existential quantification.
There have been several papers on HN talking about a shift to universal quantification to get around limitations lately.
Unfortunately discussions about the limits of first order logic have historical challenges and adding in the limits of fragments of first order logic like grounding are compounded upon those challenges with cognitive dissonance.
While understanding the abilities of multi level perceptrons is challenging, there is a path of realizing the implications of an individual perceptron as a choice function that is useful for me.
The same limits that have been known for decades still hold in the general case for those who can figure a way to control their own cognitive dissonance, but they are just lenses.
As an industry we need to find ways to avoid the traps of the Brouwer–Hilbert controversy and unsettled questions and opaque definitions about the nature of intelligence to fully exploit the advantages.
Hopefully experience will tempor the fear and enthusiasm for AGI that has made it challenging to discuss the power and constraints of ML.
I know that even discussing dropping the a priori assumption of LEM with my brother who has a PhD in complex analysis is challenging.
But the platonic ideals simply don't hold for non-trivial properties, and no matter if we are using ML or BoG Sat, the hard problems are too high in the polynomial hierarchy to make that assumption.
Interesting, a large US company with over 1000 materials scientists (there can only be a handful of those) introduced a cutting-edge AI tool and decided to make a study out of it / randomize it and gave all the credentials to some econ PHD student. Would love to know more about how this came to be. Also, why his PHD supervisor didn't get a co-author, never seen that. I'm always slightly suspicious of these very strong results without any public data / way to reproduce it. We essentially have to believe 1 guys word.
How generalizable are these findings given the rapid pace of AI advancement? The paper studies a snapshot in time with current AI capabilities, but the relationship between human expertise and AI could look very different with more advanced models. I would love to have seen the paper:
- Examine how the human-AI relationship evolved as the AI system improved during the study period
- Theorize more explicitly about which aspects of human judgment might be more vs less persistent
- Consider how their findings might change with more capable AI systems
Well damn, that’s a lot more specific and empirical than I was expecting given the title. Fascinating stuff, talk about a useful setup for studying the issue! “AI is useless to many but invaluable to some” (as mentioned in the abstract) is a great counterpoint to anti-AI luddites. No offense to any luddites on here ofc, the luddites were pretty darn woke for their time, all things considered
Well I hope it works well and fast enough. I cannot wait for my 10k cycles, 300 Wh/kg batteries. 35% efficiency solar modules in market at cheap prices and plenty of nanotech breakthroughs that were promised yet we are still waiting on
[+] [-] youoy|1 year ago|reply
> I find that AI substantially boosts materials discovery, leading to an increase in patent filing and a rise in downstream product innovation. However, the technology is effective only when paired with sufficiently skilled scientists.
I can see the point here. Today I was exploring the possibility of some new algorithm. I asked Claude to generate some part which is well know (but there are not a lot of examples on the internet) and it hallucinated some function. In spite of being bad, it was sufficiently close to the solution that I could myself "rehallucinate it" from my side, and turn it into a creative solution. Of course, the hallucination would have been useless if I was not already an expert in the field.
[+] [-] fsndz|1 year ago|reply
[+] [-] prisenco|1 year ago|reply
If a model is right 99.99% of the time (which nobody has come close to), we still need something that understands what it's doing enough to observe and catch that 0.01% where it's wrong.
Because wrong at that level is often dangerously wrong.
This is explored (in an earlier context) in the 1983 paper "Ironies of Automation".
https://en.wikipedia.org/wiki/Ironies_of_Automation
[+] [-] vatys|1 year ago|reply
[+] [-] darepublic|1 year ago|reply
[+] [-] zeeshanm|1 year ago|reply
[+] [-] m3kw9|1 year ago|reply
[+] [-] slopeloaf|1 year ago|reply
What an interesting finding and not what I was expecting. Is this an issue with the UX/tooling? Could we alleviate this with an interface that still incorporates the joy of problem solving.
I haven’t seen any research that Copilot and similar tools for programmers have a similar reduction in satisfaction. Likely with how much the tools feel like an extension of traditional auto complete, and you still spend a lot of time “programming”. You haven’t abandoned your core skill.
Related: I often find myself disabling copilot when I have a fun problem I want the satisfaction of solving myself.
[+] [-] dennisy|1 year ago|reply
Half statement, half question… I have personally stopped using AI assistance in programming as I felt it was making my mind lazy, and I stopped learning.
[+] [-] gmaster1440|1 year ago|reply
- Reduced creativity and ideation work (dropping from 39% to 16% of time)
- Increased focus on evaluating AI suggestions (rising to 40% of time)
- Feelings of skill underutilization
[+] [-] sourcepluck|1 year ago|reply
The way things seem to be going, I'd be worried management will find a way to monitor and try cut out this "security risk" in the coming months and years.
[+] [-] Animats|1 year ago|reply
So the AI is in charge, and mostly needs a bunch of lab assistants.
"Machines should think. People should work." - not a joke any more.
[+] [-] uxhacker|1 year ago|reply
[+] [-] nyrikki|1 year ago|reply
Biological neurons have many features like active dendritic compartmentalization that perceptrons cannot duplicate.
They are different with different advantages and limitations.
We have also known about the specification and frame problems for a long time also.
Note that part of the reason for the split between the symbolic camp and statistical camp in the 90s was due to more practical models being possible with existential quantification.
There have been several papers on HN talking about a shift to universal quantification to get around limitations lately.
Unfortunately discussions about the limits of first order logic have historical challenges and adding in the limits of fragments of first order logic like grounding are compounded upon those challenges with cognitive dissonance.
While understanding the abilities of multi level perceptrons is challenging, there is a path of realizing the implications of an individual perceptron as a choice function that is useful for me.
The same limits that have been known for decades still hold in the general case for those who can figure a way to control their own cognitive dissonance, but they are just lenses.
As an industry we need to find ways to avoid the traps of the Brouwer–Hilbert controversy and unsettled questions and opaque definitions about the nature of intelligence to fully exploit the advantages.
Hopefully experience will tempor the fear and enthusiasm for AGI that has made it challenging to discuss the power and constraints of ML.
I know that even discussing dropping the a priori assumption of LEM with my brother who has a PhD in complex analysis is challenging.
But the platonic ideals simply don't hold for non-trivial properties, and no matter if we are using ML or BoG Sat, the hard problems are too high in the polynomial hierarchy to make that assumption.
[+] [-] lysecret|1 year ago|reply
[+] [-] gmaster1440|1 year ago|reply
- Examine how the human-AI relationship evolved as the AI system improved during the study period
- Theorize more explicitly about which aspects of human judgment might be more vs less persistent
- Consider how their findings might change with more capable AI systems
[+] [-] 11101010001100|1 year ago|reply
https://pubs.acs.org/doi/10.1021/acs.chemmater.4c00643
were considered in the analysis?
[+] [-] iimaginary|1 year ago|reply
[+] [-] caycep|1 year ago|reply
[+] [-] bbor|1 year ago|reply
[+] [-] newyankee|1 year ago|reply