Once every 36 seconds? Even for simple transcription, that seems insane. This reads like the hustle culture memes...
2:00am wake up
2:05am cold shower
2:15am breakfast: almonds, breast milk bought off facebook, 50mg adderal
2:30am begin transcribing thoughts into the machine
I read the whole article, and It could be bragging
or a cry for help.
In any case its clear that the world of performance metrics has achived some kind of unimaginable gravity and angular momentum, and has started to orbit itself
> Within the last 24 months it’s clear that AI has become an essential coworker
I'm curious why the personification of these tools? Like, with the same logic we could call our dishwashers and washing machines employees and co-workers.
My guess is it is some form of hype-speak to raise the level of perceived importance of the technology for financial gain.
Because you don't start your dishwasher by saying to it "alright, I need you to wash my dishes. they are going to probably need some extra soaking because I burned dinner yesterday." Instead, you push the pots and pans button and off to work the robot goes.
> In fact, I find myself growing reliant on the AI to the extent that I no longer remember some of the R syntax, A sign of working at a higher level of abstraction : one of the promises of AI.
> A sign of working at a higher level of abstraction.
Working at a higher level of abstraction while loosing knowledge at the lower level also means by some degree that one is going to be reliant on the abstraction without any understanding what is abstracted away.
Hustle culture nonsense aside, I find using whisper (AI) transcription instead of typing to be super efficient.
My workflow involves setting up a whisper server, downloading the Whispering(1) app on my computer, and binding it to a shortcut on my keyboard and mouse. Whenever I want to write something down, I just hit the shortcut, dictate and it transcribes instantly. With a Nvidia GPU (1070 in my case), transcription is nearly instantaneous. Although I haven’t set it up on my MacBook yet, I suspect it will be just as fast with Apple Silicon
Learning to leverage AI as it exists today takes effort, but is usually worth it.
However, ignoring the effort undercuts the distance these products have to go. Speech is great because there is just a single interface to integrate with. Obviously I'm biased to my employer's speech product, but I'm sure there are many.
Biggest thing for me was when I saw that the lem editor[0] posted on hacker news[1] was a small editor, which has 3 top level features: common lisp API, LSP support, and copilot support.
I've installed gptel[2] in emacs, and hacking up a few tools that really make it shine. Up next is figuring out voice + AI + emacs :)
I'm probably at 20 times a day, at least for random trivia. What would be great (and maybe it already exists) is a daily email report, formatted for Anki, that helps me remember the things I've asked ChatGPT.
> Publishing data-driven blog post analysis is a key part from formatting the data to analyzing it using R and then publishing charts. All of this is now predominantly handled with prompts to an AI.
> I can generate several hundred lines of code in 5-10 minutes. With the newer models, I expect this to collapse to 1-2 minutes.
It seems main usage is to make plots with R... Which is funny as the bar chart on the page can be done in Excel/Numbers/Sheets by entering a few numbers and headers.
You have no idea how many times you use it.
Maps? Recommendation systems (the algo)? Other optimizations?
BTW, to the best of my knowledge, dictation still has a high error rate.
There is only so much AI can do because it currently lacks in certain domain knowledge.
The worst one I ever had to fix was ESPN captions of commentary for some indoor motorcross thing with dirt bikes going around a track. First, the motorbike noise, but secondly, the commentators were using the (well known to fans) nicknames for all the riders, which the AI had no idea how to transcribe, no idea who they were, and were almost impossible for me to even Google.
Like it or not "AI" has become synonymous with generative AI, when most people talk about how much they use AI they don't mean the YouTube algorithm or the neural network that handles autocorrect on their phone.
Whisper (OpenAI 2022) and other recent dictation models are rather good. Whisper automatically punctuates sentences and usually gets just a few words wrong in my experience.
Probably because AI is a misnomer. Machine learning is a better name, but even than it is pretty lacking.
What is generally called AI in common speech is pretty much a class of non-linear statistical models which require some training to generate weights which are than used to fit the model. Most people that know anything about statistics knows this, so it is fine actually. Misnomers exists in all industries and all science, and we just deal with them.
I've been thinking that a lot lately - LLMs are really nice search engines, which is great, that's a product everyone needs.
Before Google, information is hard to find. After Google, websites are incentivized to bury information behind ads and bullshit. After LLMs, the information is juiced out like lemonade.
I wonder what the next step in the incentive landscape is?
I notice you keep popping up every time there's a chance to spread FUD and hate regarding AI tech, its developers and its users, with comments that are getting more and more intolerant and unhinged. I consider you a toxic, potentially dangerous extremist, someone we all should fight against.
(I have not done the math:) If you factor in 20-30 years of education that each typical white collar human requires as an entry point into their field, and the energy that goes into that, and the energy that goes into the continued upkeep of that human, the billions of seconds you can shave off menial work by pushing it all to an increasingly efficient AI are probably easily worth it, from a purely ecological perspective.
He mentions that both the dictation model and the editing model run on his laptop, so I'm not sure this is actually using all that much resources (beyond what it took to train the models originally)
itishappy|1 year ago
metalman|1 year ago
snakeyjake|1 year ago
[deleted]
uludag|1 year ago
> Within the last 24 months it’s clear that AI has become an essential coworker
I'm curious why the personification of these tools? Like, with the same logic we could call our dishwashers and washing machines employees and co-workers.
My guess is it is some form of hype-speak to raise the level of perceived importance of the technology for financial gain.
mmmlinux|1 year ago
thefz|1 year ago
Polar opposite of what I want to happen to me.
There's joy in knowing one's tools.
chrisandchris|1 year ago
Working at a higher level of abstraction while loosing knowledge at the lower level also means by some degree that one is going to be reliant on the abstraction without any understanding what is abstracted away.
recursive|1 year ago
kobe_bryant|1 year ago
swiftcoder|1 year ago
dumbmrblah|1 year ago
My workflow involves setting up a whisper server, downloading the Whispering(1) app on my computer, and binding it to a shortcut on my keyboard and mouse. Whenever I want to write something down, I just hit the shortcut, dictate and it transcribes instantly. With a Nvidia GPU (1070 in my case), transcription is nearly instantaneous. Although I haven’t set it up on my MacBook yet, I suspect it will be just as fast with Apple Silicon
(1) https://github.com/braden-w/whispering/
You can also use an API like grok, but I'm generally wary of such services.
I'm a bit of an introvert, so I found talking out loud to be awkward at first. But now I can't go back to regular typing, given the efficiency gains.
codemac|1 year ago
However, ignoring the effort undercuts the distance these products have to go. Speech is great because there is just a single interface to integrate with. Obviously I'm biased to my employer's speech product, but I'm sure there are many.
Biggest thing for me was when I saw that the lem editor[0] posted on hacker news[1] was a small editor, which has 3 top level features: common lisp API, LSP support, and copilot support.
I've installed gptel[2] in emacs, and hacking up a few tools that really make it shine. Up next is figuring out voice + AI + emacs :)
[0]: https://lem-project.github.io/
[1]: https://news.ycombinator.com/item?id=41357409
[2]: https://github.com/karthink/gptel
mtsolitary|1 year ago
keiferski|1 year ago
elAhmo|1 year ago
> I can generate several hundred lines of code in 5-10 minutes. With the newer models, I expect this to collapse to 1-2 minutes.
It seems main usage is to make plots with R... Which is funny as the bar chart on the page can be done in Excel/Numbers/Sheets by entering a few numbers and headers.
DuctTapeAI|1 year ago
Balgair|1 year ago
cpufry|1 year ago
[deleted]
gatinsama|1 year ago
qingcharles|1 year ago
There is only so much AI can do because it currently lacks in certain domain knowledge.
The worst one I ever had to fix was ESPN captions of commentary for some indoor motorcross thing with dirt bikes going around a track. First, the motorbike noise, but secondly, the commentators were using the (well known to fans) nicknames for all the riders, which the AI had no idea how to transcribe, no idea who they were, and were almost impossible for me to even Google.
jsheard|1 year ago
85392_school|1 year ago
Artgor|1 year ago
simonw|1 year ago
runarberg|1 year ago
What is generally called AI in common speech is pretty much a class of non-linear statistical models which require some training to generate weights which are than used to fit the model. Most people that know anything about statistics knows this, so it is fine actually. Misnomers exists in all industries and all science, and we just deal with them.
85392_school|1 year ago
nashashmi|1 year ago
nunez|1 year ago
Is this supposed to be a good thing?
e-clinton|1 year ago
unknown|1 year ago
[deleted]
EGreg|1 year ago
jgalt212|1 year ago
bookofjoe|1 year ago
01HNNWZ0MV43FF|1 year ago
Before Google, information is hard to find. After Google, websites are incentivized to bury information behind ads and bullshit. After LLMs, the information is juiced out like lemonade.
I wonder what the next step in the incentive landscape is?
pie420|1 year ago
[deleted]
vouaobrasil|1 year ago
[deleted]
acchow|1 year ago
qingcharles|1 year ago
I certainly use AI all throughout the day as a force multiplier for almost any computer-related task I do. I probably ask 100 questions a day to LLMs.
t-writescode|1 year ago
01HNNWZ0MV43FF|1 year ago
Isamu|1 year ago
Because???????
elpocko|1 year ago
OtomotO|1 year ago
[deleted]
simonw|1 year ago
jstummbillig|1 year ago
swiftcoder|1 year ago
artursapek|1 year ago
orthecreedence|1 year ago