vegesm's comments

vegesm | 1 year ago | on: Calculus with Julia

Pytorch and tensorflow is pretty big. It implies all state of the art research code is not written in Julia, so it is a non-starter for any neural network based project.

vegesm | 2 years ago | on: Python 3.13 Gets a JIT

Because it only increases high single digit each release. If they keep up the 10% improvement for the next 10 release, we will reach a speedup of around 2.5 times. That's very small, considering how Python is like 10-20 times slower than JS (not even talking about C or Java like speeds).

vegesm | 3 years ago | on: The Long S

I always wondered where the integral sign was coming from. This gives a very simple explanation: it's just the letter 's', the shorthand for sum!

vegesm | 3 years ago | on: A systems model of anxiety-driven procrastination

This rings so true with me. I have a side project [1] that I try to promote and it always feels hard. Where do I start? Should I post comments in random subreddits? Should I write blog posts that may nobody read in the end? Start doing ads? I think having the multiple options kinda freezes me and it is so hard to start doing something.

[1] https://androidcalculator.com/

vegesm | 3 years ago | on: The Principles of Deep Learning Theory

I don't really get the first part of your comment. You say an NN is "just" compression + kNN and does no representational learning. But finding a compression (a transformation in other words) that makes kNN feasible on the data is exactly what people mean when they say it finds a hidden representation. It is a highly non-trivial task: e.g. simple distance in pixel space between images would get you useless results.

vegesm | 4 years ago | on: The Bitter Lesson (2019)

Actually, MobileNetV3 is a supporting example of the bitter lesson and not the other way round. The point of Sutton's essay is that it isn't worth adding inductive biases (specific loss functions, handcrafted features, special architectures) to our algorithm. Having lots of data, just put that into a generic architecture and it eventually outperforms manually tuned ones.

MobileNetV3 uses architecture search, which is a prime example of the above: even the architecture hyperparameters are derived from data. The handcrafted optimizations just concern speed and do not include any inductive biases.

vegesm | 4 years ago | on: Ask HN: Does targeted advertising work at all?

Back of a napkin calculation: showing OP 100 ads of blinds probably costs 0.1$ in the US. The cost of a blind is 100$ (I have no idea, don't live in the US). So if the probability of buying a second blind is at least 0.1% it was worth running the campaign.

The fact is, online ads are super cheap.

vegesm | 5 years ago | on: It Can Happen to You

That's true for all software development. In seven years most of your team is replaced.

vegesm | 5 years ago | on: Email from Jeff Bezos to employees

I remember my school ran a contest where you had to find obscure information on the net. Things like when the inventor of the saxophone was born? You had to submit an URL so using the library was not an option. It was insanely hard, nowadays it's just one wikipedia click away.

vegesm | 5 years ago | on: Statement of SEC Regarding Recent Market Volatility

Why would shorting Gamestop mean making it to go bankrupt? The falling stock price of a company doesn't cause a company to go bankrupt, the causality is in the other direction. Putting downward pressure on price is useful only if you want to do a hostile takeover.

vegesm | 5 years ago | on: Robinhood, in Need of Cash, Raises $1B from Its Investors

Also, don't card companies pay net30/60? So even if customers were able to pay their orders, Robinhood only receives the money in 30 days but has to pay to the clearinghouse in two days. That would wreck RH in a volatile period like this.
page 1