frogamel | 2 years ago | on: A decoder-only foundation model for time-series forecasting
frogamel's comments
frogamel | 2 years ago | on: AI Companies and Advocates Are Becoming More Cult-Like
frogamel | 2 years ago | on: Self-rewarding-lm-PyTorch: Self-Rewarding Language Model from MetaAI
1. Train model like normal
2. Evaluate model using self
3. Use eval results for DPO finetune
frogamel | 2 years ago | on: Show HN: Deep search of all ML papers
frogamel | 2 years ago | on: 'No cash accepted' signs are bad news for unbanked Americans
It's state-owned but functions as a typical retail bank that takes deposits, makes mortgages, business loans, etc.
frogamel | 2 years ago | on: Let's talk about joins
- Ability to specify whether your join should be one-to-one, many-to-one, etc. So that R will throw an error instead of quietly returning 100x as many rows as expected (which I've seen a lot in SQL pipelines).
- A direct anti_join function. Much cleaner than using LEFT JOIN... WHERE b IS NULL to replicate an anti join.
- Support for rolling joins. E.g. for each user, get the price of their last transaction. Super common but can be a pain in SQL since it requires nested subqueries or CTEs.
frogamel | 2 years ago | on: Americans are feeling much better about the economy thanks to slowing inflation
From a macro perspective, the same factors that drove inflation is what drove their big pay hikes. But most people only see that "I earned more money and now I'm being screwed by inflation". These are often the people who spend like crazy while simultaneously complaining about how bad the economy is!
frogamel | 2 years ago | on: My AI costs went from $100 to less than $1/day: Fine-tuning Mixtral with GPT4
It is remarkably easy - it takes practically zero knowledge of ML and can usually be done with less than <$1k of cloud compute costs. The issue is that for most realistic tasks you can expect to end up with something roughly on the level of GPT-3.5, and its actually really hard to compute with GPT-3.5 on a cost level, at least if you use cloud GPUs.
frogamel | 2 years ago | on: OpenAI drops ban on military tools to partner with The Pentagon
frogamel | 2 years ago | on: Against learning from dramatic events
1. I doubt many people would think of 9/11 as a "1 in 50 years" event if it didn't actually happen. If you had a year-by-country level dataset of every developed country post ww2, you'd have thousands of observations but none would have as many fatalies from terrorism as the US in 2001.
2. If a genuinely super rare event occurs (like one in thousands), its often more reasonable to think that theres been some fundamental shift in the world that you failed to recognize, rather than that you just got super lucky or unlucky to have lived through it.
frogamel | 2 years ago | on: NFL Pick-Em's LLM Bot
I would be cool to see if something like this can do can outperform a benchmark Elo system but even that is an extremely high bar that most sophisticated models can't beat.
frogamel | 2 years ago | on: Yihui Xie, RMarkdown Creator, Laid Off from RStudio
I hope this does not mean there are financial troubles within Rstudio/Posit.
frogamel | 2 years ago | on: Rabbit: LLM-First Mobile Phone
This also runs into the same core issue as voice assistants, which is that it's way easier to transmit complex thoughts/actions on a nice big visual UI than it is to convert those thoughts/actions into verbal instructions.
For example, when people order food from an online menu, most people are not fine with "get me the closest pizza available". They're usually considering complex tradeoffs between lots of factors: alternative restaurants nearby, how visually appealing is pictures of the food, health content, other options on the menu including their prices and content and how well they complement other dishes, etc. Figuring out how to express these preferences via a back-and-forth with an AI is more tedious than just looking at a visual interface and letting your subconscious process all this information into a choice.
frogamel | 2 years ago | on: Polars
Unfortunately polars was not it. Too many bugs on standard operations, unreliable interoperability with pandas (which is an issue since so many libraries require pandas dataframes as inputs), the API is also very verbose for a modern dataframe library, though it's still better than pandas.
Hopefully these will get resolved out over time but for now I had the best luck using duckdb on top of pandas, it is as fast as polars but more stable/better interoperability.
Eventually I hope the Python dataframe ecosystem gets to the same point as R's, where you have a analytics-oriented dataframe library with an intuitive API (dplyr) that can be easily used alongside a high-performance dataframe library (data.table).
frogamel | 2 years ago | on: Noteable.io is shutting down
frogamel | 2 years ago | on: Factors driving a productivity explosion
The reason productivity is up so much lately is that for the past two years:
- Employment (https://fred.stlouisfed.org/series/PAYEMS) has been increasing at 1-2% annualized rate (because unemployment rate/labor force participation rate hasn't been changing, so employment growth has come from population growth, which was 1-2% when current 18-30 year-olds were born)
- Real GDP (https://fred.stlouisfed.org/series/GDPC1) has been increasing at 3-5% annualized rate
The reality is there is a giant consumption boom the past few years, but minimal increase in employment since we were already at full employment, so by definition there will just be increase in aggregate productivity. It is unrelated to concepts like how hard are people working.
frogamel | 2 years ago | on: Greedflation: Corporate profiteering 'significantly' boosted global prices,study
In 2020-2022 my input costs (housing, food) went up ~20%, so I also needed to increase the price of my labor. This means negotiating a promotion or switching jobs, which is a lot of effort. And since I'm putting in that effort anyways, I might as well optimize it and try to get every extra dime, which ended being more than a 20% increase.
In non-inflationary periods, it's not even worth the effort and stress to optimize my salary like this.
I imagine the same was true for a lot of businesses in the supply chain. If you're some SAAS company, raising prices 20% requires renegotiating contracts, dealing with angry clients, etc. So if you're going to do that anyways you might as well try to raise it even more.
You can find recent papers from researchers about how their new transformers model is the best and SOTA, papers which claim transformers is garbage for time series and claim their own MLP variant is SOTA, other papers which claim deep learning in general underperforms compared to xgboost/lightgbm, etc.
Realistically I think time series is incredibly diverse, and results are going to be highly dependent on which dataset was cherry-picked for benchmarking. IMO this is why the idea of a time series foundation model is fundamentally flawed - transfer learning is the reason why foundation models work in language models, but most time series are overwhelmingly noise and don't provide enough context to figure out what information is actually transferrable between different time series.