Maybe I'm missing something obvious, but what is the idea behind quantizing and tokenizing time series? We tokenize text because text isn't numbers. In the case of time series, we're... turning numbers into less precise numbers? The benefit of scaling and centering is trivial and i guess all timeseries ML does it, but I don't see why we need a token after that.
matrix2596|1 year ago
lamename|1 year ago
(It depends on what you mean by "outperform" since metrics for classification and regression aren't always comparable, but I think I'm following the meaning of your comment overall)
dist-epoch|1 year ago
> We tokenize text because text isn't numbers.
Text is actually numbers. People tried inputting UTF8 directly into transformers, but it doesn't work that well. Karpathy explains why:
https://www.youtube.com/watch?v=zduSFxRajkE
prlin|1 year ago
Text can be represented by numbers but they aren't the same datatype. They don't support the same operations (addition, subtraction, multiplication, etc).
lamename|1 year ago
spyder|1 year ago
https://en.wikipedia.org/wiki/Neuro-symbolic_AI
555watch|1 year ago
intalentive|1 year ago