danger | 13 years ago | on: Predicting March Madness
danger's comments
danger | 13 years ago | on: A breakdown of how I was talked out of $100
danger | 13 years ago | on: How deep learning on GPUs wins datamining contest without feature engineering
danger | 13 years ago | on: Machine learning for the impatient: algorithms tuning algorithms
1. That the particular input features that you've chosen are somehow the only possible choice. But who's to say that you shouldn't add new features which are the square of each original feature? Or maybe some cross product terms, like the product of the ith feature times the jth feature. Or maybe some good features to add would be the distance from each point you've seen so far. Etc. Continuing down this path, you basically get to the question discussed in the OP about choosing a kernel for SVMs. This is just one example where hyperparameters come into play, and you need some method for choosing them.
2. That a linear predictor is impervious to overfitting. Consider the extreme case (which comes up often) where you have millions or billions of features and far fewer examples (e.g., if features are n-gram occurrences in text, or gene expression data). Then it's likely that there are many settings of weights that fit the data perfectly, but there's no way to tell if you're just picking up on statistical noise, or if you've learned something that will make good predictions on new data that you encounter. In both theory and practice, you need some form of regularization, and along with this comes more hyperparameters, which need to be chosen.
Finally, by your reasoning, it seems like you would always choose a 1-nearest neighbors classifier [1] (because it will always end up with 0 error under the setting your propose). But there's no reason why this is in general a good idea.
[1] http://en.wikipedia.org/wiki/K-nearest_neighbor_algorithm
danger | 13 years ago | on: Machine learning for the impatient: algorithms tuning algorithms
danger | 13 years ago | on: Machine learning for the impatient: algorithms tuning algorithms
danger | 13 years ago | on: Machine learning for the impatient: algorithms tuning algorithms
The basic idea of automatically tuning hyperparameters (the things this post discusses tuning with genetic algorithms) is cool, though, and is becoming a popular subject in machine learning research. A couple recent research papers on the topic are pretty readable:
Algorithms for Hyper-Parameter Optimization:
http://books.nips.cc/papers/files/nips24/NIPS2011_1385.pdf
Practical Bayesian Optimization of Machine Learning Algorithms:
danger | 14 years ago | on: Git repo for predicting March Madness using machine learning
danger | 14 years ago | on: March Madness for Machines, 2012 edition
This is the early announcement this year, where we're soliciting suggestions about other forms of data to include for this year.
danger | 14 years ago | on: Stanford Class: Probabilistic Graphical Models
Working with her was one of the highlights of my undergrad education, and her class was great, too.
danger | 14 years ago | on: Long-standing Google Apps bug forces users to renew domain names
http://www.google.com/support/forum/p/Google+Apps/thread?tid...
http://www.google.com/support/forum/p/Google+Apps/thread?tid...
http://www.google.com/support/forum/p/Google+Apps/thread?tid...
http://www.google.com/support/forum/p/Google+Apps/thread?tid...
I'm personally having the same problem today, when the auto renewal date is a month away. I've filed two tickets without response. It seems the only solution people have found is to remove their credit card information from Google Checkout.
danger | 14 years ago | on: Who Does Facebook Think You Are Searching For?
With the "filter[0]" in the url, all scores came back as 0 for me.
danger | 15 years ago | on: Selection Sunday: Is your March Madness prediction algorithm ready?
danger | 15 years ago | on: Selection Sunday: Is your March Madness prediction algorithm ready?
If you have other suggestions, please let us know, and we'll add it for next year (if possible). The only thing we're trying to avoid is somebody coming in with a lot of data at the last minute, beyond the point when anybody else can realistically get it incorporated into their model.
danger | 15 years ago | on: Selection Sunday: Is your March Madness prediction algorithm ready?
danger | 15 years ago | on: Selection Sunday: Is your March Madness prediction algorithm ready?
danger | 15 years ago | on: Predicting March Madness: On Tournament Structure and Bracket Scoring Rules
danger | 15 years ago | on: George Dahl: Machine Learning for March Madness
"Incorporating Side Information into Probabilistic Matrix Factorization Using Gaussian Processes." Ryan Prescott Adams, George E. Dahl, and Iain Murray. In Proceedings of the 26th Conference on Uncertainty in Artificial Intelligence, 2010.
danger | 15 years ago | on: Vicarious Systems Says Its Artificial Intelligence Is The Real Deal
Classification: For each of the twenty classes, predicting presence/absence of an example of that class in the test image.
Detection: Predicting the bounding box and label of each object from the twenty target classes in the test image.
Neither uses the full ImageNet data set. Instead, it's images from 20 classes of object, like shown here: http://pascallin.ecs.soton.ac.uk/challenges/VOC/voc2010/exam...