top | item 5770689

Unlabeled Object Recognition in Google+

144 points| sabalaba | 13 years ago |byronknoll.blogspot.com | reply

59 comments

order
[+] VLM|13 years ago|reply
My guess is this is the background research for a future product something like google glass for blind people. Leveraging 1980s text adventures, fed thru a speech synth.

"You are facing north in the center of the sidewalk, toward the intersection of main and 1st streets. Standing 40 feet in front of you is a brown dog and 72 feet in front of you there is a woman standing on the other side of the road. Twenty seven feet ahead is the front door of mcdonalds. Would you like to hear the latest google locations reviews of that restaurant? Your apartment building entrance is 157 feet ahead. 34 feet ahead and to your left a graffiti artist has tagged a brick wall with the QR code pointing to the goatse website and there is a billboard with an advertisement for the latest star trek movie to your right and 50 feet upward. Also it is dark and you are likely to be eaten by a grue."

[+] joshguthrie|13 years ago|reply
So being blind and using Google Glass would equal playing Zork all day?
[+] zepolud|13 years ago|reply
More likely it's just Hinton having fun with all the computational power he now has access to.
[+] TeMPOraL|13 years ago|reply
Such research could also be of immense value to sighted people - automatic, continous object recognition will be very useful for augmented reality applications in a Glass-like device.
[+] perlpimp|13 years ago|reply
I'd totally do moria like adventure with google glass.
[+] neilk|13 years ago|reply
I did my own tests with Google+. Some results:

- Google+ queues images for recognition. Results improved steadily over 72 hours.

- Google+ does not use OCR of text in the images. That surprised me. But perhaps it's a privacy issue.

- Google+ does use information gleaned from elsewhere on the web. Words that were associated with the same images on Flickr would turn up those very pictures on Google+.

- Oddly, Google+ does not use information associated with those images on Twitter.

- Google probably uses EXIF data married to a database of location names.

- The much-vaunted feature recognition is impressive, better than any other system, but for me did not achieve creepy levels of intuition.

http://neilk.net/blog/2013/05/23/testing-google-photos-ai/

[+] kulkarnic|13 years ago|reply
And it also doesn't seem to stem words. [flower] and [flowers] give different results (actually flowers gives no results). But I am impressed by the number of classes they have: who labels pineapples in an image corpus?
[+] NoodleIncident|13 years ago|reply
The unlabeled object recognition test is a standard test of machine learning algorithms.

Historically, error rates of around 20-25% won competitions and set records. A year or two ago, though some researchers and professors from the University of Toronto absolutely smashed those records, getting around a 16% error rate. They went and made a startup out of their tech, and got acquired by Google a few months ago.

I think that this is going to be the first of a long line of Google products integrating this sort of deep neural network technology. I wouldn't be shocked if Google in 10 years was known for something besides search, at this rate.

[+] hayksaakian|13 years ago|reply
At the end of the day though, object recognition is also search, in a sense.

If I'm flipping through my album of dog photos, or looking especially closely at dogs via google glass, maybe Google will show me an ad for dog food?

[+] eliben|13 years ago|reply
Wow, that's awesome. Imagine your photo collection of 1000s of photos, and you remember "the one with that cat" but how do you dig through the photos to actually find it? This can go a long way to a much more meaningful photo library management experience.
[+] ge0rg|13 years ago|reply
Am I the only one scared by the thought of uploading my whole photo collection to Google's servers? What about creating an offline database of object fingerprints that can classify my pictures without privacy violations?
[+] kailuowang|13 years ago|reply
Google has been building her knowledge graph for a couple of years. The goal is for computer to truly understand real world concepts rather than keywords and text. I didn't fully understand the application of it rather than some fancy cards on the search results page until yesterday when I asked Google "where did Golden retriever originate?", and Google answered "England". Google might not really understand the concept originated or golden retriever, but Google understands that "where" is asking for a place and she found a lot of mention of "England" in all the page results of "golden retrieve origin", she also understand that England is a place. So Google guessed the answer.

The Google computer has been reading about these concepts for years, now we know it can see them in pictures (and maybe even in live videos). That excites me to a degree that it becomes a little bit scary. When will that computer learn the concept of "self"?

Update: actually Google seems to understand the concept of "golden retriever", I search my photos with the word and yes, at least Google knows how golden retrievers look like.

[+] rictic|13 years ago|reply
We actually have an explicit concept of Golden Retrievers and their origins as an Animal Breed within the subset of the knowledge graph that we expose with a permissive license: http://www.freebase.com/m/01t032

The data is available for querying as well as licensed such that you can take it and build your own commercial database with it (requiring only attribution).

[+] saraid216|13 years ago|reply
Completely OT, but when did Google become female? And also why?
[+] danmaz74|13 years ago|reply
When will that computer learn the concept of "self"? It doesn't matter, as long as they didn't program it with the instinct of self-preservation...
[+] chris_wot|13 years ago|reply
When I type in "handsome", it consistently shows pictures of myself. Just goes to show that image recognition has a long way to go.
[+] danmaz74|13 years ago|reply
Beauty lies in the eye of the beholder...
[+] danso|13 years ago|reply
This is related to last year's article about Google's neural network being able to recognize cats, right?

http://www.nytimes.com/2012/06/26/technology/in-a-big-networ...

[+] neilk|13 years ago|reply
I don't think so. The neural network in Google+ was trained on labeled images and now finds similar objects in unlabeled images.

The technology discussed in that article is about deducing the existence of a common feature, in this instance a cat, from a large collection of unlabelled images.

[+] shawabawa3|13 years ago|reply
I just tried it on my photo collection and it's incredible. It even works for famous places, e.g. I searched for "western wall" and "dome of the rock" and it found them. I can't imagine how that works
[+] eliben|13 years ago|reply
Why, imagining is fun.

There are tons of photos of these places online, many of them tagged ("breaking news from the dome of the rock", or "here's me and Sam at the western wall"). Collect enough of these and you can attach knowledge to images. Then, you just have to know two images look similar, and you have your classification.

Neither of the above is easy - nay, it's very hard. But once you have those two building blocks, this technology is viable. And it's very exciting!

[+] 0x0|13 years ago|reply
For places, could it be cheating and peeking at GPS positions in the EXIf tags?
[+] eigenvector|13 years ago|reply
It appears to recognize some famous mountains, including Mt Robson (Canada), Mt Rainier (USA) and Cerro Torre (Chile/Argentina).
[+] Kequc|13 years ago|reply
I think the greater achievement came with Google image search, someone had to tag all those photos.

They wrote an algorithm that takes that data and recognises new images with it. As long as there is a way for us to tag inaccurate matches then it should be able to continue to learn. I imagine any flagged matches are being reviewed carefully.

[+] ehsanu1|13 years ago|reply
I always thought that was done using keywords on the page the image was taken from (and image captions, alt text, titles and filenames). This is reinforced by what you see when you go many pages ahead in images and see things that don't seem related to what you searched, but the keyword is there somewhere on the page.
[+] iliis|13 years ago|reply
This is seriously fun! You can actually search for "blue car" and it works. Searching for "picture" results in an error however. Same for "image". "photo" seems to return more or less everything.
[+] goblin89|13 years ago|reply
Object recognition also works on videos, judging from the fact that a recording of my cat came up in search results for “dog”. (Could be that it only looks at the first frame, though.)
[+] jbuzbee|13 years ago|reply
This is really a quite cool feature. For me, it was able to do a good job on searches for things like snow, road, dog, sunset, etc.
[+] wtdominey|13 years ago|reply
Would be a fantastic feature in desktop photo management software like Aperture or Lightroom.
[+] 0x0|13 years ago|reply
Or even google picasa!
[+] EGreg|13 years ago|reply
As a software developer, very few things blow my mind - but this just did.

How did they do that?

[+] exit|13 years ago|reply
google should advertise the sheer technical depth of the stuff they do

make search sound like scotty explaining a warp core on star trek

[+] dirkgently|13 years ago|reply
I am fine them not advertising it much. They seem to be confident about their technical superiority over hype created purely by marketing.

(In other words, I consider Google technology company, Apple a marketing one.)

[+] JCordeiro|13 years ago|reply
It's these sort of things that make technology feel like magic.