I found this article so intriguing. “A picture is worth a thousand (coherent) words: building a natural description of images".
At the same time, it shouldn’t come as totally unexpected.
After all, Google has been doing amazing things with images for a long time. Google Goggles is just an indispensible tool. Take a picture and let Googles help you decipher what it is. It’s a great replacement for a specific QR Code reader application but can be so useful in answering the questions “What is that?” or “Tell me more about that?” It’s a great tool for inquiry. In addition, Google’s Search by Image lets you get alternative views to a picture that you’ve taken or found online.
Humanity has been feeding the web images forever, tagging and describing them, and search engines are there to index them. An algorithm to describe the images just seems to me to be a natural in the progression of things.
Just the possibilities of this are kind of mind boggling. Google has a picture of my house online. I wasn’t there when they drove by but they did. Can you see comments generated like “Looks like Doug needs to cut his grass”. Or, in the next stage of commerce, identifying all the dandelions on my front lawn and sending an email to the local landscaping companies who have subscribed to a service with the message “We’ve got a live one…”
Technology serves us best when it makes out lives better.
- Immediate results from an x-ray or MRI with a detailed description not subject to human error;
- A description of a robbery suspect moments after a holdup at the local variety store;
- A description of a hit and run vehicle caught on camera, again moments after it happened.
Sometimes, we’re slow at adopting new technologies. While the process seems at its infancy now, it’s bound to mature. What will our lives be like when it does? What does this do to our privacy?
This article makes for a great classroom discussion. Can students extend the list of uses for an application like this?