About cookies on this site Our websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising. For more information, please review your options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.
Abstract
The explosion of geo-tagged images taken from mobile devices around the world is visually capturing life at amazingly high spatial-, temporal-, and semantic-density. In places like cities, which cover only 3% of the Earth's landmass, yet account for 50% of the world's population, the density of photos averages one photo per every 18 square meters per year. In densely populated cities like New York City, London, Paris, San Francisco and Tokyo the number reaches up to 40 photos. What this means is that computers are able to learn from photos not only what is where - in terms of fixed and persistent objects like landmarks and buildings, but what happens where - in terms of dynamic and transient activities and events. The key idea is to combine semantic concept modeling of photos with geo-spatio-temporal localization to perform {it geo-semantic modeling. In this talk, we describe how different semantic activities and events can be defined using a geo-semantic ontology and used to create geo-semantic probability maps of what happens where from photos. We describe how this provides a powerful capability to learn visually from historical geo-tagged photos things like: "where is a good place to go fishing," "when and where are the street fairs in a location," "where is good vantage spot for the yearly fireworks show," or even "where to meet people with particular fashion or style". © 2013 Author.