Categories
2020 3D Information Augmented Reality Social and Societal

Reality TV in 3D

I haven't regularly watched television in nearly 8 years. There isn't a television in my home or my office and when I'm near a television, it doesn't occur to me to turn it on. I am seriously out of touch with what this industry has to offer but my life doesn't lack content. It is just filled with media that I'm choosing to watch or to listen to, when there's time and interest. And, I don't use a terminal that has "channels" in the old fashion broadcast television style. 

I'm certainly not alone in making media choices on a daily or hourly basis via a device other than a TV set. In fact, the whole Web 2.0 and social media movement has provided entertainment "on demand" and just-in-time informational outlets for huge segments of society. And some televisions are already Internet-connected terminals capable of much more than only showing broadcast content.

Many consumers in 2020 will be buying and regularly using devices first introduced this year at the International Consumer Electronics Show. According to an IEEE Spectrum article on the upcoming CES, television is going to be prominently featured among the 2012 edition's announcements. Television sizes and resolutions continue to grow. But this year, as in the past two years, the theme is 3D TV. Manufacturers of displays and televisions are steadily improving what they provide for those who want to watch 3D content. The problem, some believe, isn't the idea that we need 3D, or that there is a shortage of 3D content. It's the glasses. We need solutions that don't require glasses. Stream TV's glasses-free Ultra-D 3D technology is among the popular tech topics for the past three weeks.

Perhaps one of the reasons that people are enamored with 3D is that it mimics reality, the real world. If we are looking for realism, perhaps we also want realistic content. Reality TV has grown and is not showing signs of going away. Unfortunately, from what I've seen of it, Reality TV doesn't have much to do with real Reality.

Reality TV in 3D is getting closer to what might be possible using one of the other hot segments on display at CES this year: eyewear for hands-free Augmented Reality.

Who says glasses are the problem? I and several billion other people wear glasses daily. At least a dozen companies, including Vuzix, the most well-known name in the segment, will show their latest eyewear at CES. While it will show off a lightweight dual-screen model, Vuzix has already disclosed that the next optical see-through displays they will release in 2012 will be monocles, not as shown in this illustration.

In addition to using such appliances to display digital content over the real world (AR), extending them with a couple of cameras (particularly dual cameras for stereoscopic capture) could give us Reality TV in 3 dimensions with a first person point of view. Imagine that you could tune into the life of a famous person or an animal, seeing the world from their eyes. Could this be reminiscent of the feeling we get from following prolific people on Facebook or bloggers ("life bloggers")? Only, in the scenario I'm proposing, text and photos would be replaced with a live stereo video and audio feed. Will this redefine what people consider to be entertaining or boring?

Will this be television in 2020 or just an ordinary pair of glasses?

Categories
3D Information Business Strategy News Research & Development

London’s Imperial College

On July 27, 2011 the UK Research Council awarded a £5.9m grant to the “digital city exchange” programme of Imperial College. According to the press release, the funds are to be used to establish a new research center focusing on “smart cities” technologies.

A multidisciplinary team, involving businesses, public administrations and academia, is being put in place to use the city of London as a test bed for emerging smart cities hardware, software and processes. The article in the Financial Times very perceptively puts the focus on the following statements issued by spokesperson, David Gann, the head of the innovation and entrepreneurship group at Imperial College.

"New sensors and real-world monitoring can be combined with “cloud computing” to bring greater efficiency and new services to cities. For instance, data about peak traffic periods and local sensors could be used to avoid congestion for supermarket deliveries.

“London, with all its economic and social diversity, will be a very good place to launch some of these capabilities into new cities around the world and create new jobs and growth. The act of invention is at the point of consumption.”

Another article about the grant emphasizes, as did the press release, more of an urban planning angle.

It's very exciting to have this center establishing itself, although the size of the grant does not seem in line with the ambitions and objectives as they are described, and there should be others of its kind connecting to it as well.

Categories
Augmented Reality News Research & Development

Pittsburgh Pattern Recognition

On July 22 2011, Google acquired PittPatt, the Pittsburgh Pattern Recognition Team, a privately-held spin out of CMU Robotics.

Three questions jumped out when I learned of this acquisition.

  • Why? Doesn't Google already have face recognition technology?
    Unfortunately, based on the publicly available information, it's not clear what is new or different about PittPatt's technology. Okay, so they have an SDK. There are several possible explanations for this acquisition. Maybe the previous facial recognition technology Google had acquired with Neven Vision in August 2006 then released as part of Picasa in 3rd quarter 2008 (it appeared in Picasa as early as May 2007) was insufficient. Insufficient could mean inaccurate too often, too difficult to implement in mobile, not scalable. That doesn't seem likely.
    Maybe the difference is that the PittPatt technology was working on video as well as still images. YouTube already has a face recognition algorithm, but it is not real time. For AR it would be valuable if the face recognition and tracking performs reliably in real time.
    Another possible explanation has to do with IP. Given the people who founded PittPatt, perhaps there are some intellectual properties that Google wants for itself or to which it wants to prevent a competitor to have access.
     
  • What are the hot "nearby" properties that will get a boost in their valuation as a result of Google's purchase?
    Faces are the most important attribute we have as individuals and the human brain is hard wired to search for and identify faces. Simulating what our brains do with and for faces is a fundamental computer vision challenge. Since this is not trivial and so many applications could be powered by face recognition (and when algorithms can recognize faces, other 3D objects will not be far behind), there's always a lot of resources going into developing robust, accurate algorithms.

     

     

    Many–perhaps dozens–of commercial and academic groups continually work on facial recognition and tracking technology. Someone has certainly done the landscape analysis on this topic. One of the face recognition research groups with which I've had contact is at Idiap in Martigny, Switzerland. Led by Sebastien Marcel, this research team is focusing on the use of such highly accurate facial recognition that it can be the basis for granting access. KeyLemon is an Idiap spin off using the Idiap technology for biometric authentication to personal computers. And, there is (almost certainly) a sizable group already in Google dedicated to this topic. 
     

  • What value added services or features can emerge that are not in conflict with Google's privacy policy and haven't been thought of already/implemented by Google and others?
    This is an important question that probably has a very long and complex, multi-part answer. I suspect it has a lot to do with 3D objects. What's great about studying faces is that there are so many different ones to work with and they are plastic (distort easily). When the algorithms for detecting, recognizing and tracking faces in video are available on mobile devices, we can imagine that other naturally occurring and plastic objects would not be too far behind.

I hope Eric Schmidt is proven wrong about there not being facial recognition in the future of Google Goggles and similar applications and we see what is behind the curtain in the PittPatt acquisition!

Categories
Innovation Research & Development

3D City Models and AR

Google StreetView was certainly a trail-blazing concept and it has entered the mainstream. But it was not the first service and Google isn’t the first company that had the concept to collect data about the physical world by driving a specially equipped vehicle (with one or more cameras, high performance GPS and other sensors) through space. Decades earlier, the Jet Propulsion Laboratory worked on this concept in order to permit the vehicles landing on the moon (or other spatial bodies) to record their immediate environment. Earthmine is a pioneer not only in the capture of the real world (using designs developed by the JPL) but also to explore business models based on this data sets. What do these have in common? They proved that the ambitious goal of digitally “capturing” the real world in a form that supports navigation through the data afterwards, was possible.

As the technologies developed in these projects have evolved and become more powerful–in every dimension–and competitors have emerged based on other maturing technologies, systems are detecting the physical world at higher and higher resolutions, and the data gathered produce increasingly more accurate models at lower costs.

Instead of “manually” building up a 3D model from a 2D map and/or analog data, urban environments are being scanned, measured and modeled at an amazing speed, and at lower cost than ever before. Fascinating, but to what end?

In the AR-4-Basel project, we seek to make available to AR developers accurate 3D models in order for the digital representation of the real world to serve as the basis for higher performance AR experiences. The concept is that if a developer were able to use the model when designing experiences, or the placement of content, they would have a virtual reality in which to experiment. Then, when in the real world the user’s device with a camera would automatically extract features, such as edges of buildings, roofs, and other stationary attributes of the world, and match those with the features “seen” earlier in the digital model. The digital data would be aligned more accurately and the process of augmenting the world with the desired content would be faster.

In order to determine if this is more than just a concept, I need to find and receive the assistance of 3D city model experts. Here are a few of the sites to which I’ve been in search of such knowledge:

This process is proving to be time consuming but it might yield some results before another solution to improve AR experience quality emerges!