Two robots, Vincent & Emily, are connected to each other as if deeply in love: where at the heights of romance, every motion, utterance, or external influence is shared in an acutely empathic, highly attuned ’emotional’ response:
The creation of German artists Nikolas Schmid-Pfähler and Carolin Liebl, the robots take in sound and motion data–from each other and from spectators– via sensors, which causes them to react–via gears and motors–with certain expressions. Shown in a gallery and open to the interaction of visitors, the project aims to explore the ideal of the human couple by distilling it into a more basic form. Simple lines represent bodies. Reacting to inputs replaces complicated decision-making.
Like in any relationship, miscommunication is a factor – so an intimate moment can lead to conflict, and eventual resolution. This gives a certain texture to their ‘dance of love’ that makes it hard not to anthropomorphise, or indeed relate to!
A simple plus/minus 1V signal from a beat-heavy song can be used to stimulate the motor neurons in the leg of a cockroach. This is an example of such.
Using setups like this can help us understand how neurons and muscles work, and can assist us in understanding our own nervous systems.
I’ll tell you what else this helped me understand: we’ve reached such mastery of nature that we’re now just having fun with it. I’m not sure if this is good or bad, but the above example is certainly a bit macabre.
The guys at GelSight are on to something big – or at least, magnified.
Their specially designed rubber lens lets one see details as small as two microns thick, through their patent-pending and newly perfected approach. This video demonstrates how it all works:
And this video shows off the extent of GelSight’s sensitivity:
The stuff looks really cheap to produce, but with a wide range of applications, especially for ballistics or engineering. Personally, I’d buy some just to use as a desk toy, alongside my Intelligent Putty and other cool shit.
It’s the simple combination of rubber and reflective paint that makes GelSight’s patent so valuable, albeit so simple. Yet it took two MIT alumni to spot the gap and to productize. Goes to show what other great combinations are still out there waiting to be discovered!
In summary, Mobile AR has many paths leading to it. It is this convergence of various paths that makes a true historical appraisal of this technology difficult to achieve. However, I have highlighted facets of its contributing technologies that assist in the developing picture of the implications that Mobile AR has in store. A hybridisation of a number of different technologies, Mobile AR embodies the most gainful properties of its three core technologies: This analyst sees Mobile AR as a logical progression from VR, but recognises its ideological rather than technological founding. The hardware basis of Mobile AR stems from current mobile telephony trends that exploit the growing capabilities of Smartphone devices. The VR philosophy and the mobile technology are fused through the Internet, the means for enabling context-based, live-updating content, and housing databases of developer-built and user-generated digital objects and elements, whilst connecting users across the world.
I have shown that where the interest in VR technologies dwindled due to its limited real-world applicability, Mobile Internet also lacks in comparison to Mobile AR and its massive scope for intuitive, immersive and realistic interpretations of digital information. Wearable AR computing shares VR’s weaknesses, despite keeping the user firmly grounded in physical reality. Mobile AR offers a solution that places the power of these complex systems into a mobile telephone: the ubiquitous technology of our generation. This new platform solves several problems at once, most importantly for AR developers and interested Blue-chip parties, market readiness. Developing for Mobile AR is simply the commercially sensible thing to do, since the related industries are already making the changes required for its mass-distribution.
Like most nascent technologies, AR’s success depends on its commercial viability and financial investment, thus most sensible commercial developers of AR technologies are working on projects for the entertainment and advertising industries, where their efforts can be rewarded quickly. These small-scale projects are often simple in concept, easily grasped and thus not easily forgotten. I claim here that the first Mobile AR releases will generate early interest in the technology and entertainment markets, with the effect that press reportage and word-of-mouth behaviour assist Mobile AR’s uptake. I must be careful with my claims here however, since there is no empirical evidence to suggest that this will occur for Mobile AR. Looking at the emergence of previous technologies, however, the Internet and mobile telephony grew rapidly and to massive commercial success thanks to some strong business models and advancements in their own supporting technologies. It is strongly hoped by developers like Gameware and T-Immersion that Mobile AR can enjoy this same rapid lift-off. Both technologies gained prominence once visible in the markets thanks to a market segment called early adopters. This important group gathers their information from specialist magazine sources and word of mouth. Mobile AR developers would do well to recognise the power of this group, perhaps by offering shareware versions of their AR software that encourage a form of viral transmission that exploit text messaging.
Gameware have an interesting technique for the dissemination of their HARVEE software. They share a business interest with a Bluetooth technology firm, which has donated a prototype product the Bluetooth Push Box, which scans for local mobile devices and automatically sends files to users in acceptance. Gameware’s Push Box sends their latest demo to all visitors to their Cambridge office. This same technology could be placed in public places or commercial spaces to offer localised AR advertising, interactive tourist information, or 3D restaurant menus, perhaps.
Gameware, through its Nokia projects and HARVEE development program is well placed to gain exposure on the back of a market which is set to explode as mobile offerings become commercially viable, ‘social’, powerful, multipurpose and newsworthy. Projects like HARVEE are especially interesting in terms of their wide applicability and mass-market appeal. It is its potential as a revolutionary new medium that inspires this very series.
I have been aided in this series by a connection with Gameware Development Limited, a Cambridge-based commercial enterprise working in the entertainment industry. Gameware was formed in May 2003 from Creature Labs Ltd, developing for the PC games market which produced the market leading game in Artificial Intelligence (AI), Creatures. When Gameware was formed, a strategic decision was made to move away from retail products and into the provision of technical services. They now work within the Broadcasting and Mobile Telephony space in addition to the traditional PC market. I use this business as a platform to launch into a discussion of the developments current and past that could see AR become a part of contemporary life, and just why AR is such a promising technology.
Gameware’s first explorations into AR came when they were commissioned by the BBC to develop an AR engine and software toolkit for a television show to be aired on the CBBC channel. The toolkit lets children build virtual creatures or zooks at home on their PCs which are uploaded back to the BBC and assessed:
The children with the best designs are then invited to the BAMZOOKi studio to have their virtual creatures compete against each other in a purpose-built arena comprised of real and digital elements. The zooks themselves are not real, of course, but the children can see silhouettes of digital action projected onto the arena in front of them. Each camera has an auxiliary camera pointed at AR markers on the studio ceiling, meaning each camera’s exact location in relation to the simulated events can be processed in real time. The digital creatures are stitched into the footage, and are then navigable and zoomable as if they were real studio elements. No post-production is necessary. BAMZOOKi is currently in its fourth series, with repeats aired daily:
BAMZOOKi has earned Childrens BBC some of its highest viewing figures (up to 1.2 million for the Monday shows on BBC1 and around 100,000 for each of the 20 episodes shown on digital Children’s BBC), which represents a massive milestone for AR and its emergence as a mainstream media technology. The evidence shows that there is a willing audience already receptive to contemporary AR applications. Further to the viewing figures the commercial arm of the BBC, BBC Worldwide, is in talks to distribute the BAMZOOKi format across the world, with its AR engine as its biggest USP. Gameware hold the rights required to further develop their BAMZOOKi intellectual property (IP), and are currently working on a stripped down version of their complex AR engine for the mobile telephony market.
I argue, however, that Broadcast AR is not the central application of AR technologies, merely an enabler for its wider applicability in other, more potent forms of media. Mobile AR offers a new channel of distribution for a variety of media forms, and it is its flexibility as a platform that could see it become a mainstream medium. Its successful deployment and reception is reliant on a number of cooperating factors; the innovation of its developers and the quality of the actual product being just part of the overall success the imminent release.
As well as their AR research, Gameware creates innovative digital games based on their Creatures AI engine. They recently produced Creebies; a digital game for Nokia Corp. Creebies is one of the first 3D games which incorporates AI for mobile phones. Gameware’s relationship with Nokia was strengthened when Nokia named them Pro-Developers. This is a title that grants Gameware a certain advantage: access to prototype mobile devices, hardware specifications, programming tools and their own Symbian operating system (Symbian OS) for mobile platforms. It was this development in combination with their experiences with BAMZOOKi and a long-standing collaboration with Cambridge University which has led to the idea for their HARVEE project. HARVEE stands for Handheld Augmented Reality Virtual Entertainment Engine.
Their product allows full 3D virtual objects to co-exist with real objects in physical space, viewed through the AR Device, which are animated, interactive and navigable, meaning the software can make changes to the objects as required, providing much space for interesting digital content. The applications of such a tool range from simple toy products; advertising outlets; tourist information or multiplayer game applications; to complex visualisations of weather movements; collaborating on engineering or architectural problems; or even implementing massive city-wide databases of knowledge where users might ‘tag’ buildings with their own graphical labels that might be useful to other AR users. There is rich potential here.
In HARVEE, Gameware attempt to surmount the limitations of current AR hardware in order to deliver the latest in interactive reality imaging to a new and potentially huge user base. Indeed, Nokia’s own market research suggests that AR-capable Smartphones will be owned by 25% of all consumers by 2009 (Nokia Research Centre Cambridge, non-public document). Mobile AR of the type HARVEE hopes to achieve represents not only a significant technical challenge, but also a potentially revolutionary step in mobile telephony technologies and the entertainment industry.
Gameware’s HARVEE project is essentially the creation of an SDK (Software Development Kit) which will allow developers to create content deliverable via their own Mobile AR applications. The SDK is written with the developer in mind, and does the difficult work of augmenting images and information related to the content. This simple yet flexible approach opens up a space for various types of AR content created at low cost for developers and end-users. I see Mobile AR’s visibility on the open market the only impediment to its success, and I believe that its simplicity of concept could see it become a participatory mass-medium of user-generated and mainstream commercial content.
Presently, most AR research is concerned with live video imagery and it’s processing, which allows the addition of live-rendered 3D digital images. This new augmented reality is viewable through a suitably equipped device, which incorporates a camera, a screen and a CPU capable of running specially developed software. This software is written by specialist software programmers, with knowledge of optics, 3D-image rendering, screen design and human interfaces. The work is time consuming and difficult, but since there is little competition in this field, the rare breakthroughs that do occur are as a result of capital investment: something not willingly given to developers of such a nascent technology.
What is exciting about AR research is that once the work is done, its potential is immediately seen, since in essence it is a very simple concept. All that is required from the user is their AR device and a real world target. The target is an object in the real world environment that the software is trained to identify. Typically, these are specially designed black and white cards known as markers:
These assist the recognition software in judging viewing altitude, distance and angle. Upon identification of a marker, the software will project or superimpose a virtual object or graphical overlay above the target, which becomes viewable on the screen of the AR device. As the device moves, the digital object orients in relation to the target in real-time:
The goal of some AR research is to free devices from markers, to teach AR devices to make judgements about spatial movements without fixed reference points. This is the cutting edge of AR research: markerless tracking. Most contemporary research, however, uses either marker-based or GPS information to process an environment.
Marker-based tracking is suited to local AR on a small scale, such as the Invisible Train Project (Wagner et al., 2005) in which players collaboratively keep virtual trains from colliding on a real world toy train track, making changes using their touch-screen handheld computers:
GPS tracking is best applied to large scale AR projects, such as ARQuake (Thomas et al, 2000), which exploits a scale virtual model of the University of Adelaide and a modified Quake engine to place on-campus players into a ‘first-person-shooter’. This application employs use of a headset, wearable computer, and a digital compass, which offer the effect that enemies appear to walk the corridors and ‘hide’ around corners. Players shoot with a motion-sensing arcade gun, but the overall effect is quite crude:
More data input would make the game run smoother and would provide a more immersive player experience. The best applications of AR will exploit multiple data inputs, so that large-scale applications might have the precision of marker-based applications whilst remaining location-aware.
Readers of this blog will be aware that AR’s flexibility as a platform lends applicability to a huge range of fields:
Current academic work uses AR to treat neurological conditions: AR-enabled projections have successfully cured cockroach phobia in some patients (Botella et al., 2005);
There are a wide range of civic and architectural uses: Roberts et al. (2002) have developed AR software that enables engineers to observe the locations of underground pipes and wires in situ, without the need schematics
AR offers a potentially rich resource to the tourism industry: the Virtuoso project (Wagner et al., 2005) is a handheld computer program that guides visitors around an AR enabled gallery, providing additional aural and visual information suited to each artefact;
The first commercial work in the AR space was far more playful, however: AR development in media presentations for television has led to such primetime projects as Time Commanders (Lion TV for BBC2, 2003-2005) in which contestants oversee an AR-enabled battlefield, and strategise to defeat the opposing army, and FightBox (Bomb Productions for BBC2, 2003) in which players build avatars to compete in an AR ‘beat-em-up’ that is filmed in front of a live audience; T-Immersion (2003- ) produce interactive visual installations for theme parks and trade expositions; other work is much more simple, in one case the BBC commissioned an AR remote-control virtual Dalek meant for mobile phones, due for free download from BBC Online:
The next entry in this series is a case study in AR development. If you haven’t already done so, please follow me on Twitter or grab an RSS feed to be alerted when my series continues.
Augmented Reality (AR) is a theme of computer research which deals with a combination of real world and computer generated data. AR is just one version of a Mixed Reality (MR) technology, where digital and real elements are mixed to create meaning. In essence AR is any live image that has an overlay of information that augments the meaning of these images.
Digital graphics are commonly put to work in the entertainment industry, and ‘mixing realities’ is a common motif for many of today’s media forms. There are varying degrees to which The Real and The Virtual can be combined. This is illustrated in my Mixed Reality Scale:
This is a simplified version of Milgram and Kishino’s (1994) Virtuality Continuum; simplified, because their research is purely scientific, without an explicit interest in media theory or effects, therefore not wholly applicable to my analysis. At the far left of my Mixed Reality Scale lies The Real, or physical, every-day experiential reality. For the longest time we lived solely in this realm. Then, technological innovation gave rise to the cinema, and then television. These media are located one step removed from The Real, a step closer to The Virtual, and can be considered a window on another world. This world is visually similar to our own, a fact exploited by its author to narrate believable, somewhat immersive stories. If willing, the viewer is somewhat ‘removed’ from their grounding here in physical reality, allowing them to participate in the construction of a sculpted, yet static existence. The viewer can only observe this contained reality, and cannot interact with it, a function of the viewing apparatus.
Later advancements in screen media technologies allowed the superimposition of graphical information over moving images. These were the beginnings of AR, whereby most of what is seen is real with some digital elements supplementing the image. Indeed, this simple form of AR is still in wide use today, notably in cases where extra information is required to make sense of a subject. In the case of certain televised sports, for example, a clock and a scoreboard overlay a live football match, which provides additional information that is useful to the viewer. Television viewers are already accustomed to using information that is displayed in this way:
More recently, computing and graphical power gave designers the tools to build wholly virtual environments. The Virtual is a graphical representation of raw data, and the furthest removed from physical reality on my Mixed Reality Scale. Here lies the domain of Virtual Reality (VR), a technology that uses no real elements except for the user’s human senses. The user is submersed in a seemingly separate reality, where visual, acoustic and sometimes haptic feedback serve to transpose them into this artificial, yet highly immersive space. Notice the shift from viewer to user: this is a function of the interactivity offered by digital space. VR was the forerunner to current AR research, and remains an active realm of academic study.
Computer graphics also enhanced the possibilities offered by television and cinema, forging a new point on the Mixed Reality Scale. I refer to the Augmented Virtuality (AV) approach, which uses mainly digital graphics with some real elements superimposed. For example, a newsreader reporting from a virtual studio environment is one common application. I position AV one step closer towards The Virtual to reflect the ratio of real to virtual elements:
There is an expansive realm between AV and VR technologies, media which offer the user wholly virtual constructions that hold potential for immersion and interactivity. I refer to the media of video games and desktop computers. Here the user manipulates visually depicted information for a purpose. These media are diametrically opposed to their counterpart on my scale, the cinema and television, because they are windows this time into a virtual world, actively encouraging (rather than denying) user interactivity to perform their function. Though operating in virtuality, the user remains grounded in The Real due to apparatus constraints.
Now, further technological advancements allow the fusion of real and virtual elements in ways not previously possible. Having traversed our way from The Real to The Virtual, we have now begun to make our way back. We are making a return to Augmented Reality, taking with us the knowledge to manipulate wholly virtual 3D objects and the computing power to integrate digital information into live, real world imagery. AR is deservedly close to The Real on my scale, because it requires physicality to function. This exciting new medium has the potential to change the way we perceive our world, forging a closer integration between our two binary worlds. It is this potential as an exciting and entirely new medium that has driven me to carry out the following work.
To begin, I address the science behind AR and its current applications. Next, I exploit an industry connection to inform a discussion of AR’s development as an entertainment medium. Then, I construct a methodology for analysis from previous academic thought on emergent technologies, whilst addressing the problems of doing so. I use this methodology to locate AR in its wider technologic, academic, social and economic context. This discussion opens ground for a deeper analysis of AR’s potential socio-cultural impact, which makes use of theories of media and communication and spatial enquiry. I conclude with a final critique that holds implications for the further analysis of Mixed Reality technology.