Onedotzero_adventures in motion

Yesterday I went down to the BFI Southbank for the onedotzero adventures in motion exhibition.

When I got there a talk had just started “nokia: the art of open source workshop”, which showcased the development potential of the maemo platform for high end Nokia phones,  so I sat in on it not really know what to expect. It was quite interesting but did get a bit too geeky for me, talking about abstract concepts behind the design of the N900 device rather than showing how to develop for it and more of what can be/has been done with the technology.
It did inspire me to checkout and I would like to develop for such a device but am put off by it being linux based and by not having a handset to test on. I might have to consider one of the handsets for my next upgrade, especially when they get multi touch support.

The main installation at the exhibition, the “onedotzero identity” was using the N900.

onedotzero identityonedotzero identity
Here is a demo:

I had a go on the device in the evening, it was really responsive and a good concept that has given me a few commercial ideas that I might put together a prototype for if I get some time soon.

There were also a couple of other interactive pieces around the BFI:

Glowing Pathfinder Bugs by was a really nice piece, I had a chat to the guys behind it about the technology; it used a 3d camera with 2 lenses to work out the depth of the sand, and a projector. I was quite surprised that the application was built in Director, but that’s what the developer was happy with and it could access some native stuff.
Glowing Pathfinder Bugs in action:

There was a multi touch table with the program of events and some related media on it, this was the first multi touch table I have had a chance to play with. I found it a bit hard to control but it was quite responsive, think the main problem was the design of the menu. It wasn’t very stable as I did see flash player 10 crash message.

By far the least impressive and most pointless installation was “graffonic: virtual spraypaint”.  It was basically using an irLED a laser to draw a line, but the worst part was that the display and style changed every 10 seconds so you couldn’t actually draw anything, just a line that was removed from the screen as fast as it appeared.
I have seen much better thought out and produced digital graffiti installations, like a prototype by my friend and colleague Andrew Myher that uses a wii remote as a controller, this is me playing with it:
Digital wall:

I also saw an amazing film preview while I was there: “Mary and Max”, its not on general release in the UK but I would recommend seeing it, the onedotzero site describes it perfectly:
“Mary and max is an exceptional film that deals with little people and their big issues in a compassionate and entertaining way. It is funny, warm, moving and sad all at the same time, and will once and for all dispel the notion that animation can only deal with fluffy and trivial issues.”

Flash Augmented Reality Getting Started

My post yesterday about Augmented Reality raised a couple of issues which have both been solved by one blog post I found : FlarToolkit/Flash Augmented RealityGetting Started

The above post sheds light on the questions I had about the marker pattern, there is an AIR app available to create the code Mikko also give some guidance on creating your own marker.

Mikko has saved me a lot of time by rewriting the source code I was working from to make it easier to follow and he has provided comments in English.

I’m really happy now I don’t have to dig through the code and translate the Japanese using bable fish, it’s quite hard to understand.

Augmented Reality Globe Demo


Recently I’ve been trying to get my head round augmented reality in flash.
Here is a little demo of a globe…. It works quite well most of the time but does loose the image now and then.

Augmented Reality Globe Demo from Sam Hassan on Vimeo.

I have been working from the source of this demo:
And it’s quite easy to replace the object that is used once you get it all working.

I want to post some clean code with an expiation of what I think the code is doing, so hopefully ill have time to do that soon.

One thing that I have been trying to work out is how the data for the pattern is created, I’ve searched the web and had a look @ the FlarToolKit forum but it’s mostly in Japanese, so not had much luck.
All the demos I’ve seen use one of 2 patterns, so I’m thinking either is really hard to create or the people who know how to do it don’t want to let it out.
I have tried to create the data from the bitmap data of a movie clip in the library, but can’t get the values in the correct format; can anyone shed any light on this?

Amazing innovations at Adobe

This week I have seen 2 amazing bits of technologies come from the Advanced Technologies Lab at Adobe Systems.
The bad thing is you cant get your hands on them yet unless you know someone.

I really want to get my hands on both of these, Zoetrope would be so useful is so many different ways and the video Manipulation will be fun at the start but I’m sure it will have some commercial uses.

Interactive Video Object Manipulation: just amazing video tracking and manipulation

Interactive Video Object Manipulation from Dan Goldman on Vimeo.

Zoetrope : the time machine for the web
Technology Review:
Video demo: