Microsoft Research has released Information about a 3D Reconstruction App that runs on Windows Mobile. As far as the technology is concerned hardly anything new. It’s basically the same as Autodesks 123D Catch that is available on severall platforms with the actual image processing happening on the server.
The App looks very nice though, it comes with an acceleromteter supported visual guide that helps the user to capture all required images for a successfull 3D model construction.
Fraunhofer Heinrich Hertz Institute is currently working on a complete end-to-end system for the real-time acquisition and projection of immersive, high-resolution video panoramas. The approach is based on a flexible, mirror-based multi-camera rig that uses multiple HD cameras to capture very high-resolution video panoramas. These videos are then exhibited with a flexible multi-projection system in order to provide the viewers with the immersive feeling of watching the event on-site from the very best seat.
Real-time acquisition and projection of high resolution video panoramas with large field-of-view (FOV),
Capturing with flexible rig for up to five HD cameras with FOV of 150 degrees,
Full circle rig for up to 12 HD cameras with 360 degree capturing capability,
Exhibition with flexible, high-resolution multi-projection system,
Based on CineCard PCI-card that supports up to four projectors per board (and up to 16 projectors per PC),
Unlimited cascading in both vertical and horizontal direction,
Immersive viewing experience that feels like “being there”.
Several options exist to stream the picture of a webcam or the Raspberry Pi cam from the Pi. The first is using a MJPEG stream. This is the most compatible as many applications and even browsers can display such a stream.
The second one is H264. Also H264 can be encoded on the Pis GPU it has a very high latency, at least five seconds from my experience.
And last but not least you can simply pipe the video stream over netcat to transmit it to another client.
“Motion” can serve up a MJPEG stream. Apart from that is has several other features as listed below: eg. a simple motion detection. It may run in the background as Linux daemon. Here’s a guide how to get motion going with a PS3 Eye Cam.
Taking snapshots of movement
Watch multiple video devices at the same time
Watch multiple inputs on one capture card at the same time
Live streaming webcam (using multipart/x-mixed-replace)
Real time creation of mpeg movies using libraries from ffmpeg
Take automated snapshots on regular intervals
Take automated snapshots at irregular intervals using cron
Execute external commands when detecting movement (and e.g. send SMS or email)
Motion tracking (camera follow motion – special hardware required)
Feed events to a MySQL or PostgreSQL database.
Feed video back to a video4linux loopback for real time viewing
Lots of user contributed related projects with web interfaces etc.
User configurable and user defined on screen display.
Control via browser (older versions used xml-rpc)
Automatic noise and threshold control
Motion is a daemon with low CPU consumption and small memory footprint.
It’s an augmented-reality, OCR-capable translation app, but that’s a poor description. A better one would be “magic.” World Lens looks at any printed text through the iPhones camera, reads it, translates between Spanish and English. That’s pretty impressive already — it does it in real time — but it also matches the color, font and perspective of the text, and remaps it onto the image. It’s as if the world itself has been translated.