Photo based 3D Reconstruction on Portable Devices

Microsoft Research has released Information about a 3D Reconstruction App that runs on Windows Mobile. As far as the technology is concerned hardly anything new. It’s basically the same as Autodesks 123D Catch that is available on severall platforms with the actual image processing happening on the server.

The App looks very nice though, it comes with an acceleromteter supported visual guide that helps the user to capture all required images for a successfull 3D model construction.


2D / 3D High Res 360 Degree Camera

Fraunhofer Heinrich Hertz Institute is currently working on a complete end-to-end system for the real-time acquisition and projection of immersive, high-resolution video panoramas. The approach is based on a flexible, mirror-based multi-camera rig that uses multiple HD cameras to capture very high-resolution video panoramas. These videos are then exhibited with a flexible multi-projection system in order to provide the viewers with the immersive feeling of watching the event on-site from the very best seat.


  • Real-time acquisition and projection of high resolution video panoramas with large field-of-view (FOV),
  • Capturing with flexible rig for up to five HD cameras with FOV of 150 degrees,
  • Full circle rig for up to 12 HD cameras with 360 degree capturing capability,
  • Exhibition with flexible, high-resolution multi-projection system,
  • Based on CineCard PCI-card that supports up to four projectors per board (and up to 16 projectors per PC),
  • Unlimited cascading in both vertical and horizontal direction,
  • Immersive viewing experience that feels like “being there”.


Raspberry Pi – Webcam Streaming

Streaming an external cam

Several options exist to stream the picture of a webcam or the Raspberry Pi cam from the  Pi. The first is using a MJPEG stream. This is the most compatible as many applications and even browsers can display such a stream.

The second one is H264. Also H264 can be encoded on the Pis GPU it has a very high latency, at least five seconds from my experience.

And last but not least you can simply pipe the video stream over netcat to transmit it to another client.

1. Motion

“Motion” can serve up a MJPEG stream. Apart from that is has several other features as listed below: eg. a simple motion detection. It may run in the background as Linux daemon. Here’s a guide how to get motion going with a PS3 Eye Cam.

  • Taking snapshots of movement
  • Watch multiple video devices at the same time
  • Watch multiple inputs on one capture card at the same time
  • Live streaming webcam (using multipart/x-mixed-replace)
  • Real time creation of mpeg movies using libraries from ffmpeg
  • Take automated snapshots on regular intervals
  • Take automated snapshots at irregular intervals using cron
  • Execute external commands when detecting movement (and e.g. send SMS or email)
  • Motion tracking (camera follow motion – special hardware required)
  • Feed events to a MySQL or PostgreSQL database.
  • Feed video back to a video4linux loopback for real time viewing
  • Lots of user contributed related projects with web interfaces etc.
  • User configurable and user defined on screen display.
  • Control via browser (older versions used xml-rpc)
  • Automatic noise and threshold control
  • Motion is a daemon with low CPU consumption and small memory footprint.

2. With MJPEG streamer.

A nice guide how to build and run the open source MJPEG streamer on the pi by Miguel Grindberg.

2. As a H264 Stream with VLC via RTSP

Install VLC Player on the Pi. VLC will act as the streaming server.

sudo apt-get install vlc

Run raspivid and pipe the videostream into vlc for streaming.

raspivid -o - -t 0 -n -vf -hf -w 1280 -h 720 -fps 25 -g 100 | cvlc -vvv stream:///dev/stdin --sout '#rtp{sdp=rtsp://:8554/,mux=ts}' :demux=h264

you can now display the stream in VLC using rtsp://[Raspberry IP]l:8554/as your address.

Word Lens: Augmented Reality App Translates Street Signs Instantly

It’s an augmented-reality, OCR-capable translation app, but that’s a poor description. A better one would be “magic.” World Lens looks at any printed text through the iPhones camera, reads it, translates between Spanish and English. That’s pretty impressive already — it does it in real time — but it also matches the color, font and perspective of the text, and remaps it onto the image. It’s as if the world itself has been translated.

[via ]