Has anyone seen this? I Have been doing some SLAM research recently and stumbled upon this: https://github.com/xdspacelab/openvslam
It is an open source SLAM (simultaneous localization and mapping) framework that advertises support for equirectangular videos like those recorded (or streamed) from a Theta. It supports monocular camera setups, where you use a single camera and stucture-from-motion to estimate the actual 6 dimensional trajectory of the camera.
To be clear, this is a REALLY cool. Projects like this make autonomous robotics MUCH more accessible. The computer vision technology they are using, ORB features for key framing, is a pretty cutting edge approach that has only really become popular in the past 4 years. Technology like this allows a robot to navigate and map its environment autonomously in real time. A single camera is an order of magnitude cheaper than the LIDAR / LIDAR+IMU / camera+IMU setups that are usually used for these kinds of things. I even found a report from someone who claims to have been able to succesfully use the software with a Theta V.
Anybody have experience with this? I am finally settled and think I might pick up a Theta to build a bot and test this out.