Geof Barrows of Centeye sent me some raw sensors so I can develop for them. These are great sensors due to inpart there flexibility, low external part count (you can see the board has two capacitors), and the use of chip on board technology. They are fairly easy to use.
As of right now, I am able to read the image in log mode. Once I get a lens, then I can start some image processing code. First full image (8 bit greyscale):
I am streaming pixel data via a wireless modem as the USB connection on the XMOS is VERY slow, limiting the frame rate I can use it at. I am grabbing one frame in around 32 milliseconds, but I should be able to bring that down with tuning of delay values. The whole image only takes 32K RAM, meaning I can work with the whole image, and have 32K RAM for processing capabilities. Final plan consists of a scalable architecture with add on modules to support any number of sensors. The sensors will be trained via IMU.
In the world of IMU's, I ordered another board as the other was messed up beyond recovery (port mappings). I also messed up this time (Flipped two ports...), but it is usable.
I am using Bill's DCM algorithm, and I need to tune the PI values. Current output isn't very usable: