(apologies if this is the wrong section, but I thought this is a question for developers)
Greetings all you smart, beautiful people!
I’m trying to refine a video light painting app and I’m running into problems.
Basically, I’m using the USB feed from the 7D into my computer (1056 x 704). The software routes the feed through a 1 frame loop and creates the effect of a time lapse photo.

The problem is that I get dropouts in the video feed. I’ve ruled out my computer process by running a GL texture through the loop. The GL texture is seamless.

(captured simultaneously with 2 cameras, the exposure difference is related to having only 1 neutral density filter)
I’m suspecting it has something to do with the way the camera outputs the live feed. Is there a way to eliminate these gaps? Forgive my ignorance… I have no idea of this process or how to describe it. Is it possible for someone to explain this process and describe what the challenges/solutions are to overcoming this? I know you folks can do this… you’re so deep into the code that I’m sure there’s a way. I on the other hand, have absolutely no programming skills, but I am willing to provide a little financial reward.
This function may seem a bit esoteric to consider for a feature, but if you do live VJ art… it would be a handy tool!
Thanks for listening!
lucem