(apologies if this is the wrong section, but I thought this is a question for developers)
Greetings all you smart, beautiful people!
I'm trying to refine a video light painting app and I'm running into problems.
Basically, I'm using the USB feed from the 7D into my computer (1056 x 704). The software routes the feed through a 1 frame loop and creates the effect of a time lapse photo.
(https://thumb.ibb.co/bHp4t7/Screen_Shot_2018_02_15_at_2_21_28_PM.png) (https://ibb.co/bHp4t7)
The problem is that I get dropouts in the video feed. I've ruled out my computer process by running a GL texture through the loop. The GL texture is seamless.
(https://thumb.ibb.co/nOLG6S/MLRocks.jpg) (https://ibb.co/nOLG6S)
(captured simultaneously with 2 cameras, the exposure difference is related to having only 1 neutral density filter)
I'm suspecting it has something to do with the way the camera outputs the live feed. Is there a way to eliminate these gaps? Forgive my ignorance... I have no idea of this process or how to describe it. Is it possible for someone to explain this process and describe what the challenges/solutions are to overcoming this? I know you folks can do this... you're so deep into the code that I'm sure there's a way. I on the other hand, have absolutely no programming skills, but I am willing to provide a little financial reward.
This function may seem a bit esoteric to consider for a feature, but if you do live VJ art... it would be a handy tool!
Thanks for listening!
lucem