Thanks again for the followup.
I am familiar with cnn training and back-propagation, and I still feel that given what the cam can currently do with the tf library / operations, compute should not be the primary barrier.
I understand that large networks are definitely compute intensive for backprop, but specifically I was interested in something akin to transfer learning or personalization, where most of the training work of a large network occurs on a stronger piece of hardware, but a few layers (less than 3) at the end (not that deep, not many neurons) could be trained on-device on the camera (distinct and modular from rest of much larger network, would not have to backprop through entire pipeline).
Of course, I defer to your much greater expertise and familiarity with the cam hardware, but I would like your opinion on the cam’s ability to train even very small networks on-device (which alone would not be complex enough to be useful, but combined in the aforementioned manner, is useful).