I’m looking into different solutions for a very particular problem. Without going into too much detail, I need a way to use a camera (ipad, webcam, android… preferably any camera) and translate the facial movements of someone sitting in front of that camera, onto a humanoid 3D model, that is present in my VR environment (not AR).
So… to be clear. I have a VR experience already built. There is a 3D humanoid model in it. I want a person sitting in front of a camera connected to the same machine, to control the facial animations of that 3D model so that the person in the VR environment can see them. To go further, arm and hand movements would be ideal as well.
Am I barking up the wrong tree with MARS here, or should I be looking into other solutions?