

Just a CSV file filled with random blendshape names and millions of values. No maya scene, no unreal asset, not even a simple fbx file that we could use to import and clean the motion in a 3rd-party-software like MotionBuilder.

how to transfer an actor's facial performance onto a metahuman, in real-time. Everything is fine-tuned to get the best of Unreal shaders and will leave you speechless 😍 I invite you to take a peek at it, plenty of resources out there.Īs a mocap director, my interest quickly shifted over the live-retarget capabilities, i.e.

Through the Quixel Bridge editor, you can create and customize a highly realistic avatar in a few clicks, and send it to your Unreal project. I recently started playing around with the mind-blowing MetaHuman (MH) feature from Unreal Engine. But I hope my notes will help you find your own solution. Note: I wrote this script as part of a work project and I am not able to share it. How to retarget facial animations recorded with the Live Link Face iPhone app onto an Unreal MetaHuman character, using MotionBuilder and Python.
