Motion Capture
Motion Capture in ACT3 AI lets you animate digital actors using real human performance. Record movement through a webcam, smartphone, or professional mocap hardware, then transfer natural body motions and facial expressions directly into your scenes and shots. See add mocap to an actor for the step-by-step workflow.
This feature ensures AI-generated characters retain the realism, nuance, and emotional depth that comes from human performance.
How Motion Capture Works
ACT3 AI's motion capture system captures movement data from your recording device and maps it to a digital actor's skeleton and facial rig. The platform includes AI smoothing to clean up jitter and inconsistencies automatically.
The result is a digital actor that moves and emotes like a real person — not a hand-animated character or a stiff AI-generated figure.
Capture Types
Full-Body Motion Capture — Records body movements including walking, running, gestures, and physical interactions. Use this for character blocking, action sequences, and expressive physical performance.
Facial Capture — Records facial expressions and lip movements, synchronized with dialogue. Use this for close-up dialogue scenes where emotional nuance matters.
Hybrid Capture — Combines both body and facial capture in a single session. Requires a setup with good camera placement for both body and face.
Supported Input Devices
- Smartphone camera — Compatible with the ACT3 AI mobile companion app
- Webcam — Built-in or USB, minimum 720p resolution
- Professional mocap suits — Compatible with industry-standard formats
- iPhone Face ID — Supported for high-quality facial capture on Apple devices
Better hardware produces better results, but a modern smartphone captures usable performance data for most projects.
Recording a Capture Session
- Open the Editor and select the Motion Capture Panel
- Choose the capture type: body, face, or hybrid
- Connect or select your recording device
- Position yourself or your performer in good lighting with a clear background
- Start the recording and perform the action or dialogue
- Stop the recording when done
- Review the captured motion data in the preview
Tips for good capture:
- Use good lighting with clear contrast between the performer and background
- Keep backgrounds clean and uncluttered
- Record multiple takes for complex sequences — you can select the best one
- Ensure the full body is visible for body capture; frame from shoulders up for facial capture
Applying Motion Data to a Digital Actor
- After recording, assign the captured motion to a specific digital actor
- Preview the result in the Timeline or Top-Down View
- Use AI Smoothing to clean up any jitter or artifacts
- Adjust timing and intensity if needed
- The digital actor now performs the captured movement in all renders
AI Smoothing
ACT3 AI automatically applies motion refinement to captured data:
- Reduces jitter from minor recording inconsistencies
- Smooths transitions between keyframes
- Corrects impossible joint angles
- Normalizes movement to match the actor's physical proportions
You can adjust the smoothing intensity from subtle (preserves more raw performance data) to strong (more polished but less naturalistic).
Applications
- Dialogue scenes requiring genuine facial emotion and lip sync
- Action sequences with specific choreography
- Group performances for crowd or ensemble shots
- Prototype action sequences quickly before refining
- Matching a performer's specific movement style to a recurring character
Best Practices
- Wear fitted clothing for body capture to ensure accurate tracking
- Record in a well-lit environment with no flickering lights
- Record multiple short takes rather than one long continuous session
- Combine facial capture with TTS voice tracks for fully synchronized performances — see review lipsync to check the result
- Test with draft renders before committing to final quality
Troubleshooting
Motion tracking is erratic — Improve lighting, simplify the background, or wear higher-contrast clothing.
Facial expressions not registering — Ensure your face is well-lit and fully visible in the camera frame, with no obstructions.
Capture session won't start — Check camera permissions in your browser or operating system settings.
Performance looks stiff after smoothing — Reduce the AI smoothing intensity to preserve more of the original raw motion.
NVIDIA Audio2Face Integration
ACT3 AI uses NVIDIA Audio2Face to drive facial performance from audio. When you attach TTS or recorded dialogue to a character, Audio2Face analyzes the audio waveform and generates realistic facial expressions and mouth movement synchronized to the speech.
Audio2Face produces results that go beyond simple lip sync — it generates natural blinking, brow movement, and emotional expression changes that follow the emotional quality of the speech, not just the phonemes. This results in characters that look genuinely engaged rather than mechanically mouthing words.
Marker-Less Motion Capture
ACT3 AI supports marker-less full-body motion capture from standard video footage. You do not need a motion capture suit or specialized hardware:
- Record your performer with a phone camera or webcam
- Upload the video to the Motion Capture Panel
- ACT3 AI extracts pose and facial data from the video using computer vision
- The extracted data is applied to your digital actors in Blender
- Sync the Blender scene back to ACT3 AI via the Hero Blender Sync integration
This workflow makes professional-level motion capture accessible without expensive equipment.
iPhone Motion Capture
iPhones with Face ID (TrueDepth sensor) or LiDAR provide higher-accuracy facial capture than standard cameras. Use the ACT3 AI mobile companion app to record directly and export the animation data to Blender, then sync it back to your ACT3 AI project.