The version 3.4 update focused on body movement and intelligent automation:
Refined algorithms provided more accurate matching between mouth shapes (visemes) and audio, resulting in higher-quality dialogue sequences.
This expansion allows characters' legs to respond naturally to movement, enabling actions like squatting, jumping, and bending without manual frame-by-frame adjustments. Adobe Character Animator 2020 v3.4
Powered by Adobe Sensei AI, this tool automatically generates head and body movements based on the tone and inflection of a recorded audio track.
This workflow improvement allows users to consolidate multiple lip-sync or trigger takes into a single, manageable track on the timeline. Core Functionality The version 3
Uses your webcam and microphone to track facial expressions and voice in real-time, instantly mapping them onto a 2D puppet.
Adobe Character Animator 2020 (v3.4), released in , was a major update that introduced sophisticated automation tools to the performance-based animation platform. This version bridged the gap between manual rigging and AI-driven movement, making it significantly easier to create expressive 2D characters. Key Features of v3.4 This version bridged the gap between manual rigging
Allows you to open a scene directly in After Effects or Premiere Pro with a live connection, so changes made in Character Animator update automatically in your video project. Minimum System Requirements (2020 Release)