- Overview
- Features
- Architecture
- Installation
- Usage
- Configuration
- Troubleshooting
- [How to use](#How to use)
This project enables real-time pose tracking using MediaPipe and animates a 3D avatar in Unity based on the captured pose data. The system captures body movements via a webcam, processes the data with Python, and sends the pose information to Unity using UDP for smooth and responsive avatar animations.
- Real-Time Pose Detection: Utilizes MediaPipe's advanced pose estimation to detect and track body landmarks in real-time.
- Seamless Integration with Unity: Sends processed pose data to Unity via UDP, enabling smooth avatar animations.
- Customizable Avatar: Supports a customizable avatar with a hierarchical bone structure for accurate movement replication.
The system comprises two main components:
- Python Script: Captures video from the webcam, processes pose data using MediaPipe, and sends the data to Unity via UDP.
- Unity Project: Receives pose data, maps it to the avatar's bone structure, and updates the avatar's pose in real-time.
- Python 3.7+
- Unity 2021.3 LTS or later
- Webcam for pose detection
git clone https://github.com/skill-diver/Unity_MediaPipe_Action_Tracking.git
Import Unity directory in Unity.
Navigate to the Python directory and start the tracking script:
cd Python
python main.py
Run the Unity Scene Open the scene in Unity. Click the Play button in Unity to start the simulation. The character in Unity should now move in real-time, mirroring your movements captured by the webcam.