Sirui-Xu/STARS

[ECCV 2022 oral] Official PyTorch implementation of the paper "Diverse Human Motion Prediction Guided by Multi-Level Spatial-Temporal Anchors"

33
/ 100
Emerging

This project helps create realistic, diverse predictions of how people might move next, given their recent past movements. It takes a short video or sequence of human poses and outputs multiple plausible future motion paths, making it useful for animating characters or analyzing human behavior. It's designed for researchers and developers working in fields like animation, robotics, and human-computer interaction.

No commits in the last 6 months.

Use this if you need to generate multiple realistic future movement possibilities for a human figure based on a short clip of their recent actions.

Not ideal if you need to predict the exact, singular future motion with high certainty, as this tool focuses on generating diverse outcomes.

human-motion-synthesis character-animation robotics-planning behavior-prediction virtual-reality
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 8 / 25

How are scores calculated?

Stars

75

Forks

5

Language

Python

License

MIT

Last pushed

Mar 19, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/Sirui-Xu/STARS"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.