📘Documentation | 🛠️Installation | 👀Model Zoo | 📜Papers | 🆕Update News | 🤔Reporting Issues
English | 简体中文
MMPose is an open-source toolbox for pose estimation based on PyTorch. It is a part of the OpenMMLab project.
The master branch works with PyTorch 1.6+.
mmpose.demo.mp4
Major Features
-
Support diverse tasks
We support a wide spectrum of mainstream pose analysis tasks in current research community, including 2d multi-person human pose estimation, 2d hand pose estimation, 2d face landmark detection, 133 keypoint whole-body human pose estimation, 3d human mesh recovery, fashion landmark detection and animal pose estimation. See Demo for more information.
-
Higher efficiency and higher accuracy
MMPose implements multiple state-of-the-art (SOTA) deep learning models, including both top-down & bottom-up approaches. We achieve faster training speed and higher accuracy than other popular codebases, such as HRNet. See benchmark.md for more information.
-
Support for various datasets
The toolbox directly supports multiple popular and representative datasets, COCO, AIC, MPII, MPII-TRB, OCHuman etc. See dataset_zoo for more information.
-
Well designed, tested and documented
We decompose MMPose into different components and one can easily construct a customized pose estimation framework by combining different modules. We provide detailed documentation and API reference, as well as unittests.
-
2022-10-14: MMPose v1.0.0rc0 is released. Major updates include:
-
2022-09-01: MMPose v1.0.0b0 is released!
- This release introduced major refactoring to MMPose towards better performance, extensibility and user-friendliness.
- Built upon a brand new and flexible training & test engine, which is still in progress. Welcome to try according to the documentation.
- There are BC-breaking changes. Please check the migration tutorial.
- The beta and release candidate versions will last until the end of 2022, and during the release candidate, we will develop on the
1.x
branch. And we will still maintain 0.x version still at least the end of 2023.
Below are quick steps for installation:
conda create -n open-mmlab python=3.8 pytorch==1.10.1 torchvision==0.11.2 cudatoolkit=11.3 -c pytorch -y
conda activate open-mmlab
pip install openmim
git clone -b 1.x https://github.com/open-mmlab/mmpose.git
cd mmpose
mim install -e .
Please refer to installation.md for more detailed installation and dataset preparation.
We provided a series of tutorials about the basic usage of MMPose for new users:
- About Configs
- Add New Dataset
- Keypoint Encoding & Decoding
- Inference with Existing Models
- Train and Test
- Visualization Tools
- Other Useful Tools
Results and models are available in the README.md of each method's config directory. A summary can be found in the Model Zoo page.
Supported algorithms:
- DeepPose (CVPR'2014)
- CPM (CVPR'2016)
- Hourglass (ECCV'2016)
- SimpleBaseline3D (ICCV'2017)
- Associative Embedding (NeurIPS'2017)
- SimpleBaseline2D (ECCV'2018)
- DSNT (ArXiv'2021)
- HRNet (CVPR'2019)
- IPR (ECCV'2018)
- VideoPose3D (CVPR'2019)
- HRNetv2 (TPAMI'2019)
- MSPN (ArXiv'2019)
- SCNet (CVPR'2020)
- HigherHRNet (CVPR'2020)
- RSN (ECCV'2020)
- InterNet (ECCV'2020)
- VoxelPose (ECCV'2020)
- LiteHRNet (CVPR'2021)
- ViPNAS (CVPR'2021)
- Debias-IPR (ICCV'2021)
- SimCC (ECCV'2022)
Supported techniques:
- FPN (CVPR'2017)
- FP16 (ArXiv'2017)
- Wingloss (CVPR'2018)
- AdaptiveWingloss (ICCV'2019)
- DarkPose (CVPR'2020)
- UDP (CVPR'2020)
- Albumentations (Information'2020)
- SoftWingloss (TIP'2021)
- RLE (ICCV'2021)
Supported datasets:
- AFLW [homepage] (ICCVW'2011)
- sub-JHMDB [homepage] (ICCV'2013)
- COFW [homepage] (ICCV'2013)
- MPII [homepage] (CVPR'2014)
- Human3.6M [homepage] (TPAMI'2014)
- COCO [homepage] (ECCV'2014)
- CMU Panoptic [homepage] (ICCV'2015)
- DeepFashion [homepage] (CVPR'2016)
- 300W [homepage] (IMAVIS'2016)
- RHD [homepage] (ICCV'2017)
- CMU Panoptic HandDB [homepage] (CVPR'2017)
- AI Challenger [homepage] (ArXiv'2017)
- MHP [homepage] (ACM MM'2018)
- WFLW [homepage] (CVPR'2018)
- PoseTrack18 [homepage] (CVPR'2018)
- OCHuman [homepage] (CVPR'2019)
- CrowdPose [homepage] (CVPR'2019)
- MPII-TRB [homepage] (ICCV'2019)
- FreiHand [homepage] (ICCV'2019)
- Animal-Pose [homepage] (ICCV'2019)
- OneHand10K [homepage] (TCSVT'2019)
- Vinegar Fly [homepage] (Nature Methods'2019)
- Desert Locust [homepage] (Elife'2019)
- Grévy’s Zebra [homepage] (Elife'2019)
- ATRW [homepage] (ACM MM'2020)
- Halpe [homepage] (CVPR'2020)
- COCO-WholeBody [homepage] (ECCV'2020)
- MacaquePose [homepage] (bioRxiv'2020)
- InterHand2.6M [homepage] (ECCV'2020)
- AP-10K [homepage] (NeurIPS'2021)
- Horse-10 [homepage] (WACV'2021)
Supported backbones:
- AlexNet (NeurIPS'2012)
- VGG (ICLR'2015)
- ResNet (CVPR'2016)
- ResNext (CVPR'2017)
- SEResNet (CVPR'2018)
- ShufflenetV1 (CVPR'2018)
- ShufflenetV2 (ECCV'2018)
- MobilenetV2 (CVPR'2018)
- ResNetV1D (CVPR'2019)
- ResNeSt (ArXiv'2020)
- Swin (CVPR'2021)
- HRFormer (NIPS'2021)
- PVT (ICCV'2021)
- PVTV2 (CVMJ'2022)
We will keep up with the latest progress of the community, and support more popular algorithms and frameworks. If you have any feature requests, please feel free to leave a comment in MMPose Roadmap.
We appreciate all contributions to improve MMPose. Please refer to CONTRIBUTING.md for the contributing guideline.
MMPose is an open source project that is contributed by researchers and engineers from various colleges and companies. We appreciate all the contributors who implement their methods or add new features, as well as users who give valuable feedbacks. We wish that the toolbox and benchmark could serve the growing research community by providing a flexible toolkit to reimplement existing methods and develop their own new models.
If you find this project useful in your research, please consider cite:
@misc{mmpose2020,
title={OpenMMLab Pose Estimation Toolbox and Benchmark},
author={MMPose Contributors},
howpublished = {\url{https://github.com/open-mmlab/mmpose}},
year={2020}
}
This project is released under the Apache 2.0 license.
- MMEngine: OpenMMLab foundational library for training deep learning models.
- MMCV: OpenMMLab foundational library for computer vision.
- MIM: MIM installs OpenMMLab packages.
- MMClassification: OpenMMLab image classification toolbox and benchmark.
- MMDetection: OpenMMLab detection toolbox and benchmark.
- MMDetection3D: OpenMMLab's next-generation platform for general 3D object detection.
- MMRotate: OpenMMLab rotated object detection toolbox and benchmark.
- MMSegmentation: OpenMMLab semantic segmentation toolbox and benchmark.
- MMOCR: OpenMMLab text detection, recognition, and understanding toolbox.
- MMPose: OpenMMLab pose estimation toolbox and benchmark.
- MMHuman3D: OpenMMLab 3D human parametric model toolbox and benchmark.
- MMSelfSup: OpenMMLab self-supervised learning toolbox and benchmark.
- MMRazor: OpenMMLab model compression toolbox and benchmark.
- MMFewShot: OpenMMLab fewshot learning toolbox and benchmark.
- MMAction2: OpenMMLab's next-generation action understanding toolbox and benchmark.
- MMTracking: OpenMMLab video perception toolbox and benchmark.
- MMFlow: OpenMMLab optical flow toolbox and benchmark.
- MMEditing: OpenMMLab image and video editing toolbox.
- MMGeneration: OpenMMLab image and video generative models toolbox.
- MMDeploy: OpenMMLab Model Deployment Framework.