Awesome
ComfyUI_Sapiens
You can call Using Sapiens to get seg,normal,pose,depth,mask maps.
Sapiens From: facebookresearch/sapiens
Notice-2024/12/05
- 因为YOLO所需的ultralytics 库其特定的8.3.41Pypi版本被植入挖矿脚本,因为pose要调用YOLO,最近通过pip安装了ultralytics,请务必按以下操作查看你的ultralytics版本 :
1、comfyUI便携包,在python_embeded目录下,打开CMD,运行 python -m pip show ultralytics
2、comfyUI安装版,打开CMD 运行 pip show ultralytics
只要版本不是8.3.41,就无需理会, - 如果ultralytics的版本是8.3.41,请务必执行pip unistall ultralytics 并清除python的Lib\site-packages\ultralytics文件夹和带ultralytics名称的文件夹
- 如果要继续使用8.3.41,可以使用pip install git+https://github.com/ultralytics/ultralytics.git 安装,当然我建议你使用其他版本。
- Because the ultralytics library required by YOLO has a specific 8.3.41Pypi version embedded in the mining script, and pose needs to call YOLO, ultralytics has recently been installed through pip. Please be sure to check your ultralytics version by following these steps:
- ComfyUI portable package, in the python_ embedded directory, open CMD and run: python -m pip show ultralytics
- ComfyUI installation version, open CMD and run : pip show ultralytics
As long as the version is not 8.3.41, there is no need to worry,
- If the version of ultralytics is 8.3.41, be sure to execute pip unisstall ultralytics and clear the ...Lib\site packages\ultralytics folder and folders with ultralytics names in Python
- If you want to continue using 8.3.41, you can use pip install git+ https://github.com/ultralytics/ultralytics.git Install, of course I suggest you use a different version.
Update-2024/12/01
- 基于COCOfullbody编码 ,单独使用pose模型时,可选pose的5种分离模式,分别是躯干,下肢,手,上肢,头部,对应选择seg_select 的编号分别是(21.torso,4.Left_foot,5.Left_Hand,6.Left_lower_arm,3.Face_Neck),这5种也可以自由组合,全选默认输出所有pose;
- Based on COCOfullbody encoding, when using the pose model alone, five separation modes of pose can be selected, namely ' torso, lower limbs, hands,lower_arm, and head.' The corresponding selection numbers for 'seg_ select' are (21. Torso, 4. Left_Foot, 5. Left_Hand,6.Left_lower_arm, 3. Face_Neck). These five modes can also be freely combined, and selecting all will output all poses by default;
previous update
- Fixed bug where SEG cannot be used to separate normals, poses, and depths, and added a button to save pose npy files;
- try add MPS support or no cuda user..
- Add model uninstallation code for easy connection to other nodes. Thanks to @lyxkilo's code, it is possible to convert the FP32 model to an FP16 model (smaller, around 1B SEG 2G),The first run will generate an fp16 model file with the same name, and there is no need to enable the fp16 generation button after generation.
- 修复无法利用SEG分离法线,姿态和深度的bug,新增保存姿态npy文件按钮;
- 加入模型卸载代码,便于连接其他节点,感谢@lyxkilo 的代码,通过他的代码可以将fp32的模型转为fp16模型(更小,1B seg 2G左右),首次运行会生成一个同名加fp16模型文件,生成后不需要再开启fp16生成按钮;
- seg选择人体部位的方式是数字加“,”,例如 2,1,11,注意逗号是英文符号。
1.Installation
In the ./ComfyUI /custom_node directory, run the following:
git clone https://github.com/smthemex/ComfyUI_Sapiens
2.requirements
pip install -r requirements.txt
If some modules missing, please pip install #ultralytics yolov8
3.checkpoints
3.1 base :(choice repo_id or ckpt_name)
- 3.1.1 #sapiens
only support torchscript version now,you can choice 1b,0.3b,0.6b,2b,do not changge ckpt's name!!!
只支持torchscript的版本,但是有多种模型可选,最好质量的是1b或者2B,如果模型选择全是none,会自动下载一个1B的seg,下载后不要改模型名字;
seg
pose
depth
normal
├── ComfyUI/models/sapiens/
| ├── seg/sapiens_1b_goliath_best_goliath_mIoU_7994_epoch_151_torchscript.pt2
| ├── pose/sapiens_1b_goliath_best_goliath_AP_639_torchscript.pt2
| ├── normal/sapiens_0.3b_normal_render_people_epoch_66_torchscript.pt2
| ├── depth/sapiens_1b_render_people_epoch_88_torchscript.pt2
- 3.1.2 yolo # if using pose
yolo是pose必须的,官方的太复杂,不如yolo好用;
yolov8m
├── ComfyUI/models/sapiens/
| ├── yolov8m.pt
4 Example
seg body
- 可以选人体部位27种类型 (Latest version)
- only pose 仅使用pose模型分离躯干;
Citation
- Using some ibaiGorordo's codes from ibaiGorordo
- UsingJaykumaran's codes from Jaykumaran
facebookresearch/sapiens
@article{khirodkar2024sapiens,
title={Sapiens: Foundation for Human Vision Models},
author={Khirodkar, Rawal and Bagautdinov, Timur and Martinez, Julieta and Zhaoen, Su and James, Austin and Selednik, Peter and Anderson, Stuart and Saito, Shunsuke},
journal={arXiv preprint arXiv:2408.12569},
year={2024}
}