HumanGif: Single-View Human Diffusion with Generative Prior
Paper • 2502.12080 • Published • 1
How to use Sony/humangif with Diffusers:
pip install -U diffusers transformers accelerate
import torch
from diffusers import DiffusionPipeline
# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("Sony/humangif", dtype=torch.bfloat16, device_map="cuda")
prompt = "Astronaut in a jungle, cold color palette, muted colors, detailed, 8k"
image = pipe(prompt).images[0]import torch
from diffusers import DiffusionPipeline
# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("Sony/humangif", dtype=torch.bfloat16, device_map="cuda")
prompt = "Astronaut in a jungle, cold color palette, muted colors, detailed, 8k"
image = pipe(prompt).images[0]HumanGif models are the official checkpoints for ther paper "HumanGif: Single-View Human Diffusion with Generative Prior". HumanGif can generate novel view and novel pose human images from a single input conditioned on human poses and camera poses. In this repository, we offer the codes for inference of the model. For detailed information, please refer to the paper.
The model is intended for research purposes only. Possible research areas and tasks include
Excluded uses are described below.
Note: This section is taken from the CreativeML Open RAIL-M HumanGif models.
Use Restrictions
You agree not to use the Model or Derivatives of the Model:
@article{hu2025humangif,
title={HumanGif: Single-View Human Diffusion with Generative Prior},
author={Hu, Shoukang and Narihira, Takuya and Fukuda, Kazumi and Sawata, Ryosuke and Shibuya, Takashi and Mitsufuji, Yuki},
journal={arXiv preprint arXiv:2502.12080},
year={2025}
}