feat: implement rae autoencoder. (#13046)
feat: implement three RAE encoders(dinov2, siglip2, mae)
feat: finish first version of autoencoder_rae
fix formatting
make fix-copies
initial doc
fix latent_mean / latent_var init types to accept config-friendly inputs
use mean and std convention
cleanup
add rae to diffusers script
use imports
use attention
remove unneeded class
example traiing script
input and ground truth sizes have to be the same
fix argument
move loss to training script
cleanup
simplify mixins
fix training script
fix entrypoint for instantiating the AutoencoderRAE
added encoder_image_size config
undo last change
fixes from pretrained weights
cleanups
address reviews
fix train script to use pretrained
fix conversion script review
latebt normalization buffers are now always registered with no-op defaults
Update examples/research_projects/autoencoder_rae/README.md
Co-authored-by: Sayak Paul spsayakpaul@gmail.com
- Update src/diffusers/models/autoencoders/autoencoder_rae.py
Co-authored-by: Sayak Paul spsayakpaul@gmail.com
use image url
Encoder is frozen
fix slow test
remove config
use ModelTesterMixin and AutoencoderTesterMixin
make quality
strip final layernorm when converting
_strip_final_layernorm_affine for training script
fix test
add dispatch forward and update conversion script
update training script
error out as soon as possible and add comments
Update src/diffusers/models/autoencoders/autoencoder_rae.py
Co-authored-by: dg845 58458699+dg845@users.noreply.github.com
use buffer
inline
Update src/diffusers/models/autoencoders/autoencoder_rae.py
Co-authored-by: dg845 58458699+dg845@users.noreply.github.com
remove optional
_noising takes a generator
Update src/diffusers/models/autoencoders/autoencoder_rae.py
Co-authored-by: dg845 58458699+dg845@users.noreply.github.com
fix api
rename
remove unittest
use randn_tensor
fix device map on multigpu
check if the key is missing in the original state dict and only then add to the allow_missing set
remove initialize_weights
Co-authored-by: wangyuqi wangyuqi@MBP-FJDQNJTWYN-0208.local Co-authored-by: Kashif Rasul kashif.rasul@gmail.com Co-authored-by: Sayak Paul spsayakpaul@gmail.com Co-authored-by: dg845 58458699+dg845@users.noreply.github.com
版权所有:中国计算机学会技术支持:开源发展技术委员会
京ICP备13000930号-9
京公网安备 11010802032778号
🤗 Diffusers is the go-to library for state-of-the-art pretrained diffusion models for generating images, audio, and even 3D structures of molecules. Whether you’re looking for a simple inference solution or training your own diffusion models, 🤗 Diffusers is a modular toolbox that supports both. Our library is designed with a focus on usability over performance, simple over easy, and customizability over abstractions.
🤗 Diffusers offers three core components:
Installation
We recommend installing 🤗 Diffusers in a virtual environment from PyPI or Conda. For more details about installing PyTorch, please refer to their official documentation.
PyTorch
With
pip(official package):With
conda(maintained by the community):Apple Silicon (M1/M2) support
Please refer to the How to use Stable Diffusion in Apple Silicon guide.
Quickstart
Generating outputs is super easy with 🤗 Diffusers. To generate an image from text, use the
from_pretrainedmethod to load any pretrained diffusion model (browse the Hub for 30,000+ checkpoints):You can also dig into the models and schedulers toolbox to build your own diffusion system:
Check out the Quickstart to launch your diffusion journey today!
How to navigate the documentation
Contribution
We ❤️ contributions from the open-source community! If you want to contribute to this library, please check out our Contribution guide. You can look out for issues you’d like to tackle to contribute to the library.
Also, say 👋 in our public Discord channel
. We discuss the hottest trends about diffusion models, help each other with contributions, personal projects or just hang out ☕.
Popular Tasks & Pipelines
Popular libraries using 🧨 Diffusers
Thank you for using us ❤️.
Credits
This library concretizes previous work by many different authors and would not have been possible without their great research and implementations. We’d like to thank, in particular, the following implementations which have helped us in our development and without which the API could not have been as polished today:
We also want to thank @heejkoo for the very helpful overview of papers, code and resources on diffusion models, available here as well as @crowsonkb and @rromb for useful discussions and insights.
Citation