Stars
PyTorch code and models for the DINOv2 self-supervised learning method.
PixArt-α: Fast Training of Diffusion Transformer for Photorealistic Text-to-Image Synthesis
The simplest, fastest repository for training/finetuning medium-sized GPTs.
The image prompt adapter is designed to enable a pretrained text-to-image diffusion model to generate images with image prompt.
The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.
[NeurIPS 2023] Uni-ControlNet: All-in-One Control to Text-to-Image Diffusion Models
Generative Models by Stability AI
Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable…
A playbook for systematically maximizing the performance of deep learning models.
A collection of out-of-tree LLVM passes for teaching and learning
[IJCV] FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attention
Annotate better with CVAT, the industry-leading data engine for machine learning. Used and trusted by teams at any scale, for data of any scale.
Label Studio is a multi-type data labeling and annotation tool with standardized output format
Personalize Segment Anything Model (SAM) with 1 shot in 10 seconds
Implementation of "DiffFit: Unlocking Transferability of Large Diffusion Models via Simple Parameter-Efficient Fine-Tuning"
Nightly release of ControlNet 1.1
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Everything you want to know about Google Cloud TPU
Stable Diffusion web UI
Easily turn large sets of image urls to an image dataset. Can download, resize and package 100M urls in 20h on one machine.