-
Salesforce Research
- Palo Alto
- https://azshue.github.io/
Stars
My learning notes/codes for ML SYS.
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
Solve puzzles. Improve your pytorch.
Official Repo for Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learning
verl: Volcano Engine Reinforcement Learning for LLMs
Minimal reproduction of DeepSeek R1-Zero
Build multimodal language agents for fast prototype and production
An Easy-to-use, Scalable and High-performance RLHF Framework based on Ray (PPO & GRPO & REINFORCE++ & LoRA & vLLM & RFT)
A instruction data generation system for multimodal language models.
Extend existing LLMs way beyond the original training length with constant memory usage, without retraining
MINT-1T: A one trillion token multimodal interleaved dataset.
[COLM-2024] List Items One by One: A New Data Source and Learning Paradigm for Multimodal LLMs
Official repo for Detecting, Explaining, and Mitigating Memorization in Diffusion Models (ICLR 2024)
Package to optimize Adversarial Attacks against (Large) Language Models with Varied Objectives
An open-source framework for training large multimodal models.
LAVIS - A One-stop Library for Language-Vision Intelligence
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Ongoing research training transformer models at scale
Official repository of NEFTune: Noisy Embeddings Improves Instruction Finetuning
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
GLIDE: a diffusion-based text-conditional image synthesis model
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
A framework for few-shot evaluation of language models.
The official repository of the paper "On the Exploitability of Instruction Tuning".