Stars
A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.
🦜🔗 Build context-aware reasoning applications
S-LoRA / S-LoRA
S-LoRA: Serving Thousands of Concurrent LoRA Adapters
Learn how to design large-scale systems. Prep for the system design interview. Includes Anki flashcards.
JA3 is a standard for creating SSL client fingerprints in an easy to produce and shareable way.
LLM powered retrieval engine designed to process a ton of sources to collect a comprehensive list of entities.
Opiniated RAG for integrating GenAI in your apps 🧠 Focus on your product rather than the RAG. Easy integration in existing products with customisation! Any LLM: GPT4, Groq, Llama. Any Vectorstore: …
This repo is meant to serve as a guide for Machine Learning/AI technical interviews.
Annotated version of the Mamba paper
AutoAWQ implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference. Documentation:
SkyPilot: Run AI and batch jobs on any infra (Kubernetes or 16+ clouds). Get unified execution, cost savings, and high GPU availability via a simple interface.
Large Language Model Text Generation Inference
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
MLX: An array framework for Apple silicon
A high-throughput and memory-efficient inference and serving engine for LLMs
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
FastAPI framework, high performance, easy to learn, fast to code, ready for production
Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!
Raising the Cost of Malicious AI-Powered Image Editing
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
Tacotron 2 - PyTorch implementation with faster-than-realtime inference
A collaboration friendly studio for NeRFs
A Unified Framework for Surface Reconstruction
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training