-
University of California, San Diego
- La Jolla, California
-
22:55
(UTC -07:00)
Highlights
- Pro
Stars
[TMLR] A curated list of language modeling researches for code (and other software engineering activities), plus related datasets.
Cuckoo: A Series of IE Free Riders Using LLM's Resources to Scale up Themselves.
A bibliography and survey of the papers surrounding o1
The road to hack SysML and become an system expert
A framework for the evaluation of autoregressive code generation language models.
Tools for merging pretrained large language models.
Large Language Model Text Generation Inference
GPU programming related news and material links
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama mode…
alibaba / Megatron-LLaMA
Forked from NVIDIA/Megatron-LMBest practice for training LLaMA models in Megatron-LM
MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
The ChatGPT Retrieval Plugin lets you easily find personal or work documents by asking questions in natural language.
A curated list of papers and applications on tool learning.
A repo lists papers related to LLM based agent
Train transformer language models with reinforcement learning.
A curated list of awesome resources dedicated to Scaling Laws for LLMs
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
中文nlp解决方案(大模型、数据、模型、训练、推理)
The nanoGPT-style implementation of RWKV Language Model - an RNN with GPT-level LLM performance.
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
QLoRA: Efficient Finetuning of Quantized LLMs
Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).