- Coming Soon
- @DayaneMarcos_AI
Lists (1)
Sort Name ascending (A-Z)
Stars
A generative world for general-purpose robotics & embodied AI learning.
Graph Machine Learning course, Xavier Bresson, 2023
A curated list of practical financial machine learning tools and applications.
Official code repo for the O'Reilly Book - "Hands-On Large Language Models"
A Chainlit App Used to Showcase: Async, Caching, Additional Chainlit Methods, and more!
Dynamic Metadata based RAG Framework
An open-source RAG-based tool for chatting with your documents.
Game Development Patterns with Unreal Engine 5, published by Packt
Official repo for the paper "Scaling Synthetic Data Creation with 1,000,000,000 Personas"
[NeurIPS'24 Spotlight, ICLR'25, ICML'25] To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which reduces inference latency by up to 10x for pre-filli…
Cleaned-up Dark Souls AI scripts that provide a better starting point for modding.
Universal LLM Deployment Engine with ML Compilation
Granite Code Models: A Family of Open Foundation Models for Code Intelligence
The Open Source Memory Layer For Autonomous Agents
[NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
The official implementation of RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorR…
A high-throughput and memory-efficient inference and serving engine for LLMs
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
REST: Retrieval-Based Speculative Decoding, NAACL 2024
Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads
Letta (formerly MemGPT) is the stateful agents framework with memory, reasoning, and context management.
Official implementation of "DoRA: Weight-Decomposed Low-Rank Adaptation"
Stanford NLP Python library for Representation Finetuning (ReFT)
RAFT, or Retrieval-Augmented Fine-Tuning, is a method comprising of a fine-tuning and a RAG-based retrieval phase. It is particularly suited for the creation of agents that realistically emulate a …
Chat Templates for 🤗 HuggingFace Large Language Models
tiktoken is a fast BPE tokeniser for use with OpenAI's models.