Stars
Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting yo…
A high-quality tool for convert PDF to Markdown and JSON.一站式开源高质量数据提取工具,将PDF转换成Markdown和JSON格式。
Python tool for converting files and office documents to Markdown.
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
Hands-on Deep Reinforcement Learning, published by Packt
Fully open reproduction of DeepSeek-R1
This is the code of using machine learning to play Sekiro .
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
A modular graph-based Retrieval-Augmented Generation (RAG) system
AISystem 主要是指AI系统,包括AI芯片、AI编译器、AI推理和训练框架等AI全栈底层技术
An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.
人工智能学习路线图,整理近200个实战案例与项目,免费提供配套教材,零基础入门,就业实战!包括:Python,数学,机器学习,数据分析,深度学习,计算机视觉,自然语言处理,PyTorch tensorflow machine-learning,deep-learning data-analysis data-mining mathematics data-science artificial…
📚LeetCUDA: Modern CUDA Learn Notes with PyTorch for Beginners🐑, 200+ CUDA/Tensor Cores Kernels, HGEMM, FA-2 MMA etc.🔥
Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with …
Code and documentation to train Stanford's Alpaca models, and generate the data.
Instruct-tune LLaMA on consumer hardware
An Application Framework for AI Engineering
A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)
Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Retrieval and Retrieval-augmented LLMs
A high-throughput and memory-efficient inference and serving engine for LLMs