8000 bmmcq (bmmcq) / Starred · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content
View bmmcq's full-sized avatar

Block or report bmmcq

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Showing results

Rust async runtime based on io-uring.

Rust 4,542 249 Updated Apr 22, 2025

Build and publish crates with pyo3, cffi and uniffi bindings as well as rust binaries as python packages

Rust 4,620 325 Updated Jun 16, 2025

FinGPT: Open-Source Financial Large Language Models! Revolutionize 🔥 We release the trained model on HuggingFace.

Jupyter Notebook 16,425 2,324 Updated Dec 26, 2024

A Python Interpreter written in Rust

Rust 20,164 1,318 Updated Jun 16, 2025

FlashInfer: Kernel Library for LLM Serving

Cuda 3,181 334 Updated Jun 16, 2025

Ultra fast asyncio event loop.

Cython 10,978 567 Updated Apr 17, 2025

The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!

Python 7,786 844 Updated Jun 16, 2025

The Triton Inference Server provides an optimized cloud and edge inferencing solution.

Python 9,347 1,596 Updated Jun 17, 2025

Blazingly fast LLM inference.

Rust 5,739 416 Updated Jun 16, 2025

A Datacenter Scale Distributed Inference Serving Framework

Rust 4,282 424 Updated Jun 17, 2025

🦉 OWL: Optimized Workforce Learning for General Multi-Agent Assistance in Real-World Task Automation

Python 17,000 1,996 Updated Jun 11, 2025

No fortress, purely open ground. OpenManus is Coming.

Python 46,866 8,186 Updated Jun 16, 2025

🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!

Python 22,123 2,602 Updated Apr 30, 2025

Elegant reading of real-time and hottest news

TypeScript 11,193 3,226 Updated May 10, 2025

DeepEP: an efficient expert-parallel communication library

Cuda 7,776 793 Updated Jun 16, 2025

A high-level distributed programming framework for Rust

Rust 920 50 Updated Jun 16, 2025

Dataframes powered by a multithreaded, vectorized query engine, written in Rust

Rust 34,048 2,268 Updated Jun 16, 2025

High-level tracing language for Linux

C++ 9,285 1,384 Updated Jun 16, 2025

Rust bindings for the Python interpreter

Rust 13,837 843 Updated Jun 12, 2025

A fast communication-overlapping library for tensor/expert parallelism on GPUs.

C++ 978 66 Updated May 28, 2025

Meet OceanBase, the MySQL compatible distributed database for your cloud native apps. High Performance, High Available, Low Cost, Multi-Cloud. Welcome to our community: https://discord.gg/74cF8vbNEs

C++ 9,190 1,738 Updated Jun 17, 2025

A fast GPU memory copy library based on NVIDIA GPUDirect RDMA technology

C++ 1,131 160 Updated Jun 5, 2025

Optimized primitives for collective multi-GPU communication

C++ 3,788 935 Updated May 30, 2025

Library providing helpers for the Linux kernel io_uring support

C 3,178 451 Updated Jun 16, 2025

BLAS-like Library Instantiation Software Framework

C 140 39 Updated Jun 5, 2025

LLM inference in C/C++

C++ 81,841 12,109 Updated Jun 17, 2025

Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"

Python 12,085 766 Updated Dec 17, 2024

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorR…

C++ 10,756 1,500 Updated Jun 17, 2025

A high-throughput and memory-efficient inference and serving engine for LLMs

Python 49,816 8,040 Updated Jun 17, 2025
Next
0