Stars
Notebooks and sample code for Build On Trainium
MLOps on Amazon EKS
A CLI tool to convert your codebase into a single LLM prompt with source tree, prompt templating, and token counting.
[NeurIPS 2024] VFIMamba: Video Frame Interpolation with State Space Models
Welcome to the VMware ESXi License Keys! Here, you'll find a curated collection of free license keys for various versions, including vSphere 6 and 7
A hardware-agnostic (NVIDIA's GPUs and AWS Inferentia accelerators) deployment of computer-vision models (e.g., YOLO, ViT), generate text and text-to-image (e.g., Llama3 and Stable Diffusion ) on E…
Felafax is building AI infra for non-NVIDIA GPUs
OpenAI-Compatible RESTful APIs for Amazon Bedrock
Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.
DSPy: The framework for programming—not prompting—language models
aws-neuron / upstreaming-to-vllm
Forked from vllm-project/vllmA high-throughput and memory-efficient inference and serving engine for LLMs
🤗 LeRobot: Making AI for Robotics more accessible with end-to-end learning
Data set of Finnish grey literature, containing curated Dublin Core style metadata and links to original PDF publications
A helper library to connect into Amazon SageMaker with AWS Systems Manager and SSH (Secure Shell)
Hummingbird compiles trained ML models into tensor computation for faster inference.
AutoQASM is an experimental module offering a quantum-imperative programming experience in Python for developing quantum programs.
Foundation model benchmarking tool. Run any model on any AWS platform and benchmark for performance across instance type and serving stack options.
Example of how to export longitude and latitude to a Garmin FPL file format
Control and monitor your Bitcoin Miners from Home Assistant.
An email signature generator for Civil Air Patrol.
Civil Air Patrol python utilities for CAPWATCH
The source code for the Event Manager+ suite of tools for Civil Air Patrol units
Powering AWS purpose-built machine learning chips. Blazing fast and cost effective, natively integrated into PyTorch and TensorFlow and integrated with your favorite AWS services
Infinity is a high-throughput, low-latency serving engine for text-embeddings, reranking models, clip, clap and colpali