Stars
🦉 OWL: Optimized Workforce Learning for General Multi-Agent Assistance in Real-World Task Automation
A diagnostic method for Alzheimer's disease based on multimodal CoT reasoning and self-refine
Awesome Large Reasoning Model(LRM) Safety.This repository is used to collect security-related research on large reasoning models such as DeepSeek-R1 and OpenAI o1, which are currently very popular.
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
LLMs-from-scratch项目中文翻译
[CVPR 2025 Oral] VideoEspresso: A Large-Scale Chain-of-Thought Dataset for Fine-Grained Video Reasoning via Core Frame Selection
Video Chain of Thought, Codes for ICML 2024 paper: "Video-of-Thought: Step-by-Step Video Reasoning from Perception to Cognition"
A high-throughput and memory-efficient inference and serving engine for LLMs
[NeurIPS2024] Repo for the paper `ControlMLLM: Training-Free Visual Prompt Learning for Multimodal Large Language Models'
[ACM Trans. Multimedia Comput. Commun. Appl.] S^3 Agent: Unlocking the Power of VLLM for Zero-Shot Multi-modal Sarcasm Detection
Wrong-of-Thought: An Integrated Reasoning Framework with Multi-Perspective Verification and Wrong Information (WoT)
Awesome-LLM: a curated list of Large Language Model
Github Pages template based upon HTML and Markdown for personal, portfolio-based websites.
⏰ Collaboratively track deadlines of conferences recommended by CCF (Website, Python Cli, Wechat Applet) / If you find it useful, please star this project, thanks~
Export (your) GitHub stars to a CSV file
AutoCAP: Towards Automatic Cross-lingual Alignment Planning for Zero-shot Chain-of-Thought
🦜🔗 Build context-aware reasoning applications
Large Language Models are Temporal and Causal Reasoners for Video Question Answering (EMNLP 2023)
[CVPR'24 Highlight] The official code and data for paper "EgoThink: Evaluating First-Person Perspective Thinking Capability of Vision-Language Models"