-
City University of Hong Kong
- Hong Kong, China
-
06:57
(UTC +08:00) - @FengjiZhang98
Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Stars
RL Tango: Reinforcing Generator and Verifier Together for Language Reasoning
Skywork-R1V2:Multimodal Hybrid Reinforcement Learning for Reasoning
Scaling Deep Research via Reinforcement Learning in Real-world Environments.
Multi-SWE-bench: A Multilingual Benchmark for Issue Resolving
This repository provides valuable reference for researchers in the field of multimodality, please start your exploratory travel in RL-based Reasoning MLLMs!
⚡FlashRAG: A Python Toolkit for Efficient RAG Research (WWW2025 Resource)
ReCall: Learning to Reason with Tool Call for LLMs via Reinforcement Learning
Search-R1: An Efficient, Scalable RL Training Framework for Reasoning & Search Engine Calling interleaved LLM based on veRL
A code-first agent framework for seamlessly planning and executing data analytics tasks.
Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"
This repo contains the dataset and code for the paper "SWE-Lancer: Can Frontier LLMs Earn $1 Million from Real-World Freelance Software Engineering?"
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Multimodal Large Language Models for Code Generation under Multimodal Scenarios
LLaVA-CoT, a visual language model capable of spontaneous, systematic reasoning
Codev-Bench (Code Development Benchmark), a fine-grained, real-world, repository-level, and developer-centric evaluation framework. Codev-Bench assesses whether a code completion tool can accuratel…
MapCoder: Multi-Agent Code Generation for Competitive Problem Solving
Enhancing AI Software Engineering with Repository-level Code Graph
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
A Lightweight Visual Reasoning Benchmark for Evaluating Large Multimodal Models through Complex Diagrams in Coding Tasks
Codebase for Aria - an Open Multimodal Native MoE
An LLM-powered repository agent designed to assist developers and teams in generating documentation and understanding repositories quickly.
Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)
DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence
Baselines for all tasks from Long Code Arena benchmarks 🏟️
Open-source evaluation toolkit of large multi-modality models (LMMs), support 220+ LMMs, 80+ benchmarks