- Mountain View, CA
-
02:34
(UTC -07:00) - https://yueyu1030.github.io
- in/yue-yu-b75553115
- https://scholar.google.com/citations?user=zQ3Jh6UAAAAJ&hl=en
Stars
This is the official repository for paper "MedAgentGYM: Training LLM Agents for Code-Based Medical Reasoning at Scale"
Official Code Repository for WorkForceAgent-R1
[TMLR] A curated list of language modeling researches for code (and other software engineering activities), plus related datasets.
[EMNLP'24] MedAdapter: Efficient Test-Time Adaptation of Large Language Models Towards Medical Reasoning
Official Code Repository for paper "HYDRA: Model Factorization Framework for Black-Box LLM Personalization"
Official implementation for "Law of the Weakest Link: Cross capabilities of Large Language Models"
Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".
[ACL 2024] This is the code for our paper ”RAM-EHR: Retrieval Augmentation Meets Clinical Predictions on Electronic Health Records“.
[NeurIPS 2024] A task generation and model evaluation system for multimodal language models.
[EMNLP 2024] This is the code for our paper "BMRetriever: Tuning Large Language Models as Better Biomedical Text Retrievers".
A flexible and efficient codebase for training visually-conditioned language models (VLMs)
[ICML 2024] Selecting High-Quality Data for Training Language Models
[EMNLP'24] EHRAgent: Code Empowers Large Language Models for Complex Tabular Reasoning on Electro 91F4 nic Health Records
[ACL 2024 Findings] This is the code for our paper "Knowledge-Infused Prompting: Assessing and Advancing Clinical Text Data Generation with Large Language Models".
EcoAssistant: using LLM assistant more affordably and accurately
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
This is a collection of research papers for Self-Correcting Large Language Models with Automated Feedback.
MAD: The first work to explore Multi-Agent Debate with Large Language Models :D
[ACL2023] We introduce LLM-Blender, an innovative ensembling framework to attain consistently superior performance by leveraging the diverse strengths of multiple open-source LLMs. LLM-Blender cut …
ToolQA, a new dataset to evaluate the capabilities of LLMs in answering challenging questions with external tools. It offers two levels (easy/hard) across eight real-life scenarios.
MUBen: Benchmarking the Uncertainty of Molecular Representation Models