-
Waseda University
- Tokyo, Japan
Stars
[ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning
Fully open reproduction of DeepSeek-R1
List of papers on Self-Correction of LLMs.
EasyR1: An Efficient, Scalable, Multi-Modality RL Training Framework based on veRL
The Large-scale Manipulation Platform for Scalable and Intelligent Embodied Systems
Yomitoku is an AI-powered document image analysis package designed specifically for the Japanese language.
S2ORC: The Semantic Scholar Open Research Corpus: https://www.aclweb.org/anthology/2020.acl-main.447/
The evaluation scripts of JMTEB (Japanese Massive Text Embedding Benchmark)
Benchmarking LLMs with Challenging Tasks from Real Users
Official repository for KoMT-Bench built by LG AI Research
A series of large language models trained from scratch by developers @01-ai
Arena-Hard-Auto: An automatic LLM benchmark.
Scalable toolkit for efficient model alignment
OCR, layout analysis, reading order, table recognition in 90+ languages
A terminal application to view, tail, merge, and search log files (plus JSONL).
JMultiWOZ: A Large-Scale Japanese Multi-Domain Task-Oriented Dialogue Dataset, LREC-COLING 2024
[Nature Reviews Bioengineering🔥] Application of Large Language Models in Medicine. A curated list of practical guide resources of Medical LLMs (Medical LLMs Tree, Tables, and Papers)
RealPersonaChat: A Realistic Persona Chat Corpus with Interlocutors' Own Personalities
📃Language Model based sentences scoring library
Data and tools for generating and inspecting OLMo pre-training data.
Orion-14B is a family of models includes a 14B foundation LLM, and a series of models: a chat model, a long context model, a quantized model, a RAG fine-tuned model, and an Agent fine-tuned model. …
A principled instruction benchmark on formulating effective queries and prompts for large language models (LLMs). Our paper: https://arxiv.org/abs/2312.16171
[ICML'24] Magicoder: Empowering Code Generation with OSS-Instruct