-
UTC
- Tennessee, US
- https://www.linkedin.com/in/giang-do-hust/
Stars
The codes about "Uni-MoE: Scaling Unified Multimodal Models with Mixture of Experts"
Code for this paper "Sparse Mixture of Experts as Unified Competitive Learning".
Code for this paper "On the effectiveness of discrete representations in sparse mixture of experts".
PyTorch implementation of FractalGen https://arxiv.org/abs/2502.17437
Code for this paper "SimSMoE: Toward Efficient Training Mixture of Experts via Solving Representational Collapse".
A lightweight and extensible toolbox for image classification
This is the official implementation of "Back to Optimization: Diffusion-based Zero-Shot 3D Human Pose Estimation"
[NeurIPS 24] MoE Jetpack: From Dense Checkpoints to Adaptive Mixture of Experts for Vision Tasks
Repo for ACL2023 Findings paper "Emergent Modularity in Pre-trained Transformers"
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
A repository for research on medium sized language models.
CodeSage: Code Representation Learning At Scale (ICLR 2024)
Some methods for comparing network representations in deep learning and neuroscience.
Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"
[MICCAI'23] Implicit Anatomical Rendering for Medical Image Segmentation with Stochastic Experts
Python sample codes and textbook for robotics algorithms.
This repository contains a collection of papers and resources on Reasoning in Large Language Models.
Repository for benchmarking graph neural networks (JMLR 2023)
✨✨Latest Advances on Multimodal Large Language Models
🔥🔥🔥Latest Papers, Codes and Datasets on Vid-LLMs.
A curated list of resources for Document Understanding (DU) topic