8000 junfanz1 (Junfan Zhu) Β· GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content
View junfanz1's full-sized avatar
πŸ€—
Open to Work!
πŸ€—
Open to Work!

Block or report junfanz1

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
junfanz1/README.md
Profile Views Β Β  GitHub Followers Β Β  GitHub Stars

Junfan Zhu πŸ‘‹

Resume LinkedIn X Email GitHub Instagram Facebook Douban Zhihu WeChat

πŸ€— I'm an AI/LLM Researcher | Machine Learning Engineer with 4 years as a Quant at a leading investment bank in Chicago, grounded in Computer Science (Georgia Tech) and Financial Mathematics (UChicago). Passionate about scalable model architectures and multimodal agentic reasoning, I design cutting-edge LLM systems and pipelines that push AI’s real-world impact.

πŸ’» I'm open to work! Expertise: AI Research & Large Language Models (LLM) πŸ€– β€’ Machine Learning & Deep Learning πŸ“š β€’ Full-Stack AI Applications πŸ’‘ β€’ Quantitative Finance & Algorithmic Trading πŸ’Ή

πŸš€ AI Research/Engineering Portfolio

My portfolio boasts pioneering projects in MoE & Attention for scalable LLM, reflective multi-agent orchestrations, and full-stack GenAI applications.

  • 1. AI-LLM-ML-CS-Quant-Review
    In-depth review of industry trends in AI, LLMs, Machine Learning, Computer Science, and Quantitative Finance.
  • 2. MiniGPT-and-DeepSeek-MLA-Multi-Head-Latent-Attention
    Memory-efficient multi-head latent attention in PyTorch, that leverages low-rank approximation and decoupled rotary positional embeddings, to compress key–value representations, reducing inference memory while maintaining high performance in long-context language models.
  • 3. DeepSeek-MoE-Mixture-of-Experts-in-PyTorch
    Implemented scalable 8-expert MoE model with top-k routing, expert load balancing, and capacity-aware gating; enabled parallel sparse activation and DeepSeek-R1-style distributed training scalability.
  • 4. MCP-MultiServer-Interoperable-Agent2Agent-LangGraph-AI-System
    A decoupled real-time agent architecture connecting LangGraph agents to remote tools served by custom MCP servers via SSE and STDIO, enabling a scalable multi-agent system for LLM workflows. The design supports flexible multi-server connectivity and lays the groundwork for an Agent2Agent protocol, fostering seamless, cloud-deployable interoperability across diverse AI systems.
  • 5. LangGraph-Reflection-Researcher
    Engineered LangGraph-based multi-agent system with self-reflection and retrieval-grounded alignment; integrated LangSmith trace for reasoning introspection, cutting hallucination 40% with iterative expert routing.
  • 6. Cognito-LangGraph-RAG-Chatbot
    Advanced Retrieval Augmented Generation (RAG) chatbot that utilizes LangGraph to enhance answer accuracy and minimize hallucinations in LLM outputs.
  • 7. Cursor-FullStack-AI-App
    Cursor Vibe Engineering: Full-stack micro SaaS AI application that processes GitHub URLs to generate insightful JSON reports powered by AI analytics.
  • 8. Cryptocurrency-Blockchain-FullStack
    Comprehensive decentralized blockchain platform demonstrating practical applications of core blockchain concepts through a modular, full-stack approach.
Favorite project integrating Generative AI, Humanoid Robotics (RLHF), and Low-Altitude Economy.

πŸ› οΈ Tech Stack

Python PyTorch NumPy Pandas Scikit-learn LangChain LangGraph Pydantic CUDA R MATLAB Java C++ JavaScript Solidity Django Flask Node.js SQLite PostgreSQL MySQL MongoDB Redis React HTML5 CSS3 Docker Kubernetes AWS Azure Linux Postman Git Vercel

🌏 Fun Facts

I’m a traveler ✈️, violinist 🎻, amateur bartender 🍷 and aspiring private pilot 🚁. My passport has seen more stamps than a post office! πŸ›‚ Having visited 6 continents πŸ—ΊοΈ, 60+ countries/regions 🌐, USA 50 states & 30+ national parks 🏞️.

I summited πŸ‡ΉπŸ‡Ώ Kilimanjaro Uhuru Peak (5895m) at the Roof of Africa πŸ¦’, trekked πŸ‡³πŸ‡΅ Annapurna Base Camp πŸ”, hiked πŸ‡¬πŸ‡Ή VolcΓ‘n de Fuego πŸŒ‹, traversed a desert in πŸ‡¨πŸ‡³ Inner Mongolia 🏜, and completed a marathon within 5h πŸƒ.

My expeditions have taken me to beautiful adventurous journeys, such as πŸ‡³πŸ‡΄ Longyearbyen & Barentsburg (πŸ₯Ά icebreaker β›΄) in Svalbard 🌌, πŸ‡¨πŸ‡± Rapa NuiπŸ—Ώ, πŸ‡¨πŸ‡¦ Iqaluit Nunavut πŸ‹, πŸ‡¦πŸ‡· Ushuaia 🐧, πŸ‡ΊπŸ‡Έ Unalaska & Cold Bay in Aleutian IslandsπŸ—» / UtqiaΔ‘vik Alaska ❄️, πŸ‡¨πŸ‡³ Tibet πŸŒ„, πŸ‡΅πŸ‡« Bora Bora πŸͺΈ, πŸ‡ΊπŸ‡Έ MolokaΚ»i 🏝️, πŸ‡ͺπŸ‡¨ middle of the Earth 🌎 and so on. These experiences have shaped my adaptability πŸ‘½, problem-solving skills ✍️, and global perspective 🌊.

πŸ“– Motto: "Every man carries within him a world, composed of all that he has seen and loved, and it is to this world that he constantly returns, even when he seems to be journeying and living in another different world." β€” Chateaubriand, "Voyages en Italie" πŸŒ…

πŸ“Š GitHub Stats

Junfan Zhu's GitHub Stats Top Languages GitHub Streak Total Contributions https://github.com/junfanz1

Contribution Heatmap

Pinned Loading

  1. AI-LLM-ML-CS-Quant-Review AI-LLM-ML-CS-Quant-Review Public

    In-depth review of industry trends in AI, LLMs, Machine Learning, Computer Science, and Quantitative Finance.

    110 21

  2. MCP-MultiServer-Interoperable-Agent2Agent-LangGraph-AI-System MCP-MultiServer-Interoperable-Agent2Agent-LangGraph-AI-System Public

    This project demonstrates a decoupled real-time agent architecture that connects LangGraph agents to remote tools served by custom MCP (Modular Command Protocol) servers. The architecture enables a…

    Python 8 2

  3. MoE-Mixture-of-Experts-in-PyTorch MoE-Mixture-of-Experts-in-PyTorch Public

    Implementations of a Mixture-of-Experts (MoE) architecture designed for research on large language models (LLMs) and scalable neural network designs. One implementation targets a **single-device/NP…

    Python 5

  4. LangGraph-Reflection-Researcher LangGraph-Reflection-Researcher Public

    The LangGraph project implements a "Reflection Agent" designed to iteratively refine answers to user queries using a Large Language Model (LLM) and web search. It simulates a research process where…

    Jupyter Notebook 1

  5. MiniGPT-and-DeepSeek-MLA-Multi-Head-Latent-Attention MiniGPT-and-DeepSeek-MLA-Multi-Head-Latent-Attention Public

    An efficient and scalable attention module designed to reduce memory usage and improve inference speed in large language models. Designed and implemented the Multi-Head Latent Attention (MLA) modul…

    Python 2

  6. Cursor-FullStack-AI-App Cursor-FullStack-AI-App Public

    Build E2E Micro SaaS AI application, takes in Github urls, generate json reports with AI powered insights and repo stats

    JavaScript 1

0