Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
-
Updated
Jun 3, 2025 - Python
8000
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
Unattended Lightweight Text Classifiers with LLM Embeddings
Faster, smaller BERT models in just a few lines.
Text Mining (PubMed Search) with NLP & LLM
A demo from the blog post comparing MiniLM-based models using song lyrics and Milvus for vector similarity search—an approach that works for any text content.
An AI-powered study companion that helps students understand lecture material through intelligent question answering, slide summarization, PDF summaries, and flashcard generation. Built with LangChain, Hugging Face Transformers, and Gradio — and fully powered by open-source LLMs running on your local GPU.
Advanced NLP project detecting duplicate questions on Quora using transformer-based embeddings, LSTM architectures, and ensemble models, achieving 88% accuracy with scalable solutions for real-world applications 🧠💬.
Add a description, image, and links to the minilm topic page so that developers can more easily learn about it.
To associate your repository with the minilm topic, visit your repo's landing page and select "manage topics."