view post Post 5022 Qwen releases 4 new Qwen3.5 Small models: 0.8B • 2B • 4B • 9B!Run Qwen3.5-0.8B, 2B and 4B on your phone. Run 9B on 6GB RAM.The vision reasoning LLMs perform better than models 4x their size.GGUFs to run: https://huggingface.co/collections/unsloth/qwen35Guide: https://unsloth.ai/docs/models/qwen3.5 See translation 4 replies · 🔥 14 14 🤗 13 13 🚀 1 1 + Reply
view article Article Training and Finetuning Reranker Models with Sentence Transformers v4 Mar 26, 2025 • 185
MedGemma Release Collection Collection of Gemma 3 variants for performance on medical text and image comprehension to accelerate building healthcare-based AI applications. • 9 items • Updated Jan 14 • 450
Qwen3.5 Collection Qwen3.5 is Qwen's new model family including Qwen3.5 Small: 0.8B, 2B, 4B, 9B and Qwen3.5 Medium: 35B-A3B, 27B, 122B-A10B and 397B-A17B. • 25 items • Updated 3 days ago • 90
TinyLettuce Collection This Collection contains our small, Ettin-encoder (https://arxiv.org/abs/2507.11412) based models trained on synthetic and RagTruth data. • 6 items • Updated Aug 31, 2025 • 4
pplx-embed Collection Diffusion-Pretrained Dense and Contextual Embeddings • 7 items • Updated 8 days ago • 84
view article Article **ColBERT-Zero: To Pre-train Or Not To Pre-train ColBERT models?** 15 days ago • 17
RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models Paper • 2401.00396 • Published Dec 31, 2023 • 6