The University of Edinburgh's Submission to the WMT22 Code-Mixing Shared Task (MixMT) Paper • 2210.11309 • Published Oct 20, 2022
An Expanded Massive Multilingual Dataset for High-Performance Language Technologies Paper • 2503.10267 • Published Mar 13, 2025 • 2
Tokenizer Choice For LLM Training: Negligible or Crucial? Paper • 2310.08754 • Published Oct 12, 2023 • 3
Neighborhood Contrastive Learning for Scientific Document Representations with Citation Embeddings Paper • 2202.06671 • Published Feb 14, 2022 • 2
Specialized Document Embeddings for Aspect-based Similarity of Research Papers Paper • 2203.14541 • Published Mar 28, 2022
Efficient Language Model Training through Cross-Lingual and Progressive Transfer Learning Paper • 2301.09626 • Published Jan 23, 2023 • 2
MMTEB: Massive Multilingual Text Embedding Benchmark Paper • 2502.13595 • Published Feb 19, 2025 • 43
Towards Best Practices for Open Datasets for LLM Training Paper • 2501.08365 • Published Jan 14, 2025 • 62
mOSCAR: A Large-scale Multilingual and Multimodal Document-level Corpus Paper • 2406.08707 • Published Jun 13, 2024 • 17
Semi-automatic staging area for high-quality structured data extraction from scientific literature Paper • 2309.10923 • Published Sep 19, 2023
Mining experimental data from Materials Science literature with Large Language Models: an evaluation study Paper • 2401.11052 • Published Jan 19, 2024 • 1
SuperMat: Construction of a linked annotated dataset from superconductors-related publications Paper • 2101.02455 • Published Jan 7, 2021 • 2
Automatic extraction of materials and properties from superconductors scientific literature Paper • 2210.15600 • Published Oct 26, 2022 • 2