OFA: A Framework of Initializing Unseen Subword Embeddings for Efficient Large-scale Multilingual Continued Pretraining Paper • 2311.08849 • Published Nov 15, 2023 • 6
Imp: Highly Capable Large Multimodal Models for Mobile Devices Paper • 2405.12107 • Published May 20, 2024 • 29
Learn it or Leave it: Module Composition and Pruning for Continual Learning Paper • 2406.18708 • Published Jun 26, 2024 • 1
LangSAMP: Language-Script Aware Multilingual Pretraining Paper • 2409.18199 • Published Sep 26, 2024 • 1
Tracing Multilingual Factual Knowledge Acquisition in Pretraining Paper • 2505.14824 • Published May 20 • 4
Language Mixing in Reasoning Language Models: Patterns, Impact, and Internal Causes Paper • 2505.14815 • Published May 20 • 2
Refusal Direction is Universal Across Safety-Aligned Languages Paper • 2505.17306 • Published May 22 • 2
Through a Compressed Lens: Investigating the Impact of Quantization on LLM Explainability and Interpretability Paper • 2505.13963 • Published May 20 • 1
BMIKE-53: Investigating Cross-Lingual Knowledge Editing with In-Context Learning Paper • 2406.17764 • Published Jun 25, 2024
Lost in Multilinguality: Dissecting Cross-lingual Factual Inconsistency in Transformer Language Models Paper • 2504.04264 • Published Apr 5 • 2