Fill-Mask
Transformers
PyTorch
modernbert
orionweller commited on
Commit
633e71f
·
verified ·
1 Parent(s): 76b1acb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -12,7 +12,7 @@ pipeline_tag: fill-mask
12
  # mmBERT: A Modern Multilingual Encoder
13
 
14
  [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT)
15
- [![Paper](https://img.shields.io/badge/Paper-Arxiv-red)](https://arxiv.org/abs/xxxx.xxxxx)
16
  [![Model](https://img.shields.io/badge/🤗%20Hugging%20Face-Model-blue)](https://huggingface.co/jhu-clsp/mmBERT-base)
17
  [![Collection](https://img.shields.io/badge/🤗%20Model%20Collection-blue)](https://huggingface.co/collections/jhu-clsp/mmbert-a-modern-multilingual-encoder-68b725831d7c6e3acc435ed4)
18
  [![GitHub](https://img.shields.io/badge/GitHub-Code-black)](https://github.com/jhu-clsp/mmBERT)
@@ -469,14 +469,14 @@ mmBERT was trained on a carefully curated 3T+ token multilingual dataset:
469
  If you use mmBERT in your research, please cite our work:
470
 
471
  ```bibtex
472
- @misc{marone2025mmbert,
473
- title={Learning New Languages in the Decay Phase: A Multilingual Modern Encoder Model},
474
  author={Marc Marone and Orion Weller and William Fleshman and Eugene Yang and Dawn Lawrie and Benjamin Van Durme},
475
  year={2025},
476
- eprint={xxxx.xxxxx},
477
  archivePrefix={arXiv},
478
  primaryClass={cs.CL},
479
- url={https://arxiv.org/abs/xxxx.xxxxx},
480
  }
481
  ```
482
  """
 
12
  # mmBERT: A Modern Multilingual Encoder
13
 
14
  [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT)
15
+ [![Paper](https://img.shields.io/badge/Paper-Arxiv-red)](https://arxiv.org/abs/2509.06888)
16
  [![Model](https://img.shields.io/badge/🤗%20Hugging%20Face-Model-blue)](https://huggingface.co/jhu-clsp/mmBERT-base)
17
  [![Collection](https://img.shields.io/badge/🤗%20Model%20Collection-blue)](https://huggingface.co/collections/jhu-clsp/mmbert-a-modern-multilingual-encoder-68b725831d7c6e3acc435ed4)
18
  [![GitHub](https://img.shields.io/badge/GitHub-Code-black)](https://github.com/jhu-clsp/mmBERT)
 
469
  If you use mmBERT in your research, please cite our work:
470
 
471
  ```bibtex
472
+ @misc{marone2025mmbertmodernmultilingualencoder,
473
+ title={mmBERT: A Modern Multilingual Encoder with Annealed Language Learning},
474
  author={Marc Marone and Orion Weller and William Fleshman and Eugene Yang and Dawn Lawrie and Benjamin Van Durme},
475
  year={2025},
476
+ eprint={2509.06888},
477
  archivePrefix={arXiv},
478
  primaryClass={cs.CL},
479
+ url={https://arxiv.org/abs/2509.06888},
480
  }
481
  ```
482
  """