README.md
| 1 | --- |
| 2 | tags: |
| 3 | - summarization |
| 4 | |
| 5 | language: |
| 6 | - fr |
| 7 | |
| 8 | license: apache-2.0 |
| 9 | |
| 10 | pipeline_tag: "fill-mask" |
| 11 | --- |
| 12 | A french sequence to sequence pretrained model based on [BART](https://huggingface.co/facebook/bart-large). <br> |
| 13 | BARThez is pretrained by learning to reconstruct a corrupted input sentence. A corpus of 66GB of french raw text is used to carry out the pretraining. <br> |
| 14 | Unlike already existing BERT-based French language models such as CamemBERT and FlauBERT, BARThez is particularly well-suited for generative tasks (such as abstractive summarization), since not only its encoder but also its decoder is pretrained. |
| 15 | |
| 16 | In addition to BARThez that is pretrained from scratch, we continue the pretraining of a multilingual BART [mBART](https://huggingface.co/facebook/mbart-large-cc25) which boosted its performance in both discriminative and generative tasks. We call the french adapted version [mBARThez](https://huggingface.co/moussaKam/mbarthez). |
| 17 | |
| 18 | | Model | Architecture | #layers | #params | |
| 19 | | ------------- |:-------------:| :-----:|:-----:| |
| 20 | | [BARThez](https://huggingface.co/moussaKam/barthez) | BASE | 12 | 165M | |
| 21 | | [mBARThez](https://huggingface.co/moussaKam/mbarthez) | LARGE | 24 | 458M | |
| 22 | |
| 23 | paper: https://arxiv.org/abs/2010.12321 \ |
| 24 | github: https://github.com/moussaKam/BARThez |
| 25 | |
| 26 | |
| 27 | ``` |
| 28 | @article{eddine2020barthez, |
| 29 | title={BARThez: a Skilled Pretrained French Sequence-to-Sequence Model}, |
| 30 | author={Eddine, Moussa Kamal and Tixier, Antoine J-P and Vazirgiannis, Michalis}, |
| 31 | journal={arXiv preprint arXiv:2010.12321}, |
| 32 | year={2020} |
| 33 | } |
| 34 | ``` |
| 35 | |