File size: 1,443 Bytes
1afbf8c 5d5fc17 1afbf8c 5d5fc17 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 |
---
language: fr
license: mit
datasets:
- amazon_reviews_multi
- allocine
widget:
- text: "Je pensais lire un livre nul, mais finalement je l'ai trouvé super..."
---
DistilCamemBERT-Sentiment
=========================
We present DistilCamemBERT-Sentiment which is [DistilCamemBERT](https://huggingface.co/cmarkea/distilcamembert-base) fine tuned for the sentiment analysis task for the French language. This model is construct over 2 datasets: [amazon_reviews_multi](https://huggingface.co/datasets/amazon_reviews_multi) and [allocine](https://huggingface.co/datasets/allocine) to aims minimize the biais. Inded, Amazon review are very similare beetwen the messages and relatevely short. To opposate Allocine criticims are long and rich text.
This modelisation is closely of [tblard/tf-allocine](https://huggingface.co/tblard/tf-allocine) base on [CamemBERT](https://huggingface.co/camembert-base) model. The problem of the modelizations based on CamemBERT is at the scaling moment, for the production phase for example. Indeed, inference cost can be a technological issue. To counteract this effect, we propose this modelization which **divides the inference time by 2** with the same consumption power thanks to [DistilCamemBERT](https://huggingface.co/cmarkea/distilcamembert-base).
Dataset
-------
Evaluation results
------------------
Benchmark
---------
How to use DistilCamemBERT-Sentiment
------------------------------------
|