File size: 1,443 Bytes
1afbf8c
 
 
 
 
 
 
 
 
 
 
 
 
5d5fc17
 
 
1afbf8c
 
 
 
 
 
 
 
 
 
5d5fc17
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
---
language: fr
license: mit
datasets:
- amazon_reviews_multi
- allocine
widget:
- text: "Je pensais lire un livre nul, mais finalement je l'ai trouvé super..."
---

DistilCamemBERT-Sentiment
=========================

We present DistilCamemBERT-Sentiment which is [DistilCamemBERT](https://huggingface.co/cmarkea/distilcamembert-base) fine tuned for the sentiment analysis task for the French language. This model is construct over 2 datasets: [amazon_reviews_multi](https://huggingface.co/datasets/amazon_reviews_multi) and [allocine](https://huggingface.co/datasets/allocine) to aims minimize the biais. Inded, Amazon review are very similare beetwen the messages and relatevely short. To opposate Allocine criticims are long and rich text.

This modelisation is closely of [tblard/tf-allocine](https://huggingface.co/tblard/tf-allocine) base on [CamemBERT](https://huggingface.co/camembert-base) model. The problem of the modelizations based on CamemBERT is at the scaling moment, for the production phase for example. Indeed, inference cost can be a technological issue. To counteract this effect, we propose this modelization which **divides the inference time by 2** with the same consumption power thanks to [DistilCamemBERT](https://huggingface.co/cmarkea/distilcamembert-base).

Dataset
-------
 
Evaluation results
------------------

Benchmark
---------

How to use DistilCamemBERT-Sentiment
------------------------------------