Bugpie commited on
Commit
e00ad02
·
1 Parent(s): 9f744e5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -0
README.md CHANGED
@@ -6,12 +6,47 @@ datasets:
6
  ---
7
 
8
  ## Model description
 
9
  CamemBERT is a state-of-the-art language model for French based on the RoBERTa model.
10
  It is now available on Hugging Face in six different versions with varying number of parameters, amount of pretraining data and pretraining data source domains.
 
11
  ## Intended uses & limitations
 
12
  ## How to use
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  ## Limitations and bias
 
14
  ## Training data
 
15
  OSCAR or Open Super-large Crawled Aggregated coRpus is a multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the Ungoliant architecture.
 
16
  ## Training procedure
 
17
  ## Evaluation results
 
6
  ---
7
 
8
  ## Model description
9
+
10
  CamemBERT is a state-of-the-art language model for French based on the RoBERTa model.
11
  It is now available on Hugging Face in six different versions with varying number of parameters, amount of pretraining data and pretraining data source domains.
12
+
13
  ## Intended uses & limitations
14
+
15
  ## How to use
16
+
17
+ ```python
18
+ >>> from transformers import pipeline
19
+ >>> unmasker = pipeline('fill-mask', model='bert-base-cased')
20
+ >>> unmasker("Hello I'm a [MASK] model.")
21
+
22
+ [{'sequence': "[CLS] Hello I'm a fashion model. [SEP]",
23
+ 'score': 0.09019174426794052,
24
+ 'token': 4633,
25
+ 'token_str': 'fashion'},
26
+ {'sequence': "[CLS] Hello I'm a new model. [SEP]",
27
+ 'score': 0.06349995732307434,
28
+ 'token': 1207,
29
+ 'token_str': 'new'},
30
+ {'sequence': "[CLS] Hello I'm a male model. [SEP]",
31
+ 'score': 0.06228214129805565,
32
+ 'token': 2581,
33
+ 'token_str': 'male'},
34
+ {'sequence': "[CLS] Hello I'm a professional model. [SEP]",
35
+ 'score': 0.0441727414727211,
36
+ 'token': 1848,
37
+ 'token_str': 'professional'},
38
+ {'sequence': "[CLS] Hello I'm a super model. [SEP]",
39
+ 'score': 0.03326151892542839,
40
+ 'token': 7688,
41
+ 'token_str': 'super'}]
42
+ ```
43
+
44
  ## Limitations and bias
45
+
46
  ## Training data
47
+
48
  OSCAR or Open Super-large Crawled Aggregated coRpus is a multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the Ungoliant architecture.
49
+
50
  ## Training procedure
51
+
52
  ## Evaluation results