srvmishra832 commited on
Commit
e479bb7
·
verified ·
1 Parent(s): 37102cd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -8
README.md CHANGED
@@ -7,29 +7,28 @@ tags:
7
  model-index:
8
  - name: IMDb_data_subset-MLM_with-custom_collator-distilbert-base-uncased
9
  results: []
 
 
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
-
15
  # IMDb_data_subset-MLM_with-custom_collator-distilbert-base-uncased
16
 
17
- This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
  - Loss: 3.2538
20
  - Model Preparation Time: 0.0042
21
 
22
  ## Model description
23
 
24
- More information needed
25
 
26
  ## Intended uses & limitations
27
 
28
- More information needed
29
 
30
  ## Training and evaluation data
31
 
32
- More information needed
33
 
34
  ## Training procedure
35
 
@@ -61,4 +60,4 @@ The following hyperparameters were used during training:
61
  - Transformers 4.50.0
62
  - Pytorch 2.6.0+cu124
63
  - Datasets 3.4.1
64
- - Tokenizers 0.21.1
 
7
  model-index:
8
  - name: IMDb_data_subset-MLM_with-custom_collator-distilbert-base-uncased
9
  results: []
10
+ language:
11
+ - en
12
  ---
13
 
 
 
 
14
  # IMDb_data_subset-MLM_with-custom_collator-distilbert-base-uncased
15
 
16
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on IMDb dataset.
17
  It achieves the following results on the evaluation set:
18
  - Loss: 3.2538
19
  - Model Preparation Time: 0.0042
20
 
21
  ## Model description
22
 
23
+ [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased)
24
 
25
  ## Intended uses & limitations
26
 
27
+ Mask filling
28
 
29
  ## Training and evaluation data
30
 
31
+ The IMDb dataset is tokenized, and words are masked with 0.2 probability. The resulting dataset is downsampled, resulting in 10,000 training samples and 1,000 validation samples.
32
 
33
  ## Training procedure
34
 
 
60
  - Transformers 4.50.0
61
  - Pytorch 2.6.0+cu124
62
  - Datasets 3.4.1
63
+ - Tokenizers 0.21.1