trevorkwan commited on
Commit
943c923
verified
1 Parent(s): 9031956

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +80 -57
README.md CHANGED
@@ -1,57 +1,80 @@
1
- ---
2
- library_name: transformers
3
- base_model: yikuan8/Clinical-Longformer
4
- tags:
5
- - generated_from_trainer
6
- datasets:
7
- - squad_v2
8
- model-index:
9
- - name: clinical_longformer_squadv2_maxlen320
10
- results: []
11
- ---
12
-
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
15
-
16
- # clinical_longformer_squadv2_maxlen320
17
-
18
- This model is a fine-tuned version of [yikuan8/Clinical-Longformer](https://huggingface.co/yikuan8/Clinical-Longformer) on the squad_v2 dataset.
19
-
20
- ## Model description
21
-
22
- More information needed
23
-
24
- ## Intended uses & limitations
25
-
26
- More information needed
27
-
28
- ## Training and evaluation data
29
-
30
- More information needed
31
-
32
- ## Training procedure
33
-
34
- ### Training hyperparameters
35
-
36
- The following hyperparameters were used during training:
37
- - learning_rate: 2e-05
38
- - train_batch_size: 4
39
- - eval_batch_size: 4
40
- - seed: 42
41
- - gradient_accumulation_steps: 4
42
- - total_train_batch_size: 16
43
- - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
44
- - lr_scheduler_type: linear
45
- - num_epochs: 3.0
46
- - mixed_precision_training: Native AMP
47
-
48
- ### Training results
49
-
50
-
51
-
52
- ### Framework versions
53
-
54
- - Transformers 4.48.0.dev0
55
- - Pytorch 2.0.1+cu117
56
- - Datasets 3.0.1
57
- - Tokenizers 0.21.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ base_model: yikuan8/Clinical-Longformer
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - squad_v2
8
+ model-index:
9
+ - name: clinical_longformer_squadv2_maxlen320
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # clinical_longformer_squadv2_maxlen320
17
+
18
+ This model is a fine-tuned version of [yikuan8/Clinical-Longformer](https://huggingface.co/yikuan8/Clinical-Longformer) on the squad_v2 dataset.
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ Tuning script used:
35
+
36
+ set BASE_MODEL=yikuan8/Clinical-Longformer
37
+ set OUTPUT_DIR=U:\Documents...
38
+
39
+ python run_qa.py ^
40
+ --model_name_or_path %BASE_MODEL% ^
41
+ --dataset_name squad_v2 ^
42
+ --do_train ^
43
+ --do_eval ^
44
+ --version_2_with_negative ^
45
+ --per_device_train_batch_size 4 ^
46
+ --per_device_eval_batch_size 4 ^
47
+ --gradient_accumulation_steps 4 ^
48
+ --learning_rate 2e-5 ^
49
+ --num_train_epochs 3 ^
50
+ --max_seq_length 320 ^
51
+ --doc_stride 128 ^
52
+ --weight_decay 0.01 ^
53
+ --fp16 ^
54
+ --output_dir %OUTPUT_DIR% ^
55
+ --overwrite_output_dir
56
+
57
+ ### Training hyperparameters
58
+
59
+ The following hyperparameters were used during training:
60
+ - learning_rate: 2e-05
61
+ - train_batch_size: 4
62
+ - eval_batch_size: 4
63
+ - seed: 42
64
+ - gradient_accumulation_steps: 4
65
+ - total_train_batch_size: 16
66
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
67
+ - lr_scheduler_type: linear
68
+ - num_epochs: 3.0
69
+ - mixed_precision_training: Native AMP
70
+
71
+ ### Training results
72
+
73
+
74
+
75
+ ### Framework versions
76
+
77
+ - Transformers 4.48.0.dev0
78
+ - Pytorch 2.0.1+cu117
79
+ - Datasets 3.0.1
80
+ - Tokenizers 0.21.0