varun-v-rao commited on
Commit
71334b2
1 Parent(s): f557847

Upload GPT2ForSequenceClassification

Browse files
Files changed (3) hide show
  1. README.md +23 -55
  2. pytorch_adapter.bin +1 -1
  3. pytorch_model_head.bin +1 -1
README.md CHANGED
@@ -1,74 +1,42 @@
1
  ---
2
- license: mit
3
- base_model: openai-community/gpt2-large
4
  tags:
5
- - generated_from_trainer
 
6
  datasets:
7
- - stanfordnlp/snli
8
- metrics:
9
- - accuracy
10
- model-index:
11
- - name: gpt2-large-bn-adapter-7.42M-snli-model2
12
- results:
13
- - task:
14
- name: Text Classification
15
- type: text-classification
16
- dataset:
17
- name: snli
18
- type: stanfordnlp/snli
19
- metrics:
20
- - name: Accuracy
21
- type: accuracy
22
- value: 0.8996138996138996
23
  ---
24
 
25
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
26
- should probably proofread and complete it, then remove this comment. -->
27
 
28
- # gpt2-large-bn-adapter-7.42M-snli-model2
29
 
30
- This model is a fine-tuned version of [openai-community/gpt2-large](https://huggingface.co/openai-community/gpt2-large) on the snli dataset.
31
- It achieves the following results on the evaluation set:
32
- - Loss: 0.2746
33
- - Accuracy: 0.8996
34
 
35
- ## Model description
36
 
37
- More information needed
38
 
39
- ## Intended uses & limitations
 
 
40
 
41
- More information needed
42
 
43
- ## Training and evaluation data
 
44
 
45
- More information needed
 
 
46
 
47
- ## Training procedure
48
 
49
- ### Training hyperparameters
50
 
51
- The following hyperparameters were used during training:
52
- - learning_rate: 2e-05
53
- - train_batch_size: 32
54
- - eval_batch_size: 32
55
- - seed: 8
56
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
57
- - lr_scheduler_type: linear
58
- - num_epochs: 3
59
 
60
- ### Training results
61
 
62
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
63
- |:-------------:|:-----:|:-----:|:---------------:|:--------:|
64
- | 0.3612 | 1.0 | 17168 | 0.2979 | 0.8883 |
65
- | 0.3305 | 2.0 | 34336 | 0.2790 | 0.8971 |
66
- | 0.3089 | 3.0 | 51504 | 0.2746 | 0.8996 |
67
 
68
-
69
- ### Framework versions
70
-
71
- - Transformers 4.35.2
72
- - Pytorch 2.1.1+cu121
73
- - Datasets 2.15.0
74
- - Tokenizers 0.15.0
 
1
  ---
 
 
2
  tags:
3
+ - adapter-transformers
4
+ - gpt2
5
  datasets:
6
+ - snli
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  ---
8
 
9
+ # Adapter `varun-v-rao/gpt2-large-bn-adapter-7.42M-snli-model2` for openai-community/gpt2-large
 
10
 
11
+ An [adapter](https://adapterhub.ml) for the `openai-community/gpt2-large` model that was trained on the [snli](https://huggingface.co/datasets/snli/) dataset.
12
 
13
+ This adapter was created for usage with the **[Adapters](https://github.com/Adapter-Hub/adapters)** library.
 
 
 
14
 
15
+ ## Usage
16
 
17
+ First, install `adapters`:
18
 
19
+ ```
20
+ pip install -U adapters
21
+ ```
22
 
23
+ Now, the adapter can be loaded and activated like this:
24
 
25
+ ```python
26
+ from adapters import AutoAdapterModel
27
 
28
+ model = AutoAdapterModel.from_pretrained("openai-community/gpt2-large")
29
+ adapter_name = model.load_adapter("varun-v-rao/gpt2-large-bn-adapter-7.42M-snli-model2", source="hf", set_active=True)
30
+ ```
31
 
32
+ ## Architecture & Training
33
 
34
+ <!-- Add some description here -->
35
 
36
+ ## Evaluation results
 
 
 
 
 
 
 
37
 
38
+ <!-- Add some description here -->
39
 
40
+ ## Citation
 
 
 
 
41
 
42
+ <!-- Add some description here -->
 
 
 
 
 
 
pytorch_adapter.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08023fa652d647c217646c583c93657bc7a9dd56241f92bfffa9c139ac587f03
3
  size 29739378
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feb0594539dd83bd1053c797649e34345358178402727468489c3c49859e4381
3
  size 29739378
pytorch_model_head.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7bc126484bd566e5083b7b3cbcc6ab37250d0d2c29000dc962da3e05d907b99e
3
  size 16659
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83ae65accbd77489310754f56d891179fac9cd0a0986fbae4ad30d35c84b5bf2
3
  size 16659