Update README.md
Browse files
README.md
CHANGED
|
@@ -54,7 +54,7 @@ The model was converted using [this notebook](https://nbviewer.org/urls/huggingf
|
|
| 54 |
### How to use
|
| 55 |
|
| 56 |
```sh
|
| 57 |
-
wget https://huggingface.co/mrm8488/bertin-gpt-j-6B-ES-8bit/resolve/main/utils.py -O
|
| 58 |
pip install transformers
|
| 59 |
pip install bitsandbytes-cuda111==0.26.0
|
| 60 |
```
|
|
@@ -63,28 +63,7 @@ pip install bitsandbytes-cuda111==0.26.0
|
|
| 63 |
import transformers
|
| 64 |
import torch
|
| 65 |
|
| 66 |
-
from
|
| 67 |
-
|
| 68 |
-
|
| 69 |
-
class GPTJBlock(transformers.models.gptj.modeling_gptj.GPTJBlock):
|
| 70 |
-
def __init__(self, config):
|
| 71 |
-
super().__init__(config)
|
| 72 |
-
|
| 73 |
-
convert_to_int8(self.attn)
|
| 74 |
-
convert_to_int8(self.mlp)
|
| 75 |
-
|
| 76 |
-
|
| 77 |
-
class GPTJModel(transformers.models.gptj.modeling_gptj.GPTJModel):
|
| 78 |
-
def __init__(self, config):
|
| 79 |
-
super().__init__(config)
|
| 80 |
-
convert_to_int8(self)
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
class GPTJForCausalLM(transformers.models.gptj.modeling_gptj.GPTJForCausalLM):
|
| 84 |
-
def __init__(self, config):
|
| 85 |
-
super().__init__(config)
|
| 86 |
-
convert_to_int8(self)
|
| 87 |
-
|
| 88 |
|
| 89 |
device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
| 90 |
|
|
|
|
| 54 |
### How to use
|
| 55 |
|
| 56 |
```sh
|
| 57 |
+
wget https://huggingface.co/mrm8488/bertin-gpt-j-6B-ES-8bit/resolve/main/utils.py -O Utils.py
|
| 58 |
pip install transformers
|
| 59 |
pip install bitsandbytes-cuda111==0.26.0
|
| 60 |
```
|
|
|
|
| 63 |
import transformers
|
| 64 |
import torch
|
| 65 |
|
| 66 |
+
from Utils import GPTJBlock, GPTJForCausalLM
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 67 |
|
| 68 |
device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
| 69 |
|