Update README.md
Browse files
README.md
CHANGED
|
@@ -20,11 +20,7 @@ Phi-1.5 can write poems, draft emails, create stories, summarize texts, write Py
|
|
| 20 |
|
| 21 |
## How to Use
|
| 22 |
|
| 23 |
-
Phi-1.5 has been integrated in the `transformers` version 4.37.0
|
| 24 |
-
|
| 25 |
-
* When loading the model, ensure that `trust_remote_code=True` is passed as an argument of the `from_pretrained()` function.
|
| 26 |
-
|
| 27 |
-
The current `transformers` version can be verified with: `pip list | grep transformers`.
|
| 28 |
|
| 29 |
## Intended Uses
|
| 30 |
|
|
@@ -91,8 +87,6 @@ where the model generates the text after the comments.
|
|
| 91 |
|
| 92 |
* Phi-1.5 has not been tested to ensure that it performs adequately for any production-level application. Please refer to the limitation sections of this document for more details.
|
| 93 |
|
| 94 |
-
* If you are using `transformers<4.37.0`, always load the model with `trust_remote_code=True` to prevent side-effects.
|
| 95 |
-
|
| 96 |
## Sample Code
|
| 97 |
|
| 98 |
```python
|
|
@@ -101,8 +95,8 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
| 101 |
|
| 102 |
torch.set_default_device("cuda")
|
| 103 |
|
| 104 |
-
model = AutoModelForCausalLM.from_pretrained("microsoft/phi-1_5", torch_dtype="auto"
|
| 105 |
-
tokenizer = AutoTokenizer.from_pretrained("microsoft/phi-1_5"
|
| 106 |
|
| 107 |
inputs = tokenizer('''def print_prime(n):
|
| 108 |
"""
|
|
|
|
| 20 |
|
| 21 |
## How to Use
|
| 22 |
|
| 23 |
+
Phi-1.5 has been integrated in the `transformers` version 4.37.0, please ensure that you are using a version equal or higher than it.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
|
| 25 |
## Intended Uses
|
| 26 |
|
|
|
|
| 87 |
|
| 88 |
* Phi-1.5 has not been tested to ensure that it performs adequately for any production-level application. Please refer to the limitation sections of this document for more details.
|
| 89 |
|
|
|
|
|
|
|
| 90 |
## Sample Code
|
| 91 |
|
| 92 |
```python
|
|
|
|
| 95 |
|
| 96 |
torch.set_default_device("cuda")
|
| 97 |
|
| 98 |
+
model = AutoModelForCausalLM.from_pretrained("microsoft/phi-1_5", torch_dtype="auto")
|
| 99 |
+
tokenizer = AutoTokenizer.from_pretrained("microsoft/phi-1_5")
|
| 100 |
|
| 101 |
inputs = tokenizer('''def print_prime(n):
|
| 102 |
"""
|