Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | @@ -235,7 +235,7 @@ This is a simple example of how to use **Granite-8B-Code-Instruct** model. | |
| 235 | 
             
            import torch
         | 
| 236 | 
             
            from transformers import AutoModelForCausalLM, AutoTokenizer
         | 
| 237 | 
             
            device = "cuda" # or "cpu"
         | 
| 238 | 
            -
            model_path = "granite-8b-code-instruct"
         | 
| 239 | 
             
            tokenizer = AutoTokenizer.from_pretrained(model_path)
         | 
| 240 | 
             
            # drop device_map if running on CPU
         | 
| 241 | 
             
            model = AutoModelForCausalLM.from_pretrained(model_path, device_map=device)
         | 
|  | |
| 235 | 
             
            import torch
         | 
| 236 | 
             
            from transformers import AutoModelForCausalLM, AutoTokenizer
         | 
| 237 | 
             
            device = "cuda" # or "cpu"
         | 
| 238 | 
            +
            model_path = "ibm-granite/granite-8b-code-instruct"
         | 
| 239 | 
             
            tokenizer = AutoTokenizer.from_pretrained(model_path)
         | 
| 240 | 
             
            # drop device_map if running on CPU
         | 
| 241 | 
             
            model = AutoModelForCausalLM.from_pretrained(model_path, device_map=device)
         | 

