Add link to code
Browse files
    	
        README.md
    CHANGED
    
    | @@ -36,9 +36,11 @@ The dev-set performance reported in the paper is averaged over three seeds, and | |
| 36 | 
             
            | stdev        | 0.108 | 0.035 |
         | 
| 37 | 
             
            ```
         | 
| 38 |  | 
| 39 | 
            -
            Code:  | 
| 40 |  | 
| 41 | 
            -
             | 
|  | |
|  | |
| 42 | 
             
            ```bibtex
         | 
| 43 | 
             
            @article{kurtic2022optimal,
         | 
| 44 | 
             
              title={The Optimal BERT Surgeon: Scalable and Accurate Second-Order Pruning for Large Language Models},
         | 
|  | |
| 36 | 
             
            | stdev        | 0.108 | 0.035 |
         | 
| 37 | 
             
            ```
         | 
| 38 |  | 
| 39 | 
            +
            Code: [https://github.com/neuralmagic/sparseml/tree/main/research/optimal_BERT_surgeon_oBERT](https://github.com/neuralmagic/sparseml/tree/main/research/optimal_BERT_surgeon_oBERT)
         | 
| 40 |  | 
| 41 | 
            +
            If you find the model useful, please consider citing our work.
         | 
| 42 | 
            +
             | 
| 43 | 
            +
            ## Citation info
         | 
| 44 | 
             
            ```bibtex
         | 
| 45 | 
             
            @article{kurtic2022optimal,
         | 
| 46 | 
             
              title={The Optimal BERT Surgeon: Scalable and Accurate Second-Order Pruning for Large Language Models},
         | 

