Add/update the quantized ONNX model files and README.md for Transformers.js v3
#2
by
						
whitphx
	
							HF Staff
						- opened
							
					
Applied Quantizations
	
		
	
	
		β Based on model.onnx with slimming
	
None
β³ β int8: model_int8.onnx (added but JS-based E2E test failed)
/home/ubuntu/src/tjsmigration/node_modules/.pnpm/[email protected]/node_modules/onnxruntime-node/dist/backend.js:25
            __classPrivateFieldGet(this, _OnnxruntimeSessionHandler_inferenceSession, "f").loadModel(pathOrBuffer, options);
                                                                                           ^
Error: Could not find an implementation for ConvInteger(10) node with name '/vision_model/embeddings/patch_embedding/Conv_quant'
    at new OnnxruntimeSessionHandler (/home/ubuntu/src/tjsmigration/node_modules/.pnpm/[email protected]/node_modules/onnxruntime-node/dist/backend.js:25:92)
    at Immediate.<anonymous> (/home/ubuntu/src/tjsmigration/node_modules/.pnpm/[email protected]/node_modules/onnxruntime-node/dist/backend.js:67:29)
    at process.processImmediate (node:internal/timers:485:21)
Node.js v22.16.0
β³ β
 uint8: model_uint8.onnx (added)
β³ β
 q4: model_q4.onnx (added)
β³ β
 q4f16: model_q4f16.onnx (added)
β³ β
 bnb4: model_bnb4.onnx (added)
Xenova
	
				
		changed pull request status to
		merged
			

