Model does not run with VLLM
									2
	#3 opened 11 months ago
		by
		
				
 aswad546
							
						aswad546
	
Any idea when the evaluation data will be in for this model? would like to know how the performance differ from unquantized version of the model
#2 opened about 1 year ago
		by
		
				
 jahhs0n
							
						jahhs0n
	
Any chance your team is working on a 4-bit Llama-3.2-90B-Vision-Instruct-quantized.w4a16 version?
									2
	#1 opened about 1 year ago
		by
		
				
 mrhendrey
							
						mrhendrey
	
