Jan-v2-VL-low 8-bit MLX

This is an 8-bit quantized MLX conversion of janhq/Jan-v2-VL-low.

Model Description

Jan-v2-VL is an 8-billion parameter vision-language model designed for long-horizon, multi-step tasks in real software environments. This "low" variant is optimized for faster inference while maintaining strong performance on agentic automation and UI control tasks.

Key Features:

  • Vision-language understanding for browser and desktop applications
  • Screenshot grounding and tool call capabilities
  • Stable multi-step execution with minimal performance drift
  • Error recovery and intermediate state maintenance

Quantization

This model was converted to MLX format with 8-bit quantization using MLX-VLM by Prince Canuma.

Conversion command:

mlx_vlm.convert --hf-path janhq/Jan-v2-VL-low --quantize --q-bits 8 --mlx-path Jan-v2-VL-low-8bit-mlx

Usage

Installation

pip install mlx-vlm

Python

from mlx_vlm import load, generate
from mlx_vlm.prompt_utils import apply_chat_template
from mlx_vlm.utils import load_config

# Load the model
model_path = "mlx-community/Jan-v2-VL-low-8bit-mlx"
model, processor = load(model_path)
config = load_config(model_path)

# Prepare input
image = ["path/to/image.jpg"]
prompt = "Describe this image."

# Apply chat template
formatted_prompt = apply_chat_template(
    processor, config, prompt, num_images=len(image)
)

# Generate output
output = generate(model, processor, formatted_prompt, image, verbose=False)
print(output)

Command Line

mlx_vlm.generate --model mlx-community/Jan-v2-VL-low-8bit-mlx --max-tokens 100 --prompt "Describe this image" --image path/to/image.jpg

Intended Use

This model is designed for:

  • Agentic automation and UI control
  • Stepwise operation in browsers and desktop applications
  • Screenshot grounding and tool calls
  • Long-horizon multi-step task execution

License

This model is released under the Apache 2.0 license.

Original Model

For more information, please refer to the original model: janhq/Jan-v2-VL-low

Acknowledgments

Downloads last month
23
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for mlx-community/Jan-v2-VL-low-8bit-mlx

Quantized
(4)
this model