Jan-v2-VL-f32-AIO-GGUF

Jan-v2-VL is an 8-billion-parameter vision-language model designed for long-horizon, multi-step task execution in real software environments such as browsers and desktop applications. It tightly integrates language reasoning with visual perception to follow complex instructions, maintain intermediate states, and recover from minor execution errors, enabling stable and continuous task completion. The model is evaluated with a benchmark focused on long-horizon execution, confirming its strength in steady, low-drift stepwise operation. Jan-v2-VL comes in three variantsโ€”low (efficiency and lower latency), medium (balanced latency and quality), and high (deeper reasoning with higher think time)โ€”to best suit different performance needs. It supports agentic automation and UI control via screenshot grounding and tool calls and is optimized for local deployment with popular frameworks like vLLM and llama.cpp, offering a powerful tool for automation and interaction tasks in complex visual and textual environments.

Models

Jan-v2-VL-high

File Name Quant Type File Size
Jan-v2-VL-high-BF16.gguf BF16 16.4 GB
Jan-v2-VL-high-F16.gguf F16 16.4 GB
Jan-v2-VL-high-F32.gguf F32 32.8 GB
Jan-v2-VL-high-Q3_K_L.gguf Q3_K_L 4.43 GB
Jan-v2-VL-high-Q3_K_M.gguf Q3_K_M 4.12 GB
Jan-v2-VL-high-Q3_K_S.gguf Q3_K_S 3.77 GB
Jan-v2-VL-high-Q4_0.gguf Q4_0 4.77 GB
Jan-v2-VL-high-Q4_1.gguf Q4_1 5.25 GB
Jan-v2-VL-high-Q4_K_M.gguf Q4_K_M 5.03 GB
Jan-v2-VL-high-Q4_K_S.gguf Q4_K_S 4.8 GB
Jan-v2-VL-high-Q5_0.gguf Q5_0 5.72 GB
Jan-v2-VL-high-Q5_1.gguf Q5_1 6.19 GB
Jan-v2-VL-high-Q5_K_M.gguf Q5_K_M 5.85 GB
Jan-v2-VL-high-Q5_K_S.gguf Q5_K_S 5.72 GB
Jan-v2-VL-high-Q6_K.gguf Q6_K 6.73 GB
Jan-v2-VL-high-Q8_0.gguf Q8_0 8.71 GB
Jan-v2-VL-high-mmproj-bf16.gguf mmproj-bf16 1.16 GB
Jan-v2-VL-high-mmproj-f16.gguf mmproj-f16 1.16 GB
Jan-v2-VL-high-mmproj-f32.gguf mmproj-f32 2.31 GB
Jan-v2-VL-high-mmproj-q8_0.gguf mmproj-q8_0 752 MB
Jan-v2-VL-high.i1-IQ1_M.gguf i1-IQ1_M 2.26 GB
Jan-v2-VL-high.i1-IQ1_S.gguf i1-IQ1_S 2.12 GB
Jan-v2-VL-high.i1-IQ2_M.gguf i1-IQ2_M 3.05 GB
Jan-v2-VL-high.i1-IQ2_S.gguf i1-IQ2_S 2.86 GB
Jan-v2-VL-high.i1-IQ2_XS.gguf i1-IQ2_XS 2.7 GB
Jan-v2-VL-high.i1-IQ2_XXS.gguf i1-IQ2_XXS 2.49 GB
Jan-v2-VL-high.i1-IQ3_M.gguf i1-IQ3_M 3.9 GB
Jan-v2-VL-high.i1-IQ3_S.gguf i1-IQ3_S 3.79 GB
Jan-v2-VL-high.i1-IQ3_XS.gguf i1-IQ3_XS 3.63 GB
Jan-v2-VL-high.i1-IQ3_XXS.gguf i1-IQ3_XXS 3.37 GB
Jan-v2-VL-high.i1-IQ4_NL.gguf i1-IQ4_NL 4.79 GB
Jan-v2-VL-high.i1-IQ4_XS.gguf i1-IQ4_XS 4.56 GB
Jan-v2-VL-high.i1-Q2_K.gguf i1-Q2_K 3.28 GB
Jan-v2-VL-high.i1-Q2_K_S.gguf i1-Q2_K_S 3.08 GB
Jan-v2-VL-high.i1-Q3_K_L.gguf i1-Q3_K_L 4.43 GB
Jan-v2-VL-high.i1-Q3_K_M.gguf i1-Q3_K_M 4.12 GB
Jan-v2-VL-high.i1-Q3_K_S.gguf i1-Q3_K_S 3.77 GB
Jan-v2-VL-high.i1-Q4_0.gguf i1-Q4_0 4.79 GB
Jan-v2-VL-high.i1-Q4_1.gguf i1-Q4_1 5.25 GB
Jan-v2-VL-high.i1-Q4_K_M.gguf i1-Q4_K_M 5.03 GB
Jan-v2-VL-high.i1-Q4_K_S.gguf i1-Q4_K_S 4.8 GB
Jan-v2-VL-high.i1-Q5_K_M.gguf i1-Q5_K_M 5.85 GB
Jan-v2-VL-high.i1-Q5_K_S.gguf i1-Q5_K_S 5.72 GB
Jan-v2-VL-high.i1-Q6_K.gguf i1-Q6_K 6.73 GB
Jan-v2-VL-high.imatrix.gguf imatrix 5.35 MB

Jan-v2-VL-med

File Name Quant Type File Size
Jan-v2-VL-med-BF16.gguf BF16 16.4 GB
Jan-v2-VL-med-F16.gguf F16 16.4 GB
Jan-v2-VL-med-F32.gguf F32 32.8 GB
Jan-v2-VL-med-Q3_K_L.gguf Q3_K_L 4.43 GB
Jan-v2-VL-med-Q3_K_M.gguf Q3_K_M 4.12 GB
Jan-v2-VL-med-Q3_K_S.gguf Q3_K_S 3.77 GB
Jan-v2-VL-med-Q4_0.gguf Q4_0 4.77 GB
Jan-v2-VL-med-Q4_1.gguf Q4_1 5.25 GB
Jan-v2-VL-med-Q4_K_M.gguf Q4_K_M 5.03 GB
Jan-v2-VL-med-Q4_K_S.gguf Q4_K_S 4.8 GB
Jan-v2-VL-med-Q5_0.gguf Q5_0 5.72 GB
Jan-v2-VL-med-Q5_1.gguf Q5_1 6.19 GB
Jan-v2-VL-med-Q5_K_M.gguf Q5_K_M 5.85 GB
Jan-v2-VL-med-Q5_K_S.gguf Q5_K_S 5.72 GB
Jan-v2-VL-med-Q6_K.gguf Q6_K 6.73 GB
Jan-v2-VL-med-Q8_0.gguf Q8_0 8.71 GB
Jan-v2-VL-med-mmproj-bf16.gguf mmproj-bf16 1.16 GB
Jan-v2-VL-med-mmproj-f16.gguf mmproj-f16 1.16 GB
Jan-v2-VL-med-mmproj-f32.gguf mmproj-f32 2.31 GB
Jan-v2-VL-med-mmproj-q8_0.gguf mmproj-q8_0 752 MB
Jan-v2-VL-med.i1-IQ1_M.gguf i1-IQ1_M 2.26 GB
Jan-v2-VL-med.i1-IQ1_S.gguf i1-IQ1_S 2.12 GB
Jan-v2-VL-med.i1-IQ2_M.gguf i1-IQ2_M 3.05 GB
Jan-v2-VL-med.i1-IQ2_S.gguf i1-IQ2_S 2.86 GB
Jan-v2-VL-med.i1-IQ2_XS.gguf i1-IQ2_XS 2.7 GB
Jan-v2-VL-med.i1-IQ2_XXS.gguf i1-IQ2_XXS 2.49 GB
Jan-v2-VL-med.i1-IQ3_M.gguf i1-IQ3_M 3.9 GB
Jan-v2-VL-med.i1-IQ3_S.gguf i1-IQ3_S 3.79 GB
Jan-v2-VL-med.i1-IQ3_XS.gguf i1-IQ3_XS 3.63 GB
Jan-v2-VL-med.i1-IQ3_XXS.gguf i1-IQ3_XXS 3.37 GB
Jan-v2-VL-med.i1-IQ4_NL.gguf i1-IQ4_NL 4.79 GB
Jan-v2-VL-med.i1-IQ4_XS.gguf i1-IQ4_XS 4.56 GB
Jan-v2-VL-med.i1-Q2_K.gguf i1-Q2_K 3.28 GB
Jan-v2-VL-med.i1-Q2_K_S.gguf i1-Q2_K_S 3.08 GB
Jan-v2-VL-med.i1-Q3_K_L.gguf i1-Q3_K_L 4.43 GB
Jan-v2-VL-med.i1-Q3_K_M.gguf i1-Q3_K_M 4.12 GB
Jan-v2-VL-med.i1-Q3_K_S.gguf i1-Q3_K_S 3.77 GB
Jan-v2-VL-med.i1-Q4_0.gguf i1-Q4_0 4.79 GB
Jan-v2-VL-med.i1-Q4_1.gguf i1-Q4_1 5.25 GB
Jan-v2-VL-med.i1-Q4_K_M.gguf i1-Q4_K_M 5.03 GB
Jan-v2-VL-med.i1-Q4_K_S.gguf i1-Q4_K_S 4.8 GB
Jan-v2-VL-med.i1-Q5_K_M.gguf i1-Q5_K_M 5.85 GB
Jan-v2-VL-med.i1-Q5_K_S.gguf i1-Q5_K_S 5.72 GB
Jan-v2-VL-med.i1-Q6_K.gguf i1-Q6_K 6.73 GB
Jan-v2-VL-med.imatrix.gguf imatrix 5.35 MB

Jan-v2-VL-low

File Name Quant Type File Size
Jan-v2-VL-low-BF16.gguf BF16 16.4 GB
Jan-v2-VL-low-F16.gguf F16 16.4 GB
Jan-v2-VL-low-F32.gguf F32 32.8 GB
Jan-v2-VL-low-Q3_K_L.gguf Q3_K_L 4.43 GB
Jan-v2-VL-low-Q3_K_M.gguf Q3_K_M 4.12 GB
Jan-v2-VL-low-Q3_K_S.gguf Q3_K_S 3.77 GB
Jan-v2-VL-low-Q4_0.gguf Q4_0 4.77 GB
Jan-v2-VL-low-Q4_1.gguf Q4_1 5.25 GB
Jan-v2-VL-low-Q4_K_M.gguf Q4_K_M 5.03 GB
Jan-v2-VL-low-Q4_K_S.gguf Q4_K_S 4.8 GB
Jan-v2-VL-low-Q5_0.gguf Q5_0 5.72 GB
Jan-v2-VL-low-Q5_1.gguf Q5_1 6.19 GB
Jan-v2-VL-low-Q5_K_M.gguf Q5_K_M 5.85 GB
Jan-v2-VL-low-Q5_K_S.gguf Q5_K_S 5.72 GB
Jan-v2-VL-low-Q6_K.gguf Q6_K 6.73 GB
Jan-v2-VL-low-Q8_0.gguf Q8_0 8.71 GB
Jan-v2-VL-low-mmproj-bf16.gguf mmproj-bf16 1.16 GB
Jan-v2-VL-low-mmproj-f16.gguf mmproj-f16 1.16 GB
Jan-v2-VL-low-mmproj-f32.gguf mmproj-f32 2.31 GB
Jan-v2-VL-low-mmproj-q8_0.gguf mmproj-q8_0 752 MB

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
3,715
GGUF
Model size
8B params
Architecture
qwen3vl
Hardware compatibility
Log In to view the estimation

1-bit

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for prithivMLmods/Jan-v2-VL-f32-AIO-GGUF

Quantized
(8)
this model