rnj-1-GGUF

rnj-1 from EssentialAI is an 8.3B-parameter dense language model family trained from scratch on 8.4T tokens with a Gemma 3-like architecture featuring global attention, YaRN for 32K context extension, and 128K vocabulary, optimized for code generation, STEM tasks, math, science, and agentic workflows under Apache 2.0 license. The base rnj-1 and instruct-tuned rnj-1-instruct excel in benchmarks like HumanEval+, MBPP+, LiveCodeBench v6 (competing with larger models), SWE-bench Verified (20.8% bash-only, surpassing Gemini 2.0 Flash), MultiPL-E multilingual code, Enamel performance optimization, Berkeley Function Calling Leaderboard, GPQA-Diamond, and fill-in-the-middle (86.21% HE-FIM-Python), with strong pass@N scaling potential for specialization. Supporting vLLM/SGLang inference, llama.cpp quantization, Transformers tool-calling (Hermes parser), FIM, and integration with Cline, Claude Code, mini-SWE-agent for real-world coding assistants, PR fixes, and data analysis, it processes up to 24M token batches efficiently post mid-training/SFT phases using Muon optimizer.

rnj-1 [GGUF]

File Name Quant Type File Size File Link
rnj-1.BF16.gguf BF16 16.6 GB Download
rnj-1.F16.gguf F16 16.6 GB Download
rnj-1.Q8_0.gguf Q8_0 8.84 GB Download

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
501
GGUF
Model size
8B params
Architecture
gemma3
Hardware compatibility
Log In to view the estimation

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for prithivMLmods/rnj-1-GGUF

Base model

EssentialAI/rnj-1
Quantized
(3)
this model