-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathModelfile
More file actions
30 lines (26 loc) · 1.05 KB
/
Modelfile
File metadata and controls
30 lines (26 loc) · 1.05 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
# NL2Shell Ollama Modelfile
#
# NOTE: Qwen3.5's hybrid DeltaNet architecture is not yet supported by Ollama's
# GGUF loader. Once Ollama adds qwen35 architecture support, this Modelfile will
# work with the GGUF files on HuggingFace:
#
# ollama create nl2shell -f Modelfile
#
# Until then, Vox CLI defaults to qwen2.5-coder:0.5b as a fallback.
# The fine-tuned model can still be used via Python transformers:
#
# from transformers import AutoModelForCausalLM, AutoTokenizer
# model = AutoModelForCausalLM.from_pretrained("AryaYT/nl2shell-0.8b")
#
# For v3 training, consider Qwen2.5-Coder-1.5B as an Ollama-compatible base.
FROM gguf/Qwen3.5-0.8B.Q4_K_M.gguf
SYSTEM "You are an expert shell programmer. Given a natural language request, output ONLY the corresponding shell command. No explanations, no markdown, no code fences, no comments. Just the raw command."
PARAMETER temperature 0.1
PARAMETER num_predict 256
PARAMETER stop "<|im_end|>"
TEMPLATE """<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assistant
"""