GODsStrongestSoldier commited on
Commit
05c3d7d
·
verified ·
1 Parent(s): 45dc069

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +167 -1
README.md CHANGED
@@ -11,4 +11,170 @@ datasets:
11
  - nvidia/Nemotron-RL-instruction_following
12
  - nvidia/Nemotron-RL-agent-calendar_scheduling
13
  - nvidia/Nemotron-RL-instruction_following-structured_outputs
14
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  - nvidia/Nemotron-RL-instruction_following
12
  - nvidia/Nemotron-RL-agent-calendar_scheduling
13
  - nvidia/Nemotron-RL-instruction_following-structured_outputs
14
+ ---
15
+ Nvidia.Agentic.Coder-4B-GGUF
16
+
17
+ 📌 Model Overview
18
+
19
+ Model Name: WithinUsAI/Nvidia.Agentic.Coder-4B-GGUF
20
+ Organization: Within Us AI
21
+ Model Type: Code LLM (Agentic, Instruction-Following)
22
+ Parameter Size: 4B
23
+ Format: GGUF (quantized for local inference)
24
+ Primary Use: Agentic coding, tool-using workflows, software engineering reasoning
25
+
26
+ This model is part of the Within Us AI ecosystem focused on building agentic, reasoning-driven coding systems designed to think, act, and verify like real engineers. 
27
+
28
+
29
+
30
+ 🧬 Architecture & Lineage
31
+
32
+ * Base Family: NVIDIA Nemotron-style 4B class models (inferred lineage from naming + ecosystem alignment)
33
+ * Format Conversion: GGUF quantization for efficient local inference
34
+ * Training Approach:
35
+ * Instruction-tuned for coding tasks
36
+ * Agentic workflow emphasis (multi-step reasoning, tool usage)
37
+ * Likely merged / fine-tuned using Within Us AI proprietary pipelines
38
+
39
+ Related ecosystem models include:
40
+
41
+ * NVIDIA-Nemotron-3-Nano-4B
42
+ * Other 4B agentic coders and merges in the same class 
43
+
44
+
45
+
46
+ ⚙️ Key Capabilities
47
+
48
+ 🧑‍💻 Code Intelligence
49
+
50
+ * Multi-language code generation
51
+ * Bug fixing and refactoring
52
+ * Structured output generation
53
+
54
+ 🤖 Agentic Behavior
55
+
56
+ * Step-by-step reasoning
57
+ * Task decomposition
58
+ * Tool-calling alignment (design goal)
59
+
60
+ 🧠 Reasoning Focus
61
+
62
+ * Instruction-following with logical chaining
63
+ * Designed for evaluation-style datasets (tests-as-truth philosophy)
64
+
65
+
66
+
67
+ 📦 GGUF Quantization
68
+
69
+ GGUF allows efficient local inference with tools like:
70
+
71
+ * llama.cpp
72
+ * LM Studio
73
+ * Ollama (GGUF-compatible builds)
74
+
75
+ Typical quantizations for 4B GGUF models include:
76
+
77
+ * Q2_K (~1.8GB)
78
+ * Q3_K (~2.0–2.3GB)
79
+ * Q4_K (~2.5GB, recommended balance) 
80
+
81
+
82
+
83
+ 🚀 Intended Use
84
+
85
+ ✅ Ideal Use Cases
86
+
87
+ * Local AI coding assistants
88
+ * Autonomous coding agents
89
+ * SWE-bench style evaluation
90
+ * Tool-augmented workflows
91
+ * Offline developer copilots
92
+
93
+ ⚠️ Limitations
94
+
95
+ * Smaller 4B parameter size limits deep reasoning vs larger models
96
+ * Performance depends heavily on prompt structure
97
+ * Tool-use requires external orchestration (not built-in runtime)
98
+
99
+
100
+
101
+ 🛠️ Usage Example (llama.cpp)
102
+
103
+ ./main -m Nvidia.Agentic.Coder-4B.Q4_K.gguf \
104
+ -p "Write a Python function to parse JSON logs and extract errors." \
105
+ -n 512
106
+
107
+
108
+
109
+ 🧪 Training Philosophy (Within Us AI)
110
+
111
+ Within Us AI focuses on:
112
+
113
+ * Agentic AI systems
114
+ * Test-driven training (tests-as-truth)
115
+ * Diff-first patching workflows
116
+ * Secure and auditable code generation
117
+ * Evaluation-first development pipelines 
118
+
119
+
120
+
121
+ 📊 Evaluation
122
+
123
+ No formal benchmark results published yet.
124
+
125
+ Expected strengths:
126
+
127
+ * Strong instruction adherence
128
+ * Lightweight agentic reasoning
129
+ * Efficient local deployment
130
+
131
+
132
+
133
+ 📚 Datasets & Training Sources
134
+
135
+ This model follows the Within Us AI methodology:
136
+
137
+ * Proprietary datasets created by Within Us AI
138
+ * May include third-party datasets for training (no ownership claimed)
139
+ * Emphasis on:
140
+ * Code reasoning traces
141
+ * Agentic workflows
142
+ * Evaluation-driven samples
143
+
144
+
145
+
146
+ 📜 License
147
+
148
+ License Type: Custom / Other (Within Us AI License)
149
+
150
+ Terms:
151
+
152
+ * Within Us AI created the fine-tuning, merging, and training methodology
153
+ * Base model architecture originates from third-party LLM ecosystems (e.g., NVIDIA / Nemotron class)
154
+ * Third-party datasets may be used without claiming ownership
155
+ * Full credit and acknowledgment belong to original dataset and base model creators
156
+
157
+
158
+
159
+ 🙏 Acknowledgements
160
+
161
+ Special thanks to:
162
+
163
+ * NVIDIA Nemotron ecosystem contributors
164
+ * Open-source GGUF tooling community
165
+ * Dataset creators across Hugging Face
166
+ * The broader open-source AI research community
167
+
168
+
169
+
170
+ 🔗 Links
171
+
172
+ * Model: https://huggingface.co/WithinUsAI/Nvidia.Agentic.Coder-4B-GGUF
173
+ * Organization: https://huggingface.co/WithinUsAI
174
+
175
+
176
+
177
+ 🧩 Closing Note
178
+
179
+ This model is a compact engineer in a bottle 🧪
180
+ Not the biggest brain in the room, but fast, focused, and built to act, not just