gss1147's picture
Update README.md
29c5663 verified
---
license: other
base_model:
- Qwen/Qwen3.5-4B
library_name: llama.cpp
tags:
- gguf
- qwen
- qwen3.5
- code
- coder
- conversational
- text-generation
- withinusai
language:
- en
datasets:
- WithinUsAI/Python_GOD_Coder_50k
- reedmayhew/gemini-3.1-pro-2048-reasoning-1100x
- m-a-p/Code-Feedback
- crownelius/Opus-4.6-Reasoning-2100x-formatted
- crownelius/Opus4.6-No-Reasoning-260x
- crownelius/Creative_Writing_Multiturn_Enhanced
- HuggingFaceH4/llava-instruct-mix-vsft
- Roman1111111/gemini-3-pro-10000x-hard-high-reasoning
model_type: gguf
inference: false
---
# WithIn-Us-Coder-4B.gguf
**WithIn-Us-Coder-4B.gguf** is a GGUF release from **WithIn Us AI**, built for local inference and coding-focused assistant use cases. It is based on **Qwen/Qwen3.5-4B** and distributed in quantized GGUF formats for efficient deployment in llama.cpp-compatible runtimes.
## Model Summary
This model is intended as a coding-oriented conversational assistant with emphasis on:
- code generation
- code reasoning
- implementation planning
- debugging assistance
- instruction following
- general assistant-style chat for development workflows
This repository currently provides the following GGUF variants:
- `WithIn-Us-Coder-4B.Q4_K_M.gguf`
- `WithIn-Us-Coder-4B.Q5_K_M.gguf`
## Creator
**WithIn Us AI** is the creator of this model release, including the model packaging, fine-tuning / merging concept, process, naming, and GGUF distribution.
## Base Model
This model is based on:
- **Qwen/Qwen3.5-4B**
Credit and appreciation go to the original creators of the base LLM architecture and weights.
## Training Data
The current repository metadata lists the following datasets as part of the model’s training / fine-tuning lineage:
- `WithinUsAI/Python_GOD_Coder_50k`
- `reedmayhew/gemini-3.1-pro-2048-reasoning-1100x`
- `m-a-p/Code-Feedback`
- `crownelius/Opus-4.6-Reasoning-2100x-formatted`
- `crownelius/Opus4.6-No-Reasoning-260x`
- `crownelius/Creative_Writing_Multiturn_Enhanced`
- `HuggingFaceH4/llava-instruct-mix-vsft`
- `Roman1111111/gemini-3-pro-10000x-hard-high-reasoning`
**Attribution note:**
WithIn Us AI does not claim ownership over third-party base models or third-party datasets. Full credit, thanks, and attribution belong to the original model and dataset creators.
## Intended Use
This model is intended for:
- local coding assistants
- offline development help
- code explanation
- bug-fixing support
- prompt-based code generation
- experimentation in llama.cpp and GGUF-compatible environments
### Suggested Use Cases
- generating Python, JavaScript, C++, and other programming language snippets
- explaining code blocks
- rewriting or improving functions
- brainstorming implementation strategies
- creating scaffolding and prototypes
- assisting with debugging and refactoring
## Out-of-Scope Use
This model is not guaranteed to be reliable for:
- high-stakes legal advice
- medical advice
- financial decision-making
- autonomous execution without review
- security-critical production decisions without human verification
Users should always validate generated code before deployment.
## Quantization Formats
This repository currently includes:
- **Q4_K_M** for smaller memory footprint and faster local inference
- **Q5_K_M** for improved quality while remaining efficient
Choose the quant level based on your hardware budget and quality needs.
## Prompting Notes
As a coding-focused conversational model, best results usually come from prompts that are:
- specific
- structured
- explicit about language, framework, or goal
- clear about desired output format
Example prompt style:
> Write a Python function that parses a CSV file, removes duplicate rows by email, and saves the cleaned result. Include error handling and comments.
## Limitations
Like other language models, this model may:
- hallucinate APIs or library behavior
- generate insecure or inefficient code
- make reasoning mistakes
- produce outdated patterns
- require prompt iteration for best results
Human review is strongly recommended, especially for production code.
## License
This repository uses a **custom WithIn Us AI license approach**.
- The base model may be subject to its original upstream license and terms.
- Third-party datasets remain the property of their respective creators / licensors.
- WithIn Us AI claims authorship of the fine-tuning / merging concept, process, packaging, naming, and release structure for this model distribution.
- This repository does **not** claim ownership over third-party datasets or the underlying upstream base model.
You can include a `LICENSE` file in this repository with the exact custom terms you want enforced.
## Acknowledgments
Special thanks to:
- **Qwen** for the base model
- all third-party dataset creators listed above
- the open-source GGUF / llama.cpp ecosystem
- the broader Hugging Face community
## Files
Current repository files include:
- `WithIn-Us-Coder-4B.Q4_K_M.gguf`
- `WithIn-Us-Coder-4B.Q5_K_M.gguf`
## Disclaimer
This model may generate incorrect, biased, insecure, or incomplete outputs.
Use responsibly, validate important results, and review all generated code before real-world use.