| --- |
| base_model: Qwen/Qwen2.5-Coder-14B |
| library_name: peft |
| pipeline_tag: text-generation |
| license: apache-2.0 |
| tags: |
| - code |
| - coding-assistant |
| - lora |
| - refactor |
| - bug-fix |
| - optimization |
| - async |
| - concurrency |
| - security |
| - logging |
| - networking |
| - transformers |
| --- |
| |
| # EdgePulse Coder 14B (LoRA) |
|
|
| **EdgePulse Coder 14B** is a production-grade coding assistant fine-tuned using LoRA on top of **Qwen2.5-Coder-14B**. |
| It is designed to handle real-world software engineering workflows with high reliability and correctness. |
|
|
| --- |
|
|
| ## Model Details |
|
|
| ### Model Description |
|
|
| EdgePulse Coder 14B focuses on **practical developer tasks**, trained on a large, strictly validated dataset covering: |
|
|
| - Bug fixing |
| - Code explanation |
| - Refactoring |
| - Optimization |
| - Async & concurrency correction |
| - Logging & observability |
| - Security & defensive coding |
| - Networking & I/O handling |
| - Multi-file context reasoning |
| - Test generation and impact analysis |
|
|
| The model is optimized for **IDE usage**, **CLI workflows**, and **Cursor-like streaming environments**. |
|
|
| --- |
|
|
| - **Developed by:** EdgePulseAI |
| - **Shared by:** EdgePulseAI |
| - **Model type:** Large Language Model (Code-focused) |
| - **Language(s):** Python, JavaScript, TypeScript, Bash (primary), general programming concepts |
| - **License:** Apache-2.0 |
| - **Finetuned from:** Qwen/Qwen2.5-Coder-14B |
|
|
| --- |
|
|
| ## Model Sources |
|
|
| - **Base Model:** https://huggingface.co/Qwen/Qwen2.5-Coder-14B |
| - **Website:** https://EdgePulseAi.com |
|
|
| --- |
|
|
| ## Uses |
|
|
| ### Direct Use |
|
|
| EdgePulse Coder 14B can be used directly for: |
|
|
| - Code explanation |
| - Bug fixing |
| - Refactoring existing code |
| - Generating tests |
| - Improving logging and error handling |
| - Fixing async / concurrency bugs |
| - Secure coding suggestions |
| - Network & I/O robustness |
|
|
| ### Downstream Use |
|
|
| - IDE assistants (VS Code / Cursor-style tools) |
| - CI/CD automation |
| - Code review bots |
| - Developer copilots |
| - Internal engineering tools |
|
|
| ### Out-of-Scope Use |
|
|
| - Medical or legal advice |
| - Autonomous system control |
| - High-risk decision making without human review |
|
|
| --- |
|
|
| ## Bias, Risks, and Limitations |
|
|
| - The model may occasionally produce syntactically correct but logically incorrect code. |
| - Security-sensitive code should always be reviewed by humans. |
| - Performance depends on correct prompt framing and context size. |
|
|
| ### Recommendations |
|
|
| - Use human review for production deployments. |
| - Combine with static analysis and testing tools. |
| - Prefer structured prompts for multi-file tasks. |
|
|
| --- |
|
|
| ## How to Get Started with the Model |
|
|
| ```python |
| from transformers import AutoTokenizer, AutoModelForCausalLM |
| from peft import PeftModel |
| |
| base_model = "Qwen/Qwen2.5-Coder-14B" |
| adapter_model = "edgepulse-ai/EdgePulse-Coder-14B-LoRA" |
| |
| tokenizer = AutoTokenizer.from_pretrained(base_model) |
| |
| model = AutoModelForCausalLM.from_pretrained( |
| base_model, |
| device_map="auto", |
| torch_dtype="auto" |
| ) |
| |
| model = PeftModel.from_pretrained(model, adapter_model) |
| model.eval() |
| |
| prompt = "Fix this bug:\n\ndef add(a,b): return a-b" |
| inputs = tokenizer(prompt, return_tensors="pt").to(model.device) |
| |
| output = model.generate(**inputs, max_new_tokens=128) |
| print(tokenizer.decode(output[0], skip_special_tokens=True)) |