An AI Agent Skill for running DeepSeek V3.2 LLM via Atlas Cloud API β coding, reasoning, analysis at $0.26/M input tokens. Access DeepSeek's powerful language models directly from your terminal with streaming support, file context injection, and dedicated code analysis modes.
- Features
- Model Variants
- Installation
- Quick Start
- Usage
- Agent Skill Integration
- Use Cases
- API Reference
- Configuration
- Examples
- Pricing
- Troubleshooting
- License
- Chat Completion β Send prompts and receive high-quality responses from DeepSeek V3.2, one of the most capable open-weight LLMs
- Streaming Output β Real-time token streaming via Server-Sent Events for responsive interactive sessions
- Code Generation β Dedicated code mode with optimized system prompts for writing, reviewing, and debugging code
- Reasoning & Analysis β Access DeepSeek R1-0528 for complex multi-step reasoning tasks with chain-of-thought
- 164K Context Window β Process massive documents, entire codebases, or long conversation histories
- OpenAI-Compatible API β Drop-in replacement using the familiar chat completions API format
- File Context Injection β Read local files and inject them as context for code review, analysis, or Q&A
- Agent Skill β Works with 15+ AI coding agents including Claude Code, Cursor, OpenAI Codex, GitHub Copilot, Gemini CLI, Windsurf, OpenCode, Kiro, and more
- Multiple Model Variants β Choose between speed, capability, and cost across 7 model options
- Temperature Control β Fine-tune response creativity from deterministic (0.0) to highly creative (2.0)
DeepSeek offers a powerful lineup of language models accessible through Atlas Cloud:
| Model | Context | Input Price | Output Price | Description |
|---|---|---|---|---|
| DeepSeek V3.2 | 164K | $0.26/M | $0.38/M | Latest flagship model β best overall performance |
| DeepSeek V3.2 Fast | 164K | $0.26/M | $0.38/M | Optimized for lower latency with equivalent quality |
| DeepSeek V3.2 Speciale | 164K | $0.26/M | $0.38/M | Specialized variant with enhanced instruction following |
| DeepSeek V3.2 Exp | 164K | $0.26/M | $0.38/M | Experimental features and cutting-edge capabilities |
| DeepSeek R1-0528 | 164K | $0.50/M | $2.18/M | Reasoning model with chain-of-thought for complex tasks |
| DeepSeek V3.1 | 164K | $0.26/M | $0.38/M | Previous generation with proven production reliability |
| DeepSeek OCR | 164K | $0.26/M | $0.38/M | Optimized for optical character recognition tasks |
deepseek-ai/deepseek-v3.2 β DeepSeek V3.2
deepseek-ai/deepseek-v3.2-fast β DeepSeek V3.2 Fast
deepseek-ai/deepseek-v3.2-speciale β DeepSeek V3.2 Speciale
deepseek-ai/deepseek-v3.2-exp β DeepSeek V3.2 Exp
deepseek-ai/deepseek-r1-0528 β DeepSeek R1-0528 (Reasoning)
deepseek-ai/deepseek-v3.1 β DeepSeek V3.1
deepseek-ai/deepseek-ocr β DeepSeek OCR
- Node.js 18+ and npm
- An Atlas Cloud API key (get one here)
git clone https://github.com/thoughtincode/deepseek-llm-skill.git
cd deepseek-llm-skill
npm install
npm run build
npm linkcp .env.example .env
# Edit .env and add your Atlas Cloud API key# Simple chat
deepseek "What is the time complexity of merge sort?"
# Stream the response in real-time
deepseek "Explain quantum entanglement" --stream
# Code review mode
deepseek "Review this code for bugs" --file ./src/main.ts --code
# Use the reasoning model for complex problems
deepseek "Prove that the square root of 2 is irrational" --model deepseek-ai/deepseek-r1-0528Send a prompt and receive a complete response:
# Simple question
deepseek "What are the main differences between REST and GraphQL?"
# Detailed analysis
deepseek "Compare the performance characteristics of B-trees vs. LSM-trees for database indexing"
# Creative writing
deepseek "Write a short story about an AI that discovers music" --temperature 1.2Enable real-time token streaming for responsive output:
# Stream a long response
deepseek "Write a comprehensive guide to Docker networking" --stream
# Stream with custom temperature
deepseek "Generate 10 creative startup ideas for AI in healthcare" --stream --temperature 0.9Activate code-focused mode with an optimized system prompt:
# Generate code
deepseek "Write a Python function to find the longest palindromic substring" --code
# Debug code
deepseek "Why does this function return undefined for empty arrays?" --code --file ./utils.js
# Architecture advice
deepseek "Design a microservices architecture for an e-commerce platform" --codeRead local files and include them as context:
# Code review
deepseek "Review this code for security vulnerabilities" --file ./src/auth.ts
# Explain code
deepseek "Explain what this code does step by step" --file ./algorithm.py
# Suggest improvements
deepseek "How can I optimize this database query?" --file ./queries.sql
# Multiple concerns
deepseek "Check for memory leaks and race conditions" --file ./server.go --codeCustomize the AI's behavior with system prompts:
# Technical writer persona
deepseek "Document the API endpoints in this file" --system "You are a senior technical writer. Write clear, concise API documentation with examples." --file ./routes.ts
# Security auditor
deepseek "Audit this code" --system "You are a security expert. Focus on OWASP Top 10 vulnerabilities." --file ./app.js
# Tutor mode
deepseek "Explain recursion" --system "You are a patient CS teacher. Use simple analogies and build up from basics."# Control response length
deepseek "Summarize this paper" --file ./paper.txt --max-tokens 500
# Deterministic output
deepseek "Convert this JSON to a TypeScript interface" --file ./data.json --temperature 0
# Use specific model
deepseek "Analyze this dataset" --model deepseek-ai/deepseek-v3.2-fast --file ./data.csvThis tool is designed to work as an AI agent skill compatible with 15+ coding agents including Claude Code, Cursor, OpenAI Codex, GitHub Copilot, Gemini CLI, Windsurf, OpenCode, Kiro, and more. Once installed and linked, your AI agent can invoke it directly when you ask for code analysis, review, or DeepSeek-powered tasks.
- Install the skill (see Installation)
- Your AI agent reads the skill configuration file (e.g.,
CLAUDE.md,.cursorrules, or equivalent) to understand available commands - When you ask for code review, analysis, or invoke DeepSeek, the agent constructs and runs the
deepseekcommand
You: Use DeepSeek to review my authentication module
Claude: I'll analyze your auth module with DeepSeek.
> deepseek "Review this authentication module for security vulnerabilities,
race conditions, and best practices compliance" --file ./src/auth.ts --code
[DeepSeek analysis output...]
You: Explain this complex algorithm using DeepSeek
Claude: I'll have DeepSeek explain the algorithm step by step.
> deepseek "Explain this algorithm in detail, including time and space
complexity analysis" --file ./src/algorithm.ts --code --stream
[Streaming explanation...]
You: Use the reasoning model to solve this optimization problem
Claude: I'll use DeepSeek R1 for this complex reasoning task.
> deepseek "Find the optimal solution for this scheduling problem
and prove its correctness" --model deepseek-ai/deepseek-r1-0528
--file ./problem.txt
[Chain-of-thought reasoning output...]
# Comprehensive review
deepseek "Perform a thorough code review. Check for bugs, security issues, performance problems, and suggest improvements" --file ./src/api.ts --code
# PR review
deepseek "Review this diff for potential issues" --file ./changes.diff --code# Debug error
deepseek "This function throws 'Cannot read property of undefined' on line 42. Why?" --file ./broken.js --code
# Trace logic error
deepseek "The output is incorrect for edge cases. Find the bug" --file ./calculator.py --code# Generate docs
deepseek "Generate comprehensive JSDoc comments for all exported functions" --file ./lib.ts --code
# API documentation
deepseek "Create OpenAPI/Swagger documentation for these endpoints" --file ./routes.ts --code# Analyze CSV structure
deepseek "Analyze this dataset. Describe the schema, identify patterns, and suggest queries" --file ./sales.csv
# SQL optimization
deepseek "Optimize these SQL queries for better performance" --file ./queries.sql --code# Technical translation
deepseek "Translate this README to Japanese while preserving technical accuracy" --file ./README.md
# Code comments
deepseek "Translate all comments in this file from Chinese to English" --file ./legacy.py --codePOST https://api.atlascloud.ai/v1/chat/completions
Headers:
Authorization: Bearer <ATLAS_API_KEY>
Content-Type: application/json
Request Body:
{
"model": "deepseek-ai/deepseek-v3.2",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant."
},
{
"role": "user",
"content": "Explain the CAP theorem"
}
],
"temperature": 0.7,
"max_tokens": 4096,
"stream": false,
"top_p": 1.0
}Response (non-streaming):
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1710000000,
"model": "deepseek-ai/deepseek-v3.2",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "The CAP theorem states that..."
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 25,
"completion_tokens": 150,
"total_tokens": 175
}
}Response (streaming):
data: {"id":"chatcmpl-abc123","choices":[{"delta":{"content":"The"},"index":0}]}
data: {"id":"chatcmpl-abc123","choices":[{"delta":{"content":" CAP"},"index":0}]}
data: {"id":"chatcmpl-abc123","choices":[{"delta":{"content":" theorem"},"index":0}]}
data: [DONE]
| Parameter | Type | Required | Default | Description |
|---|---|---|---|---|
model |
string | Yes | β | Model identifier (see Model Variants) |
messages |
array | Yes | β | Array of message objects with role and content |
temperature |
number | No | 0.7 |
Sampling temperature (0.0 - 2.0) |
max_tokens |
number | No | 4096 |
Maximum tokens in response |
stream |
boolean | No | false |
Enable Server-Sent Events streaming |
top_p |
number | No | 1.0 |
Nucleus sampling threshold |
| Variable | Required | Description |
|---|---|---|
ATLAS_API_KEY |
Yes | Your Atlas Cloud API key |
Usage: deepseek [options] <prompt>
Arguments:
prompt The prompt or question to send
Options:
--model <model> Model identifier (default: deepseek-ai/deepseek-v3.2)
--system <prompt> System prompt to set behavior
--temperature <temp> Sampling temperature 0.0-2.0 (default: 0.7)
--max-tokens <n> Maximum response tokens (default: 4096)
--stream Enable streaming output
--file <path> Read file and include as context
--code Enable code-focused mode with optimized system prompt
--top-p <p> Nucleus sampling threshold (default: 1.0)
-h, --help Show help
-v, --version Show version
# Write a complete implementation
deepseek "Implement a thread-safe LRU cache in Rust with O(1) get/put operations" --code
# Convert between languages
deepseek "Convert this Python code to idiomatic Go" --file ./script.py --code
# Write tests
deepseek "Write comprehensive unit tests for this module using Jest" --file ./utils.ts --code
# Refactor
deepseek "Refactor this code to use the strategy pattern" --file ./handler.ts --code --stream# Complex reasoning with R1
deepseek "Analyze the time complexity of this algorithm and suggest optimizations" \
--model deepseek-ai/deepseek-r1-0528 --file ./sort.py --code
# Architecture analysis
deepseek "Evaluate this microservices architecture for scalability bottlenecks" --file ./arch.md
# Data interpretation
deepseek "Interpret these benchmark results and explain the performance differences" --file ./benchmarks.json# Summarize documents
deepseek "Summarize the key points of this document in bullet points" --file ./report.pdf --max-tokens 500
# Generate commit messages
deepseek "Write a conventional commit message for these changes" --file ./changes.diff
# Email drafting
deepseek "Draft a professional email declining this meeting request while suggesting alternatives" --temperature 0.5# Writing assistance
deepseek "Continue this story from where it left off, maintaining the same tone and style" \
--file ./story.txt --temperature 1.0
# Brainstorming
deepseek "Generate 20 unique feature ideas for a fitness app targeting seniors" --temperature 1.2 --streamAll pricing is per million tokens through Atlas Cloud:
| Model | Input (per M tokens) | Output (per M tokens) |
|---|---|---|
| DeepSeek V3.2 | $0.26 | $0.38 |
| DeepSeek V3.2 Fast | $0.26 | $0.38 |
| DeepSeek V3.2 Speciale | $0.26 | $0.38 |
| DeepSeek V3.2 Exp | $0.26 | $0.38 |
| DeepSeek R1-0528 | $0.50 | $2.18 |
| DeepSeek V3.1 | $0.26 | $0.38 |
| DeepSeek OCR | $0.26 | $0.38 |
Cost Comparison: DeepSeek V3.2 at $0.26/M input tokens is one of the most cost-effective frontier-class LLMs available, offering GPT-4 class performance at a fraction of the cost.
API Key Not Found
Error: ATLAS_API_KEY environment variable is not set
Solution: Copy .env.example to .env and add your API key, or export it directly:
export ATLAS_API_KEY=your_key_hereContext Too Long
Error: Maximum context length exceeded
Solution: The model supports 164K tokens, but very large files may exceed this. Try:
- Sending only relevant portions of the file
- Using
--max-tokensto limit output - Splitting large files into sections
Rate Limited
Error: Rate limit exceeded (429)
Solution: Wait a few seconds and retry. Consider upgrading your Atlas Cloud tier for higher rate limits.
Stream Connection Dropped
Error: Stream connection closed unexpectedly
Solution: This can happen with very long responses. Try without --stream flag, or reduce --max-tokens.
File Not Found
Error: File not found: ./src/main.ts
Solution: Ensure the file path is correct relative to your current working directory. Use absolute paths if needed.
- Open an issue on GitHub
- Check Atlas Cloud documentation
This workflow is optimized for Atlas Cloud. Move from experiment to enterprise-ready scale.
- Production-Ready: Hosted DeepSeek deployments at only $0.26/M input tokens
- Zero Maintenance: Serverless architecture β focus on your product, not the servers
- Enterprise Security: SOC I & II Certified | HIPAA Compliant
- 25% Bonus: First recharge up to $100
This project is licensed under the MIT License β see the LICENSE file for details.
Contributions are welcome! Please feel free to submit a Pull Request.
- Fork the repository
- Create your feature branch (
git checkout -b feature/amazing-feature) - Commit your changes (
git commit -m 'Add amazing feature') - Push to the branch (
git push origin feature/amazing-feature) - Open a Pull Request
- DeepSeek for the groundbreaking open-weight language models
- Atlas Cloud for the API infrastructure
- Claude Code for the skill integration framework, and the broader AI coding agent ecosystem for cross-platform compatibility
Built with DeepSeek via Atlas Cloud