-
Notifications
You must be signed in to change notification settings - Fork 684
Add LlamaGate model provider #2365
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Adds LlamaGate (https://llamagate.dev) as a new model provider with: LLM Models (12): - Llama 3.1 8B Instruct, Llama 3.2 3B - DeepSeek R1 8B, DeepSeek R1 Distill Qwen 7B - Qwen 3 8B, Mistral 7B v0.3 - Qwen 2.5 Coder 7B, CodeLlama 7B, DeepSeek Coder 6.7B - Qwen 3 VL 8B (Vision), OpenThinker 7B, Dolphin 3 8B Embedding Models (2): - Nomic Embed Text - Qwen3 Embedding 8B Provider details: - OpenAI-compatible API at https://api.llamagate.dev/v1 - Competitive pricing: $0.02-$0.55 per 1M tokens - All models are open-weights (Apache 2.0, MIT, Llama, Gemma)
Summary of ChangesHello @hkd987, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request significantly enhances Dify's model offerings by introducing LlamaGate as a new, fully integrated provider. This addition broadens the selection of available open-source Large Language Models and embedding models, all accessible through a familiar OpenAI-compatible API. The integration aims to provide Dify users with more diverse, performant, and cost-effective options for their AI applications, backed by LlamaGate's competitive pricing and open-weight model philosophy. Highlights
🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console. Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request introduces LlamaGate as a new model provider, which is a great addition. The implementation correctly leverages the OpenAI-compatible base classes, simplifying the integration. However, I've identified a few critical issues where method parameters are not correctly passed to the superclass methods, which could lead to incorrect behavior or loss of functionality. Additionally, there are some opportunities for improvement in configuration and exception handling to enhance maintainability and correctness. Please see the detailed comments below.
- Add missing `user` parameter to LLM _invoke() call - Add missing `input_type` parameter to text embedding _invoke() call - Add 4 missing models to _position.yaml (codellama-7b, deepseek-coder-6.7b, openthinker-7b, dolphin3-8b) - Replace `raise ex` with bare `raise` to preserve stack traces 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
| @@ -0,0 +1,37 @@ | |||
| author: llamagate | |||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Author should be langgenius if you want to submit plugin to this repo.
crazywoola
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
See comments
Add LlamaGate Model Provider
Adds LlamaGate as a new model provider plugin for Dify.
Provider Details
https://api.llamagate.dev/v1)api_keycredentialModels Included
LLM Models (12):
Embedding Models (2):
Features
Checklist
dify_pluginis in requirements.txt