Update Mistral-Large-3.md for AMD GPU#209
Update Mistral-Large-3.md for AMD GPU#209jiacao-amd wants to merge 2 commits intovllm-project:mainfrom
Conversation
Summary of ChangesHello @jiacao-amd, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request significantly enhances the documentation for the Mistral-Large-3 model by adding a dedicated section for AMD GPU support. It provides a complete, step-by-step guide for users to deploy and benchmark the model on AMD MI300X and MI355 GPUs, covering everything from Docker setup to vLLM server configuration and performance testing. This addition aims to make the model more accessible and performant for users with AMD hardware. Highlights
🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console. Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
There was a problem hiding this comment.
Code Review
This pull request adds documentation for running the Mistral-Large-3 model on AMD GPUs. The instructions are clear, but I've identified a significant security concern in the provided docker run command, which mounts the entire host filesystem into the container. I've also suggested a couple of minor text changes to improve clarity and consistency regarding the model name and Docker image versioning. Addressing the security issue is highly recommended.
Mistral/Mistral-Large-3.md
Outdated
| --group-add video \ | ||
| --cap-add=SYS_PTRACE \ | ||
| --security-opt seccomp=unconfined \ | ||
| -v /:/work \ |
There was a problem hiding this comment.
Mounting the host's root directory (/) into the container with -v /:/work is a significant security risk. It grants the container unrestricted access to the entire host filesystem, which could lead to a full system compromise if the container is breached. It is strongly recommended to only mount specific directories that are necessary. For example, you could mount the current directory with -v $(pwd):/work or a specific data directory with -v /path/to/your/data:/work.
| -v /:/work \ | |
| -v $(pwd):/work \ |
Mistral/Mistral-Large-3.md
Outdated
|
|
||
| ## AMD GPU Support | ||
|
|
||
| Please follow the steps here to install and run Mistral-Large-3-675B-Instruct-2512 models on AMD MI300X and MI355 GPU. |
There was a problem hiding this comment.
For clarity and consistency with the commands that follow, it's better to use the full model name including the Hugging Face organization, mistralai/Mistral-Large-3-675B-Instruct-2512.
| Please follow the steps here to install and run Mistral-Large-3-675B-Instruct-2512 models on AMD MI300X and MI355 GPU. | |
| Please follow the steps here to install and run mistralai/Mistral-Large-3-675B-Instruct-2512 models on AMD MI300X and MI355 GPU. |
Mistral/Mistral-Large-3.md
Outdated
|
|
||
| Please follow the steps here to install and run Mistral-Large-3-675B-Instruct-2512 models on AMD MI300X and MI355 GPU. | ||
| ### Step 1: Prepare Docker Environment | ||
| Pull the latest vllm docker: |
There was a problem hiding this comment.
The instruction says to pull the 'latest' vLLM docker image, but the command that follows specifies a fixed version (v0.14.1). This is contradictory and could be confusing. It would be clearer to state that a specific version is required for these instructions.
| Pull the latest vllm docker: | |
| Pull the required vLLM docker image for ROCm: |
…model Signed-off-by: jiacao-amd <jiahui.cao@amd.com> add uv launch support Signed-off-by: jiacao-amd <jiahui.cao@amd.com>
8ddecc5 to
451114d
Compare
|
@jiacao-amd Please revise all of your command to up-to-date with the latest vLLM version. There is environment variables that have been deprecated many versions ago. For this documentation, we don't need to attach the benchmark script. Please merge the ROCm installation step and server command with the main context. Open subheaders to split between two hardware similar to the suggestion mentioned here https://github.com/vllm-project/recipes/pull/219/changes#r2804656457 |
Signed-off-by: jiacao-amd <jiahui.cao@amd.com>
22f3a01 to
76d0959
Compare
Done! |
…model