-
Notifications
You must be signed in to change notification settings - Fork 235
Readme improvements: Update model's name and improve CUDA_VISIBLE_DEVICES section #1691
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Reviewer's GuideThis PR refines documentation by replacing outdated model shortnames with the current one across examples and enriches the CUDA usage guide with explicit behavior and examples when CUDA_VISIBLE_DEVICES is empty. File-Level Changes
Tips and commandsInteracting with Sourcery
Customizing Your ExperienceAccess your dashboard to:
Getting Help
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Summary of Changes
Hello @mbortoli, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!
This pull request focuses on enhancing the clarity and accuracy of the project's documentation. It addresses an outdated model name reference and provides essential guidance on configuring GPU/CPU device selection using the CUDA_VISIBLE_DEVICES
environment variable, improving the user experience for ramalama
users.
Highlights
- Documentation Accuracy: Updated references to the
granite-moe3
model togranite3-moe
across theREADME.md
and relevant man pages (ramalama-bench.1.md
,ramalama-perplexity.1.md
), reflecting a likely name change in the Ollama registry. - CUDA Device Management Clarity: Added detailed instructions and warnings in
docs/ramalama-cuda.7.md
regarding the behavior of theCUDA_VISIBLE_DEVICES
environment variable. This clarifies that setting it to an empty string (export CUDA_VISIBLE_DEVICES=""
) forces CPU usage and provides guidance on how to unset it to revert to GPU usage.
Using Gemini Code Assist
The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.
Invoking Gemini
You can request assistance from Gemini at any point in your pull request via creating an issue comment (i.e. comment on the pull request page) using either /gemini <command>
or @gemini-code-assist <command>
. Below is a summary of the supported commands.
Feature | Command | Description |
---|---|---|
Code Review | /gemini review |
Performs a code review for the current pull request in its current state. |
Pull Request Summary | /gemini summary |
Provides a summary of the current pull request in its current state. |
Comment | @gemini-code-assist | Responds in comments when explicitly tagged, both in issue comments and review comments. |
Help | /gemini help |
Displays a list of available commands. |
Customization
To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/
folder in the base of the repository. Detailed instructions can be found here.
Limitations & Feedback
Gemini Code Assist is currently in preview and may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments to provide feedback.
You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.
Footnotes
-
Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution. ↩
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request updates the model name from granite-moe3
to granite3-moe
across the documentation. It also clarifies the behavior of the CUDA_VISIBLE_DEVICES
environment variable, explaining how to force CPU usage and how to revert to using GPUs. The changes improve the documentation's accuracy and clarity.
You need to sign your commits. git rebase -i origin |
81849c5
to
3e89975
Compare
@rhatdan Sorry about that, first time ever contributing to a project. |
@mbortoli no problem and thanks. you still have not signed your commit. git commit -a --amend -s |
36c54b5
to
f628815
Compare
- Corrected the model name under the Benchmark section; previous name was not available in Ollama's registry. - Added instructions to switch between CPU-only mode and using all available GPUs via CUDA_VISIBLE_DEVICES. Signed-off-by: Mario Antonio Bortoli Filho <[email protected]>
f628815
to
b5826c9
Compare
@rhatdan done! my github email was set to private and I didn't know they needed to match |
Thanks @mbortoli |
granite-moe3
is not found on Ollama, the name probably changed in the repository.Reproducing 1:
On bash:
ramalama bench granite-moe3
Should show:
Error: Manifest for granite-moe3:latest was not found in the Ollama registry
Reproducing 2:
Will show:
ggml_cuda_init: failed to initialize CUDA: no CUDA-capable device is detected
And when executing
unset CUDA_VISIBLE_DEVICES
it returns to the standard behavior, using all GPUs.Summary by Sourcery
Correct the model shortname in documentation and clarify CUDA_VISIBLE_DEVICES behavior.
Documentation: