Skip to content

Make model argument mandatory #1574

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Jun 21, 2025
Merged

Make model argument mandatory #1574

merged 1 commit into from
Jun 21, 2025

Conversation

ericcurtin
Copy link
Member

@ericcurtin ericcurtin commented Jun 21, 2025

To be consistent with "ramalama run" experience. Inferencing
servers that have implemented model-swapping require this. In the
case of servers like llama-server that only load one server, any
value is sufficient.

Copy link
Contributor

sourcery-ai bot commented Jun 21, 2025

Reviewer's Guide

This PR enforces a mandatory MODEL parameter by adding it as a required positional argument in the CLI, removes internal model inference methods, and updates the chat request to use args.MODEL directly.

Class diagram for updated Chat class structure

classDiagram
    class Chat {
        - args
        - url
        - models_url
        - conversation_history
        + __init__(args)
        + handle_args()
        + _make_request_data()
    }
    %% Removed methods
    %% - model(index=0)
    %% - get_models()
Loading

File-Level Changes

Change Details Files
Enforce MODEL as a required CLI argument
  • Add positional “MODEL” argument with completer
  • Retain existing ARGS as optional parameters
ramalama/cli.py
Remove internal model caching and retrieval
  • Drop self.models initialization in constructor
  • Remove model() method
  • Remove get_models() helper
ramalama/chat.py
Use provided MODEL instead of dynamic lookup
  • Replace calls to self.model() with self.args.MODEL
ramalama/chat.py

Tips and commands

Interacting with Sourcery

  • Trigger a new review: Comment @sourcery-ai review on the pull request.
  • Continue discussions: Reply directly to Sourcery's review comments.
  • Generate a GitHub issue from a review comment: Ask Sourcery to create an
    issue from a review comment by replying to it. You can also reply to a
    review comment with @sourcery-ai issue to create an issue from it.
  • Generate a pull request title: Write @sourcery-ai anywhere in the pull
    request title to generate a title at any time. You can also comment
    @sourcery-ai title on the pull request to (re-)generate the title at any time.
  • Generate a pull request summary: Write @sourcery-ai summary anywhere in
    the pull request body to generate a PR summary at any time exactly where you
    want it. You can also comment @sourcery-ai summary on the pull request to
    (re-)generate the summary at any time.
  • Generate reviewer's guide: Comment @sourcery-ai guide on the pull
    request to (re-)generate the reviewer's guide at any time.
  • Resolve all Sourcery comments: Comment @sourcery-ai resolve on the
    pull request to resolve all Sourcery comments. Useful if you've already
    addressed all the comments and don't want to see them anymore.
  • Dismiss all Sourcery reviews: Comment @sourcery-ai dismiss on the pull
    request to dismiss all existing Sourcery reviews. Especially useful if you
    want to start fresh with a new review - don't forget to comment
    @sourcery-ai review to trigger a new review!

Customizing Your Experience

Access your dashboard to:

  • Enable or disable review features such as the Sourcery-generated pull request
    summary, the reviewer's guide, and others.
  • Change the review language.
  • Add, remove or edit custom review instructions.
  • Adjust other review settings.

Getting Help

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Summary of Changes

Hello @ericcurtin, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request modifies the ramalama chat command to enforce the explicit provision of a model argument. This change standardizes the command-line interface for consistency with other ramalama commands and simplifies the internal chat client by removing its responsibility for model discovery, which is beneficial for environments with dynamic model management.

Highlights

  • Mandatory Model Argument: The ramalama chat command now requires a MODEL argument to be explicitly provided by the user, making it a mandatory positional argument. This aligns the command's behavior with ramalama run.
  • Simplified Model Handling: The internal ramalama/chat.py client code has been streamlined by removing the logic responsible for discovering and managing available models. The chat client now directly uses the model provided via command-line arguments.
  • Support for Inferencing Servers: This change provides necessary support for inferencing servers that implement model-swapping, as they often require the model to be specified upfront rather than being inferred or auto-discovered.
Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point in your pull request via creating an issue comment (i.e. comment on the pull request page) using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in issue comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist is currently in preview and may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments to provide feedback.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

Copy link
Contributor

@sourcery-ai sourcery-ai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hey @ericcurtin - I've reviewed your changes and they look great!


Sourcery is free for open source - if you like our reviews please consider sharing them ✨
Help me be more useful! Please click 👍 or 👎 on each comment and I'll use the feedback to improve your reviews.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request makes the MODEL argument for the ramalama chat command mandatory, which aligns it with the ramalama run command. The changes are straightforward and correctly implement the intended behavior by updating the argument parser and removing the now-obsolete dynamic model fetching logic. My feedback includes a suggestion to remove a now-unused class attribute and to add system tests to verify this new mandatory argument behavior and prevent future regressions.

@@ -921,6 +921,7 @@ def chat_parser(subparsers):
)
parser.add_argument("--prefix", type=str, help="prefix for the user prompt", default=default_prefix())
parser.add_argument("--url", type=str, default="http://127.0.0.1:8080", help="the host to send requests to")
parser.add_argument("MODEL", completer=local_models) # positional argument
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

This change correctly makes the MODEL argument mandatory for the chat command. To ensure this behavior is maintained and to prevent future regressions, it would be beneficial to add a system test.

You could add a new test case that verifies:

  1. Running ramalama chat without a model name fails with an appropriate error message.
  2. Running ramalama chat my-model proceeds past argument parsing (it might fail later due to connection errors, which is fine for this test).

This would improve the test coverage for the chat command.

To be consistent with "ramalama run" experience. Inferencing
servers that have implemented model-swapping require this. In the
case of servers like llama-server that only load one server, any
value is sufficient.

Signed-off-by: Eric Curtin <[email protected]>
@rhatdan
Copy link
Member

rhatdan commented Jun 21, 2025

LGTM

@rhatdan rhatdan merged commit fe095f1 into main Jun 21, 2025
18 of 21 checks passed
@ericcurtin ericcurtin deleted the specify-model branch June 22, 2025 12:53
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants