-
-
Notifications
You must be signed in to change notification settings - Fork 19
Description
Feature and its Use Cases
Description
Currently, DocPilot uses a hardcoded model for processing medical documentation (summarization and prescription generation). While functional, this limits users from leveraging the specialized capabilities of the broader Google AI ecosystem.
With the release of the Gemini 3 series and the high-performance Gemma open-weight models, the application would benefit significantly from allowing users to choose a model that fits their specific latency and reasoning requirements.
Problem
The current implementation: Does not allow switching between Gemini 3 Flash (speed) and Gemini 3 Pro (complex reasoning).
Lacks support for Gemma models, which are highly efficient for straightforward summarization tasks.
Forces a "one size fits all" approach, which may not be ideal for varied clinical workflows.
Proposed Solution
Add a configuration option in the UI (e.g., in a "Model Settings" or "AI Configuration" section) that allows the user to select their preferred model.
Specifically:
-
UI Component: A dropdown or selection menu to choose between:
* Gemma 3 (Open Weights) (Default) * Gemini 3 Pro * Gemini 2.5/3 Flash -
Dynamic API Calls: Update the AI service layer to pass the selected model_id to the Gemini API client during inference.
Benefits
Flexibility: Users can prioritize speed (Flash/Gemma) or accuracy (Pro) depending on the complexity of the medical case.
Cost/Resource Optimization: Gemma models provide a generous and efficient alternative for standard tasks.
Additional Context (After updates)
- Response by Gemma-3-27B
Code of Conduct
- I have joined the Discord server and will post updates there
- I have searched existing issues to avoid duplicates

