You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
What problem or use case are you trying to solve?
While running, I would like to use groq/llama3 for my agent, but I run into issues with rate limiting ( It limits me to 6000 tpm, but opendevin will make requests larger then that ), I would love it to do the normal retry 3 or 5 times, but then on failer attempt a secondary model, and even better, have unlimited fallback models, so we could do stuff like groq/llama3 => openai/gpt4o => google/gemini, so we can get the benifit of the cheaper models, but the power of being able to have more context
Describe the UX of the solution you'd like
in the settings dialog add an "Add an fallback model" and put in the model and credentials
Do you have thoughts on the technical implementation?
Describe alternatives you've considered
The alternative I am considering, but will be a bit hacky is creating an external service, that proxy's the llm requests, and does the fall through its self, It will work as a work around, but having it in the app could allow for more advanced features in the future, like prioritizing different models for different tasks
Additional context
The text was updated successfully, but these errors were encountered:
What problem or use case are you trying to solve?
While running, I would like to use groq/llama3 for my agent, but I run into issues with rate limiting ( It limits me to 6000 tpm, but opendevin will make requests larger then that ), I would love it to do the normal retry 3 or 5 times, but then on failer attempt a secondary model, and even better, have unlimited fallback models, so we could do stuff like groq/llama3 => openai/gpt4o => google/gemini, so we can get the benifit of the cheaper models, but the power of being able to have more context
Describe the UX of the solution you'd like
in the settings dialog add an "Add an fallback model" and put in the model and credentials
Do you have thoughts on the technical implementation?
Describe alternatives you've considered
The alternative I am considering, but will be a bit hacky is creating an external service, that proxy's the llm requests, and does the fall through its self, It will work as a work around, but having it in the app could allow for more advanced features in the future, like prioritizing different models for different tasks
Additional context
The text was updated successfully, but these errors were encountered: