Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: When using DeepSeek in the [Ai Profive Configuration] option, the default configuration is incorrect: 'others/deepseek-chat' should be 'deepseek/deepseek-chat' .(Including Fix) #5978

Open
1 task done
Simon-CHOU opened this issue Jan 2, 2025 · 1 comment
Labels
bug Something isn't working

Comments

@Simon-CHOU
Copy link

Is there an existing issue for the same bug?

  • I have checked the existing issues.

Describe the bug and reproduction steps

Step 1: After Docker starts fresh, access http://localhost:3000/ to configure the AI Provider.

docker pull docker.all-hands.dev/all-hands-ai/runtime:0.17-nikolaik

docker run -it --rm --pull=always \
  -e SANDBOX_RUNTIME_CONTAINER_IMAGE=docker.all-hands.dev/all-hands-ai/runtime:0.17-nikolaik \
  -e LOG_ALL_EVENTS=true \
  -v /var/run/docker.sock:/var/run/docker.sock \
  -v ~/.openhands-state:/.openhands-state \
  -p 3000:3000 \
  --add-host host.docker.internal:host-gateway \
  --name openhands-app \
  docker.all-hands.dev/all-hands-ai/openhands:0.17

Step 2: Select others/deepseek-chat.
Step 3: Configure the API Key.
Step 4: Configure the GitHub token and enter a code repository.
Step 5: Send any message.

Expected: Normal response.
Actual: The chat box reports an error as follows:
completion(model='huggingface/starcoder',..) Learn more: https://docs.litellm.ai/docs/providers

Fix Method:
Go to [Ai Profive Configuration] -> [Customer Mode], change others/deepseek-chat to deepseek/deepseek-chat, save, and restart the session.

Referrence: https://docs.litellm.ai/docs/providers/deepseek

OpenHands Installation

Docker command in README

OpenHands Version

all-hands-ai/openhands:0.17

Operating System

WSL on Windows

Logs, Errors, Screenshots, and Additional Context

Default Configuration which is wrong
LLM Profiver Default Value for deepseek
Chat Error with default config other/deep
error-screenshot

Fix Configuration according to https://docs.litellm.ai/docs/providers/deepseek
LLM Profiver Fixed Value for deepseek

Chat works well after fixed config.
well-screenshot

@Simon-CHOU Simon-CHOU added the bug Something isn't working label Jan 2, 2025
@mamoodi
Copy link
Collaborator

mamoodi commented Jan 2, 2025

Thanks! Looks like we need to revisit those basic settings dropdowns.

SmartManoj added a commit to SmartManoj/Kevin that referenced this issue Jan 3, 2025
LiteLLM lists OpenAI models without the openai/ prefix
SmartManoj added a commit to SmartManoj/litellm that referenced this issue Jan 3, 2025
krrishdholakia pushed a commit to BerriAI/litellm that referenced this issue Jan 3, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants