uv tool run opik harbor trials start -p ./examples/tasks/k8s-hello-world/ \
--agent "terminus-2" \
--agent-kwarg "api_base=$OPENAI_API_BASE" \
--model "hosted_vllm/$MODEL" \
--environment-type "docker"
OPIK: Opik tracking enabled for Harbor
Starting trial: k8s-hello-world__wv9bPao
Task: k8s-hello-world
Agent: AgentName.TERMINUS_2
Environment: docker
Trials directory: trials
OPIK: Started logging traces to the "Default Project" project at http://localhost:5173/api/v1/session/redirect/projects/?trace_id=019b0fca-cb86-78be-b98d-55df45ce7566&path=aHR0cDovL2xvY2FsaG9zdDo1MTczL2FwaS8=.
OPIK: Experiment service setup for 'harbor-20251211-154147'
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
SUMMARIZATION: Short summary failed: Model hosted_vllm/ hit max_tokens limit. Response was truncated. Consider increasing max_tokens if possible.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
SUMMARIZATION: Short summary failed: Model hosted_vllm/ hit max_tokens limit. Response was truncated. Consider increasing max_tokens if possible.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Even fallback chat failed:
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
OPIK: Using dataset 'harbor-default' for Harbor source
OPIK: Created experiment 'harbor-20251211-154147' for dataset 'harbor-default'
Trial completed!
Trial name: k8s-hello-world__wv9bPao
Task: k8s-hello-world
Started: 2025-12-11 15:41:47.284271
Finished: 2025-12-11 15:58:25.096918
Error: AgentTimeoutError
Message: Agent execution timed out after 900.0 seconds
Created
December 12, 2025 00:05
-
-
Save Mistobaan/38bd80a5b2b1212be312dd7bc3cd001a to your computer and use it in GitHub Desktop.
vllm custom server endpoint 400 error
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] Error in preprocessing prompt inputs
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] Traceback (most recent call last):
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_chat.py", line 235, in create_chat_completion
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] ) = await self._preprocess_chat(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 1133, in _preprocess_chat
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] prompt_inputs = await self._tokenize_prompt_input_async(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 990, in _tokenize_prompt_input_async
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] async for result in self._tokenize_prompt_inputs_async(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 1011, in _tokenize_prompt_inputs_async
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] yield await self._normalize_prompt_text_to_input(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 881, in _normalize_prompt_text_to_input
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] return self._validate_input(request, input_ids, input_text)
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 962, in _validate_input
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] raise ValueError(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] ValueError: This model's maximum context length is 12288 tokens. However, your request has 12704 input tokens. Please reduce the length of the input messages.
(APIServer pid=4) INFO: 172.20.6.231:57174 - "POST /v1/chat/completions HTTP/1.1" 400 Bad Request
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment