Skip to content

Instantly share code, notes, and snippets.

@Mistobaan
Created December 12, 2025 00:05
Show Gist options
  • Select an option

  • Save Mistobaan/38bd80a5b2b1212be312dd7bc3cd001a to your computer and use it in GitHub Desktop.

Select an option

Save Mistobaan/38bd80a5b2b1212be312dd7bc3cd001a to your computer and use it in GitHub Desktop.
uv tool run opik harbor trials start -p ./examples/tasks/k8s-hello-world/ \
        --agent "terminus-2" \
        --agent-kwarg "api_base=$OPENAI_API_BASE" \
  --model "hosted_vllm/$MODEL" \
  --environment-type "docker"

OPIK: Opik tracking enabled for Harbor
Starting trial: k8s-hello-world__wv9bPao
Task: k8s-hello-world
Agent: AgentName.TERMINUS_2
Environment: docker
Trials directory: trials
OPIK: Started logging traces to the "Default Project" project at http://localhost:5173/api/v1/session/redirect/projects/?trace_id=019b0fca-cb86-78be-b98d-55df45ce7566&path=aHR0cDovL2xvY2FsaG9zdDo1MTczL2FwaS8=.
OPIK: Experiment service setup for 'harbor-20251211-154147'
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

SUMMARIZATION: Short summary failed: Model hosted_vllm/ hit max_tokens limit. Response was truncated. Consider increasing max_tokens if possible.

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

SUMMARIZATION: Short summary failed: Model hosted_vllm/ hit max_tokens limit. Response was truncated. Consider increasing max_tokens if possible.

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.


Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Even fallback chat failed: 
Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

Failed to retrieve model info for 'hosted_vllm/': This model isn't mapped yet. model=hosted_vllm/, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.. Using fallback context limit: 1000000

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.

OPIK: Using dataset 'harbor-default' for Harbor source
OPIK: Created experiment 'harbor-20251211-154147' for dataset 'harbor-default'

Trial completed!
Trial name: k8s-hello-world__wv9bPao
Task: k8s-hello-world
Started: 2025-12-11 15:41:47.284271
Finished: 2025-12-11 15:58:25.096918
Error: AgentTimeoutError
Message: Agent execution timed out after 900.0 seconds

vllm custom server endpoint 400 error

(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] Error in preprocessing prompt inputs
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] Traceback (most recent call last):
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]   File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_chat.py", line 235, in create_chat_completion
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]     ) = await self._preprocess_chat(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]   File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 1133, in _preprocess_chat
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]     prompt_inputs = await self._tokenize_prompt_input_async(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]   File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 990, in _tokenize_prompt_input_async
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]     async for result in self._tokenize_prompt_inputs_async(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]   File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 1011, in _tokenize_prompt_inputs_async
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]     yield await self._normalize_prompt_text_to_input(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]   File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 881, in _normalize_prompt_text_to_input
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]     return self._validate_input(request, input_ids, input_text)
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]   File "/usr/local/lib/python3.12/site-packages/vllm/entrypoints/openai/serving_engine.py", line 962, in _validate_input
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257]     raise ValueError(
(APIServer pid=4) ERROR 12-11 23:58:03 [serving_chat.py:257] ValueError: This model's maximum context length is 12288 tokens. However, your request has 12704 input tokens. Please reduce the length of the input messages.
(APIServer pid=4) INFO:     172.20.6.231:57174 - "POST /v1/chat/completions HTTP/1.1" 400 Bad Request
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment