INFO 2025-10-05T17:18:09 +312ms service=default version=0.14.3 args=["run","Olá, você está funcionando com o LM Studio? Responda em português."] opencode INFO 2025-10-05T17:18:09 +18ms service=project directory=/Users/neog fromDirectory INFO 2025-10-05T17:18:09 +29ms service=config path=/Users/neog/.config/opencode/config.json loading INFO 2025-10-05T17:18:09 +1ms service=config path=/Users/neog/.config/opencode/opencode.json loading INFO 2025-10-05T17:18:09 +0ms service=config path=/Users/neog/.config/opencode/opencode.jsonc loading INFO 2025-10-05T17:18:09 +8ms service=config path=/Users/neog/opencode.json loading INFO 2025-10-05T17:18:09 +13ms service=plugin path=opencode-copilot-auth@0.0.3 loading plugin INFO 2025-10-05T17:18:09 +19ms service=plugin path=opencode-anthropic-auth@0.0.2 loading plugin INFO 2025-10-05T17:18:09 +45ms service=bus type=* subscribing INFO 2025-10-05T17:18:09 +0ms service=bus type=session.updated subscribing INFO 2025-10-05T17:18:09 +0ms service=bus type=message.updated subscribing INFO 2025-10-05T17:18:09 +0ms service=bus type=message.part.updated subscribing INFO 2025-10-05T17:18:09 +0ms service=format init INFO 2025-10-05T17:18:09 +0ms service=bus type=file.edited subscribing INFO 2025-10-05T17:18:09 +5ms service=session id=ses_64a9e7fe1ffepYksmgQLX1xCtL version=0.14.3 projectID=global directory=/Users/neog title=New session - 2025-10-05T17:18:09.952Z time={"created":1759684689952,"updated":1759684689952} created INFO 2025-10-05T17:18:09 +0ms service=lsp serverIds=typescript, vue, eslint, gopls, ruby-lsp, pyright, elixir-ls, zls, csharp, rust, clangd, svelte, jdtls enabled LSP servers INFO 2025-10-05T17:18:09 +6ms service=bus type=session.updated publishing INFO 2025-10-05T17:18:09 +2ms service=bus type=message.part.updated subscribing INFO 2025-10-05T17:18:09 +0ms service=bus type=session.error subscribing INFO 2025-10-05T17:18:09 +1ms service=session.prompt session=ses_64a9e7fe1ffepYksmgQLX1xCtL prompt INFO 2025-10-05T17:18:09 +5ms service=bus type=message.updated publishing INFO 2025-10-05T17:18:09 +1ms service=bus type=message.part.updated publishing INFO 2025-10-05T17:18:09 +1ms service=bus type=session.updated publishing INFO 2025-10-05T17:18:09 +2ms service=models.dev file={} refreshing INFO 2025-10-05T17:18:09 +5ms service=provider init INFO 2025-10-05T17:18:09 +2ms service=provider providerID=opencode found INFO 2025-10-05T17:18:09 +0ms service=provider providerID=lmstudio found INFO 2025-10-05T17:18:09 +0ms service=provider providerID=lmstudio modelID=qwen/qwen3-next-80b getModel INFO 2025-10-05T17:18:09 +1ms service=provider status=started providerID=lmstudio getSDK INFO 2025-10-05T17:18:10 +123ms service=provider status=completed duration=124 providerID=lmstudio getSDK INFO 2025-10-05T17:18:10 +2ms service=provider providerID=lmstudio modelID=qwen/qwen3-next-80b found INFO 2025-10-05T17:18:10 +1ms service=session.prompt session=ses_64a9e7fe1ffepYksmgQLX1xCtL sessionID=ses_64a9e7fe1ffepYksmgQLX1xCtL locking INFO 2025-10-05T17:18:10 +809ms service=bus type=message.updated publishing INFO 2025-10-05T17:18:10 +10ms service=session.prompt session=ses_64a9e7fe1ffepYksmgQLX1xCtL process INFO 2025-10-05T17:18:10 +5ms service=session.prompt session=ses_64a9e7fe1ffepYksmgQLX1xCtL type=start part ERROR 2025-10-05T17:18:11 +777ms service=session.prompt session=ses_64a9e7fe1ffepYksmgQLX1xCtL error={"error":"The number of tokens to keep from the initial prompt is greater than the context length. Try to load the model with a larger context length, or provide a shorter input"} stream error INFO 2025-10-05T17:18:11 +1ms service=session.prompt session=ses_64a9e7fe1ffepYksmgQLX1xCtL type=start-step part INFO 2025-10-05T17:18:11 +5ms service=bus type=session.updated publishing INFO 2025-10-05T17:18:11 +1ms service=bus type=message.part.updated publishing INFO 2025-10-05T17:18:11 +0ms service=session.prompt session=ses_64a9e7fe1ffepYksmgQLX1xCtL type=error part ERROR 2025-10-05T17:18:11 +1ms service=session.prompt session=ses_64a9e7fe1ffepYksmgQLX1xCtL error=The number of tokens to keep from the initial prompt is greater than the context length. Try to load the model with a larger context length, or provide a shorter input process INFO 2025-10-05T17:18:11 +3ms service=bus type=message.updated publishing INFO 2025-10-05T17:18:11 +0ms service=session.prompt session=ses_64a9e7fe1ffepYksmgQLX1xCtL sessionID=ses_64a9e7fe1ffepYksmgQLX1xCtL unlocking ERROR 2025-10-05T17:18:11 +8ms service=default name=TypeError message="The number of tokens to keep from the initial prompt is greater than the context length. Try to load the model with a larger context length, or provide a shorter input" is not an Object. (evaluating '"name" in input') stack=TypeError: "The number of tokens to keep from the initial prompt is greater than the context length. Try to load the model with a larger context length, or provide a shorter input" is not an Object. (evaluating '"name" in input') at isInstance (/$bunfs/root/index.js:61386:26) at process (/$bunfs/root/index.js:103626:56) at processTicksAndRejections (native:7:39) fatal INFO 2025-10-05T17:18:11 +2ms service=bus type=session.idle publishing