Using third‑party LLM APIs in VS Code (Qwen via Together/DeepInfra)
Overview
A developer is replacing a flat-fee assistant with pay‑per‑use API models in VS Code, specifically Qwen Coder 2.5 via Together or DeepInfra, for occasional code generation and PR review. The goal is minimal setup while avoiding vendor lock‑in. For teams, this means treating the editor as a client of LLM endpoints and planning for keys, context sizing, and latency trade‑offs.
All Sources
Story Timeline
Using third‑party LLM APIs in VS Code (Qwen via Together/DeepInfra)
A developer is replacing a flat-fee assistant with pay‑per‑use API models in VS Code, specifically Qwen Coder 2.5 via Together or DeepInfra, for occasional code generation and PR review. The goal is minimal setup while avoiding vendor lock‑in. For teams, this means treating the editor as a client of LLM endpoints and planning for keys, context sizing, and latency trade‑offs.