terminal
howtonotcode.com
Vertex AI logo

Vertex AI

Platform

Vertex AI simplifies the machine learning workflow on Google Cloud.

article 4 storys calendar_today First seen: 2025-12-30 update Last seen: 2026-03-03 open_in_new Website menu_book Wikipedia

Resources

Links to check for updates: homepage, feed, or git repo.

home Homepage

Stories

Showing 1-4 of 4

Google’s Gemini 3.1 Flash-Lite targets high-volume, low-latency workloads

Google released Gemini 3.1 Flash-Lite, a faster, cheaper model aimed at high-volume developer workloads and signaling a broader shift to lighter LLMs for routine backend and data tasks. Google’s launch of [Gemini 3.1 Flash-Lite](https://thenewstack.io/google-gemini-3-1-flash-lite/) emphasizes low-latency responses for tasks where cost is critical, with preview access via the Gemini API in Google AI Studio and enterprise access in Vertex AI, alongside industry moves like OpenAI’s GPT-5.3 Instant toward lighter models ([context and availability](https://www.thedeepview.com/articles/openai-google-target-lighter-models)). Independent coverage pegs Flash-Lite at $0.25/million input tokens and $1.5/million output tokens—about one-eighth the price of Gemini 3.1 Pro—and notes support for four “thinking” levels to trade speed for reasoning when needed ([pricing and modes](https://simonwillison.net/2026/Mar/3/gemini-31-flash-lite/#atom-everything)). For backend/data teams, this sweet spot makes Flash-Lite a strong default for translation, content moderation, summarization, and structured generation (dashboards/simulations), reserving heavier models for only the hardest requests ([use cases](https://www.thedeepview.com/articles/openai-google-target-lighter-models)). If your pipelines push files, mind Gemini’s surface-specific limits across Apps (including NotebookLM notebooks), API, and enterprise tools—think up to 10 files per prompt, 100MB per file/ZIP with caveats, strict video caps, and code folder/GitHub repo constraints—so ingestion doesn’t silently truncate or fail ([file-handling constraints](https://www.datastudios.org/post/gemini-file-upload-support-explained-supported-formats-size-constraints-and-document-handling-acr)). Zooming out, the race to lighter models (OpenAI’s GPT-5.3 Instant and Alibaba’s Qwen Small Model Series) underscores a clear pattern: push routine throughput to cheaper, faster tiers and escalate to heavyweight reasoning only on ambiguity or failure ([trend snapshot](https://www.thedeepview.com/articles/openai-google-target-lighter-models)).

calendar_today 2026-03-03
google gemini-31-flash-lite gemini-api google-ai-studio vertex-ai

Claude Code 2.1.x lands practical speedups and governed multi‑agent workflows

Anthropic pushed a rapid series of Claude Code 2.1 updates (v2.1.26–v2.1.31) that cut RAM on session resume, add page‑level PDF reads, support MCP servers without dynamic registration, enable PR‑based session bootstraps, and ship many reliability fixes [Reddit summary](https://www.reddit.com/r/ClaudeAI/comments/1qvgdc5/claude_code_v21262130_what_changed/)[^1] and [official v2.1.31 notes](https://github.com/anthropics/claude-code/releases/tag/v2.1.31)[^2]. Practitioners also highlight 2.1’s skill hot‑reload, lifecycle hooks, and forked sub‑agents as a foundation for governed, observable multi‑agent workflows—positioning Claude Code as a lightweight "agent OS" for real projects [deep dive](https://medium.com/@richardhightower/build-agent-skills-faster-with-claude-code-2-1-release-6d821d5b8179)[^3]. [^1]: Adds: community changelog for v2.1.26–30 covering performance, MCP, GitHub/PR workflows, and PDF handling. [^2]: Adds: official v2.1.31 fixes (PDF lockups, sandbox FS errors, streaming temperature override, tool routing prompts, provider labels) and hard limits (100 pages, 20MB). [^3]: Adds: perspective on skill hot‑reload, lifecycle hooks, and forked sub‑agents enabling governed multi‑agent patterns.

calendar_today 2026-02-04
claude-code anthropic mcp-model-context-protocol github slack