terminal
howtonotcode.com
DeepMind logo

DeepMind

Company

DeepMind is a leading artificial intelligence research company known for its work in developing advanced AI technologies. It is primarily for researchers, developers, and organizations interested in cutting-edge AI solutions. A key use case is its development of AI systems that can solve complex problems, such as AlphaGo, which defeated a world champion Go player.

article 3 storys calendar_today First seen: 2025-12-30 update Last seen: 2026-02-17 open_in_new Website menu_book Wikipedia

Resources

Links to check for updates: homepage, feed, or git repo.

home Homepage

Stories

Showing 1-3 of 3

DeepMind’s delegation framework meets practical Agent Skills for safer, cheaper coding agents

DeepMind outlined a principled framework for safely delegating work across AI agents while developers show that SKILL.md-based agent skills and tooling make coding agents more efficient and dependable. Google DeepMind’s [Intelligent AI Delegation](https://arxiv.org/abs/2602.11865) proposes an adaptive task-allocation framework—covering role boundaries, transfer of authority, accountability, and trust—for delegating work across AI agents and humans, with explicit mechanisms for recovery from failures. On the ground, a hands-on walkthrough of Agent Skills shows how a SKILL.md plus progressive disclosure architecture can reduce context bloat and improve code consistency in tools like Claude Code, with clear patterns for discovery, on-demand instruction loading, and resource access ([guide](https://levelup.gitconnected.com/why-do-my-ai-agents-perform-better-than-yours-eb6a93369366)). For observability and reproducibility, Simon Willison adds [Chartroom and datasette-showboat](https://simonwillison.net/2026/Feb/17/chartroom-and-datasette-showboat/#atom-everything), a CLI-driven approach for agents to emit runnable Markdown artifacts that demonstrate code and data outputs—useful for audits, PR reviews, and postmortems.

calendar_today 2026-02-17
deepmind anthropic claude-code showboat agent-skills

Gemini Deep Think: research gains, CLI workflows, and model-extraction risks

Google’s Gemini Deep Think is graduating from contests to real research and developer workflows, but its growing capability is also attracting copycat extraction and criminal abuse that teams must plan around. Google DeepMind details how Gemini Deep Think, guided by experts, is tackling professional math and science problems using an agent (Aletheia) that iteratively generates, verifies, revises, and even browses to avoid spurious citations, with results improving as inference-time compute scales and outperforming prior Olympiad-level benchmarks ([Google DeepMind](https://deepmind.google/blog/accelerating-mathematical-and-scientific-discovery-with-gemini-deep-think/?_bhlid=c06248275cf06add0c919aabac361f98ed7c1e95)). A broader industry pulse notes the release’s framing and early user anecdotes around “Gemini 3 Deep Think” appearing in the wild ([Simon Willison’s Weblog](https://simonwillison.net/2026/Feb/12/gemini-3-deep-think/#atom-everything)). For context on user expectations, this differs from Google Search’s ranking-first paradigm—Gemini aims for single-response reasoning rather than surfacing diverse sources ([DataStudios](https://www.datastudios.org/post/why-does-gemini-give-different-answers-than-google-search-reasoning-versus-ranking-logic)). For day-to-day engineering, a terminal-native Gemini CLI is emerging to integrate AI directly into developer workflows—writing files, chaining commands, and automating tasks without browser context switching, which can accelerate prototyping, code generation, and research summarization in-place ([Gemini CLI guide](https://atalupadhyay.wordpress.com/2026/02/12/gemini-cli-from-first-steps-to-advanced-workflows/)). Security posture must catch up: Google reports adversaries tried to clone Gemini via high-volume prompting (>100,000 prompts in one session) to distill its behavior, and separate threat intel highlights rising criminal use of Gemini for phishing, malware assistance, and reconnaissance—underscoring the need for rate limits, monitoring, and policy controls around model access and outputs ([Ars Technica](https://arstechnica.com/ai/2026/02/attackers-prompted-gemini-over-100000-times-while-trying-to-clone-it-google-says/), [WebProNews](https://www.webpronews.com/from-experimentation-to-exploitation-how-cybercriminals-are-weaponizing-googles-own-ai-tools-against-the-digital-world/)).

calendar_today 2026-02-12
google-deepmind google gemini-deep-think gemini-cli google-search

Update: Shift from Bigger LLMs to Tool-Using Agents

New coverage moves from high-level trend to concrete examples: agentic systems with persistent memory, tool-grounded actions, and human-in-the-loop controls. The video highlights vendor moves (e.g., Anthropic’s Claude/Claude Code updates and DeepMind’s agent-first roadmap) as evidence that reliability/cost gains now come from tools, memory, and planning rather than scaling base models.

calendar_today 2025-12-30
agents tool-use memory enterprise-ai rag