terminal
howtonotcode.com
Google DeepMind logo

Google DeepMind

Company

Google DeepMind is a leading artificial intelligence research lab focused on developing advanced AI technologies. It is designed for researchers, developers, and organizations interested in cutting-edge AI solutions. A key use case is its application in healthcare, gaming, and energy efficiency.

article 4 storys calendar_today First seen: 2025-12-30 update Last seen: 2026-02-17 open_in_new Website menu_book Wikipedia

Resources

Links to check for updates: homepage, feed, or git repo.

home Homepage

code Git repo

Stories

Showing 1-4 of 4

DeepMind’s delegation framework meets practical Agent Skills for safer, cheaper coding agents

DeepMind outlined a principled framework for safely delegating work across AI agents while developers show that SKILL.md-based agent skills and tooling make coding agents more efficient and dependable. Google DeepMind’s [Intelligent AI Delegation](https://arxiv.org/abs/2602.11865) proposes an adaptive task-allocation framework—covering role boundaries, transfer of authority, accountability, and trust—for delegating work across AI agents and humans, with explicit mechanisms for recovery from failures. On the ground, a hands-on walkthrough of Agent Skills shows how a SKILL.md plus progressive disclosure architecture can reduce context bloat and improve code consistency in tools like Claude Code, with clear patterns for discovery, on-demand instruction loading, and resource access ([guide](https://levelup.gitconnected.com/why-do-my-ai-agents-perform-better-than-yours-eb6a93369366)). For observability and reproducibility, Simon Willison adds [Chartroom and datasette-showboat](https://simonwillison.net/2026/Feb/17/chartroom-and-datasette-showboat/#atom-everything), a CLI-driven approach for agents to emit runnable Markdown artifacts that demonstrate code and data outputs—useful for audits, PR reviews, and postmortems.

calendar_today 2026-02-17
deepmind anthropic claude-code showboat agent-skills

Gemini Deep Think: research gains, CLI workflows, and model-extraction risks

Google’s Gemini Deep Think is graduating from contests to real research and developer workflows, but its growing capability is also attracting copycat extraction and criminal abuse that teams must plan around. Google DeepMind details how Gemini Deep Think, guided by experts, is tackling professional math and science problems using an agent (Aletheia) that iteratively generates, verifies, revises, and even browses to avoid spurious citations, with results improving as inference-time compute scales and outperforming prior Olympiad-level benchmarks ([Google DeepMind](https://deepmind.google/blog/accelerating-mathematical-and-scientific-discovery-with-gemini-deep-think/?_bhlid=c06248275cf06add0c919aabac361f98ed7c1e95)). A broader industry pulse notes the release’s framing and early user anecdotes around “Gemini 3 Deep Think” appearing in the wild ([Simon Willison’s Weblog](https://simonwillison.net/2026/Feb/12/gemini-3-deep-think/#atom-everything)). For context on user expectations, this differs from Google Search’s ranking-first paradigm—Gemini aims for single-response reasoning rather than surfacing diverse sources ([DataStudios](https://www.datastudios.org/post/why-does-gemini-give-different-answers-than-google-search-reasoning-versus-ranking-logic)). For day-to-day engineering, a terminal-native Gemini CLI is emerging to integrate AI directly into developer workflows—writing files, chaining commands, and automating tasks without browser context switching, which can accelerate prototyping, code generation, and research summarization in-place ([Gemini CLI guide](https://atalupadhyay.wordpress.com/2026/02/12/gemini-cli-from-first-steps-to-advanced-workflows/)). Security posture must catch up: Google reports adversaries tried to clone Gemini via high-volume prompting (>100,000 prompts in one session) to distill its behavior, and separate threat intel highlights rising criminal use of Gemini for phishing, malware assistance, and reconnaissance—underscoring the need for rate limits, monitoring, and policy controls around model access and outputs ([Ars Technica](https://arstechnica.com/ai/2026/02/attackers-prompted-gemini-over-100000-times-while-trying-to-clone-it-google-says/), [WebProNews](https://www.webpronews.com/from-experimentation-to-exploitation-how-cybercriminals-are-weaponizing-googles-own-ai-tools-against-the-digital-world/)).

calendar_today 2026-02-12
google-deepmind google gemini-deep-think gemini-cli google-search

UK/NY AI rules meet adversarial safety: what backend/data teams must change

AI governance is shifting from voluntary guidelines to binding obligations while labs formalize adversarial and constitutional safety methods, raising new requirements for evaluation, logging, and incident reporting. The UK is proposing mandatory registration, pre‑release safety testing, and incident reporting for frontier models enforced via the AI Safety Institute, moving beyond voluntary pledges [Inside the Scramble to Tame AI: Why the UK’s New Regulatory Push Could Reshape the Global Tech Order](https://www.webpronews.com/inside-the-scramble-to-tame-ai-why-the-uks-new-regulatory-push-could-reshape-the-global-tech-order/)[^1]. New York is advancing transparency and impact‑assessment bills for high‑risk AI decisions [Albany’s AI Reckoning: Inside New York’s Ambitious Bid to Become America’s Toughest Regulator of Artificial Intelligence](https://www.webpronews.com/albanys-ai-reckoning-inside-new-yorks-ambitious-bid-to-become-americas-toughest-regulator-of-artificial-intelligence/)[^2], while labs push adversarial reasoning and constitutional alignment to harden model behavior [Inside Adversarial Reasoning: How AI Labs Are Teaching Models to Think by Fighting Themselves](https://www.webpronews.com/inside-adversarial-reasoning-how-ai-labs-are-teaching-models-to-think-by-fighting-themselves/)[^3] [Thoughts on Claude's Constitution](https://windowsontheory.org/2026/01/27/thoughts-on-claudes-constitution/ct assessments, and penalties. [^3]: Explains adversarial debate/self‑play and automated red‑teaming as next‑gen training/eval methods. [^4]: An OpenAI researcher’s critique of Anthropic’s Claude Constitution and implications for alignment practice.

calendar_today 2026-02-09
openai anthropic google-deepmind meta uk-ai-safety-institute

Update: Google DeepMind AGI roadmap and agentic systems

In a new video, Demis Hassabis lays out the clearest public roadmap to AGI yet, explicitly centering on agentic systems that plan, use tools, and work across modalities. New vs prior: he more clearly sequences milestones (improving tool-use reliability and long‑horizon planning before higher autonomy) and positions Gemini and Project Astra as stepping stones rather than endpoints.

calendar_today 2025-12-30
agi agents multimodal google deepmind