terminal
howtonotcode.com
Dataiku logo

Dataiku

Company

Dataiku is a French-American artificial intelligence (AI) and machine learning company which was founded in 2013 in Paris, France. In December 2019, Dataiku announced that CapitalG—the late-stage growth venture capital fund financed by Alphabet Inc.—joined Dataiku as an investor and that it had achieved unicorn status. As of 2021, Dataiku is valued at $4.6 billion. Dataiku currently employs more than 1,000 people worldwide between offices in New York, Denver, Washington DC, Los Angeles, Paris, L

article 1 story calendar_today First seen: 2026-02-11 update Last seen: 2026-02-11 open_in_new Website menu_book Wikipedia

Resources

Links to check for updates: homepage, feed, or git repo.

home Homepage

Stories

Showing 1-1 of 1

LLM safety erosion: single-prompt fine-tuning and URL preview data leaks

Enterprise fine-tuning and common chat UI features can quickly undermine LLM safety and silently exfiltrate data, so treat agentic AI security as a lifecycle with zero‑trust controls and gated releases. Microsoft’s GRP‑Obliteration shows a single harmful prompt used with GRPO can collapse guardrails across several model families, reframing safety as an ongoing process rather than a one‑time alignment step [InfoWorld](https://www.infoworld.com/article/4130017/single-prompt-breaks-ai-safety-in-15-major-language-models-2.html)[^1] and is reinforced by a recap urging teams to add safety evaluations to CI/CD pipelines [TechRadar](https://www.techradar.com/pro/microsoft-researchers-crack-ai-guardrails-with-a-single-prompt)[^2]. Separately, researchers demonstrate that automatic URL previews can exfiltrate sensitive data via prompt‑injected links, and a practical release checklist outlines SDLC gates to verify value, trust, and safety before launching agents [WebProNews](https://www.webpronews.com/the-silent-leak-how-url-previews-in-llm-powered-tools-are-quietly-exfiltrating-sensitive-data/)[^3] [InfoWorld](https://www.infoworld.com/article/4105884/10-essential-release-criteria-for-launching-ai-agents.html)[^4]. [^1]: Adds: original reporting on Microsoft’s GRP‑Obliteration results and cross‑model safety degradation. [^2]: Adds: lifecycle framing and guidance to integrate safety evaluations into CI/CD. [^3]: Adds: concrete demonstration of URL‑preview data exfiltration via prompt injection (OpenClaw case study). [^4]: Adds: actionable release‑readiness checklist for AI agents (metrics, testing, governance).

calendar_today 2026-02-10
microsoft azure gpt-oss deepseek-r1-distill google