terminal
howtonotcode.com
YouTube logo

YouTube

Platform

A video sharing platform for uploading, viewing, and sharing videos.

article 46 storys calendar_today First seen: 2025-12-30 update Last seen: 2026-03-03 open_in_new Website menu_book Wikipedia

Resources

Links to check for updates: homepage, feed, or git repo.

home Homepage

Stories

Showing 41-46 of 46

OpenClaw rockets to GitHub’s top spot—security and ops readiness now in focus

OpenClaw, an open-source legal AI project, has surged to GitHub’s most-starred status while raising fresh security and governance questions for teams considering adoption. A [WebProNews report](https://www.webpronews.com/openclaws-meteoric-rise-on-github-how-an-open-source-legal-ai-project-dethroned-react-as-the-most-starred-software-repository/) says OpenClaw has overtaken React in stars, propelled by its structured legal datasets and AI tooling that promise to democratize access and fuel model training. The New Stack urges caution on provenance and security in “is it safe?” coverage, flagging supply-chain and governance risks before production use ([read more](https://thenewstack.io/openclaw-github-stars-security/)). A March update video highlights Docker support, cron job fixes, and how-to-upgrade guidance—plus references to Claude 4.6 “Adaptive Thinking”—signaling quickening operational maturity and clearer integration touchpoints ([watch](https://www.youtube.com/watch?v=4K1JRI7xA08&pp=ygUSQ2xhdWRlIENvZGUgdXBkYXRl)).

calendar_today 2026-03-03
openclaw github claude docker security

Coding Benchmarks Shake-up: Qwen 3.5, MiniMax M2.5, and a SWE-bench Reality Check

Open models like Alibaba’s Qwen 3.5 and MiniMax M2.5 post strong coding-agent results, but OpenAI’s audit of SWE-bench Verified shows contamination and flawed tests that can mislead real-world adoption. Alibaba’s Qwen 3.5 family uses a sparse MoE design (397B total/17B active), ships open weights under Apache 2.0, and shows strong instruction following and competitive coding scores in public benchmarks, with setup guidance and comparisons to frontier models detailed in this deep-dive guide [Qwen 3.5: The Complete Guide](https://techie007.substack.com/p/qwen-35-the-complete-guide-benchmarks). MiniMax’s latest model claims state-of-the-art coding and agentic performance, faster task completion, and ultra-low runtime cost (about $1/hour at 100 tok/s), alongside reported scores on coding and browsing evaluations [MiniMax-M2.5 on Hugging Face](https://huggingface.co/unsloth/MiniMax-M2.5). OpenAI, however, reports that many SWE-bench Verified tasks have broken tests and that major models were trained on benchmark solutions, halting its use of the metric and urging caution in interpreting scores [OpenAI Abandons SWE-bench Verified](https://blockchain.news/news/openai-abandons-swe-bench-verified-contamination-flawed-tests). For quick, low-cost trials of multiple “top models,” a short explainer points to an Alibaba Cloud coding plan bundling popular options [This $3 AI Coding Plan Gives You Every Top Model You Need](https://www.youtube.com/watch?v=Qnz7S-5fzWo&pp=ygUXbmV3IEFJIG1vZGVsIGZvciBjb2RpbmfSBwkJrgoBhyohjO8%3D).

calendar_today 2026-03-03
qwen-35 alibaba alibaba-cloud minimax-m25 openai

From vibe coding to agentic engineering: PEV, context, and evals that ship

Production teams are moving from vibe coding to agentic engineering that plans, executes, and verifies work with tight context and evals. A practical guide to agentic engineering argues for a Plan → Execute → Verify loop, with humans acting as architects and supervisors while agents plan, write, test, and ship; it cites real adoption signals like TELUS time-savings, Zapier-wide usage, and Stripe’s weekly PR throughput ([guide](https://www.nxcode.io/resources/news/agentic-engineering-complete-guide-vibe-coding-ai-agents-2026)). Context discipline is emerging as a make-or-break factor: a new study shows repo-level AGENTS.md/CLAUDE.md files can degrade agent performance, pushing teams toward slimmer, task-scoped context that’s validated in CI ([AGENTS.md breakdown](https://www.youtube.com/watch?v=miDg-3rSJlQ&t=75s&pp=ygURU1dFLWJlbmNoIHJlc3VsdHM%3D), [DevOps context engineering](https://devops.com/context-engineering-is-the-key-to-unlocking-ai-agents-in-devops-2/)). Architecturally, vibe coding is “already dead” at scale; production agents enforce planning, tests, PR gates, and continuous evals before code lands ([Stripe agent deep dive](https://www.youtube.com/watch?v=V5A1IU8VVp4&pp=ygUYQUkgY29kaW5nIGFnZW50IHdvcmtmbG93)). For hands-on operating patterns—self-checks, context management, and when to escalate to humans—see this practitioner’s playbook ([effective coding agents](https://hackernoon.com/how-to-use-ai-coding-agents-effectively?source=rss)).

calendar_today 2026-03-03
stripe zapier telus claude-code openai-codex

Pragmatic agentic coding workflow using Claude Code

A YouTube walkthrough shows a pragmatic agentic coding workflow to build software end-to-end with coding agents like Claude Code. This [walkthrough video](https://www.youtube.com/watch?v=goOZSXmrYQ4&t=2320s&pp=ygUXY29kaW5nIGFnZW50IGV2YWx1YXRpb24%3D) demonstrates building from scratch by delegating coding tasks to an agent, iterating on outputs, and keeping the process lean to avoid overengineering. For engineering leads, it offers a concrete pattern for when to let the agent scaffold and when to step in for targeted reviews and tests, helping teams move faster while keeping baseline quality guardrails.

calendar_today 2026-02-24
claude-code youtube agentic-workflows ai-coding-agents sdlc

From vibe coding to agentic engineering: test-first orchestration

Engineering teams are shifting from vibe coding to disciplined agentic engineering that treats AI as test-driven collaborators and demands spec-first oversight. In a concise critique of “prompt DJ” development, [Roger Wong](https://rogerwong.me/2026/02/agentic-engineering) summarizes Addy Osmani’s call for agentic engineering—engineers orchestrate coding agents, act as architects and reviewers, and enforce spec-first discipline instead of accepting whatever the model returns. [Simon Willison’s](https://simonwillison.net/guides/agentic-engineering-patterns/first-run-the-tests/#atom-everything) “First run the tests” pattern operationalizes this by making a test suite the entry point for any agent, turning TDD into a four‑word prompt and letting agents learn a codebase through its tests. Hands-on workflows show how to scale this in practice, from a [complete greenfield agentic setup](https://www.youtube.com/watch?v=goOZSXmrYQ4&pp=ygUYQUkgY29kaW5nIGFnZW50IHdvcmtmbG93) to [advanced agent teams comparing Claude Code and Codex](https://www.youtube.com/watch?v=7BXZ-qR5cPE&pp=ygUYQUkgY29kaW5nIGFnZW50IHdvcmtmbG93), while case studies like [DumbQuestion.ai](https://dev.to/jagostoni/dumbquestionai--2ee) underline the need for structured backlogs and cost-aware multi‑model choices.

calendar_today 2026-02-24
openai codex claude-code openrouter agentic-engineering

AI IDEs go agentic: Cursor "demos" and Windsurf Cascade

AI IDEs are shifting from code suggestions to autonomous agents that run, test, and showcase changes, led by Cursor’s new demo-first experience and Windsurf’s Cascade engine. Cursor now emphasizes "demos, not diffs," with agents that can run the software they build and send video evidence of their changes ([YouTube](https://www.youtube.com/watch?v=XbZvC4KTH68&pp=ygURQ3Vyc29yIElERSB1cGRhdGU%3D)). Meanwhile, Windsurf’s agentic Cascade engine promises project-aware, multi-file edits on a familiar VS Code foundation with simple onboarding and settings import ([TechCompanyNews guide](https://www.techcompanynews.com/how-to-use-windsurf-step-by-step-guide-for-beginners/)). The direction is clear: AI IDEs are moving from inline suggestions to autonomous, runnable workflows. Operational maturity remains a concern: users report surprise auto-updates ([automatic updater](https://forum.cursor.com/t/cursor-automatic-updater/152697)), Windows update failures ([Windows updates failing](https://forum.cursor.com/t/updates-on-windows-are-failing-still-antivirus/152819)), and visibility issues before approval in a recent build ([v2.5.20 diffs visibility](https://forum.cursor.com/t/modified-code-changes-not-visible-before-approval-cursor-v2-5-20/152760)), alongside UI changes like replacing "Keep All" with auto-approve ([discussion](https://forum.cursor.com/t/the-loss-of-keep-all-the-addition-of-auto-approve/152780)). Community threads also cite rate limits even on paid plans ([Reddit](https://www.reddit.com/r/cursor/comments/1rdfk9p/what_would_make_you_switch_from_cursor_to_another/)) and a practical auth fix for a Windsurf codex plugin by clearing a local token file ([Reddit fix](https://www.reddit.com/r/codex/comments/1rdddu3/windsurf_codex_plugin_issue/)). Teams are sketching an "AI builder stack" that pairs an agentic IDE with project tracking, instant deploy previews, and AI QA to close the loop from change to validation ([HackerNoon](https://hackernoon.com/the-ai-builder-stack-linear-cursor-vercel-and-qatech?source=rss)). New native entrants like macOS-focused G-Rump hint at a widening field and specialization opportunities ([Swift forums](https://forums.swift.org/t/g-rump-a-native-macos-ai-coding-agent-looking-for-early-feedback/84953)).

calendar_today 2026-02-24
cursor windsurf codeium visual-studio-code linear