4.09.2026

Goodbye, Llama? Meta launches new proprietary AI model
Muse Spark

At its core, Muse Spark is a natively multimodal reasoning model. Unlike previous iterations that "stitched" vision and text together, Muse Spark was rebuilt from the ground up to integrate visual information across its internal logic. This architectural shift enables "visual chain of thought," allowing the model to annotate dynamic environments—identifying the components of a complex espresso machine or correcting a user's yoga form via side-by-side video analysis.

4.08.2026

AI joins the 8-hour work day as GLM ships 5.1 open source LLM, beating Opus 4.6 and GPT-5.4 on SWE-Bench Pro

Z.ai, also known as Zhupai AI, a Chinese AI startup best known for its powerful, open source GLM family of models, has unveiled GLM-5.1 today under a permissive MIT License, allowing for enterprises to download, customize and use it for commercial purposes.

The new GLM-5.1 is designed to work autonomously for up to eight hours on a single task, marking a definitive shift from vibe coding to agentic engineering.

4.07.2026

Anthropic, OpenAI, and Microsoft Just Agreed on One File Format.
It Changes Everything.

In this video, I share the inside scoop on how to build agent-readable skills that actually compound:

 • Why the description field is where most skills go to die

 • How agent-first design changes handoffs and contracts

 • What three-tier skill architecture looks like for teams

 • Where community repositories fill the domain-specific gap



4.06.2026

Karpathy shares 'LLM Knowledge Base' architecture that bypasses RAG with an evolving markdown library maintained by AI

Karpathy proposes something simpler and more loosely, messily elegant than the typical enterprise solution of a vector database and RAG pipeline. 

Instead, he outlines a system where the LLM itself acts as a full-time "research librarian"—actively compiling, linting, and interlinking Markdown (.md) files, the most LLM-friendly and compact data format.

By diverting a significant portion of his "token throughput" into the manipulation of structured knowledge rather than boilerplate code, Karpathy has surfaced a blueprint for the next phase of the "Second Brain"—one that is self-healing, auditable, and entirely human-readable.

4.03.2026

Google releases Gemma 4 under Apache 2.0

Gemma 4 arrives as four distinct models organized into two deployment tiers. The "workstation" tier includes a 31B-parameter dense model and a 26B A4B Mixture-of-Experts model — both supporting text and image input with 256K-token context windows. The "edge" tier consists of the E2B and E4B, compact models designed for phones, embedded devices, and laptops, supporting text, image, and audio with 128K-token context windows.

Microsoft takes on AI rivals with three new foundational models

Microsoft AI, the tech giant’s research lab, announced the release of three foundational AI models on Thursday that can generate text, voice, and images.

The release signals Microsoft’s continued push to build out its own stack of multimodal AI models — and compete with rival AI labs — even though it remains tied to OpenAI.

4.02.2026

Hermes Just Solved the Biggest Problem With OpenClaw

Hermes is a specialized AI agent framework that many are calling the OpenClaw killer. And in this video, I walk through what Hermes is, how I'm using it, why and where it's better and different than OpenClaw, and how you can think about using it in your business as well.