4.20.2026

Mozilla launches Thunderbolt AI client with focus on
self-hosted infrastructure

The newly announced Thunderbolt is being sold as a front-end client for users and businesses who want to run their own self-hosted AI infrastructure without relying on cloud-based third-party services.

Thunderbolt is built on top of Haystack, an existing open source AI framework that lets users build custom, modular AI pipelines from user-chosen components. Thunderbolt acts as what Mozilla calls a “sovereign AI client” on top of that underlying infrastructure.

4.17.2026

Anthropic releases Claude Opus 4.7, narrowly retaking lead for most powerful generally available LLM

Anthropic is publicly releasing its most powerful large language model yet, Claude Opus 4.7, today — as it continues to keep an even more powerful successor, Mythos, restricted to a small number of external enterprise partners for cybersecurity testing and patching vulnerabilities in the software said enterprises use (which Mythos exposed rapidly).

The big headlines are that Opus 4.7 exceeds its most direct rivals — OpenAI's GPT-5.4, released in early March 2026, scarcely more than a month ago; and Google's latest flagship model Gemini 3.1 Pro from February — on key benchmarks including agentic coding, scaled tool-use, agentic computer use, and financial analysis.

4.16.2026

Adobe’s new Firefly AI Assistant wants to run Photoshop, Premiere, Illustrator and more from one prompt

Adobe today launched its most ambitious AI offensive to date, unveiling the Firefly AI Assistant — a new agentic creative tool that can orchestrate complex, multi-step workflows across the company's entire Creative Cloud suite from a single conversational interface — alongside a raft of new video, image, and collaboration features designed to position the company at the center of the rapidly evolving AI-powered content creation landscape.

4.15.2026

Claude Code 2.0: New Desktop/CLI App

We’re looking at what feels like an unofficial “Claude Code 2.0” release, with a completely redesigned desktop app that transforms the experience into something much closer to a full IDE.



4.14.2026

Gemma 4 + Ollama = FREE Claude Code Setup!

What if you could get a Claude Code-level AI coding assistant for FREE running locally on your machine? In this video, I show you how to combine Gemma 4 with Ollama to build a powerful, privacy-first, local AI coding setup that behaves surprisingly close to Claude Code.



4.13.2026

Anthropic limits access to Mythos, its new cybersecurity AI model

Anthropic has launched a new cybersecurity AI model to a select group of customers, including Amazon, Apple, and Microsoft, days after details about the project were leaked online.\

Anthropic said the software can identify cyber vulnerabilities at a scale beyond human capacity, but it could also develop ways to exploit these vulnerabilities, which bad actors could use. The company said the model could “reshape” cyber security practices and does not plan a broad release.

4.10.2026

Hermes Agent Full Setup Tutorial: How to Setup Your First
AI Agent (Gemma 4)

In this video I walk through the full setup of Hermes Agent from scratch. We plug it into Gemma 4 running locally through Ollama and set up self-hosted Firecrawl for private web search. By the end, you have a fully local, fully private AI agent connected to Telegram with no paid APIs required.



4.09.2026

Goodbye, Llama? Meta launches new proprietary AI model
Muse Spark

At its core, Muse Spark is a natively multimodal reasoning model. Unlike previous iterations that "stitched" vision and text together, Muse Spark was rebuilt from the ground up to integrate visual information across its internal logic. This architectural shift enables "visual chain of thought," allowing the model to annotate dynamic environments—identifying the components of a complex espresso machine or correcting a user's yoga form via side-by-side video analysis.

4.08.2026

AI joins the 8-hour work day as GLM ships 5.1 open source LLM, beating Opus 4.6 and GPT-5.4 on SWE-Bench Pro

Z.ai, also known as Zhupai AI, a Chinese AI startup best known for its powerful, open source GLM family of models, has unveiled GLM-5.1 today under a permissive MIT License, allowing for enterprises to download, customize and use it for commercial purposes.

The new GLM-5.1 is designed to work autonomously for up to eight hours on a single task, marking a definitive shift from vibe coding to agentic engineering.

4.07.2026

Anthropic, OpenAI, and Microsoft Just Agreed on One File Format.
It Changes Everything.

In this video, I share the inside scoop on how to build agent-readable skills that actually compound:

 • Why the description field is where most skills go to die

 • How agent-first design changes handoffs and contracts

 • What three-tier skill architecture looks like for teams

 • Where community repositories fill the domain-specific gap



4.06.2026

Karpathy shares 'LLM Knowledge Base' architecture that bypasses RAG with an evolving markdown library maintained by AI

Karpathy proposes something simpler and more loosely, messily elegant than the typical enterprise solution of a vector database and RAG pipeline. 

Instead, he outlines a system where the LLM itself acts as a full-time "research librarian"—actively compiling, linting, and interlinking Markdown (.md) files, the most LLM-friendly and compact data format.

By diverting a significant portion of his "token throughput" into the manipulation of structured knowledge rather than boilerplate code, Karpathy has surfaced a blueprint for the next phase of the "Second Brain"—one that is self-healing, auditable, and entirely human-readable.

4.03.2026

Google releases Gemma 4 under Apache 2.0

Gemma 4 arrives as four distinct models organized into two deployment tiers. The "workstation" tier includes a 31B-parameter dense model and a 26B A4B Mixture-of-Experts model — both supporting text and image input with 256K-token context windows. The "edge" tier consists of the E2B and E4B, compact models designed for phones, embedded devices, and laptops, supporting text, image, and audio with 128K-token context windows.

Microsoft takes on AI rivals with three new foundational models

Microsoft AI, the tech giant’s research lab, announced the release of three foundational AI models on Thursday that can generate text, voice, and images.

The release signals Microsoft’s continued push to build out its own stack of multimodal AI models — and compete with rival AI labs — even though it remains tied to OpenAI.

4.02.2026

Hermes Just Solved the Biggest Problem With OpenClaw

Hermes is a specialized AI agent framework that many are calling the OpenClaw killer. And in this video, I walk through what Hermes is, how I'm using it, why and where it's better and different than OpenClaw, and how you can think about using it in your business as well.



4.01.2026

Running local models on Macs gets faster with Ollama’s
MLX support

Ollama, a runtime system for operating large language models on a local computer, has introduced support for Apple’s open source MLX framework for machine learning. Additionally, Ollama says it has improved caching performance and now supports Nvidia’s NVFP4 format for model compression, making for much more efficient memory usage in certain models.