14.3.26

I Told an AI to Build a Full Newsletter. One Sentence. No Code.

I used to spend half a day on client newsletters. Last week I did it in 15 minutes — and I'll show you exactly what I typed.


๐Ÿค– What Claude Code Actually Is

Claude Code is an extension you install inside VS Code — a free code editor. Once it's set up, you have an AI agent sitting inside your project that can read files, write code, call APIs, fix its own errors, and run automations — all through a chat window.

You talk to it. It builds things. That's the whole loop. You don't need to know how to code. You need to know what you want.


๐Ÿงช What I Actually Tested

I set up a project, gave the agent a brand logo and a color guide, connected a few API keys (one for research, one for Gmail), and typed one prompt. That was it.

What came back:

  • Research pulled from live sources
  • Three branded infographics generated with AI images
  • Full HTML newsletter formatted to the brand's colors and fonts
  • Email sent directly to a list via Gmail
  • Sources linked at the bottom, clickable

It hit errors along the way. An API endpoint had changed. A Gmail formatting issue broke the layout. Both times, it found the problem, explained what happened, fixed the tool, and kept running. I didn't touch anything.


๐Ÿ—️ How the Whole Thing Is Organized

The system behind this is called

Component What It Is
Pipeline The instructions, written in plain English — like a step-by-step recipe
Tools The specific actions the agent can take (research, generate image, send email)
Orchestrator Claude Code itself, reading the pipeline and running each tool in order

You also give it a file called CLAUDE.md — a standing brief that tells the agent what the project is, where things live, and what it's supposed to do. Every time you open the project, it reads that first.

Once you've built and tested a pipeline, you can schedule it to run on its own — every Monday morning, every time a form is submitted, whatever you need. At that point the agent isn't running live. The code it built is. It runs predictably, like any other automation.


๐Ÿ’ผ What This Means for VAs and Small Business Owners

You don't need to learn to code. You need to be able to describe a process clearly.

If you manage newsletters, reports, client updates, or social content for clients — build a pipeline, test it once, let it run. If you spend hours every week on the same repetitive tasks in your own business, same idea.

What makes this different from something like Zapier is what happens during the build. The agent adapts. It hits an error, investigates, fixes the tool, and updates the instructions so the same thing doesn't break again. The final automation is more reliable — and you got there faster than mapping every node by hand.


๐Ÿ’ฐ If You're Thinking About Offering This as a Service

Businesses aren't paying for a pipeline or a blueprint. They're paying for a solved problem.

Don't open with "do you want AI automation?" Ask: where are you losing the most time or money right now? Find that, build the thing that fixes it, and price it on what it's worth — hours saved, errors cut, costs gone — not on how long you spent building it.

A pipeline that saves a client 20 hours a week isn't a half-day job. Over a year, that's tens of thousands of dollars in time they're not paying someone else to cover. Price it that way.


๐Ÿš€ Where to Start

Tool Claude Code — runs inside VS Code (free to download)
Cost Paid Claude subscription — $17/month on the Pro plan
Setup time ~1 hour, most of it connecting API keys

Pick one task you do every single week. Something repetitive, something predictable. Let the agent build the first version, watch it run, fix what doesn't land, and build from there.

The hardest part isn't the tool. It's deciding what to automate first. 


This Paper Says We've Been Fine-Tuning the Hard Way

I was reading a paper that dropped in March 2026 and about three paragraphs in, I had to stop. The claim seemed too simple: you can adapt a large AI model to a specific task without gradient descent at all. Just random sampling and a vote.


What Fine-Tuning Actually Is

When an AI model gets trained, it learns general patterns from a massive dataset. Fine-tuning is the step where you take that general model and push it toward a specific task — coding, reasoning, following instructions — using reinforcement learning or optimization algorithms.

Methods like PPO and GRPO (types of reinforcement learning commonly used to fine-tune large language models) work well. But they're expensive, require careful setup, and involve a lot of iteration.


The Paper's Core Claim

Title: Neural Thickets: Diverse Task Experts Are Dense Around Pretrained Weights Authors: Yulu Gan and Phillip Isola Published: March 12, 2026 · arXiv: 2603.12228

The idea: in large, well-pretrained models, the weight space around the original parameters is already densely packed with useful task-specific solutions. The authors call these clusters "neural thickets."

In smaller models, good solutions are scattered — you need gradient-based search to find them. In large models, they're close to where you already are.


The Method — Remarkably Simple

They call it RandOpt. Here's how it works:

  1. Take the pretrained model weights
  2. Randomly sample N small perturbations of those weights
  3. Evaluate each perturbation on your task
  4. Keep the top K performers
  5. Combine them with a majority vote

No gradients. No reward model. No RL training loop. Perturb, evaluate, vote.

๐Ÿ’ก Code at github.com/sunrainyg/RandOpt


What the Results Showed

RandOpt kept up with PPO, GRPO, and evolutionary strategies on the tasks they tested — and this held on large-scale contemporary models.

That stopped me. These optimization methods have entire research communities, years of papers, and significant infrastructure built around them. The idea that randomly sampling neighbors of the original weights and taking a vote can match them says something about what's already sitting inside a well-trained model — waiting, not absent.

The way they frame it: small models have sparse expert solutions, so you need search to find them. Large models have dense ones. The pretrained weights aren't just a starting point. They're already rich.


What I Took From This

This paper shifted something in how I think about pretraining. We usually assume the pretrained model is a rough starting point and fine-tuning is where the real work happens. This flips that. If the solution space is already dense around the initial weights, the pretrained model is doing more work than we give it credit for.

There's also a question I don't have a full answer to yet: if random sampling with ensemble voting matches expensive RL fine-tuning, what does that mean for how we should be spending compute? I'm still working through the paper and the code, but it's the kind of result that sits with you.

Full paper: arXiv 2603.12228

12.3.26

Google Just Replaced Five AI Search Tools With One

Have you ever tried searching through a client’s content library where videos are in one folder, PDFs in another, and audio recordings scattered everywhere?

That’s the reality for most content libraries.

Until now, AI search tools struggled with this kind of setup because each type of content needed a different system to process it.

But that may be changing.

Gemini Embedding 2 — recently released by Google — can search across text, images, audio, video, and PDFs at the same time, without converting everything first.

For anyone managing knowledge bases, course content, research archives, or client media libraries, this could be a major shift.


What an Embedding Model Actually Does

Before explaining why this matters, it helps to understand what an embedding model is.

When AI systems search through content, they don’t read information the same way humans do. Instead, they convert content into numerical representations that capture meaning.

For example:

  • A sentence about a cat

  • A photo of a cat

Both produce similar number patterns, which allows the AI to recognize that they are related.

That’s how modern AI-powered search works.

The tool responsible for converting content into these numerical representations is called an embedding model.


The Problem With Older AI Search Systems

Until recently, every type of content required a different embedding system.

Typical setups looked like this:

  • Text → processed by a text embedding model

  • Images → processed by image models such as CLIP or SigLIP

  • Audio → first transcribed using systems like Whisper

  • Video → broken into frames or transcripts

  • PDFs → converted into plain text

This created several issues:

  • Multiple models to manage

  • Several conversion steps

  • More chances for things to break

  • Slower search performance

In many cases, five different pipelines were required just to search one content library.


What Gemini Embedding 2 Changes

Gemini Embedding 2 solves this by creating one shared search space for multiple content types.

Instead of converting everything separately, the model processes different media formats directly and places them into the same semantic search system.

That means a single query can return results from:

  • Documents

  • Images

  • Audio clips

  • Video files

  • PDFs

All at once.

For example, you could:

  • Upload a photo and find related videos

  • Submit a voice recording and find matching documents

  • Search inside PDF files without converting them


Supported Input Types

Gemini Embedding 2 currently supports multiple media types in one system:

Text
Up to roughly 8,000 words

Images
Up to six images in one request

Audio
Raw audio files — no transcription required

Video
Clips up to two minutes long

PDFs
Original files can be processed without converting to plain text

All of this works through one model instead of multiple specialized ones.


Combining Multiple Inputs in One Search

One interesting feature is the ability to combine different types of input into a single query.

For example, you might have:

  • A photo of a product

  • A text description of what you want

Both can be submitted together, and the system generates one combined embedding representing the meaning of both inputs.

This allows searches that were previously impossible using single-modality tools.


Easy Integration for Developers

Another surprising detail is how quickly developers can start using it.

Gemini Embedding 2 launched with support for popular AI development frameworks, including:

  • LangChain

  • LlamaIndex

  • ChromaDB

  • QDrant

Because many AI applications are already built on these frameworks, developers can integrate the model without building a new infrastructure from scratch.

It’s available through:

  • Google AI Studio (free tier for experimentation)

  • Vertex AI (enterprise deployment)


Why This Matters for Virtual Assistants and Content Managers

Think about the kinds of content many clients manage.

A podcast brand might have:

  • Audio episodes

  • Show notes

  • PDFs

  • Promotional images

A course creator may have:

  • Video lessons

  • Slide decks

  • Written summaries

A consultant might maintain:

  • Recorded calls

  • Presentations

  • Research reports

Searching across all of that in a single step has been extremely difficult.

With models like Gemini Embedding 2, developers can build search tools where one query instantly returns:

  • the right video segment

  • the correct slide

  • the relevant document section

All from one search bar.


The Bigger Picture

You probably won’t interact with Gemini Embedding 2 directly.

Instead, it will power the next generation of search tools used in:

  • knowledge management systems

  • research databases

  • course platforms

  • internal company search tools

But knowing that technology like this exists helps you understand what’s becoming possible.

That knowledge can make a big difference when clients start asking about AI-powered search, automation, or content organization systems.


If you manage content libraries, research archives, or client knowledge bases, this is a technology worth paying attention to.

The tools many teams will rely on in the near future are already being built on models like this. 

Karpathy doesn't use a fancy app to manage his research. He uses a folder, Obsidian, and an AI — and I want to copy it. He posted about ...