10.6.25

Amperity Launches Chuck Data: A Vibe-Coding AI Agent for Customer Data Engineering

 Amperity Introduces Chuck Data: An AI Agent to Automate Customer Data Engineering with Natural Language

Seattle-based customer data platform (CDP) startup Amperity Inc. has entered the AI agent arena with the launch of Chuck Data, a new autonomous assistant built specifically to tackle customer data engineering tasks. The tool aims to empower data engineers by reducing their reliance on manual coding and enabling natural language-driven workflows, a concept Amperity calls "vibe coding."

Chuck Data is trained on vast volumes of customer information derived from over 400 enterprise brands, giving it a "critical knowledge" base. This foundation enables the agent to perform tasks like identity resolution, PII (Personally Identifiable Information) tagging, and data profiling with minimal developer input.

A Natural Language AI for Complex Data Tasks

Amperity’s platform is well-known for its ability to ingest data from disparate systems — from customer databases to point-of-sale terminals — and reconcile inconsistencies to form a cohesive customer profile. Chuck Data extends this capability by enabling data engineers to communicate using plain English, allowing them to delegate repetitive, error-prone coding tasks to an intelligent assistant.

With direct integration into Databricks environments, Chuck Data leverages native compute resources and large language model (LLM) endpoints to execute complex data engineering workflows. From customer identity stitching to compliance tagging, the agent promises to significantly cut down on time and manual effort.

Identity Resolution at Scale

One of Chuck Data’s standout features is its use of Amperity’s patented Stitch identity resolution algorithm. This powerful tool can combine fragmented customer records to produce unified profiles — a key requirement for enterprises aiming to understand and engage their audiences more effectively.

To promote adoption, Amperity is offering free access to Stitch for up to 1 million customer records. Enterprises with larger datasets can join a research preview program or opt for paid plans with unlimited access, supporting scalable, AI-powered data unification.

PII Tagging and Compliance: A High-Stakes Task

As AI-driven personalization becomes more prevalent, the importance of data compliance continues to grow. Liz Miller, analyst at Constellation Research, emphasized that automating PII tagging is crucial, but accuracy is non-negotiable.

“When PII tagging is not done correctly and compliance standards cannot be verified, it costs the business not just money, but also customer trust,” said Miller.

Chuck Data aims to prevent such issues by automating compliance tasks with high accuracy, minimizing the risk of mishandling sensitive information.

Evolving the Role of the CDP

According to Michael Ni, also from Constellation Research, Chuck Data represents the future of customer data platforms — transforming from static data organizers into intelligent systems embedded within the data infrastructure.

“By running identity resolution and data preparation natively in Databricks, Amperity demonstrates how the next generation of CDPs will shift core governance tasks to the data layer,” said Ni. “This allows the CDP to focus on real-time personalization and business decision-making.”

The End of Manual Data Wrangling?

Derek Slager, CTO and co-founder of Amperity, said the goal of Chuck Data is to eliminate the “repetitive and painful” aspects of customer data engineering.

“Chuck understands your data and helps you get stuff done faster, whether you’re stitching identities or tagging PII,” said Slager. “There’s no orchestration, no UI gymnastics – it’s just fast, contextual, and command-driven.”


With Chuck Data, Amperity is betting big on agentic AI to usher in a new era of intuitive, fast, and compliant customer data management — one where data engineers simply describe what they want, and AI does the rest.

OpenAI Surpasses $10 Billion in Annual Recurring Revenue as ChatGPT Adoption Skyrockets

 OpenAI has crossed a significant financial milestone, achieving an annual recurring revenue (ARR) run rate of $10 billion as of mid-2025. This growth marks a nearly twofold increase from the $5.5 billion ARR reported at the end of 2024, underscoring the explosive rise in demand for generative AI tools across industries and user demographics.

According to insiders familiar with the company’s operations, this growth is largely fueled by the surging popularity of ChatGPT and a steady uptick in the use of OpenAI’s APIs and enterprise services. ChatGPT alone now boasts between 800 million and 1 billion users globally, with approximately 500 million active users each week. Of these, 3 million are paid business subscribers, reflecting robust interest from corporate clients.


A Revenue Surge Driven by Strategic Products and Partnerships

OpenAI’s flagship products—ChatGPT and its developer-facing APIs—are at the heart of this momentum. The company has successfully positioned itself as a leader in generative AI, building tools that range from conversational agents and writing assistants to enterprise-level automation and data analysis platforms.

Its revenue model is primarily subscription-based. Businesses pay to access advanced features, integration capabilities, and support, while developers continue to rely on OpenAI’s APIs for building AI-powered products. With both individual and corporate users increasing rapidly, OpenAI’s ARR has climbed steadily.


Strategic Acquisitions Fuel Growth and Innovation

To further bolster its capabilities, OpenAI has made key acquisitions in 2025. Among the most significant are:

  • Windsurf (formerly Codeium): Acquired for $3 billion, Windsurf enhances OpenAI’s position in the AI coding assistant space, providing advanced code completion and debugging features that rival GitHub Copilot.

  • io Products: A startup led by Jony Ive, the legendary former Apple designer, was acquired for $6.5 billion. This move signals OpenAI’s intent to enter the consumer hardware market with devices optimized for AI interaction.

These acquisitions not only broaden OpenAI’s product ecosystem but also deepen its influence in software development and design-forward consumer technology.


Setting Sights on $12.7 Billion ARR and Long-Term Profitability

OpenAI’s trajectory shows no signs of slowing. Company forecasts project ARR reaching $12.7 billion by the end of 2025, a figure that aligns with investor expectations. The firm recently closed a major funding round led by SoftBank, bringing its valuation to an estimated $300 billion.

Despite a substantial operating loss of $5 billion in 2024 due to high infrastructure and R&D investments, OpenAI is reportedly aiming to become cash-flow positive by 2029. The company is investing heavily in building proprietary data centers, increasing compute capacity, and launching major infrastructure projects like “Project Stargate.”


Navigating a Competitive AI Landscape

OpenAI’s aggressive growth strategy places it ahead of many competitors in the generative AI space. Rival company Anthropic, which developed Claude, has also made strides, recently surpassing $3 billion in ARR. However, OpenAI remains the market leader, not only in revenue but also in market share and influence.

As the company scales, challenges around compute costs, user retention, and ethical deployment remain. However, with solid financial backing and an increasingly integrated suite of products, OpenAI is positioned to maintain its leadership in the AI arms race.


Conclusion

Reaching $10 billion in ARR is a landmark achievement that cements OpenAI’s status as a dominant force in the AI industry. With a growing user base, major acquisitions, and a clear roadmap toward long-term profitability, the company continues to set the pace for innovation and commercialization in generative AI. As it expands into hardware and deepens its enterprise offerings, OpenAI’s influence will likely continue shaping the next decade of technology.

Ether0: The 24B-Parameter Scientific Reasoning Model Accelerating Molecular Discovery

 FutureHouse has unveiled Ether0, a 24 billion-parameter open-source reasoning model specialized for chemistry tasks. Built on Mistral 24B and fine-tuned through chain-of-thought reinforcement learning, Ether0 accepts natural-language prompts and generates molecule structures in SMILES notation, excelling particularly in drug-like compound design.

Why Ether0 Matters

While general-purpose LLMs possess extensive chemical knowledge, they falter at molecule manipulation—incorrect atom counts, implausible rings, or inaccurate compound names. Ether0 addresses these deficiencies by learning from reinforcement signals grounded in chemical validity rather than mimicry, significantly boosting accuracy in molecule generation.

Training Methodology

  • Base Model & Datasets: Starts with Mistral 24B Instruct.

  • Fine-tuning: Trains chains of thought and correct answers through supervised learning, separating specialists per task.

  • Reinforcement Learning: Specialized models trained on molecular tasks across ~50K examples each.

  • Distillation: Merges specialist reasoning into a generalized model, further refined with reinforcement over multiple tasks.

This modular workflow enables data efficiency, with Ether0 surpassing frontier models like GPT‑4.1 and DeepSeek‑R1 on chemistry problems while using substantially less data than traditional methods.

Capabilities and Limits

Ether0 accurately handles tasks such as:

  • Converting formulas (e.g., C₂₇H₃₇N₃O₄) to valid molecules.

  • Designing compounds by functional groups, solubility, pKa, smell, or receptor binding.

  • Proposing retrosynthesis steps and reaction outcomes.

However, it falters in:

  • Naming via IUPAC or common names.

  • Reasoning on molecular conformations.

  • General conversational chemistry outside strict molecule output.

The model develops unique behaviors—blending languages and inventing new terms (e.g., “reductamol”)—reflecting deeper reasoning at the cost of clarity in some reasoning traces.

Safety & Governance

Ether0 is released under an Apache 2.0 license and includes safeguards: refusal on controlled compounds, missiles-toxins filters, and rejection of explicit malicious content. This safety post-processing is critical given its open-weight deployment.

Community & Future Vision

Built by a FutureHouse team supported by Eric Schmidt and VoltagePark, Ether0 is part of a broader quest to automate scientific discovery via AI agents. The code, reward models, benchmarks, and model weights are available on GitHub and Hugging Face. Next steps include integrating Ether0 into Phoenix—FutureHouse’s chemistry agent—as a foundational block toward a generalized scientific reasoning engine 


Key Takeaways

  1. Domain-specific reasoning: Demonstrates how reinforcement-tuned LLMs can learn scientific tasks beyond pretraining.

  2. Data-efficient training: Delivers strong performance using ~50K task-specific examples, far fewer than traditional AI training regimes.

  3. Open-source advancement: Enables scientific and developer communities to build upon Ether0 in drug design and other chemistry domains.

  4. Transparent reasoning traces: Offers insight into LLM ‘thought processes’, facilitating interpretability in scientific AI.

9.6.25

Google Open‑Sources a Full‑Stack Agent Framework Powered by Gemini 2.5 & LangGraph

 Google has unveiled an open-source full-stack agent framework that combines Gemini 2.5 and LangGraph to create conversational agents capable of multi-step reasoning, iterative web search, self-reflection, and synthesis—all wrapped in a React-based frontend and Python backend 


🔧 Architecture & Workflow

The system integrates these components:

  • React frontend: User interface built with Vite, Tailwind CSS, and Shadcn UI.

  • LangGraph backend: Orchestrates agent workflow using FastAPI for API handling and Redis/PostgreSQL for state management 

  • Gemini 2.5 models: Power each stage—dynamic query generation, reflection-based reasoning, and final answer synthesis.


🧠 Agent Reasoning Pipeline

  1. Query Generation
    The agent kicks off by generating targeted web search queries via Gemini 2.5.

  2. Web Research
    Uses Google Search API to fetch relevant documents.

  3. Reflective Reasoning
    The agent analyzes results for "knowledge gaps" and determines whether to continue searching—essential for deep, accurate answers 

  4. Iterative Looping
    It refines queries and repeats the search-reflect cycle until satisfactory results are obtained.

  5. Final Synthesis
    Gemini consolidates the collected information into a coherent, citation-supported answer.


🚀 Developer-Friendly

  • Hot-reload support: Enables real-time updates during development for both frontend and backend 

  • Full-stack quickstart repo: Available on GitHub with Docker‑Compose setup for local deployment using Gemini and LangGraph 

  • Robust infrastructure: Built with LangGraph, FastAPI, Redis, and PostgreSQL for scalable research applications.


🎯 Why It Matters

This framework provides a transparent, research-grade AI pipeline: query ➞ search ➞ reflect ➞ iterate ➞ synthesize. It serves as a foundation for building deeper, more reliable AI assistants capable of explainable and verifiable reasoning—ideal for academic, enterprise, or developer research tools 


⚙️ Getting Started

To get hands-on:

  • Clone the Gemini Fullstack LangGraph Quickstart from GitHub.

  • Add .env with your GEMINI_API_KEY.

  • Run make dev to start the full-stack environment, or use docker-compose for production setup 

This tooling lowers the barrier to building research-first agents, making multi-agent workflows more practical for developers.


✅ Final Takeaway

Google’s open-source agent stack is a milestone: it enables anyone to deploy intelligent agents capable of deep research workflows with citation transparency. By combining Gemini's model strength, LangGraph orchestration, and a polished React UI, this stack empowers users to build powerful, self-improving research agents faster.

Enable Function Calling in Mistral Agents Using Standard JSON Schema

 This updated tutorial guides developers through enabling function calling in Mistral Agents via the standard JSON Schema format Function calling allows agents to invoke external APIs or tools (like weather or flight data services) dynamically during conversation—extending their reasoning capabilities beyond text generation.


🧩 Why Function Calling?

  • Seamless tool orchestration: Enables agents to perform actions—like checking bank interest rates or flight statuses—in real time.

  • Schema-driven clarity: JSON Schema ensures function inputs and outputs are well-defined and type-safe.

  • Leverage MCP Orchestration: Integrates with Mistral's Model Context Protocol for complex workflows 


🛠️ Step-by-Step Implementation

1. Define Your Function

Create a simple API wrapper, e.g.:

python
def get_european_central_bank_interest_rate(date: str) -> dict: # Mock implementation returning a fixed rate return {"date": date, "interest_rate": "2.5%"}

2. Craft the JSON Schema

Define the function parameters so the agent knows how to call it:

python
tool_def = { "type": "function", "function": { "name": "get_european_central_bank_interest_rate", "description": "Retrieve ECB interest rate", "parameters": { "type": "object", "properties": { "date": {"type": "string"} }, "required": ["date"] } } }

3. Create the Agent

Register the agent with Mistral's SDK:

python
agent = client.beta.agents.create( model="mistral-medium-2505", name="ecb-interest-rate-agent", description="Fetch ECB interest rate", tools=[tool_def], )

The agent now recognizes the function and can decide when to invoke it during a conversation.

4. Start Conversation & Execute

Interact with the agent using a prompt like, "What's today's interest rate?"

  • The agent emits a function.call event with arguments.

  • You execute the function and return a function.result back to the agent.

  • The agent continues based on the result.

This demo uses a mocked example, but any external API can be plugged in—flight info, weather, or tooling endpoints 


✅ Takeaways

  • JSON Schema simplifies defining callable tools.

  • Agents can autonomously decide if, when, and how to call your functions.

  • This pattern enhances Mistral Agents’ real-time capabilities across knowledge retrieval, action automation, and dynamic orchestration.

Google’s MASS Revolutionizes Multi-Agent AI by Automating Prompt and Topology Optimization

 Designing multi-agent AI systems—where several AI "agents" collaborate—has traditionally depended on manual tuning of prompt instructions and agent communication structures (topologies). Google AI, in partnership with Cambridge researchers, is aiming to change that with their new Multi-Agent System Search (MASS) framework. MASS brings automation to the design process, ensuring consistent performance gains across complex domains.


🧠 What MASS Actually Does

MASS performs a three-stage automated optimization that iteratively refines:

  1. Block-Level Prompt Tuning
    Fine-tunes individual agent prompts via local search—sharpening their roles (think “questioner”, “solver”).

  2. Topology Optimization
    Identifies the best agent interaction structure. It prunes and evaluates possible communication workflows to find the most impactful design.

  3. Workflow-Level Prompt Refinement
    Final tuning of prompts once the best network topology is set.

By alternating prompt and topology adjustments, MASS achieves optimization that surpasses previous methods which tackled only one dimension 


🏅 Why It Matters

  • Benchmarked Success: MASS-designed agent systems outperform AFlow and ADAS on challenging benchmarks like MATH, LiveCodeBench, and multi-hop question-answering 

  • Reduced Manual Overhead: Designers no longer need to trial-and-error their way through thousands of prompt-topology combinations.

  • Extended to Real-World Tasks: Whether for reasoning, coding, or decision-making, this framework is broadly applicable across domains.


💬 Community Reactions

Reddit’s r/machinelearningnews highlighted MASS’s leap beyond isolated prompt or topology tuning:

“Multi-Agent System Search (MASS) … reduces manual effort while achieving state‑of‑the‑art performance on tasks like reasoning, multi‑hop QA, and code generation.” linkedin.com

 


📘 Technical Deep Dive

Originating from a February 2025 paper by Zhou et al., MASS represents a methodological advance in agentic AI

  • Agents are modular: designed for distinct roles through prompts.

  • Topology defines agent communication patterns: linear chain, tree, ring, etc.

  • MASS explores both prompt and topology spaces, sequentially optimizing them across three stages.

  • Final systems demonstrate robustness not just in benchmarks but as a repeatable design methodology.


🚀 Wider Implications

  • Democratizing Agent Design: Non-experts in prompt engineering can deploy effective agent systems from pre-designed searches.

  • Adaptability: Potential for expanding MASS to dynamic, real-world settings like real-time planning and adaptive workflows.

  • Innovation Accelerator: Encourages research into auto-tuned multi-agent frameworks for fields like robotics, data pipelines, and interactive assistants.


🧭 Looking Ahead

As Google moves deeper into its “agentic era”—with initiatives like Project Mariner and Gemini's Agent Mode—MASS offers a scalable blueprint for future AS/AI applications. Expect to see frameworks that not only generate prompts but also self-optimize their agent networks for performance and efficiency.

7.6.25

Alibaba's Qwen3-Embedding and Qwen3-Reranker: Redefining Multilingual Embedding and Ranking Standards linkedin.com +3

 Alibaba's Qwen team has unveiled two groundbreaking models: Qwen3-Embedding and Qwen3-Reranker, aiming to revolutionize multilingual text embedding and relevance ranking. These models are designed to address the complexities of multilingual natural language processing (NLP) tasks, offering enhanced performance and versatility.

Key Features and Capabilities

  • Multilingual Proficiency:
    Both models support an impressive array of 119 languages, making them among the most versatile open-source offerings available today. 

  • Model Variants:
    Available in three sizes—0.6B, 4B, and 8B parameters—these models cater to diverse deployment needs, balancing efficiency and performance. 

  • State-of-the-Art Performance:
    Qwen3-Embedding and Qwen3-Reranker have achieved top rankings on multiple benchmarks, including MTEB, MMTEB, and MTEB-Code, outperforming leading models like Gemini. 

  • Versatile Applications:
    These models are optimized for a range of tasks such as semantic retrieval, classification, retrieval-augmented generation (RAG), sentiment analysis, and code search. 

Technical Innovations

The Qwen3 models are built upon a dense transformer-based architecture with causal attention, enabling them to produce high-fidelity embeddings by extracting hidden states corresponding to specific tokens. The training pipeline incorporates large-scale weak supervision and supervised fine-tuning, ensuring robustness and adaptability across various applications. 

Open-Source Commitment

In line with Alibaba's commitment to fostering open research, the Qwen3-Embedding and Qwen3-Reranker models are released under the Apache 2.0 license. They are accessible on platforms like Hugging Face, GitHub, and ModelScope, providing researchers and developers with the tools to innovate and build upon these models. 

Implications for the AI Community

The introduction of Qwen3-Embedding and Qwen3-Reranker marks a significant advancement in the field of multilingual NLP. By offering high-performance, open-source models capable of handling complex tasks across numerous languages, Alibaba empowers the AI community to develop more inclusive and effective language processing tools.

References:

  1. Qwen GitHub

  2. Qwen3 Embedding Collection : Qwen3-Embedding Collection

  3. Hugging Face Collection

Rime's Arcana TTS Model Elevates Sales by 15% with Personalized Voice AI

 In the evolving landscape of AI-driven customer engagement, Rime's innovative text-to-speech (TTS) model, Arcana, is making significant strides. By enabling the creation of highly personalized and natural-sounding voices, Arcana has demonstrated a remarkable 15% increase in sales for prominent brands such as Domino's and Wingstop. 

Revolutionizing Voice AI with Personalization

Traditional TTS systems often rely on a limited set of pre-recorded voices, lacking the flexibility to cater to diverse customer demographics. Arcana addresses this limitation by allowing users to generate an "infinite" variety of voices based on specific characteristics. By inputting simple text prompts describing desired attributes—such as age, gender, location, and interests—businesses can create voices that resonate more deeply with their target audiences. 

For example, a company can request a voice like "a 30-year-old female from California who is into software," resulting in a unique and relatable voice profile. This level of customization enhances the authenticity of customer interactions, fostering stronger connections and driving engagement.

Technical Advancements Behind Arcana

Arcana's success stems from its multimodal and autoregressive architecture, trained on real conversational data rather than scripted voice actor recordings. This approach enables the model to produce speech that is not only natural-sounding but also contextually appropriate and emotionally nuanced. 

The model's capabilities extend to various speech styles, including whispering and sarcasm, and support for multiple languages. Such versatility ensures that businesses can tailor their communication strategies to diverse markets and customer preferences.

Enterprise Applications and Offerings

Designed for high-volume, business-critical applications, Arcana empowers enterprises to craft unique voice experiences without the need for human agents. For organizations seeking ready-made solutions, Rime offers eight flagship voice profiles, each with distinct characteristics to suit different brand personas. 

Implications for the Future of Customer Engagement

The demonstrated impact of Arcana on sales performance underscores the potential of personalized voice AI in transforming customer engagement strategies. By delivering voices that mirror the diversity and individuality of customers, businesses can create more meaningful and effective interactions.

As AI technology continues to advance, the integration of sophisticated TTS models like Arcana is poised to become a cornerstone of customer-centric marketing and communication efforts.

Mistral AI Releases Codestral Embed – A High‑Performance Model for Scalable Code Retrieval and Semantics

 Mistral AI has introduced Codestral Embed, a powerful code embedding model purpose-built for scalable retrieval and semantic understanding in software development environments. Positioned as a companion to its earlier generative model, Codestral 22B, this release marks a notable advancement in intelligent code search and analysis.


🔍 Why Codestral Embed Matters

  • Semantic Code Retrieval:
    The model transforms snippets and entire files into rich vector representations that capture deep syntax and semantic relationships. This allows developers to search codebases more meaningfully beyond simple text matching.

  • Scalable Performance:
    Designed to work efficiently across large code repositories, Codestral Embed enables fast, accurate code search — ideal for enterprise-grade tools and platforms.

  • Synergy with Codestral Generation:
    Complementing Mistral’s existing code generation model, this pipeline combines retrieval and generation: find the right snippets with Codestral Embed, then synthesize or augment code with Codestral 22B.


⚙️ Technical and Deployment Highlights

  1. Dedicated Embedding Architecture:
    Trained specifically on code, the model learns fine-grained semantic nuances, including API usage patterns, refactoring structures, and cross-library contexts.

  2. Reranking Capabilities:
    Likely enhanced with a reranker head—mirroring embeds + reranker designs popular for academic/state-of-the-art code search systems. This design improves relevance assumptions and developer satisfaction.

  3. Enterprise-Ready APIs:
    Mistral plans to offer easy-to-integrate APIs, enabling organizations to embed the model in IDEs, CI pipelines, and self-hosted code search systems.

  4. Open and Accessible:
    True to Mistral's open-access ethos, expect code, weights, and documentation to be released under permissive terms — fostering community-driven development and integration.


🧰 Use Cases

  • Code Search Tools:
    Improve developer efficiency by enabling intelligent search across entire codebases, identifying functionally similar snippets and patterns.

  • Automated Code Review:
    Find redundant, outdated, or potentially buggy code sections via semantic similarity — rather than just matching strings.

  • Intelligent IDE Assistance:
    Real-time contextual suggestions and refactoring tools powered by deep understanding of project-specific coding patterns.

  • Knowledge Distillation:
    Build searchable "FAQ" repositories with trusted best-practices code combined with Code embed for alignment and retrieval.


📈 Implications for Developers & Teams

  • Efficiency Boost: Semantic embedding accelerates code discovery and repurposing, reducing context-switching and redundant development work.

  • Better Code Quality:
    Context-aware search helps surface anti-patterns, duplicate logic, and outdated practices.

  • Scalability at Scale:
    Designed for enterprise settings, large monorepos, and self-managed environments.

  • Ecosystem Growth:
    Open access means third parties can build plugins, integrate with SIEMs, LSPs, and continue innovating — expanding utility.


✅ Final Takeaway

Codestral Embed is a strategic addition to Mistral’s AI-powered code suite. By unlocking scalable, semantic code search and analysis, it empowers developers and organizations to traverse complex codebases with greater insight and speed. Paired with Codestral 22B, it reflects a complete retrieval-augmented generation pipeline — poised to elevate code intelligence tooling across the industry.

  Anthropic Enhances Claude Code with Support for Remote MCP Servers Anthropic has announced a significant upgrade to Claude Code , enablin...