Showing posts with label Databricks. Show all posts
Showing posts with label Databricks. Show all posts

6.8.25

OpenAI Unveils GPT-OSS: Two Apache-Licensed Open-Weight Models Aimed at Reasoning, Agents, and Real-World Deployment

 OpenAI has released GPT-OSS, a pair of open-weight language models designed for strong reasoning and agentic workflows—gpt-oss-120b and gpt-oss-20b—marking the company’s most significant “open” move since GPT-2. Both models are distributed under Apache 2.0 (with an accompanying GPT-OSS usage policy), positioning them for commercial use, customization, and local deployment. 

What’s in the release

  • Two sizes, one family. The larger gpt-oss-120b targets top-tier reasoning; gpt-oss-20b is a lighter option for edge and on-prem use. OpenAI says 120b achieves near-parity with o4-mini on core reasoning benchmarks, while 20b performs similarly to o3-mini—a notable claim for open-weight models. 

  • Hardware footprint. OpenAI highlights efficient operation for the 120b model (single 80 GB GPU) and 20b running with as little as 16 GB memory in edge scenarios, enabling local inference and rapid iteration without costly infrastructure. 

  • Licensing & model card. The company published a model card and licensing details (Apache 2.0 + usage policy), clarifying intended use, evaluations, and limitations. 

Why this matters

For years, OpenAI prioritized API-only access to frontier systems. GPT-OSS signals a strategic broadening toward open-weight distribution, meeting developers where they build—local, cloud, or hybrid—and competing more directly with leaders like Llama and DeepSeek. Early coverage underscores the shift: outlets note this is OpenAI’s first open-weight release since GPT-2 and frame it as both an ecosystem and competitive move. 

Where you can run it (day one)

OpenAI launched with unusually wide partner support, making GPT-OSS easy to try in existing MLOps stacks:

  • Hugging Face: downloadable weights and a welcome post with implementation details. 

  • AWS SageMaker JumpStart: curated deployment templates for OSS-20B/120B. 

  • Azure AI Foundry & Windows AI Foundry: managed endpoints and tooling for fine-tuning and inference. 

  • Databricks: native availability with 131k-context serving options and enterprise controls. 

  • NVIDIA: performance tuning for GB200 NVL72 systems; NVIDIA cites up to ~1.5M tokens/sec rack-scale throughput for the 120B variant. 

Developer ergonomics: Harmony & agents

OpenAI also published Harmony, a response format and prompt schema that GPT-OSS models are trained to follow. Harmony standardizes conversation structure, reasoning output, and function-calling/tool-use—useful for building agents that require predictable JSON and multi-step plans. If you’re serving via common runtimes (Hugging Face, vLLM, Ollama), the formatting is handled for you; custom servers can adopt the schema from the public repo. 

Safety posture

OpenAI says GPT-OSS went through Preparedness Framework testing, including trials where a maliciously fine-tuned 120B model was evaluated for risky capabilities. The company reports that such variants did not reach high-capability thresholds, presenting a measured step forward in open-model safety practices. 

How it stacks up (early read)

Early reports highlight the significance of the move and the headline performance claims—near-o4-mini for 120B and o3-mini-like results for 20B—alongside the practical win of local, customizable models under a permissive license. Analysts also point out the competitive context: GPT-OSS arrives as open-weight ecosystems (Llama, DeepSeek, Qwen, Kimi) surge in adoption. 

What to build first

  • Agent backends that rely on structured tool use and local policy control (Harmony + Apache 2.0 helps here). 

  • Sovereign/air-gapped deployments in regulated environments using on-prem GPUs or edge hardware, especially with the 20B model. 

  • Cost-sensitive RAG and analytics where fine-tuning and local inference can beat per-token API economics—now supported across major clouds and MLOps platforms.  

The takeaway

GPT-OSS is OpenAI’s clearest embrace of the open-weight ecosystem to date: credible reasoning performance, permissive licensing, broad partner availability, and practical tooling for agents. If your roadmap calls for customizable, locally deployable models with strong reasoning, GPT-OSS belongs on your shortlist—whether you’re targeting laptops, single-GPU servers, or GB200-class scale.

10.6.25

Amperity Launches Chuck Data: A Vibe-Coding AI Agent for Customer Data Engineering

 Amperity Introduces Chuck Data: An AI Agent to Automate Customer Data Engineering with Natural Language

Seattle-based customer data platform (CDP) startup Amperity Inc. has entered the AI agent arena with the launch of Chuck Data, a new autonomous assistant built specifically to tackle customer data engineering tasks. The tool aims to empower data engineers by reducing their reliance on manual coding and enabling natural language-driven workflows, a concept Amperity calls "vibe coding."

Chuck Data is trained on vast volumes of customer information derived from over 400 enterprise brands, giving it a "critical knowledge" base. This foundation enables the agent to perform tasks like identity resolution, PII (Personally Identifiable Information) tagging, and data profiling with minimal developer input.

A Natural Language AI for Complex Data Tasks

Amperity’s platform is well-known for its ability to ingest data from disparate systems — from customer databases to point-of-sale terminals — and reconcile inconsistencies to form a cohesive customer profile. Chuck Data extends this capability by enabling data engineers to communicate using plain English, allowing them to delegate repetitive, error-prone coding tasks to an intelligent assistant.

With direct integration into Databricks environments, Chuck Data leverages native compute resources and large language model (LLM) endpoints to execute complex data engineering workflows. From customer identity stitching to compliance tagging, the agent promises to significantly cut down on time and manual effort.

Identity Resolution at Scale

One of Chuck Data’s standout features is its use of Amperity’s patented Stitch identity resolution algorithm. This powerful tool can combine fragmented customer records to produce unified profiles — a key requirement for enterprises aiming to understand and engage their audiences more effectively.

To promote adoption, Amperity is offering free access to Stitch for up to 1 million customer records. Enterprises with larger datasets can join a research preview program or opt for paid plans with unlimited access, supporting scalable, AI-powered data unification.

PII Tagging and Compliance: A High-Stakes Task

As AI-driven personalization becomes more prevalent, the importance of data compliance continues to grow. Liz Miller, analyst at Constellation Research, emphasized that automating PII tagging is crucial, but accuracy is non-negotiable.

“When PII tagging is not done correctly and compliance standards cannot be verified, it costs the business not just money, but also customer trust,” said Miller.

Chuck Data aims to prevent such issues by automating compliance tasks with high accuracy, minimizing the risk of mishandling sensitive information.

Evolving the Role of the CDP

According to Michael Ni, also from Constellation Research, Chuck Data represents the future of customer data platforms — transforming from static data organizers into intelligent systems embedded within the data infrastructure.

“By running identity resolution and data preparation natively in Databricks, Amperity demonstrates how the next generation of CDPs will shift core governance tasks to the data layer,” said Ni. “This allows the CDP to focus on real-time personalization and business decision-making.”

The End of Manual Data Wrangling?

Derek Slager, CTO and co-founder of Amperity, said the goal of Chuck Data is to eliminate the “repetitive and painful” aspects of customer data engineering.

“Chuck understands your data and helps you get stuff done faster, whether you’re stitching identities or tagging PII,” said Slager. “There’s no orchestration, no UI gymnastics – it’s just fast, contextual, and command-driven.”


With Chuck Data, Amperity is betting big on agentic AI to usher in a new era of intuitive, fast, and compliant customer data management — one where data engineers simply describe what they want, and AI does the rest.

 OpenAI has released GPT-OSS , a pair of open-weight language models designed for strong reasoning and agentic workflows— gpt-oss-120b and ...