With $13B in fresh funding, Anthropic is safe - at least until OpenAI finishes “reorganizing” Statsig.

HOT TAKE

The Tabular Shift

Chatting with Nikolaos in this week's podcast got me thinking: XGBoost isn’t dead yet - but its days as the tabular data king are numbered.

Are you still using it, or have you switched to models like TabPFN, FT-Transformers, or SAINT?

PRESENTED WITH JFROG

Coming up: MLOps Days London

We’re teaming up with JFrog for the next MLOps Days - this time in London on Thursday, September 18. Expect sharp talks, good conversation, and the usual mix of practitioners sharing what actually works in production.

The lineup is strong: Heiko Hotz (Google), Yuval Fernbach (JFrog), and Michelle Conway (Lloyds Banking Group) are on deck, covering everything from prompt automation for voice assistants to securing ML systems and getting GenAI live in financial services.

Spots are limited, so register now.

HIDDEN GEMS

Curated finds to help you stay ahead

Closing the agentic loop with MCP, showing how autonomous agents iterate, verify, and complete tasks end-to-end across coding, data, and operational workflows.

The Infra Stack Reset highlights 10 key trends reshaping infrastructure spend in 2025, signaling a shift toward leaner, more economical platforms amid tightening budgets.

Text-to-speech model benchmarks from Coval, with a live comparison interface for testing latency and accuracy to help developers refine voice agent performance.

Inside vLLM explores how a high-throughput LLM inference engine is built, covering design, advanced features, scaling strategies, and distributed serving infrastructure.

MLOPS COMMUNITY

Distilling 200+ Hours of NeurIPS: What’s Next for AI

XGBoost’s crown is wobbling. NeurIPS signals the first tabular models that edge it out - not cheap yet, but real - plus a push toward LMs trained on actual databases.

  • Tabular shift: early wins over XGBoost on tables, with cost tradeoffs, and new efforts to build public tabular corpora for pretraining.

  • Agents with brakes: useful for decision support, not execution - reliability comes from external verifiers and knowledge graphs.

  • Frontier still leads: despite small-model buzz, top systems remain ahead; RL-style reasoning continues to improve them.

Bottom line: practical gains arrive when speed meets safeguards.

The Era of AI Agents in Marketing

Agents now run the SDR playbook at machine speed - spotting intent, drafting founder-voice notes, and spinning up account-specific pages before humans step in.

  • Workflow: detect trials or job posts, crawl LinkedIn and public signals, craft multichannel outreach, route replies to a unified inbox, and ping AEs in Slack for handoff.

  • Measurement: UTMs show rising ChatGPT/Gemini referrals; replace brittle MQL/SQL rules with AI-qualified scoring from multi-signal context, plus LLM-assisted negative keyword pruning.

Result: agents handle the grind, humans build trust - response rates climb.

Securing AI Agents: The Future of MCP Authentication & Authorization

When your AI agent can pull sensitive Salesforce or Workday data, security isn’t optional - it’s urgent.

  • Why the naive “one admin key for everything” approach collapses the moment you scale

  • How OAuth 2.1 with token exchange creates user-scoped, auditable access without storing credentials

  • The shift toward Proof-of-Possession tokens to stop stolen keys from being reused

This architecture isn’t just theory - it’s becoming the enterprise standard for secure, scalable AI agent deployments.

IN PERSON EVENTS
MEME OF THE WEEK
Job of the week

Databricks is hiring senior GenAI machine learning engineers to advance its applied AI products, including the Databricks Assistant and AI/BI Genie. The role focuses on improving LLM performance and expanding GenAI capabilities across their platform in 2025.

Responsibilities

  • Develop and deploy LLM-based features across Databricks products

  • Design and optimize ML pipelines for rapid experimentation

  • Collaborate with research and engineering teams on applied AI projects

  • Build scalable backend systems supporting GenAI applications

Requirements

  • 2-8 years in machine learning engineering or academic research

  • Expertise in Python, PyTorch/TensorFlow, and scalable ML systems

  • Experience with LLM fine-tuning and prompt optimization

  • Strong skills in model deployment, testing, and monitoring

Find more roles on our new jobs board - and if you want to post a role, get in touch.

ML CONFESSIONS

Fixed by Proxy

Pushed a config change without double-checking the params and broke part of the pipeline. Next standup, the infra lead asks, ‘Anyone been in that section recently?’ I kept my camera off and stayed quiet. A senior engineer offers to check, spends an hour poking around, and announces it’s fixed. I sat there pretending to take notes, heart rate finally dropping back to normal.

Share your confession here.

Keep reading