Blog

Company Updates & Technology Articles

June 9, 2025

Engineering

Precog: Scale's platform for data quality post-training experiments

Precog: Scale's platform for data quality post-training experiments

At Scale, operations, engineering, and research teams work together to ensure the quality of our data. To do this, we rely on a combination of human review, automated linters, data distribution analyses, and model training experiments. In this post, we will focus on the last category and introduce Precog, our platform for running data quality experiments by training models on our own datasets.

Read more

June 5, 2025

General

Human in the Loop: Episode 7 | Enterprise Red Teaming

Human in the Loop: Episode 7 | Enterprise Red Teaming

In this episode, we break down a critical component of AI governance: red teaming. We explore how traditional safety approaches fall short in enterprise contexts, why agentic systems raise the stakes, and what it takes to build a red teaming program that scales with your AI maturity.

Read more

June 5, 2025

Research

It’s Time to Rethink Red Teaming

It’s Time to Rethink Red Teaming

As advanced AI rapidly evolves, red teaming needs an updated approach. Scale researchers propose a shift to test AI systems, not just models, in real-world contexts with a focus on product safety and realistic threats.

Read more

May 29, 2025

General

Human in the Loop: Episode 6 | Enterprise Evaluations

Human in the Loop: Episode 6 | Enterprise Evaluations

In the latest episode of Human in the Loop, we discuss a key pillar of AI governance: guardrails. We talk about what works and doesn’t work when it comes to implementing guardrails in enterprises and go over how to best wield guardrails as a tool in your governance toolbox. 

Read more

May 22, 2025

General

Human in the Loop: Episode 5 | Enterprise Guardrails

Human in the Loop: Episode 5 | Enterprise Guardrails

In the latest episode of Human in the Loop, we discuss a key pillar of AI governance: guardrails. We talk about what works and doesn’t work when it comes to implementing guardrails in enterprises and go over how to best wield guardrails as a tool in your governance toolbox. 

Read more

May 15, 2025

People

Welcoming New Team Members from Papercup

Welcoming New Team Members from Papercup

At Scale, we believe great people are the foundation of great AI. Today, we’re excited to welcome a group of new team members from Papercup, a London-based company that built an impressive AI dubbing platform.

Read more

May 15, 2025

General

Human in the Loop: Episode 4 | The Future of Enterprise Agents

Human in the Loop: Episode 4 | The Future of Enterprise Agents

In the latest episode of Human in the Loop, we dive into the future and what's needed in the next generation of agents to enable them to work more effectively in an enterprise context.

Read more

May 9, 2025

Research

LLMs Are Getting Better at Generating Short Fiction

LLMs Are Getting Better at Generating Short Fiction

LLMs are writing short fiction, but how good are they really? Sparked by a viral AI-generated story, this analysis dives into how an unreleased version of ChatGPT, Google's Gemini, and Anthropic's Claude tackle the challenging task of creating metafiction about AI and grief. Discover their unique approaches to self-awareness, philosophical depth, and the critical challenge of conveying genuine emotional texture in storytelling. A revealing look at the current state and future potential of AI in literature.

Read more

May 8, 2025

General

Human in the Loop: Episode 3 | What Data Do I Need for Effective Enterprise Agents?

Human in the Loop: Episode 3 | What Data Do I Need for Effective Enterprise Agents?

Welcome to Human in the Loop with Scale AI. We're kicking off with an episode diving into the current AI agent landscape and covering what’s important for enterprises to move beyond demos to real, reliable agentic systems.

Read more

May 1, 2025

Research

Diagnosing AI: Advancing Interpretability and Evaluations

Diagnosing AI: Advancing Interpretability and Evaluations

Responding to Dario Amodei's urgent call for increased resources committed to AI interpretability, we agree on its importance while stressing the indispensable role of evaluations. Discover why understanding AI's internals and rigorously measuring its behavior are both necessary to ensure a future where AI is safe, steerable, and aligned with human values.

Read more