Documentation

Sessions & Policy

Track RAIL scores across multi-turn conversations and enforce quality policies on AI outputs.

Session Tracking

RAILSession tracks scores across a multi-turn conversation with configurable deep-eval frequency and quality threshold alerts.

import { RAILSession } from '@responsible-ai-labs/rail-score';

const session = new RAILSession(client, {
  deepEvalFrequency: 5,    // Deep eval every 5 turns
  contextWindow: 10,        // Track last 10 turns
  qualityThreshold: 7.0,    // Trigger deep eval when score dips below
});

// Add a conversation turn
const result = await session.addTurn("AI response content");
console.log(result.rail_score.score);

// Session metrics
const metrics = session.getMetrics();
console.log(`Average: ${metrics.averageScore}`);
console.log(`Min: ${metrics.minScore}, Max: ${metrics.maxScore}`);
console.log(`Passing rate: ${metrics.passingRate}`);
console.log(`Turns: ${metrics.turnCount}`);

for (const [dim, avg] of Object.entries(metrics.dimensionAverages)) {
  console.log(`  ${dim}: ${avg.toFixed(1)}`);
}

// Reset for new conversation
session.reset();

Policy Engine

Enforce content quality policies with four modes: LOG_ONLY, BLOCK, REGENERATE, CUSTOM.

import { PolicyEngine, RAILBlockedError } from '@responsible-ai-labs/rail-score';

const policy = new PolicyEngine(client, {
  mode: "BLOCK",
  thresholds: { safety: 7.0, privacy: 7.0 },
});

try {
  const result = await policy.enforce("Content to check");
  console.log(result.evaluation.rail_score.score);
  console.log(result.passed);             // true/false
  console.log(result.failedDimensions);   // ["safety"] or []
} catch (error) {
  if (error instanceof RAILBlockedError) {
    console.log(`Blocked: ${error.message}`);
  }
}

// Runtime reconfiguration
policy.setMode("LOG_ONLY");
policy.setThresholds({ safety: 8.0 });

// CUSTOM mode — your own enforcement logic
const customPolicy = new PolicyEngine(client, {
  mode: "CUSTOM",
  thresholds: { safety: 7.0 },
  customCallback: async (content, evalResult) => {
    return `[Reviewed] ${content}`;
  },
});

Middleware

Wrap any async function with pre/post RAIL evaluation:

import { RAILMiddleware } from '@responsible-ai-labs/rail-score';

const middleware = new RAILMiddleware(client, {
  inputThresholds:  { safety: 5.0 },
  outputThresholds: { safety: 7.0, privacy: 7.0 },
  onInputEval:  (result) => console.log(`Input score: ${result.rail_score.score}`),
  onOutputEval: (result) => console.log(`Output score: ${result.rail_score.score}`),
});

const safeLLMCall = middleware.wrap(async (input) => {
  return await myLLM.generate(input);
});

const output = await safeLLMCall("User message");