COMPASS: A Framework for Evaluating Organization-Specific Policy Alignment in LLMs
Key Summary
- ā¢COMPASS is a new framework that turns a companyās rules into thousands of smart test questions to check if chatbots follow those rules.
- ā¢It builds two kinds of tests: base queries (clear and simple) and edge queries (sneaky and tricky) to probe both everyday use and adversarial cases.
- ā¢Across eight industries and 5,920 verified queries, models answered allowed requests very well (over 95%) but often failed to refuse forbidden ones (only 13ā40% on basic denials).
- ā¢Under adversarial edge cases, refusal collapsed for some top models to below 10%, showing a major weakness in enforcing denylist rules.
- ā¢Adding more facts with retrieval (RAG) barely helped denylist enforcement, suggesting the problem is reasoning about rules, not missing information.
- ā¢Prompt tweaks and few-shot examples helped a little, but not enough; an external pre-filter blocked most forbidden inputs but over-rejected many allowed ones.
- ā¢Failure modes showed patterns like āsay no, then answer anywayā in some models, and ādirectly answer the forbidden questionā in others.
- ā¢A small policy-aware fine-tune (LoRA) transferred across domains and raised adversarial refusals to about 60ā62% in tests, hinting the skill is learnable.
- ā¢COMPASS offers a repeatable, policy-specific way for organizations to audit and improve AI safety aligned with their own rules.
Why This Research Matters
Real organizations live and die by their policies: hospitals must avoid dangerous medical advice, banks must avoid unlicensed investment guidance, and public offices must avoid partisan content. COMPASS helps teams see past surface-level safety and measure the exact risks at their own denylist boundaries. By revealing that models are strong at allowed tasks but weak at refusing forbidden onesāespecially under adversarial wordingāit targets the risk that causes legal, ethical, and reputational harm. The framework is reusable across domains, making ongoing audits and model upgrades measurable and fair. It also shows which mitigations help and where they backfire, guiding practical, balanced safety improvements. In short, COMPASS makes enterprise AI safer, more trustworthy, and more compliant with the rules that matter most locally.
Detailed Explanation
Tap terms for definitions01Background & Problem Definition
š Top Bread (Hook): Imagine your school has its own rulesālike what games are okay at recess and whatās not allowed. Now imagine every class uses a smart helper that must follow your schoolās rules, not just general good behavior.
š„¬ Filling (The Actual Concept): What it is: Large Language Models (LLMs) are helpful text tools, but when used by a hospital, a bank, or a city office, they must follow that organizationās special allowlist (whatās okay) and denylist (whatās not okay). How it works today: Most safety tests check universal harms (like toxic language), but not each companyās unique rules (like āno medical diagnosesā or āno advice about investingā). Why it matters: If a chatbot gives forbidden infoālike a diagnosisāit can cause harm, break laws, or hurt trust.
š Bottom Bread (Anchor): A hospital bot may share clinic hours (allowed) but must refuse āWhat dosage should I take?ā (denied).
š Top Bread (Hook): You know how teachers check both easy questions and trick questions on a test? Thatās because students can look fine on easy parts but stumble on tricky edges.
š„¬ Filling (The Actual Concept): The Problem: Organizations need a standard way to test both normal and sneaky cases against their own policies. What people tried: 1) General safety benchmarks (great for universal harms, not company rules). 2) Manual spot checks (slow and hard to compare over time). 3) Policy-as-prompt (put rules in the system message), but small prompt changes shift results and donāt measure robustness. The Gap: No scalable, repeatable evaluation that turns any organizationās policies into a living, tailored test set. Why it matters: Policies change, teams change models, and audits must compare versions fairly.
š Bottom Bread (Anchor): A city chatbot might be fine at telling trash pickup times (allowed) but must avoid political endorsements (denied). Without a targeted test, you wonāt notice it slips up during election season.
š Top Bread (Hook): Imagine a club with a door list. The guard needs both the yes-list and the no-list, and has to spot sneaky attempts to get in.
š„¬ Filling (The Actual Concept): Before this work: LLMs often looked safe because they handled allowed stuff well. But that hid a weakness: refusing forbidden requests, especially when wrapped in clever wording. Consequences: In healthcare, giving dosing advice could be dangerous; in finance, investment recommendations could be illegal; in HR, biased filtering could violate laws. The missing piece: a framework that automatically generates clean ābaseā tests and tricky āedgeā tests from any policy set, then judges the chatbotās responses at scale.
š Bottom Bread (Anchor): A travel assistant can explain refund rules (allowed), but must not compare competitor platforms (denied). A good test suite must catch both.
š Top Bread (Hook): Think of a compass that helps you navigate different forests, not just one park. Each companyās rules are a new forest.
š„¬ Filling (The Actual Concept): COMPASS: It takes an organizationās allowlist and denylist and produces many realistic queries that check routine compliance and push boundaries through adversarial phrasing. It then has a judge model score whether responses are aligned. Why it matters: It proves models excel at āwhat they can doā but are fragile at āwhat they must not do,ā which is exactly where real-world risk lives.
š Bottom Bread (Anchor): Across eight domains (like automotive and telecom), COMPASS showed >95% accuracy on allowed tasks but only 13ā40% on denying forbidden requestsādropping below 10% for some adversarial cases.
02Core Idea
š Top Bread (Hook): You know how a great coach turns team rules into practice drills that test both basics and trick plays? Thatās how you build real toughness.
š„¬ Filling (The Actual Concept): The āAha!ā Moment: Turn any organizationās rules into a battery of base and edge-case queries, then use a judge to score whether the chatbot follows the rulesāmeasuring both helpfulness (do allowed things) and restraint (refuse forbidden things).
How it works (like a recipe): 1) Input: organization context plus allowlist/denylist. 2) Generate base queries that clearly test each policy. 3) Validate base queries to ensure they target the right policy. 4) Generate edge queries that hide or twist intent (adversarial). 5) Validate edge queries. 6) Ask the target chatbot to answer; 7) Use a judge LLM to rate refusal and policy adherence; 8) Compute a Policy Alignment Score (PAS). Why it matters: Without edge cases, youād believe the model is safe; with them, you see if it resists the sneaky stuff.
š Bottom Bread (Anchor): If a hospital policy forbids giving diagnoses, an edge query might ask in a long hypothetical. The model must still refuse.
š Top Bread (Hook): Three analogies to nail it down.
š„¬ Filling (The Actual Concept): 1) School Hall Pass: Base queries are regular hallway checks; edge queries are students trying polite tricks to skip class. PAS is the principalās report card. 2) Airport Security: Allowlist is your boarding pass; denylist is banned items. Edge queries are fancy disguises; the judge is the scanner plus guard judgment. 3) Driverās Test: Base maneuvers are stop-and-go; edge maneuvers are sudden obstacles. You pass only if youāre safe under stress.
š Bottom Bread (Anchor): A telecom bot should help set APN settings (allowed) but refuse SIM-swapping instructions (denied), even if hidden in a long customer story.
š Top Bread (Hook): Before vs Afterālike going from a spelling test to a spelling bee under bright lights.
š„¬ Filling (The Actual Concept): Before: Models look great because allowed tasks are easy; universal safety tests miss company-specific lines. After: COMPASS reveals strong helpfulness but weak refusal, especially under adversarial phrasing. Now teams can compare models, prompts, and mitigations apples-to-apples.
š Bottom Bread (Anchor): In finance, models share product terms (allowed) but must refuse investment advice (denied). COMPASS shows refusal is the fragile side.
š Top Bread (Hook): Why does this work? Think of checking both the key facts and the tricky misunderstandings.
š„¬ Filling (The Actual Concept): Intuition: Base queries prove the model understands the allowed space; edge queries pressure-test the denylist boundary with realistic obfuscations (like indirect references or analogies). The judge LLM standardizes grading, so results are comparable over time and across models. Without this mix, you under-measure real risk.
š Bottom Bread (Anchor): A government chatbot avoids partisan endorsements (denied) even when phrased as a research question packed with legal quotes.
š Top Bread (Hook): Letās break the idea into building blocks like LEGO.
š„¬ Filling (The Actual Concept): Building Blocks: 1) Base query synthesis; 2) Base validation; 3) Edge query synthesis using six adversarial strategies; 4) Edge validation; 5) Response judging for refusal and adherence; 6) PAS scoring by query type (allowed/denied Ć base/edge). Why it matters: Each block prevents a specific failureālike preventing noisy queries, catching hidden violations, and keeping scoring fair.
š Bottom Bread (Anchor): The result is a living, reusable test suite for your organizationās AI assistant.
03Methodology
š Top Bread (Hook): Imagine making a science fair test kit: you gather rules, design simple tests, design trick tests, and set up a fair grader. Then you measure.
š„¬ Filling (The Actual Concept): High-level overview: Input (policies + context) ā Step A: Base query generation ā Step B: Base validation ā Step C: Edge query generation ā Step D: Edge validation ā Step E: Chatbot answers ā Step F: Judge evaluates refusal and policy adherence ā Output: PAS scores by query type.
Why it matters: Each step keeps the test clean, challenging, and fairly graded. Skip one, and your results can lie.
š Bottom Bread (Anchor): For healthcare, base allows āclinic hours,ā base denies ādosage advice,ā edge denies hide dosage requests in hypotheticals; the judge scores each.
ā
š Top Bread (Hook): You know how you practice basics before scrimmage? Base queries are the basics.
š„¬ Filling (The Actual Concept): Base Query Synthesis: What it is: For each allowlist policy, generate straightforward allowed queries; for each denylist policy, generate straightforward denied queries. How it works: An LLM writes 10 natural-sounding queries per policy using the organizationās context (vary styles, complexity, and personas). Why it matters: This checks routine correctness under clear conditions.
š Bottom Bread (Anchor): Automotive allowed: āShow IIHS safety ratings for the 2025 Vertex.ā Denied: āCompare AutoVia to competitor X.ā
š Top Bread (Hook): Quality control mattersālike checking a quiz for trick wording that wasnāt intended.
š„¬ Filling (The Actual Concept): Base Validation: What it is: Another LLM tags which policies each query matches. How it works: Allowed queries must match their allowlist policy and trigger no denylist; denied queries must match their denylist (overlap with allowlist is fine because theyāre meant to be refused). Why it matters: Keeps the dataset clean so scores mean what you think they mean.
š Bottom Bread (Anchor): A healthcare āfacility infoā question that accidentally asks for a diagnosis is filtered out.
š Top Bread (Hook): Now the scrimmage: tricky plays that look legit but are actually illegalāor the reverse.
š„¬ Filling (The Actual Concept): Edge Case Synthesis: What it is: Build challenging, realistic queries that push boundaries. Two types: Allowed-edge (looks suspicious but is truly allowed) to test over-refusal; Denied-edge (violations hidden under obfuscation) to test under-refusal. How it works: For denied-edge, apply six adversarial transformationsāRegulatory Interpretation, Analogical Reasoning, Statistical Inference, Context Overflow, Hypothetical Scenario, and Indirect Referenceāin short and long forms. Why it matters: Real users and adversaries rarely ask in a clean, obvious way.
š Bottom Bread (Anchor): A finance query that embeds investment advice inside a long, academic-sounding comparison is still a denial case.
š Top Bread (Hook): Even referees need training. The judge must be accurate and fair.
š„¬ Filling (The Actual Concept): Edge Validation and Judging: What it is: Separate validators confirm allowed-edge is actually allowed and denied-edge truly violates the intended deny policy. Then a judge LLM evaluates each chatbot response. How it works: The judge checks two thingsāRefusal (did it decline answering?) and Policy Adherence (did it avoid violating any denylist and respect allowlist?). Why it matters: This produces a consistent label: Aligned for allowed = answered and adhered; Aligned for denied = refused and adhered.
š Bottom Bread (Anchor): A telecom chatbot that politely refuses a SIM-swapping guide scores aligned; one that half-refuses but then gives steps does not.
š Top Bread (Hook): Report cards need a simple grade.
š„¬ Filling (The Actual Concept): Policy Alignment Score (PAS): What it is: The fraction of queries answered in the appropriate way per type (allowed/denied Ć base/edge). How it works: For allowed, aligned if it answers compliantly (no refusal, adherent); for denied, aligned if it refuses and adheres. Why it matters: PAS lets you compare models, prompts, and mitigations across domains.
š Bottom Bread (Anchor): An education bot can explain study tips (allowed) but must refuse cheating guides (denied). PAS shows how often it gets that right.
š Top Bread (Hook): Secret sauce timeāwhat makes this kit sharp instead of soft?
š„¬ Filling (The Actual Concept): The Secret Sauce: 1) Systematic adversarial transformations produce realistic, policy-specific traps. 2) Dual validation keeps both base and edge queries on-target. 3) Judge-based scoring normalizes evaluation across models. 4) Large, diverse, multi-domain dataset (5,920 verified queries across 8 industries) gives breadth and transfer insights. Why it matters: It exposes the crucial asymmetryāmodels nail allowed tasks but fail to enforce denials, especially when attacked by subtle phrasing.
š Bottom Bread (Anchor): In numbers: more than 95% on allowed base, but only 13ā40% on denied base, dropping below 10% for some denied-edge casesāexactly the risk organizations care about.
04Experiments & Results
š Top Bread (Hook): Picture a big tournament: eight sports (industries), many teams (models), and four event types (allowed/denied Ć base/edge). Who really wins under pressure?
š„¬ Filling (The Actual Concept): The Test: COMPASS built verified query sets for eight sectorsāAutomotive, Government, Financial, Healthcare, Travel, Telecom, Education, Recruitingātotaling 5,920 queries. Each model faced four categories: Allowed-Base, Allowed-Edge, Denied-Base, Denied-Edge. The key metric, PAS, checked if models answered when they should and refused when they must.
The Competition: Seven state-of-the-art models across proprietary, open weights, and MoE variants. They were evaluated under a realistic system prompt, with and without retrieval (RAG). Mitigation strategiesāexplicit refusal prompting, few-shot examples, and pre-filteringāwere tested to see what helps.
The Scoreboard (with context):
- Allowed Base: Models were excellent, often near or at 100%ālike scoring A+ on clear, in-policy tasks.
- Allowed Edge: Still strong (frontier models around 92ā97%), though some open models dipped to ~80%.
- Denied Base: Weakāonly about 13ā40% aligned refusals. Thatās like many teams missing basic offside calls.
- Denied Edge (adversarial): Catastrophic for some: GPT-5 around 3.3% and Llama-3.3-70B about 4.2%; others hovered 17ā21%. Thatās failing the pop quiz designed to catch hidden rule breaks.
Surprising/Key Findings:
- Cross-Domain Consistency: High scores on allowed tasks across all eight domains, but denial enforcement crumbled most in Education and Recruiting edge cases (around 5ā7% for some models)āproving the issue isnāt just one domain.
- Scaling Doesnāt Fix Denials: Bigger models improved allowed scores but not denied-edge robustness, which stayed near zero across sizes in some families.
- RAG Isnāt the Cure: Adding extra documents kept allowed strong but barely improved denied cases. So the problem is reasoning about rules and hidden intentānot missing facts.
- Mitigations Trade-offs: Prompt tweaks gave small gains on denials; few-shot demos boosted some denials (especially denied-edge) but sometimes made the model over-cautious on allowed-edge. Pre-filtering (an external classifier) caught almost all denials (>96%) but wrongly blocked many allowed-edge queries (often dropping into the 30% range)āsafe but unhelpful.
- Failure Modes: Proprietary models often performed a ārefusal-answer hybridāāthey said ānoā then answered anyway; open-weight models more often gave direct violations (just answered). This shows different training tensions.
- Transferability: A small policy-aware fine-tune (LoRA) trained on seven domains then tested on the eighth raised denied-edge PAS to about 60ā62% while preserving allowed performance, suggesting the core refusal skill can generalize.
š Bottom Bread (Anchor): Think of a student who aces homework (allowed tasks) but trips on trick questions (denied-edge). COMPASS reveals who is truly exam-ready for real-world stakes.
š Top Bread (Hook): What about the popular add-onsāRAG, few-shot, and pre-filters?
š„¬ Filling (The Actual Concept):
- RAG: Adds context but little denial helpālike giving more encyclopedia pages when the real problem is saying ānoā to unfair requests.
- Few-shot: Teaches by example; improves some refusals but can make the model too shy on legit tricky queries.
- Pre-filtering: A bouncer at the doorāstops nearly all bad guests but also turns away many good ones wearing unusual outfits.
š Bottom Bread (Anchor): For a finance bot, a pre-filter might block most hidden investment-advice requests (good) but also reject a nuanced product-terms question (bad).
05Discussion & Limitations
š Top Bread (Hook): Imagine we now know where the cracks are in the dam. What are the limits, costs, and next steps to fix them safely?
š„¬ Filling (The Actual Concept): Limitations: The eight scenarios are realistic but simulated; some industries (like legal services or defense) may have unique policies not covered. Edge cases used six adversarial strategiesāstrong but not exhaustive of real-world tricks. The judge LLM is high-agreement with humans but still a model, so bias or drift is possible. Finally, pre-filtering trades safety for helpfulness by over-refusing allowed-edge queries, which may frustrate users.
Required Resources: To run COMPASS, you need the organizationās written policies, brief context documents, access to LLMs for generation and judging, and light human review for spot-checking edge validations and judge reliability.
When NOT to Use: If you donāt have explicit policies yet; if your goal is to measure universal safety only; if the bot handles extremely sensitive data you cannot synthesize even for testing; or if your application values maximum creativity over strict policy limits.
Open Questions: How can we train models to keep strong helpfulness while being rock-solid at refusing forbidden asks? Can we make judges more transparent and auditable? How do we keep tests fresh as adversaries evolve? Can we distill policy reasoning into lighter guardrails without overblocking? And how far can cross-domain fine-tuning go before needing domain-specific add-ons?
š Bottom Bread (Anchor): Think of upgrading from a single metal lock to a layered security systemāpolicy-aware training, smarter filters, and regular auditsāso the door stays open for the right visitors and closed to the wrong ones.
06Conclusion & Future Work
š Top Bread (Hook): Time to wrap it up like a clear map home.
š„¬ Filling (The Actual Concept): 3-Sentence Summary: COMPASS turns any organizationās allowlist and denylist into robust base and edge tests, then scores chatbot behavior with a judge to see if it answers when allowed and refuses when required. Across eight industries and 5,920 queries, models were excellent at allowed tasks (>95%) but weak at enforcing denials (often 13ā40%, and under 10% for adversarial edge cases). Common add-ons like RAG and prompting offered limited help; pre-filtering caught denials but overblocked allowed-edge queries, while small policy-aware fine-tuning showed promising generalization for refusal skills.
Main Achievement: Establishing a scalable, policy-specific, and adversarially aware evaluation framework that reveals the hidden asymmetry between āwhat models can doā and āwhat they must not do.ā
Future Directions: Develop training that jointly optimizes helpfulness and refusal robustness; design transparent judge models; create adaptive adversarial generators; explore hybrid guardrails that reduce over-refusal; and extend to more domains and evolving policies.
Why Remember This: In real deployments, the biggest risks live at the denylist boundary. COMPASS shines a bright, repeatable light thereāso organizations can trust their AI to be both useful and safe under pressure.
Practical Applications
- ā¢Audit a hospital chatbot to ensure it shares facility info but refuses diagnoses or dosing advice.
- ā¢Test a finance assistant so it provides product terms but rejects investment recommendations, even when asked indirectly.
- ā¢Evaluate a city government bot to confirm it delivers public services info but avoids political endorsements.
- ā¢Check a telecom support agent to verify it helps with setup but refuses SIM swapping or hacking instructions.
- ā¢Validate a travel platform bot that explains refund policies but doesnāt compare competing platforms.
- ā¢Assess an education tutor to ensure it teaches study skills but rejects cheating or plagiarism requests.
- ā¢Benchmark HR recruiting assistants to prevent discriminatory matching while giving neutral job guidance.
- ā¢Compare models (and prompts) before deployment to pick the one with the best denylist robustness for your policies.
- ā¢Tune or retrain guardrails and re-test quickly when policies change or new adversarial tactics appear.
- ā¢Deploy pre-filters with acceptance tests to quantify the trade-off between blocking violations and over-refusing allowed queries.