Ebisu: Benchmarking Large Language Models in Japanese Finance
Key Summary
- ā¢EBISU is a new test that checks how well AI models understand Japanese finance, a language and domain where hints and special terms are common.
- ā¢It has two expert-built tasks: JF-ICR for spotting hidden agreement or refusal in investor Q&A, and JF-TE for finding and ranking nested financial terms in disclosures.
- ā¢Japanese is agglutinative and head-final, so key meaning often appears at the end of sentences, making intent and negation hard for AIs trained mostly on English.
- ā¢Across 22 models, even top systems struggled: scale helped a bit, but Japanese- or finance-specific training did not reliably improve results.
- ā¢On JF-ICR, models often mistook polite indirect refusals for weak agreement, showing cultural and pragmatic blind spots.
- ā¢On JF-TE, accuracy dropped when terms were longer, nested, or mixed scripts (kanji, hiragana, katakana), indicating boundary and variant handling issues.
- ā¢Annotation quality was high, with strong agreement between finance-trained Japanese annotators, ensuring reliable ground truth.
- ā¢EBISU reveals that simply making models bigger or feeding them more finance text is not enough; they must learn Japanese-specific morphosyntax and pragmatics.
- ā¢All data and evaluation code are publicly released, inviting progress on linguistically and culturally grounded financial NLP.
- ā¢This matters in real markets because foreign investors hold a large share of Japanese equities and need trustworthy AI tools for Japanese disclosures.
Why This Research Matters
Japanese companies communicate with polite indirectness and complex terms, and global investors increasingly rely on AI to read their disclosures. Misreading a soft refusal as a yes can mislead investment strategies or risk assessments. Extracting the wrong term boundaries can break data pipelines that feed dashboards and compliance tools. EBISU exposes exactly these failure points so researchers can build models that respect Japanese morphosyntax and communication norms. Better models mean fairer, clearer access to Japanese markets for international participants. Ultimately, that helps reduce misunderstanding, improve transparency, and support healthier, data-driven financial ecosystems.
Detailed Explanation
Tap terms for definitions01Background & Problem Definition
š Hook: Imagine trying to understand a classmate who answers questions very politely and indirectly. They rarely say āno,ā but the real answer is hidden at the end of the sentence. Thatās how Japanese business talk often worksāespecially in finance.
š„¬ The Concept (EBISU ā the world before and why we need it):
- What it is: Before this research, AI models were judged mostly with English-style tests or simple question formats that didnāt capture the way Japanese finance uses indirect wording and nested, mixed-script terms.
- How it worked (before):
- Benchmarks favored short, explicit questions and clear labels that are easy to score across many languages.
- Finance tests focused on numbers, tables, or direct Q&A with obvious answers.
- Japanese datasets often measured sentiment or exam-style facts because theyāre cheaper to build.
- Why it matters: Without tests that reflect Japanese language structure and corporate norms, models look smarter than they really are. They might miss an indirect refusal (which could mean āwe wonāt do thatā) or misread a long, nested term, leading to bad decisions.
š Anchor: Think of a model reading, āWe will consider dividends if conditions allowā and thinking itās a yes. In real Japanese investor calls, that may actually be a polite ānot now.ā
š Hook: You know how some puzzles have hidden clues at the very end? In Japanese, the crucial clue often lives at the sentence end.
š„¬ The Concept (The Problem ā language meets finance):
- What it is: Japanese is agglutinative (glues word pieces) and head-final (the ābossā word is at the end). Refusals can be polite and indirect, and financial terms often nest inside each other across different scripts.
- How it works:
- Meaning leans heavily on sentence-final endings that show mood, negation, and commitment.
- Companies use cautious wording (āconsidering,ā āreviewingā) where intent is implied, not stated.
- Financial terms appear as long compounds with nested pieces, mixing kanji, hiragana, and katakana.
- Why it matters: Models trained on English-style explicitness can miss the real stance or the exact term boundaryātwo essentials for finance understanding.
š Anchor: If a CEO says, āWe have no plan to split the stock,ā thatās a clear no. But āWeāll watch market conditions and continue discussionsā might be a soft no that an English-leaning model misreads as yes.
š Hook: Picture a library where books sometimes have titles-within-titles. If you grab only part of the title, you get the wrong book.
š„¬ The Concept (Failed attempts and the gap):
- What it is: Existing multilingual and finance benchmarks missed two tough Japanese needs: (1) reading between the lines for commitment/refusal, and (2) precisely extracting and ranking nested terms in mixed scripts.
- How it works (why they fell short):
- They prefer fixed-choice, short-context tasks for easy scoring across languages.
- They rarely test sentence-final intent signals or indirect refusals.
- They under-test term boundary and nested structure recognition.
- Why it matters: Without these, we canāt trust AI to parse investor Q&A or disclosure notes where the real meaning hides in structure and style.
š Anchor: A model thatās great at English sentiment might still fail to spot that āwe refrain from commentingā signals non-commitment in Japanese finance.
š Hook: Think of a referee who also explains the rules of the game. Thatās what this paper does for AI tests in Japanese finance.
š„¬ The Concept (EBISU fills the gap):
- What it is: EBISU is a benchmark with two expert-annotated tasks that mirror real Japanese finance: JF-ICR (implicit commitment/refusal in Q&A) and JF-TE (extraction and ranking of nested financial terms from disclosures).
- How it works:
- Curates real investor Q&A and EDINET disclosure notes.
- Uses finance-trained native-level annotators with strict guidelines and double-annotation.
- Scores models on intent accuracy (JF-ICR) and term extraction + ranking (JF-TE) with metrics suited to each job.
- Why it matters: Now we can see where models truly struggleāand fix it.
š Anchor: EBISU is like a driving test on real city streets, not just a parking lot. It checks turn signals (intent) and lane lines (term boundaries) the way they actually appear in Tokyo traffic.
02Core Idea
š Hook: You know how a magicianās trick makes you look in the wrong place? In Japanese finance, the important signal often comes at the end of the sentence or is wrapped inside a longer term, so models keep looking in the wrong place.
š„¬ The Concept (The Aha!):
- What it is: The key insight is to test models exactly where Japanese finance hides meaningāat sentence ends for intent and inside nested mixed-script termsāusing two focused tasks with expert ground truth.
- How it works:
- JF-ICR asks: Did the model catch hidden commitment or refusal in high-context Q&A?
- JF-TE asks: Did the model find the longest correct financial terms and rank the nested pieces properly?
- Together, they expose whether models genuinely grasp Japanese morphosyntax and pragmatics, not just vocabulary.
- Why it matters: Without this, models can sound fluent but still misunderstand what companies actually commit to or which exact terms disclosures define.
š Anchor: Itās like testing if a student can read both the main headline and the fine printābecause in Japanese finance, the fine print often carries the punchline.
š Hook (Analogy 1): Imagine listening to a friend who answers in hints. You must catch their last few words to know if itās a yes or a no. š Hook (Analogy 2): Think of Russian nesting dolls. Financial terms can be long dolls with smaller ones inside; you must identify both the big doll and the important smaller ones. š Hook (Analogy 3): Picture three alphabets dancing together (kanji, hiragana, katakana). Picking the right dance partner at the right moment is crucial to get the meaning.
š„¬ The Concept (Before vs After):
- What it is: Before, tests rewarded models for short, explicit, cross-language-friendly answers; after EBISU, tests reward models that handle Japanese-specific intent and term structure.
- How it works:
- Old: QA with obvious choices and simple spans.
- New: Real Q&A with indirect stances; disclosures with nested terms and script variants.
- Scoring checks both the gist (intent) and the exact building blocks (term boundaries and rankings).
- Why it matters: It shifts evaluation from surface-level fluency to culturally and linguistically correct understanding.
š Anchor: A model that once āaced the quizā now has to pass the real interviewāin Japanese corporate style.
š Hook: You know how the answer key in math isnāt just the number, but how you got it? EBISU checks both the conclusion and the language steps.
š„¬ The Concept (Why it worksāintuition without math):
- What it is: The benchmark targets the exact pressure points where Japanese differs from English: sentence-final cues for commitment and nested termhood in mixed scripts.
- How it works:
- Intent labels capture graded commitment from strong yes to strong no.
- Term extraction treats the longest valid span as primary, then ranks inner candidates.
- Metrics match each target: accuracy for intent; F1 for exact maximal spans; HitRate@K for ranking nested terms.
- Why it matters: If a model truly understands Japanese finance, it will do well where meaning actually lives.
š Anchor: EBISU isnāt asking, āDo you know the word?ā Itās asking, āDo you know what was promised, and can you point to the exact term that defines it?ā
š Hook: Imagine building with LEGO pieces where the last brick locks the structure, and smaller bricks hide inside bigger ones.
š„¬ The Concept (Building blocks):
- What it is: EBISU has two blocksāJF-ICR and JF-TEāplus expert annotation and careful metrics.
- How it works (step-by-step):
- Gather real investor Q&A and EDINET note entries.
- Finance-trained annotators apply strict rules and double-check each other.
- Convert to instruction-style prompts so models answer consistently.
- Score intent with accuracy; score terms with F1 (exact maximal spans) and HitRate@K (nested rankings).
- Why it matters: Each block probes a different āhidden meaningā channelāpragmatics for intent and morphology/orthography for terms.
š Anchor: Together, the blocks test whether the model can read both the tone of a managerās answer and the dictionary of a disclosure note, the Japanese way.
03Methodology
š Hook: Imagine grading a cooking contest where dishes come from a cuisine you donāt usually eat. To be fair, you need judges who know that cuisine, and rules that check the flavors that really matter.
š„¬ The Concept (High-level recipe):
- What it is: EBISU turns real Japanese finance text into two testing pipelines.
- How it works (At a high level): Input (Q&A or disclosure note) ā Expert curation and annotation ā Instruction-style prompt ā Model prediction ā Task-specific scoring (JF-ICR Accuracy; JF-TE F1 + HitRate@K) ā Analysis.
- Why it matters: Each step keeps the test close to real Japanese finance, not a simplified classroom exercise.
š Anchor: Itās like using actual restaurant mealsāserved by native chefsāto judge a cooking robot.
ā Step A: Data collection and curation ā š Hook: You know how you pick only the clearest puzzle pieces at first so you donāt waste time? Thatās what curation does here. š„¬ The Concept:
- What it is: Select single-topic, single-turn investor Q&A from big Japanese companies (2023ā2026) and finely scoped EDINET disclosure notes (2025) that define or explain financial concepts.
- How it works:
- Filter out multi-part or context-heavy Q&A to reduce ambiguity (leaving 94 Q&A pairs).
- Extract 202 note-level entries from Annual Securities Reports where definitions and rules live.
- Keep the original Japanese scripts to preserve real-world difficulty.
- Why it matters: Clean inputs make labels trustworthy, which makes scores meaningful. š Anchor: If you test map-reading, you use real maps with the actual symbolsāno cartoon maps.
ā Step B: Expert annotation ā š Hook: Think of two referees independently scoring the same game, and a head referee breaking ties. š„¬ The Concept:
- What it is: Finance-trained, native-level annotators label intent (five levels) and mark both maximal and nested financial term spans; disagreements are adjudicated.
- How it works:
- Clear rules decide what counts as strong/weak commitment or refusal, and what counts as a financial term or nested term.
- Each item gets two independent labels; a senior expert resolves conflicts.
- Agreement is measured (Macro-F1, Cohenās kappa, Krippendorffās alpha) to ensure reliability.
- Why it matters: High agreement shows the ground truth is solid, not guessy. š Anchor: Like two judges giving similar scores in a dance contestāif they often agree, the scoring guide is clear.
ā Step C: Instruction-style prompts ā š Hook: You know how a teacher gives the same worksheet format to every student to keep grading fair? š„¬ The Concept:
- What it is: Convert each example into a consistent instruction format so different models understand the task the same way.
- How it works:
- JF-ICR prompt: Given Question + Company Response, pick exactly one label from {+2, +1, 0, -1, -2}.
- JF-TE prompt: Given a disclosure note, output JSON with each maximal term and a ranked list of nested terms.
- Why it matters: Uniform instructions reduce random format effects and let us compare models apples-to-apples. š Anchor: Everyone gets the same test paper, not one person a text and another a comic.
ā Step D: Evaluation metrics ā š Hook: Imagine three rulers: one checks yes/no answers, one checks exact puzzle-piece shape, and one checks if the right pieces appear in your top guesses. š„¬ The Concept (Multi-class classification):
- What it is: Sorting answers into more than two groups; JF-ICR uses five intent classes.
- How it works:
- The model chooses among {+2, +1, 0, -1, -2}.
- We compute Accuracy: percent of exact matches.
- This spotlights how well the model distinguishes graded commitment and refusal.
- Why it matters: Without multi-class grading, a weak refusal and a strong refusal would get lumped together, hiding important nuance. š Anchor: Like grading from A to F, not just pass/fail.
š Hook: Imagine a tape measure that demands the exact edge of a shape. š„¬ The Concept (F1 for maximal terms):
- What it is: F1 checks whether the model finds the exact longest financial term spans in JF-TE.
- How it works:
- Precision: of the terms you predicted, how many were exactly right.
- Recall: of the true terms, how many you found.
- F1 balances both.
- Why it matters: If the model clips off part of a long term or grabs extra words, F1 catches it. š Anchor: If the gold term is āę½åØę Ŗå¼čŖæę“å¾1ę Ŗå½ććå½ęē“å©ē,ā predicting only āå½ęē“å©ēā is incomplete and should lose points.
š Hook: Think of guessing top songsādid the right song make your top 1, top 5, or top 10? š„¬ The Concept (HitRate@K):
- What it is: For nested terms in JF-TE, HitRate@K checks whether the true terms appear in the modelās top K ranked candidates.
- How it works:
- Rank nested terms under each maximal term from most to least likely.
- Check K = 1, 5, 10 to see how quickly correct terms surface.
- Average over notes.
- Why it matters: Ranking quality matters when many plausible inner terms exist; it rewards surfacing the right ones early. š Anchor: If a note defines several nested pieces, a good model lists the most important ones at the top.
ā Step E: Model evaluation pipeline ā š Hook: Picture a race where all runners start at the same line, on the same track, with the same referee. š„¬ The Concept:
- What it is: A unified harness runs 22 modelsāopen and proprietaryāunder the same settings (temperature 0, same max output length) and hardware rules.
- How it works:
- Use LM Evaluation Harness; serve models via APIs, TogetherAI, or local vLLM on H100 GPUs.
- Standardize generation length (1,024 tokens) and deterministic decoding.
- Collect predictions and compute the metrics.
- Why it matters: Fairness and repeatability let us trust comparisons. š Anchor: If one sprinter used a downhill track and another used flat ground, you couldnāt compare timesāstandardization fixes that.
ā The Secret Sauce ā š Hook: You know how a microscope is powerful because it zooms exactly where the tiny details hide? š„¬ The Concept:
- What it is: EBISUās cleverness is its focus on Japanese-specific meaning zones: sentence-final intent signals and nested, mixed-script terms.
- How it works:
- Intent labels map the real spectrum of Japanese corporate commitment.
- Exact-span scoring forces precise boundary recognition.
- Ranking metrics stress prioritization among many lookalike inner terms.
- Why it matters: These targets are exactly where general scaling and English-focused training are weakestāso EBISU reveals what needs real fixing. š Anchor: Itās the difference between asking, āDo you speak Japanese?ā and āCan you understand how Japanese CFOs politely say ānoā and which exact term in a note changes the accounting?ā
04Experiments & Results
š Hook: Imagine a spelling bee where even champions stumble on words from a language with different scripts and rules. Thatās what happened here.
š„¬ The Concept (The test and why):
- What it is: 22 modelsāopen, closed, Japanese-specific, and finance-specificāwere tested on JF-ICR (intent accuracy) and JF-TE (maximal-term F1 and nested-term HitRate@K).
- How it works:
- JF-ICR: Does the model label answers from strong yes to strong no correctly?
- JF-TE: Does it find the exact longest terms and rank inner terms well (top-1, top-5, top-10)?
- Why it matters: These scores tell us if models truly understand Japanese finance, not just words.
š Anchor: Accuracy is like the test score; F1 is like getting the exact puzzle shape right; HitRate@K checks if your best guesses include the correct pieces.
ā The competition ā
- Open-source general: Llama, Qwen, DeepSeek, Mistral, Kimi.
- Proprietary: GPT-5, GPT-4o, Gemini 3 Flash, Claude Sonnet 4.5.
- Finance-only: FinMA-7B (English-focused finance).
- Japanese-focused: Swallow, Japanese-StableLM, Nekomata; Japanese finance: nekomata-14b-pfn-qfin.
ā The scoreboard (with context) ā
- JF-ICR (Accuracy): Best model Llama-4-Scout-17B at ~0.606 accuracyālike scoring a C when an A is 0.9+; many models stayed near chance on hard cases.
- JF-TE (HitRate highlights): Llama-3.3-70B-Instruct did best but modestly (HR@1 ā 0.128, HR@5 ā 0.366, HR@10 ā 0.511)ālike finding the right treasure in your top 10 guesses half the time.
- Some strong general models (e.g., Llama-3.3-70B-Instruct) outperformed larger or proprietary systems on average EBISU score, showing size and secrecy arenāt enough.
ā Surprising findings ā
- Scale helps a bit but not a lot: Within families, bigger models did better, but gains were limited; Japanese pragmatics still tripped them up.
- Domain/language pretraining isnāt a free win: Japanese-adapted or finance-adapted models didnāt reliably beat general models at the same size; in some cases, continued finance pretraining hurt JF-TE F1.
- Systematic bias in intent: English-centric models tended to over-read commitment, mislabeling indirect refusals as weak yesesāshowing cultural-pragmatic bias.
- Term complexity hurts ranking: Longer, deeper compound terms and mixed scripts reduced HitRate@1 sharplyāmodels struggled with exact boundaries and variant forms more than with missing vocabulary.
- Zeroed metrics on some closed models for JF-TE: A few proprietary systems produced near-zero F1/HitRate for maximal terms under exact-match rules, suggesting difficulty with precise span extraction in this strict setup.
ā What the numbers mean ā
- A top JF-ICR accuracy around 0.61 means models miss roughly 4 in 10 stancesārisky in finance.
- HR@10 near 0.51 on JF-TE means that even with 10 guesses, models include the correct nested terms only about half the time.
- Together, these show todayās models are not yet safe to rely on for Japanese investor Q&A interpretation or meticulous term extraction without human review.
š Anchor: Think of EBISU as a pop quiz where champions canāt just spell; they must understand the joke at the end of the sentence and pick the exact right word inside a longer titleāmany still miss both.
05Discussion & Limitations
š Hook: Even the best athletes have weak spotsāknowing them helps training.
š„¬ The Concept (Limitations):
- What it is: EBISU focuses on two vital tasks but not the entire world of finance NLP; its data sources are limited in companies, years, and formats, and tasks use short spans.
- How it works:
- JF-ICR uses 94 Q&A pairs from 4 companies (2023ā2026), which may not cover all industries or styles.
- JF-TE uses 202 note entries from EDINET 2025 filingsāstrong for definitions, weaker for narratives.
- No official train/test split yet due to size; future releases plan splits.
- Metrics like accuracy and exact-match F1 are strict and may penalize near-misses.
- Why it matters: Results are solid but not the last word; future expansions can broaden coverage and long-context reasoning.
š Anchor: Itās like testing a car on city streets and not highways yet; still useful, but not everything.
š„¬ The Concept (Required resources):
- What it is: Running EBISU fairly needs a unified evaluation harness, GPU resources, and access to APIs or open weights.
- How it works: Deterministic decoding, standard max lengths, and consistent prompts keep the race fair; annotation guidelines and code are public.
- Why it matters: Others can reproduce and extend results.
š Anchor: A shared game rulebook means everyone can replay the match and check the score.
š„¬ The Concept (When not to use):
- What it is: Donāt treat EBISU scores as a green light for high-stakes automation.
- How it works: EBISU measures intent recognition and term extraction, not full financial reasoning, risk forecasting, or compliance interpretation.
- Why it matters: Production use needs extra validation, human oversight, and broader tests.
š Anchor: A great practice test score doesnāt mean you can drive a race car without a coach.
š„¬ The Concept (Open questions):
- What it is: Several paths remain open.
- How it works:
- Can models learn sentence-final Japanese pragmatics via targeted finetuning or architectural tweaks?
- Can character-level or mixed-script tokenization improve term boundary detection?
- Will multi-turn discourse and long-context evaluation change the picture?
- Can culturally-aware instruction data reduce over-commitment bias?
- Why it matters: These are the likely levers to turn modest gains into trustworthy performance.
š Anchor: The next seasonās training plan: practice end-of-sentence reading, nested-term drills, longer plays, and culture-aware coaching.
06Conclusion & Future Work
š Hook: You know how a good teacher writes tests that match the real skills you need? EBISU does that for Japanese finance.
š„¬ The Concept (3-sentence summary):
- What it is: EBISU is a benchmark that tests two things Japanese finance cares about most: reading indirect commitment/refusal (JF-ICR) and pinpointing nested financial terms (JF-TE).
- How it works: Expert-annotated data, instruction-style prompts, and strict metrics (accuracy, F1, HitRate@K) reveal whether models grasp Japanese morphosyntax and pragmatics.
- Why it matters: Across 22 models, even leaders struggled; scaling helped a little, but language/domain pretraining wasnāt a silver bulletāshowing where future progress must focus.
š Anchor: Itās the difference between speaking Japanese and understanding what a CFO really meant.
ā Main achievement ā EBISU cleanly isolates Japanese-specific linguistic and cultural challenges in finance and shows, with reliable data and metrics, that todayās LLMs still fall short on them.
ā Future directions ā
- Train models to attend to sentence-final cues and layered negation.
- Improve mixed-script and compound-aware tokenization for term boundaries.
- Add long-context and multi-turn discourse tasks.
- Build culturally-aware instruction sets to reduce over-commitment bias.
- Provide larger, broader datasets with train/validation/test splits.
ā Why remember this ā Because EBISU changes how we judge AI in Japanese finance: not by fluent-sounding answers, but by whether the model catches a polite ānoā and identifies the exact terms that matter in disclosures. Thatās the kind of understanding markets actually need.
Practical Applications
- ā¢Evaluate vendor LLMs for Japanese investor relations workflows using EBISU scores before procurement.
- ā¢Fine-tune models on sentence-final Japanese pragmatics to reduce over-commitment bias in Q&A analysis.
- ā¢Deploy EBISU as a gating test in data pipelines that auto-summarize Japanese disclosures for global investors.
- ā¢Train term extractors with mixed-script tokenization and boundary-aware objectives informed by JF-TE.
- ā¢Build analyst assistants that flag likely indirect refusals and hedged statements during earnings calls.
- ā¢Augment compliance tools to detect exact finance term spans in notes for precise rule mapping.
- ā¢Create active-learning loops where EBISU-style hard cases are prioritized for human correction and retraining.
- ā¢Benchmark model updates over time with EBISU to track real improvements beyond English-centric metrics.
- ā¢Design curriculum-style prompts that teach models to weigh sentence-final auxiliaries in Japanese.
- ā¢Integrate EBISU checks into RAG systems to ensure retrieved passages align with exact term boundaries.