Skip to content

Test Studio

The Test Studio provides a comprehensive interface for managing test sets, running tests, and analyzing results directly from the web UI.

The Test Studio consists of two main tabs:

  1. Test Sets: Create and manage reusable collections of test documents
  2. Test Executions: Execute tests, view results, and compare test runs

The accelerator automatically deploys four benchmark datasets from HuggingFace as ready-to-use test sets during stack deployment:

  1. RealKIE-FCC-Verified: 75 FCC invoice documents
  2. OmniAI-OCR-Benchmark: 293 diverse document images across 9 formats
  3. DocSplit-Poly-Seq: 500 multi-page packets with 13 document types
  4. Fake-W2-Tax-Forms: 2,000 synthetic US W-2 tax form images with 45-field ground truth

All datasets are deployed automatically with zero manual steps required. Each test set has a corresponding managed configuration version (e.g., fake-w2, docsplit) that is auto-selected in Test Studio when the test set is chosen. See Configuration — Managed Configuration Versions for details.


Source: https://huggingface.co/datasets/amazon-agi/RealKIE-FCC-Verified

This dataset contains 75 invoice documents sourced from the Federal Communications Commission (FCC).

During stack deployment, the system automatically:

  1. Downloads Dataset Metadata from HuggingFace parquet file (75 documents)
  2. Downloads PDFs directly from HuggingFace’s pdfs/ directory
  3. Uploads PDFs to s3://TestSetBucket/realkie-fcc-verified/input/
  4. Extracts Ground Truth from json_response field (already in accelerator format!)
  5. Uploads Baselines to s3://TestSetBucket/realkie-fcc-verified/baseline/
  6. Registers Test Set in DynamoDB with metadata
  • Fully Automatic: Complete deployment during stack creation with zero user effort
  • Direct PDF Downloads: PDFs are downloaded directly from HuggingFace’s repository (no image conversion needed)
  • Complete Ground Truth: Structured invoice attributes (Agency, Advertiser, GrossTotal, PaymentTerms, AgencyCommission, NetAmountDue, LineItems)
  • Benchmark Ready: 75 FCC invoice documents ideal for extraction evaluation

Use with: config_library/unified/realkie-fcc-verified/config.yaml


Source: https://huggingface.co/datasets/getomni-ai/ocr-benchmark

This dataset contains 293 pre-selected document images across 9 diverse document formats, filtered from the OmniAI OCR benchmark dataset.

ClassCountDescription
BANK_CHECK52Bank checks with MICR encoding
COMMERCIAL_LEASE_AGREEMENT52Commercial property leases
CREDIT_CARD_STATEMENT11Account statements with transactions
DELIVERY_NOTE8Shipping/delivery documents
EQUIPMENT_INSPECTION11Inspection reports with checkpoints
GLOSSARY31Alphabetized term lists
PETITION_FORM51Election petition forms
REAL_ESTATE59Real estate transaction data
SHIFT_SCHEDULE18Employee scheduling documents

During stack deployment, the system automatically:

  1. Downloads Metadata from HuggingFace (metadata.jsonl)
  2. Downloads Images for 293 pre-selected image IDs
  3. Converts to PNG and uploads to s3://TestSetBucket/ocr-benchmark/input/
  4. Extracts Ground Truth from true_json_output field
  5. Uploads Baselines to s3://TestSetBucket/ocr-benchmark/baseline/
  6. Registers Test Set in DynamoDB with format distribution metadata
  • Multi-Format: 9 different document types for comprehensive testing
  • Nested Schemas: Complex JSON schemas with nested objects and arrays
  • Pre-Selected: 293 images filtered for formats with >5 samples per schema
  • Deterministic: Same images deployed every time for reproducible benchmarks

Use with: config_library/unified/ocr-benchmark/config.yaml


Both datasets share these deployment characteristics:

  • Fully Automatic: Complete deployment during stack creation with zero user effort

  • Version Control: Dataset version pinned in CloudFormation, updateable via parameter

  • Smart Updates: Skips re-download on stack updates unless version changes

  • Single Public Source: Everything from HuggingFace - fully reproducible anywhere

  • First Deployment: Adds ~5-10 minutes to stack deployment (downloads PDFs and metadata)

  • Stack Updates: Near-instant (skips if version unchanged)

develop

  • Version Updates: Re-downloads and re-processes when DatasetVersion changes
  • First Deployment: Adds ~15-20 minutes to stack deployment (downloads all three datasets)
  • Stack Updates: Near-instant (skips if versions unchanged)
  • Version Updates: Re-downloads and re-processes when DatasetVersion changes =======
  • First Deployment: Adds ~5-10 minutes to stack deployment (downloads PDFs and metadata)
  • Stack Updates: Near-instant (skips if version unchanged)

develop

  • Version Updates: Re-downloads and re-processes when DatasetVersion changes

All test sets are immediately available after stack deployment:

  1. Navigate to Test Executions tab
  2. Select the test set from the Select Test Set dropdown:
    • “RealKIE-FCC-Verified” for invoice extraction testing
    • “OmniAI-OCR-Benchmark” for multi-format document testing
    • “DocSplit-Poly-Seq” for document splitting and classification testing
    • “Fake-W2-Tax-Forms” for W-2 tax form extraction testing
  3. Enter a description in the Context field
  4. Click Run Test to start processing
  5. Monitor progress and view results when complete

RealKIE-FCC-Verified is ideal for:

  • Evaluating extraction accuracy on invoice documents
  • Comparing different model configurations
  • Testing prompt engineering improvements

OmniAI-OCR-Benchmark is ideal for:

  • Testing classification across diverse document types
  • Evaluating extraction on complex nested schemas
  • Benchmarking multi-format document processing pipelines

DocSplit-Poly-Seq is ideal for:

  • Evaluating document splitting and classification accuracy
  • Testing multi-document packet processing capabilities
  • Benchmarking page-level classification across diverse document types
  • Assessing document boundary detection in complex packets

OmniAI-OCR-Benchmark is ideal for:

  • Testing classification across diverse document types
  • Evaluating extraction on complex nested schemas
  • Benchmarking multi-format document processing pipelines

DocSplit Dataset: https://huggingface.co/datasets/amazon/doc_split
Documents Source: https://huggingface.co/datasets/jordyvl/rvl_cdip_n_mp

The DocSplit dataset contains 500 multi-page packet PDFs created by combining pages from 13 different document types. Documents are sourced from the RVL-CDIP-N-MP dataset. Each packet contains multiple subdocuments of different types to test classification and document splitting capabilities.

DocSplit-Poly-Seq (Multi Category Documents Concatenation Sequentially): Creates document packets by first determining a target page count (5-20 pages), then sequentially selecting documents from different categories without repetition. For each selected document, all of its pages are included while preserving the original page ordering, and this process continues until the target page count is reached.

This benchmark simulates the most common real-world scenario where heterogeneous documents are assembled into packets, as observed in medical claims processing where prescription records, laboratory results, and insurance forms are concatenated. The varying document types test models’ ability to detect inter-document boundaries based on content and structural transitions, a fundamental requirement for accurate packet splitting.

The dataset includes 13 document types spanning common business and administrative documents:

  • invoice, email, form, letter, memo, resume
  • budget, news article, scientific publication, specification
  • questionnaire, handwritten, language (non-English documents)
MetricValue
Total Document Packets500
Total Pages7,330
Total Sections2,027
Avg Pages/Packet14.7
Avg Pages/Sections3.62
Avg Sections/Packet4.1
Avg Unique Document Type/Packet3.67

During stack deployment, the system automatically:

  1. Downloads Dataset from HuggingFace (data.tar.gz containing source PDFs)
  2. Creates Packet PDFs by merging pages from source documents based on bundled manifest
  3. Uploads Packets to s3://TestSetBucket/docsplit/input/
  4. Generates Ground Truth with document class and page split information
  5. Uploads Baselines to s3://TestSetBucket/docsplit/baseline/
  6. Registers Test Set in DynamoDB with metadata and document type distribution
  • Multi-Document Packets: Each PDF contains 2-10 distinct documents of different types
  • Splitting Evaluation: Tests ability to correctly split multi-document packets into individual sections
  • Classification Diversity: 13 document types provide comprehensive classification testing
  • Variable Page Counts: Packets range from 5 to 20 pages with varying complexity
  • Ground Truth Included: Complete page-level classification and splitting information

Use with: config_library/unified/rvl-cdip/config.yaml

This test set enables evaluation of:

  • Page-Level Classification: Accuracy of classifying each page to correct document type
  • Document Splitting: Accuracy of identifying document boundaries within packets
  • Split Order: Accuracy of maintaining correct page order within each split section

DocSplit-Poly-Seq is ideal for:

  • Evaluating document splitting and classification accuracy
  • Testing multi-document packet processing capabilities
  • Benchmarking page-level classification across diverse document types
  • Assessing document boundary detection in complex packets

HuggingFace Source: https://huggingface.co/datasets/singhsays/fake-w2-us-tax-form-dataset Original Source: https://www.kaggle.com/datasets/mcvishnu1/fake-w2-us-tax-form-dataset (CC0: Public Domain)

This dataset contains 2,000 synthetically generated US W-2 tax form images with comprehensive structured ground truth. The forms contain fake data (names, IDs, addresses, financial figures) with only real city, state, and zip codes used.

SplitCountDescription
Train1,800Training set images
Test100Test set images
Validation100Validation set images

Each document includes structured ground truth in gt_parse JSON format covering all standard W-2 boxes:

CategoryFieldsExamples
Employer InfoEIN, name, street address, city/state/zipbox_b_employer_identification_number, box_c_employer_name
Employee InfoSSN, name, street address, city/state/zipbox_a_employee_ssn, box_e_employee_name
ControlControl numberbox_d_control_number
Federal WagesWages, SS wages, Medicare wages, SS tips, allocated tipsbox_1_wages, box_3_social_security_wages, box_5_medicare_wages
Federal TaxesFederal tax, SS tax, Medicare taxbox_2_federal_tax_withheld, box_4_social_security_tax_withheld
BenefitsDependent care, nonqualified plansbox_10_dependent_care_benefits, box_11_nonqualified_plans
Codes (12a-d)Code letter + value (4 entries)box_12a_code, box_12a_value
Checkboxes (13)Statutory employee, retirement plan, third-party sick paybox_13_statutary_employee, box_13_retirement_plan
State/Local (×2)State, state ID, state wages, state tax, local wages, local tax, localitybox_15_1_state, box_16_1_state_wages, box_20_1_locality

During stack deployment, the system automatically:

  1. Downloads Parquet Files from HuggingFace (all 3 splits: train, test, validation)
  2. Extracts Images from parquet image column (JPG format, 612×792px)
  3. Uploads Images to s3://TestSetBucket/fake-w2/input/
  4. Converts Ground Truth from gt_parse JSON to accelerator inference_result format
  5. Uploads Baselines to s3://TestSetBucket/fake-w2/baseline/
  6. Registers Test Set in DynamoDB with metadata
  • Comprehensive Ground Truth: 45 structured fields per document covering all W-2 boxes
  • Large Scale: 2,000 documents enable statistically significant benchmarking
  • Synthetic = No PII: Fake data eliminates privacy concerns for testing and sharing
  • Multiple Data Types: Mix of string identifiers (SSN, EIN), monetary values (wages, taxes), codes, and checkboxes
  • Dual State/Local Entries: Each form includes two state/local tax jurisdictions for array extraction testing
  • CC0 License: Public domain — no attribution or redistribution restrictions

Use with: config_library/unified/fake-w2/config.yaml

Fake-W2-Tax-Forms is ideal for:

  • Benchmarking W-2 tax form extraction accuracy at scale
  • Evaluating numeric precision on monetary fields (wages, taxes)
  • Testing structured form data extraction with nested/repeating sections
  • Assessing image quality impact on OCR and extraction accuracy
  • Comparing model performance across 2,000 documents for statistical significance

All datasets share these deployment characteristics: OmniAI-OCR-Benchmark is ideal for:

  • Testing classification across diverse document types
  • Evaluating extraction on complex nested schemas
  • Benchmarking multi-format document processing pipelines

DocSplit-Poly-Seq is ideal for:

  • Evaluating document splitting and classification accuracy
  • Testing multi-document packet processing capabilities
  • Benchmarking page-level classification across diverse document types
  • Assessing document boundary detection in complex packets

All datasets share these deployment characteristics:

  • Location: src/lambda/test_set_resolver/index.py
  • Purpose: Handles GraphQL operations for test set management
  • Features: Creates test sets, scans TestSetBucket for direct uploads, validates file matching, manages test set status
  • Location: src/lambda/test_set_zip_extractor/index.py
  • Purpose: Extracts and validates uploaded zip files
  • Features: S3 event triggered extraction, file validation, status updates
  • Location: src/lambda/test_runner/index.py
  • Purpose: Initiates test runs and queues file processing jobs
  • Features: Test validation, SQS message queuing, fast response optimization
  • Location: src/lambda/test_file_copier/index.py
  • Purpose: Handles asynchronous file copying and processing initiation
  • Features: SQS message processing, file copying, status management
  • Location: src/lambda/test_results_resolver/index.py
  • Purpose: Handles GraphQL queries for test results and comparisons, plus asynchronous cache updates
  • Features:
    • Result retrieval with cached metrics
    • Comparison logic and metrics aggregation
    • Dual event handling (GraphQL + SQS)
    • Asynchronous cache update processing
    • Progress-aware status updates
  • Type: AWS SQS Queue
  • Purpose: Decouples heavy metric calculations from synchronous API calls
  • Features:
    • Encrypted message storage
    • 15-minute visibility timeout for long-running calculations
    • Automatic retry handling
  • Location: src/api/schema.graphql
  • Operations: getTestSets, addTestSet, addTestSetFromUpload, deleteTestSets, getTestRuns, startTestRun, compareTestRuns
  • Location: src/ui/src/components/test-studio/TestStudioLayout.jsx
  • Purpose: Main container with two-tab navigation and global state management
  • Location: src/ui/src/components/test-studio/TestSets.jsx
  • Purpose: Manage test set collections
  • Features: Pattern-based creation, zip upload, direct upload detection, dual polling (3s active, 30s discovery)
  • Location: src/ui/src/components/test-studio/TestExecutions.jsx
  • Purpose: Unified interface combining TestRunner and TestResultsList
  • Features: Test execution, results viewing, comparison, export, delete operations
components/
└── test-studio/
├── TestStudioLayout.jsx
├── TestSets.jsx
├── TestExecutions.jsx
├── TestRunner.jsx
├── TestResultsList.jsx
├── TestResults.jsx
├── TestComparison.jsx
├── TestRunnerStatus.jsx
├── DeleteTestModal.jsx
└── index.js
  1. Pattern-based: Define file patterns (e.g., *.pdf) with bucket type selection
    • Input Bucket: Scan main processing bucket for matching files
    • Test Set Bucket: Scan dedicated test set bucket for matching files
    • Description: Optional description field to document the test set purpose
  2. Zip Upload: Upload zip containing input/ and baseline/ folders
    • Description: Optional description field to document the test set purpose
  3. Direct Upload: Files uploaded directly to TestSetBucket are auto-detected
my-test-set/
├── input/
│ ├── document1.pdf
│ └── document2.pdf
└── baseline/
├── document1.pdf/
│ └── [ground truth files]
└── document2.pdf/
└── [ground truth files]
  • Each input file must have corresponding baseline folder
  • Baseline folder name must match input filename exactly
  • Status: COMPLETED (valid), FAILED (validation errors), PROCESSING (uploading)
  1. UI Zip Upload: S3 event → Lambda extraction → Validation → Status update
  2. Direct S3 Upload: Detected via refresh button or automatic polling
  1. Select test set from dropdown
  2. Optional: Select configuration version to use for processing
  3. Optional: Enter number of files to limit processing (useful for quick testing)
  4. Optional: Add context description for the test run
  5. Click “Run Test” (single test execution only)
  6. Monitor progress via TestRunnerStatus
  7. View results in integrated listing

The Test Studio supports running tests with specific configuration versions:

  • Version Selection: Choose from available configuration versions (e.g., default, Production, v1)
  • Version Tracking: Test results display which configuration version was used
  • Version Comparison: Compare test runs across different configuration versions
  • Context Generation: Test context automatically includes the selected version information

For full details on configuration versioning, see configuration-versions.md.

  • QUEUED: File copying jobs queued in SQS
  • RUNNING: Files being copied and processed
  • COMPLETED: Test finished successfully
  • FAILED: Errors during processing
  • Filter and paginate test runs
  • Multi-select for comparison
  • Navigate to detailed results view
  • Delete and export functionality
  • Reusable collections with file patterns across multiple buckets
  • Dual bucket support (Input Bucket and Test Set Bucket)
  • Optional description field for documenting test set purpose
  • Zip upload with automatic extraction
  • Direct upload detection via dual polling
  • File structure validation with error reporting
  • Single test concurrency prevention
  • Optional file count limiting for quick testing
  • Real-time status monitoring
  • Global state persistence across navigation
  • SQS-based asynchronous processing
  • Comprehensive metrics display including:
    • Test run metadata: Configuration version, duration, context, file counts
    • Overall accuracy and confidence metrics
    • Cost metrics: Total cost and average cost per page
    • Accuracy breakdown (precision, recall, F1-score, false alarm rate, false discovery rate)
    • Field-Level Metrics: Per-field extraction performance table with columns: Field Name, Accuracy, Precision, Recall, TP, FP, TN, FN
    • Average Document Split Classification Metrics:
      • Page Level Accuracy (average across documents)
      • Split Accuracy Without Order (average across documents)
      • Split Accuracy With Order (average across documents)
      • Total Pages, Total Splits (sums across documents)
      • Correctly Classified Pages, Correctly Split counts (sums across documents)
    • Cost breakdown by service and context
  • Side-by-side test comparison with all metrics including configuration versions
  • Export capabilities (JSON/CSV downloads include all metrics)
  • Integrated delete operations

Test Studio uses Stickler’s BulkStructuredModelEvaluator for accurate metric aggregation across multiple documents:

How It Works:

  1. Individual Evaluation: Each document is evaluated with include_confusion_matrix=True to capture detailed field-level metrics
  2. Storage: Raw Stickler comparison results are stored in S3 at {doc_path}/evaluation/results.json under the stickler_comparison_result field
  3. Aggregation: When viewing test results, the system:
    • Scans DynamoDB for all documents in the test run (PK pattern: doc#{test_run_id}*)
    • Loads evaluation results from S3
    • Extracts stickler_comparison_result from each document
    • Uses aggregate_from_comparisons() to compute aggregate metrics
  4. Fallback: Athena-based aggregation remains available for backward compatibility with older data

Benefits:

  • More Accurate: Uses Stickler’s confusion matrix for precise field-level metrics
  • Consistent: Same evaluation engine for single documents and bulk aggregation
  • Efficient: No Athena queries needed for new data
  • Cost Effective: Reduces Athena query costs

Test results include detailed per-field extraction performance metrics displayed in an interactive table:

Displayed Columns:

  1. Field Name: The name of the extracted field
  2. Accuracy: (TP + TN) / (TP + FP + TN + FN) - Overall correctness
  3. Precision: TP / (TP + FP) - Accuracy of positive predictions
  4. Recall: TP / (TP + FN) - Coverage of actual positives
  5. TP (True Positives): Correctly extracted values
  6. FP (False Positives): Incorrectly extracted values
  7. TN (True Negatives): Correctly identified as absent
  8. FN (False Negatives): Missed extractions

Features:

  • Searchable: Filter fields by name to quickly find specific metrics
  • Sortable: Click any column header to sort by that metric
  • Expandable Section: Collapsed by default to keep results view clean
  • Paginated: 10 fields per page for easy navigation
  • Resizable Columns: Adjust column widths as needed

How It Works:

  • Backend stores confusion matrix values (TP, FP, TN, FN) from Stickler aggregation
  • UI calculates Accuracy, Precision, and Recall on-the-fly from these values
  • Metrics displayed with 3 decimal precision (e.g., 0.850)

Use Cases:

  • Identify which fields have low extraction accuracy
  • Compare field-level performance across test runs
  • Prioritize prompt engineering efforts on problematic fields
  • Track improvement in specific fields after configuration changes