Close Menu
    Trending
    • Future of Business Analytics in This Evolution of AI | by Advait Dharmadhikari | Jun, 2025
    • You’re Only Three Weeks Away From Reaching International Clients, Partners, and Customers
    • How Brain-Computer Interfaces Are Changing the Game | by Rahul Mishra | Coding Nexus | Jun, 2025
    • How Diverse Leadership Gives You a Big Competitive Advantage
    • Making Sense of Metrics in Recommender Systems | by George Perakis | Jun, 2025
    • AMD Announces New GPUs, Development Platform, Rack Scale Architecture
    • The Hidden Risk That Crashes Startups — Even the Profitable Ones
    • Systematic Hedging Of An Equity Portfolio With Short-Selling Strategies Based On The VIX | by Domenico D’Errico | Jun, 2025
    Finance StarGate
    • Home
    • Artificial Intelligence
    • AI Technology
    • Data Science
    • Machine Learning
    • Finance
    • Passive Income
    Finance StarGate
    Home»AI Technology»Benchmarking OCR APIs on Real-World Documents
    AI Technology

    Benchmarking OCR APIs on Real-World Documents

    FinanceStarGateBy FinanceStarGateMarch 5, 2025No Comments15 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    With the speedy developments in Giant Language Fashions (LLMs) and Imaginative and prescient-Language Fashions (VLMs), many imagine OCR has develop into out of date. If LLMs can “see” and “learn” paperwork, why not use them straight for textual content extraction?

    The reply lies in reliability. Are you able to all the time be a 100% certain of the veracity of textual content output that LLMs interpret from a doc/picture? We put this to check with a easy experiment. We requested colleagues to make use of any LLM of their option to extract an inventory of passenger names (10) from a pattern PDF flight ticket.

    đź’ˇ

    The outcomes have been fairly attention-grabbing – Claude 3 Opus could not learn the PDF in any respect, Claude 3.5 Sonnet missed one passenger’s title, the output from ChatGPT o3-mini was utterly made up (100% hallucination), ChatPDF missed half of the passengers on the checklist.

    Solely NotebookLM and Deepseek received the checklist of names utterly proper!

    Whereas LLMs can interpret and summarize paperwork, they lack the precision and structured output required for essential enterprise purposes the place 100% information accuracy is essential. Moreover, LLMs require important computational sources, making them expensive and impractical for large-scale doc processing, particularly in enterprise and edge deployments.

    OCR, however, is optimized for effectivity, working on low-power gadgets whereas delivering constant outcomes. When accuracy is non-negotiable whether or not in monetary data, authorized contracts, or regulatory compliance, OCR stays essentially the most reliable answer.

    In contrast to LLMs, OCR APIs present confidence scores and bounding bins, permitting builders to detect uncertainties in extracted textual content. This stage of management is essential for companies that can’t afford incorrect or hallucinated information. That’s why OCR APIs proceed to be broadly utilized in doc automation workflows, AI-driven information extraction, and enterprise purposes.

    To evaluate the state of OCR in 2025, we benchmarked 9 of the preferred OCR APIs, masking industrial options, open-source OCR engines, and doc processing frameworks. Our aim is to offer an goal, data-driven comparability that helps builders and enterprises select the most effective instrument for his or her wants.


    Methodology

    Dataset Choice:

    To make sure a complete analysis of OCR APIs or OCR fashions in real-world eventualities, we chosen datasets that embody a various vary of doc varieties and challenges generally encountered in sensible purposes. Our dataset selections embody:

    • Frequent Enterprise Paperwork: Kinds, invoices, and monetary statements containing structured textual content.
    • Receipts: Printed transaction slips with various fonts, noise, and light textual content.
    • Low-Decision Photos: Paperwork captured below suboptimal circumstances, mimicking real-world scanning and pictures limitations.
    • Handwritten Textual content: Samples with totally different handwriting types to check handwriting recognition capabilities.
    • Blurred or Distorted Textual content: Photos with movement blur or compression artifacts to evaluate OCR robustness.
    • Rotated or Skewed Textual content: Paperwork scanned or photographed at an angle, requiring superior textual content alignment dealing with.
    • Tabular Knowledge: Paperwork containing structured tabular info, difficult for OCR fashions to protect format integrity.
    • Dense Textual content: Textual content-heavy paperwork, equivalent to account opening types, to guage efficiency in high-content areas.

    To make sure our benchmark covers all these real-world challenges, we choose the next datasets:

    1. STROIE (link to dataset)
    1. FUNSD (link to dataset)

    These datasets present a complete testbed for evaluating OCR efficiency throughout sensible and actual life eventualities.

    Fashions Choice

    To judge OCR efficiency throughout totally different eventualities, we embody a mixture of industrial APIs, open-source OCR fashions, and doc processing frameworks. This ensures a balanced comparability between proprietary options and freely accessible options. The fashions utilized in our benchmark are:

    • Standard Business OCR APIs:
      • Google Cloud Imaginative and prescient AI
      • Azure AI Doc Intelligence
      • Amazon Textract
    • Standard Open-Supply OCR APIs:
      • Surya
      • PaddleOCR
      • RapidOCR
      • Extractous
    • Standard Open-Supply Doc Processing Frameworks:

    To exhibit how every OCR API processes a picture, we offer code snippets for working OCR utilizing each industrial APIs and open-source frameworks. These examples present methods to load a picture, apply OCR, and extract the textual content, providing a sensible information for implementation and comparability. Beneath are the code snippets for every mannequin:

    1. Google Cloud Imaginative and prescient AI: First step is to arrange a brand new Google Cloud Undertaking. Within the Google Cloud Console, navigate to APIs & Providers → Library, seek for Imaginative and prescient API, and click on Allow. Go to APIs & Providers → Credentials, click on Create Credentials → Service Account, title it (e.g., vision-ocr-service), and click on Create & Proceed. Assign the Proprietor (or Editor) function and click on Finished. Now, in Service Accounts, choose the account, go to Keys → Add Key → Create New Key, select JSON, and obtain the .json file.

    Required Packages:

    pip set up google-cloud-vision
    
    from google.cloud import imaginative and prescient
    from google.oauth2 import service_account
    
    credentials = service_account.Credentials.from_service_account_file("/content material/ocr-nanonets-cea4ddeb1dd2.json") #path to the json file downloaded
    
    shopper = imaginative and prescient.ImageAnnotatorClient(credentials=credentials)
    
    def detect_text(image_path):
       """Detects textual content in a picture utilizing Google Cloud Imaginative and prescient AI."""
       with open(image_path, 'rb') as image_file:
           content material = image_file.learn()
    
       picture = imaginative and prescient.Picture(content material=content material)
       response = shopper.text_detection(picture=picture)
      
       texts = response.text_annotations
    
       if texts:
           return texts[0].description
       else:
           return "No textual content detected."
    
       if response.error.message:
           increase Exception(f"Error: {response.error.message}")
    
    # Substitute along with your picture path
    image_path = "/content material/drive/MyDrive/OCR_datasets/STROIE/test_data/img/X00016469670.jpg"
    print(detect_text(image_path))
    
    1. Azure AI Doc Intelligence: Create an Azure Account (Azure Portal) to get $200 free credit for 30 days. Within the Azure Portal, go to Create a Useful resource, seek for Azure AI Doc Intelligence (Type Recognizer), and click on Create. Select a Subscription, Useful resource Group, Area (nearest to you), set Pricing Tier to Free (if accessible) or Customary, then click on Evaluation + Create → Create. As soon as created, go to the Azure AI Doc Intelligence useful resource, navigate to Keys and Endpoint, and duplicate the API Key and Endpoint.

    Required Packages:

    pip set up azure-ai-documentintelligence
    
    from azure.ai.documentintelligence import DocumentIntelligenceClient
    from azure.core.credentials import AzureKeyCredential
    import io
    # Substitute along with your Azure endpoint and API key
    AZURE_ENDPOINT = "https://your-region.api.cognitive.microsoft.com/"
    AZURE_KEY = "your-api-key"
    
    shopper = DocumentIntelligenceClient(AZURE_ENDPOINT, AzureKeyCredential(AZURE_KEY))
    def extract_text(image_path):
       """Extracts textual content from a picture utilizing Azure AI Doc Intelligence."""
       with open(image_path, "rb") as image_file:
           image_data = image_file.learn()
           
       poller = shopper.begin_analyze_document("prebuilt-read", doc=image_data)
       
       outcome = poller.outcome()
       extracted_text = []
       for web page in outcome.pages:
           for line in web page.traces:
               extracted_text.append(line.content material)
    
       print("Detected textual content:")
       print("n".be a part of(extracted_text))
       
    image_path = image_path
    extract_text(image_path)
    
    
    1. Amazon Textract: Create an AWS Account (AWS Sign-Up) to entry Amazon Textract’s free-tier (1,000 pages/month for 3 months). Within the AWS Administration Console, go to IAM (Id & Entry Administration) → Customers → Create Consumer, title it (e.g., textract-user), and choose Programmatic Entry. Below Permissions, connect AmazonTextractFullAccess and AmazonS3ReadOnlyAccess (if utilizing S3). Click on Create Consumer and duplicate the Entry Key ID and Secret Entry Key.

    Required Packages:

    pip set up boto3
    

    Set Surroundings Variables:

    export AWS_ACCESS_KEY_ID="your-access-key"
    export AWS_SECRET_ACCESS_KEY="your-secret-key"
    export AWS_REGION="your-region"
    
    import boto3
    textract = boto3.shopper("textract", region_name="us-east-1")
    
    def extract_text(image_path):
       """Extracts textual content from a picture utilizing Amazon Textract."""
       with open(image_path, "rb") as image_file:
           image_bytes = image_file.learn()
    
       response = textract.detect_document_text(Doc={"Bytes": image_bytes})
       extracted_text = []
       for merchandise in response["Blocks"]:
           if merchandise["BlockType"] == "LINE":
               extracted_text.append(merchandise["Text"])
       print("Detected textual content:")
       print("n".be a part of(extracted_text))
       
    image_path = image_path
    extract_text(image_path)
    
    
    1. Surya :  Use pip set up surya-ocr to obtain the mandatory packages. Then create a python file with the next code and run it in terminal.
    from PIL import Picture
    from surya.recognition import RecognitionPredictor
    from surya.detection import DetectionPredictor
    
    picture = Picture.open(image_path)
    langs = ["en"]
    
    recognition_predictor = RecognitionPredictor()
    detection_predictor = DetectionPredictor()
    
    predictions = recognition_predictor([image], [langs], detection_predictor)
    
    
    1. PaddleOCR : Use “pip set up paddleocr paddlepaddle” to put in the required packages. Then create a python file with the next code and run it in terminal.
    from paddleocr import PaddleOCR
    
    ocr = PaddleOCR(use_angle_cls=True, lang="en")
    
    outcome = ocr.ocr(image_path, cls=True)
    
    1. RapidOCR : Use “pip set up rapidocr_onnxruntime” to put in the required packages. Then create a python file with the next code and run it in terminal.
    from rapidocr_onnxruntime import RapidOCR
    
    engine = RapidOCR()
    
    img_path = image_path
    outcome, elapse = engine(img_path)
    
    
    1. Extractous: Use “sudo apt set up tesseract-ocr tesseract-ocr-deu” to put in the required packages. Then create a python file with the next code and run it in terminal.
    from extractous import Extractor, TesseractOcrConfig
    extractor = Extractor().set_ocr_config(TesseractOcrConfig().set_language("en"))
    
    outcome, metadata = extractor.extract_file_to_string(image_path)
    
    print(outcome)
    
    1. Marker: Use “pip set up marker-pdf” to put in the required packages. Then in terminal use the next code.
    !marker_single image_path --output_dir saving_directory --output_format json
    
    
    1. Unstructured-IO: Use “pip set up “unstructured[image]”” to put in the required packages. Then create a python file with the next code and run it in terminal.
    from unstructured.partition.auto import partition
    
    components = partition(filename=image_path)
    
    print("nn".be a part of([str(el) for el in elements]))
    

    Analysis Metrics

    To evaluate the effectiveness of every OCR mannequin, we consider each accuracy and efficiency utilizing the next metrics:

    • Character Error Fee (CER): Measures the ratio of incorrect characters (insertions, deletions, and substitutions) to the full characters within the floor reality textual content. Decrease CER signifies higher accuracy.
    • Phrase Error Fee (WER): Much like CER however operates on the phrase stage, calculating errors relative to the full variety of phrases. It helps assess how properly fashions acknowledge full phrases.
    • ROUGE Rating: A textual content similarity metric that compares OCR output with the bottom reality primarily based on overlapping n-grams, capturing each precision and recall.

    For efficiency analysis, we measure:

    • Inference Time (Latency per Picture): The time taken by every mannequin to course of a single picture, indicating pace and effectivity in real-world purposes.

    Value Analysis:

    • For industrial OCR APIs, price is decided by their pricing fashions, sometimes primarily based on the variety of processed pages or photographs. 
    • For open-source OCR APIs, whereas there aren’t any direct utilization prices, we assess computational overhead by measuring reminiscence utilization throughout inference.

    Benchmarking Outcomes

    For the reason that datasets used—STROIE (totally different receipt photographs) and FUNSD (enterprise paperwork with tabular layouts)—comprise various format types, the extracted textual content varies throughout fashions primarily based on their potential to protect construction. This variation impacts the Phrase Error Fee (WER) and Character Error Fee (CER), as these metrics rely upon the place of phrases and characters within the output.

    A excessive error fee signifies {that a} mannequin struggles to keep up the chronological order of textual content, particularly in complicated layouts and tabular codecs.

    1. Phrase Error Fee

    WER of every mannequin on the FUNSD and STROIE datasets is introduced under. These outcomes spotlight how properly every mannequin preserves phrase order throughout totally different doc layouts.

    Phrase Error Fee on the FUNSD dataset
    Phrase Error Fee on the STROIE dataset

    2. Character Error Fee

    CER of every mannequin on the FUNSD and STROIE datasets is introduced under. These outcomes point out how precisely every mannequin captures character-level textual content whereas dealing with totally different doc layouts.

    Character Error Fee on the FUNSD dataset
    Character Error Fee on the STROIE dataset

    Why are the WER and CER metrics for Surya and Marker so excessive on the STROIE dataset?

    STROIE’s intricate layouts make OCR troublesome. Surya tries to fill gaps by inserting further phrases, resulting in excessive WER and CER, even after post-processing. Marker, which makes use of Surya for OCR and outputs markdown textual content, inherits these points. The markdown formatting additional misaligns textual content, worsening the error charges.

    Variation in Mannequin Efficiency Throughout Datasets

    OCR fashions carry out in a different way primarily based on dataset construction. Google Cloud Imaginative and prescient AI and Azure AI Doc Intelligence deal with various layouts higher, whereas open-source fashions like RapidOCR and Surya battle with structured codecs, resulting in extra errors.

    For the reason that fashions battle to protect layouts, resulting in excessive WER and CER, we use one other metric—ROUGE Rating—to evaluate textual content similarity between the mannequin’s output and the bottom reality. In contrast to WER and CER, ROUGE focuses on content material similarity reasonably than phrase place. Which means no matter format preservation, a excessive ROUGE rating signifies that the extracted textual content intently matches the reference, whereas a low rating suggests important content material discrepancies.

    3. ROUGE Rating

    ROUGE Rating of every mannequin on the FUNSD and STROIE datasets is introduced under. These outcomes mirror the content material similarity between the extracted textual content and the bottom reality, no matter format preservation.

    ROUGE Rating on the FUNSD dataset
    ROUGE Rating on the STROIE dataset

    The ROUGE scores reveal that Google Cloud Imaginative and prescient AI persistently outperforms different fashions throughout each FUNSD (75.0%) and STROIE (87.8%), indicating superior textual content extraction. Surya and Marker, which depend on the identical backend, present comparable efficiency, although Marker barely lags on STROIE (70.3%). Extractous and Unstructured-IO rating the bottom in each datasets, suggesting weaker textual content coherence. PaddleOCR and Azure AI Doc Intelligence obtain balanced outcomes, making them aggressive options. The general pattern highlights the energy of economic APIs, whereas open-source fashions exhibit combined efficiency.

    đź’ˇ

    For those who’d prefer to run the fashions your self and compute the analysis scores, you need to use this GitHub repository. The repository consists of scripts for working OCR APIs on the datasets, calculating WER, CER, and ROUGE scores.

    4. Latency per picture

    Latency per picture for every mannequin is introduced under. This measures the time taken by every mannequin to carry out OCR on one picture, offering insights into their effectivity and processing pace.

    Latency per picture or Inference time per picture

    The latency evaluation reveals that Google Cloud Imaginative and prescient AI, Amazon Textract, and Extractous preserve a very good steadiness between pace and accuracy. Surya and Paddle OCR exhibit notably larger inference instances, suggesting potential inefficiencies. Open-source fashions like Speedy OCR and Marker fluctuate in efficiency, with some providing aggressive speeds whereas others lag behind. Azure AI Doc Intelligence additionally reveals average latency, making it a viable alternative relying on the use case.

    5. Value or reminiscence utilization per picture

    For industrial APIs, we current the utilization price (price per 1000 photographs processed). For open-source fashions, the metric signifies reminiscence consumption as a proxy for price, offering insights into their useful resource effectivity.

    OCR API Value per 1,000 Pages
    Google Cloud Imaginative and prescient AI $1.50
    Amazon Textract $1.50
    Azure AI Doc Intelligence $0.50

    Reminiscence utilization per picture

    Amongst open-source fashions, Marker and Unstructured-IO have considerably larger reminiscence consumption, which can impression deployment in resource-constrained environments. Surya and Extractous strike a steadiness between efficiency and reminiscence effectivity. PaddleOCR and RapidOCR are essentially the most light-weight choices, making them preferrred for low-memory eventualities.

    Conclusion

    Primarily based on the analysis throughout latency, inference time, and ROUGE rating, no single mannequin dominates in all points. Nonetheless, some fashions stand out in particular areas:

    • Finest Latency & Inference Time: Extractous and Amazon Textract exhibit the quickest response instances, making them preferrred for real-time purposes.
    • Finest ROUGE Rating (Accuracy): Google Cloud Imaginative and prescient AI and Azure AI Doc Intelligence obtain the very best accuracy in textual content recognition, making them sturdy candidates for purposes requiring exact OCR.
    • Finest Reminiscence Effectivity: RapidOCR and PaddleOCR eat the least reminiscence, making them extremely appropriate for low-resource environments.

    Finest Mannequin Total

    Contemplating a steadiness between accuracy, pace, and effectivity, Google Cloud Imaginative and prescient AI emerges as the most effective total performer. It offers sturdy accuracy with aggressive inference time. Nonetheless, for open sourced fashions, PaddleOCR and RapidOCR provide the most effective trade-off between accuracy, pace and reminiscence effectivity.

    Leaderboard of Finest OCR APIs primarily based on totally different efficiency metrics:

    Metric Finest Mannequin Rating / Worth
    Highest Accuracy (ROUGE Rating) Google Cloud Imaginative and prescient AI Finest ROUGE Rating
    Finest Structure Dealing with (Least WER & CER) Google Cloud Imaginative and prescient AI Lowest WER & CER
    Quickest OCR (Lowest Latency) Extractous Lowest Processing Time
    Reminiscence Environment friendly RapidOCR Least Reminiscence Utilization
    Most Value-Efficient amongst Business APIs Azure AI Doc Intelligence Lowest Value Per Web page

    LLM vs. Devoted OCR: A Case Research

    To grasp how OCR fashions examine to Giant Language Fashions (LLMs) in textual content extraction, we examined a difficult picture utilizing each LLaMa 3.2 11B Vision  and RapidOCR, a small however devoted OCR mannequin.

    Outcomes:

    1. LLaMa 3.2 11B Imaginative and prescient
      • Struggled with faint textual content, failing to reconstruct sure phrases.
      • Misinterpreted some characters and added hallucinated phrases.
      • Took considerably longer to course of the picture.
      • Used quite a lot of compute sources.
    2. RapidOCR
      • Precisely extracted a lot of the textual content regardless of the troublesome circumstances.
      • Ran effectively on very low compute sources.

    Is OCR Nonetheless Related As we speak?

    With the rise of multimodal LLMs able to decoding photographs and textual content, some imagine OCR could develop into out of date. Nonetheless, the fact is extra nuanced.

    For those who or your finish clients should be 100% certain of information you are extracting from paperwork or photographs, OCR nonetheless is your greatest wager for now! Confidence scores and bounding bins from OCR APIs can be utilized to deduce when the output just isn’t dependable.

    With LLMs you’ll be able to by no means be 100% certain of the veracity of the textual content output due to hallucinations and the insecurity scores.

    Who Nonetheless Wants OCR?

    • Enterprises Dealing with Excessive-Quantity Paperwork: Banks, authorized corporations, and insurance coverage corporations depend on OCR for automated doc processing at scale.
    • Governments and Compliance: Passport scanning, tax data, and regulatory filings nonetheless require OCR for structured extraction.
    • AI-Powered Knowledge Pipelines: Many companies combine OCR with NLP pipelines to transform paperwork into structured information earlier than making use of AI fashions.
    • Multilingual and Low-Useful resource Language Functions: OCR stays important for digitizing uncommon scripts the place LLMs lack coaching information.

    Why Ought to Enterprises Nonetheless Care About OCR When Everybody Desires LLMs?

    1. Accuracy and Reliability: LLMs generate hallucinations, whereas OCR ensures exact textual content extraction, making it essential for authorized, monetary, and authorities purposes.
    2. Velocity and Value Effectivity: OCR is light-weight and works on edge gadgets, whereas LLMs require excessive compute sources and cloud inference prices.
    3. The longer term just isn’t OCR vs. LLMs—it’s OCR and LLMs: OCR can extract clear textual content, and LLMs can then course of and interpret it for insights. AI-powered OCR fashions will proceed to enhance, integrating LLM reasoning for higher post-processing.

    Remaining Ideas

    Whereas LLMs have expanded the chances of textual content extraction from photographs, OCR stays indispensable for structured, high-accuracy textual content retrieval and can all the time be essential for dependable doc processing. Somewhat than changing OCR, LLMs will complement it, bringing higher understanding, context, and automation to extracted information.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleAre “AI Safety Experts” Fear-Mongering for their own Cause? | by Andreas Maier | Mar, 2025
    Next Article The Urgent Need for Intrinsic Alignment Technologies for Responsible Agentic AI
    FinanceStarGate

    Related Posts

    AI Technology

    Powering next-gen services with AI in regulated industries 

    June 13, 2025
    AI Technology

    The problem with AI agents

    June 12, 2025
    AI Technology

    Inside Amsterdam’s high-stakes experiment to create fair welfare AI

    June 11, 2025
    Add A Comment

    Comments are closed.

    Top Posts

    You’re Not Too Small for a CRM — Here’s Why It Matters

    April 27, 2025

    Designing Pareto-optimal GenAI workflows with syftr

    May 28, 2025

    SoftBank, ZutaCore and Foxconn Join on Rack-Integrated Solution with Liquid Cooling for NVIDIA H200s

    February 28, 2025

    Reactions to President Trump’s Joint Address to Congress

    March 6, 2025

    Agentic AI 103: Building Multi-Agent Teams

    June 12, 2025
    Categories
    • AI Technology
    • Artificial Intelligence
    • Data Science
    • Finance
    • Machine Learning
    • Passive Income
    Most Popular

    JPMorgan Chase Will Allow Clients to Buy Bitcoin

    May 20, 2025

    LLaMa 4: Running Locally in Under an Hour | by Anthony Maio | Digital Children | Apr, 2025

    April 8, 2025

    Spend Less on Business Travel Forever With This $50 AI-Powered App

    February 11, 2025
    Our Picks

    Unveiling NSA: A Paradigm Shift in Transformer Architecture for Long-Context AI from DeepSeek | by Aria | Feb, 2025

    February 20, 2025

    How To Easily Determine The Right Amount Of Stock Exposure

    April 18, 2025

    Practical SQL Puzzles That Will Level Up Your Skill

    March 4, 2025
    Categories
    • AI Technology
    • Artificial Intelligence
    • Data Science
    • Finance
    • Machine Learning
    • Passive Income
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 Financestargate.com All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.