Share via

Azure Foundry Content Understanding Returning No Info

Andre E 20 Reputation points
2026-03-26T11:43:14.2866667+00:00

Hi all - i'm having difficulty getting Foundry's Content Understanding to return any real information from API or SDK calls - can verify all the calls are successful and getting a return, it's just not the actual content. Works fine on Doc Intelligence. Here's a summary

Issue: The Content Understanding REST API (/contentunderstanding/analyzers/prebuilt-layout:analyze, API version 2025-11-01) returns empty markdown ("```text\n\n```") with zero pages, zero words, and zero paragraphs for valid PDF documents. The Document Intelligence REST API (/documentintelligence/documentModels/prebuilt-layout:analyze, API version 2024-11-30) on the same resource, same subscription key returns full OCR — 2,780 characters, 2 pages, 401 words with bounding boxes...

Repro steps:

  1. Base64-encode any valid PDF (tested with a 55KB, 2-page document)
  2. POST to /contentunderstanding/analyzers/prebuilt-layout:analyze?api-version=2025-11-01 with body {"inputs": [{"data": "<base64>", "mimeType": "application/pdf"}]}
  3. Poll Operation-Location → status "Succeeded"
  4. Result: contents[0].markdown = "```text\n\n```", contents[0].pages = []
  5. Repeat same PDF via /documentintelligence/documentModels/prebuilt-layout:analyze?api-version=2024-11-30 with raw binary body + Content-Type: application/pdf
  6. Result: analyzeResult.content = 2,780 chars of real OCR text, pages = 2, words = 401

Tested analyzers (all return empty): prebuilt-layout, prebuilt-read, prebuilt-documentSearch

Confirmed via: Official Python SDK (azure-ai-contentunderstanding 1.0.1) — same empty result. Also confirmed model deployments exist (gpt-4.1, gpt-4.1-mini, text-embedding-3-large) and default model mappings are configured (verified via client.get_defaults()).

Impact: Content Understanding prebuilt analyzers are non-functional for document OCR on this resource. We've fallen back to the Document Intelligence API as a workaround

Azure Document Intelligence in Foundry Tools
0 comments No comments

Answer accepted by question author
  1. Q&A Assist
    2026-03-26T11:43:36.2833333+00:00
    AI generated content. This question contains an answer created with AI from Q&A Assist. Learn more

    Content Understanding and Document Intelligence are different services, even if they share similar analyzer names like prebuilt-layout. The behavior described is consistent with Content Understanding’s design and limitations rather than a generic OCR failure.

    Key points from the service behavior and limitations:

    1. Content Understanding document analyzers are grounded and schema-driven
      • Content Understanding “forces grounding – anchoring outputs in the text of the input documents – and will not return answers if they cannot be grounded.” In some cases, “the value of the field may be missing.”
      • Extraction quality is “heavily dependent on the way you name the fields and description of the fields.”
      • This means that, unlike Document Intelligence’s prebuilt-layout, Content Understanding may return very sparse or empty structured output (including markdown) if it cannot confidently ground content according to its internal models and expectations.
    2. Different engine and purpose from Document Intelligence
      • Document Intelligence prebuilt-layout is optimized for general OCR and layout extraction (pages, words, bounding boxes) and will typically return text for any reasonably legible PDF.
      • Content Understanding prebuilt analyzers are designed for higher-level content extraction, RAG, and domain-specific scenarios, not as a drop-in replacement for raw OCR. The documentation positions them for “content ingestion in search and retrieval-augmented generation (RAG) workflows” and intelligent document processing, with strong emphasis on grounding and contextualization.
      • As a result, a document that works fine with Document Intelligence OCR can legitimately produce little or no usable output in Content Understanding if the system cannot confidently ground or structure the content.
    3. Service limits and document behavior apply
      • Content Understanding’s document behavior is governed by the documented service limits and behaviors for different document types. For example, the Azure Content Understanding skill in Azure AI Search calls the same document analyzer and notes that “all documented service behaviors for different document types apply to its output,” and that DOCX vs PDF can produce different results.
      • If the PDF’s layout, resolution, or embedded text quality is marginal for the underlying layout model, Content Understanding may effectively return empty content even though Document Intelligence still returns OCR.
    4. RAG and prebuilt analyzers focus
      • RAG-oriented analyzers like prebuilt-documentSearch are optimized to “extract paragraphs, tables, and figure descriptions” and produce markdown for search and retrieval. They are not guaranteed to mirror Document Intelligence’s low-level OCR output.
      • Domain-specific analyzers (finance, tax, procurement, contracts) similarly focus on extracting structured fields, not raw text.
    5. What this implies for the scenario
      • The described behavior (empty markdown, zero pages/words) on Content Understanding while Document Intelligence returns full OCR is consistent with:
        • Different extraction objectives (grounded, schema- and context-driven vs raw OCR), and
        • Document/field characteristics that do not meet Content Understanding’s grounding and extraction thresholds.
      • The current workaround of continuing to use Document Intelligence for OCR is aligned with the documented positioning: Content Understanding is not documented as a full replacement for all Document Intelligence OCR scenarios.

    Given the current documentation, the practical guidance is:

    • Continue using Document Intelligence prebuilt-layout for general-purpose OCR and layout extraction where full text is required.
    • Use Content Understanding prebuilt analyzers when the goal is:
      • RAG ingestion (markdown paragraphs, tables, figure descriptions) or
      • Domain-specific structured extraction (finance/tax, procurement/contracts) or
      • Agentic workflows that rely on grounded, structured outputs.
    • When designing Content Understanding custom analyzers, pay close attention to field naming and descriptions, since extraction quality depends on them and the system will omit values it cannot confidently ground.

    If the same PDF consistently returns empty content across multiple Content Understanding analyzers and the document is clearly legible and simple, this would be a candidate for a support ticket, but the fallback to Document Intelligence is currently the documented and appropriate approach for robust OCR.


    References:

    0 comments No comments

0 additional answers

Sort by: Most helpful

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.