Extracts, embeds, and stores multimodal PDF elements — text with SentenceTransformers and images with CLIP — in vector database for unified semantic search.Extracts, embeds, and stores multimodal PDF elements — text with SentenceTransformers and images with CLIP — in vector database for unified semantic search.

How to Extract and Embed Text and Images from PDFs for Unified Semantic Search

2025/10/27 12:43

PDFs are rich with both text and visual content — from descriptive paragraphs to illustrations and tables. This example builds an end-to-end flow that parses, embeds, and indexes both, with full traceability to the original page.

In this example, we splits out both text and images, linking them back to page metadata, and enabling unified semantic search. We’ll use CocoIndex to define the flow, SentenceTransformers for text embeddings, and CLIP for image embeddings — all stored in Qdrant for retrieval.

\

🔍 What It Does

This flow automatically:

  • Extracts page texts and images from PDF files
  • Skips tiny or redundant images
  • Creates consistent thumbnails (up to 512×512)
  • Chunks and embeds text with SentenceTransformers (all-MiniLM-L6-v2)
  • Embeds images with CLIP (openai/clip-vit-large-patch14)
  • Stores both embeddings and metadata (e.g., where the text or image came from) in Qdrant, enabling unified semantic search across text and image modalities

📦 Prerequisite:

Run Qdrant

If you don’t have Qdrant running locally, start it via Docker:

docker run -p 6333:6333 -p 6334:6334 qdrant/qdrant

📁 Input Data

We’ll use a few sample PDFs (board game manuals). Download them into the source_files directory:

./fetch_manual_urls.sh

Or, feel free to drop in any of your own PDFs.

⚙️ Run the Flow

Install dependencies:

pip install -e .

Then build your index (sets up tables automatically on first run):

cocoindex update --setup main

Or Run in CocoInsight

cocoindex server -ci main

Define the flow

Flow definition

Let’s break down what happens inside the PdfElementsEmbedding flow

@cocoindex.flow_def(name="PdfElementsEmbedding") def multi_format_indexing_flow( flow_builder: cocoindex.FlowBuilder, data_scope: cocoindex.DataScope ) -> None: data_scope["documents"] = flow_builder.add_source( cocoindex.sources.LocalFile( path="source_files", included_patterns=["*.pdf"], binary=True ) ) text_output = data_scope.add_collector() image_output = data_scope.add_collector()

We defines the flow, added source and added data collectors.

For flow definition, the decorator

@cocoindex.flow_def(name="PdfElementsEmbedding")

\ marks the function as a CocoIndex flow definition, registering it as part of the data indexing system. When executed via CocoIndex runtime, it orchestrates data ingestion, transformation, and collection.When started, CocoIndex’s runtime executes the flow in either one-time update or live update mode, and incrementally embedding only what changed.

Process Each Document

Extract PDF documents

We iterate each document row and run a custom transformation which extract pdf elements.

with data_scope["documents"].row() as doc: doc["pages"] = doc["content"].transform(extract_pdf_elements)

Extract PDF Elements

Define dataclass for structured extraction - we want to extract a list of PdfPage each has page number, text, and list of images.

@dataclass class PdfImage: name: str data: bytes @dataclass class PdfPage: page_number: int text: str images: list[PdfImage]

Next, define a CocoIndex function called extract_pdf_elements that extracts both text and images from a PDF file, returning them as structured, page-wise data objects.

@cocoindex.op.function() def extract_pdf_elements(content: bytes) -> list[PdfPage]: """ Extract texts and images from a PDF file. """ reader = PdfReader(io.BytesIO(content)) result = [] for i, page in enumerate(reader.pages): text = page.extract_text() images = [] for image in page.images: img = image.image if img is None: continue # Skip very small images. if img.width < 16 or img.height < 16: continue thumbnail = io.BytesIO() img.thumbnail(IMG_THUMBNAIL_SIZE) img.save(thumbnail, img.format or "PNG") images.append(PdfImage(name=image.name, data=thumbnail.getvalue())) result.append(PdfPage(page_number=i + 1, text=text, images=images)) return result

The extract_pdf_elements function reads a PDF file from bytes and extracts both text and images from each page in a structured way. Using pypdf, it parses every page to retrieve text content and any embedded images, skipping empty or very small ones to avoid noise.

Each image is resized to a consistent thumbnail size (up to 512×512) and converted into bytes for downstream use. The result is a clean, per-page data structure (PdfPage) that contains the page number, extracted text, and processed images — making it easy to embed and index PDFs for multimodal search.

\

Process Each Page

Once we have the pages, we will process each page

  1. Chunk the text

Takes each PDF page’s text and splits it into smaller, overlapping chunks

with doc["pages"].row() as page: page["chunks"] = page["text"].transform( cocoindex.functions.SplitRecursively( custom_languages=[ cocoindex.functions.CustomLanguageSpec( language_name="text", separators_regex=[ r"\n(\s*\n)+", r"[\.!\?]\s+", r"\n", r"\s+", ], ) ] ), language="text", chunk_size=600, chunk_overlap=100, )

  1. Process each chunk

Embed and collect the metadata we need, each chunk has embedding, original text, and reference the filename and page in where it came from.

with page["chunks"].row() as chunk: chunk["embedding"] = chunk["text"].call(embed_text) text_output.collect( id=cocoindex.GeneratedField.UUID, filename=doc["filename"], page=page["page_number"], text=chunk["text"], embedding=chunk["embedding"], )

  1. Process each image

We use CLIP to embed the image and collects the data, embedding and metadata filename, page_number.

with page["images"].row() as image: image["embedding"] = image["data"].transform(clip_embed_image) image_output.collect( id=cocoindex.GeneratedField.UUID, filename=doc["filename"], page=page["page_number"], image_data=image["data"], embedding=image["embedding"], )

When we collect image outputs, we also want to preserve relevant metadata alongside the embedding. For each image, we not only collect the image embedding and binary data, but also metadata such as the original file name and page number. This association makes it possible to trace embedded images back to their source document and page during retrieval or exploration.

Export to Qdrant

Finally we export the collected info to the target store

text_output.export( "text_embeddings", cocoindex.targets.Qdrant( connection=qdrant_connection, collection_name=QDRANT_COLLECTION_TEXT, ), primary_key_fields=["id"], ) image_output.export( "image_embeddings", cocoindex.targets.Qdrant( connection=qdrant_connection, collection_name=QDRANT_COLLECTION_IMAGE, ), primary_key_fields=["id"], )

🧭 Explore with CocoInsight (Free Beta)

Use CocoInsight to visually trace your data lineage and debug the flow.

It connects locally with zero data retention.

Start your local server:

cocoindex server -ci main

Then open the UI 👉 https://cocoindex.io/cocoinsight

💡 Why This Matters

Traditional document search only scratches the surface — it’s text-only and often brittle across document layouts. This flow gives you multimodal recall, meaning you can:

  • Search PDFs by text or image similarity
  • Retrieve related figures, diagrams, or captions
  • Build next-generation retrieval systems across rich content formats

Compare with ColPali Vision Model (OCR Free)

We also had an example for ColPali

To compare two approaches:

| Aspect | ColPali Multi-Vector Image Grids | Separate Text and Image Embeddings | |----|----|----| | Input | Whole page as an image grid (multi-vector patches) | Text extracted via OCR + images processed separately | | Embedding | Multi-vector patch-level embeddings preserving spatial context | Independent text and image vectors | | Query Matching | Late interaction between text token embeddings and image patches | Separate embedding similarity / fusion | | Document Structure Handling | Maintains layout and visual cues implicitly | Layout structure inferred by heuristics | | OCR Dependence | Minimal to none; model reads text visually | Heavy dependence on OCR (for scanned PDF) and text extraction | | Use Case Strength | Document-heavy, visual-rich formats | General image and text data, simpler layouts | | Complexity | Higher computational cost, more complex storage | Simpler, fewer compute resources needed |

In essence, ColPali excels in deep, integrated vision-text understanding, ideal for visually complex documents, while splitting text and images for separate embeddings is more modular but prone to losing spatial and semantic cohesion. The choice depends on your document complexity, precision needs, and resource constraints.

Support us

⭐ Star CocoIndex on GitHub and share with your community if you find it useful!

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

UK Looks to US to Adopt More Crypto-Friendly Approach

UK Looks to US to Adopt More Crypto-Friendly Approach

The post UK Looks to US to Adopt More Crypto-Friendly Approach appeared on BitcoinEthereumNews.com. The UK and US are reportedly preparing to deepen cooperation on digital assets, with Britain looking to copy the Trump administration’s crypto-friendly stance in a bid to boost innovation.  UK Chancellor Rachel Reeves and US Treasury Secretary Scott Bessent discussed on Tuesday how the two nations could strengthen their coordination on crypto, the Financial Times reported on Tuesday, citing people familiar with the matter.  The discussions also involved representatives from crypto companies, including Coinbase, Circle Internet Group and Ripple, with executives from the Bank of America, Barclays and Citi also attending, according to the report. The agreement was made “last-minute” after crypto advocacy groups urged the UK government on Thursday to adopt a more open stance toward the industry, claiming its cautious approach to the sector has left the country lagging in innovation and policy.  Source: Rachel Reeves Deal to include stablecoins, look to unlock adoption Any deal between the countries is likely to include stablecoins, the Financial Times reported, an area of crypto that US President Donald Trump made a policy priority and in which his family has significant business interests. The Financial Times reported on Monday that UK crypto advocacy groups also slammed the Bank of England’s proposal to limit individual stablecoin holdings to between 10,000 British pounds ($13,650) and 20,000 pounds ($27,300), claiming it would be difficult and expensive to implement. UK banks appear to have slowed adoption too, with around 40% of 2,000 recently surveyed crypto investors saying that their banks had either blocked or delayed a payment to a crypto provider.  Many of these actions have been linked to concerns over volatility, fraud and scams. The UK has made some progress on crypto regulation recently, proposing a framework in May that would see crypto exchanges, dealers, and agents treated similarly to traditional finance firms, with…
Share
BitcoinEthereumNews2025/09/18 02:21
Why is Bitcoin (BTC) Trading Lower Today?

Why is Bitcoin (BTC) Trading Lower Today?

The post Why is Bitcoin (BTC) Trading Lower Today? appeared on BitcoinEthereumNews.com. Bitcoin BTC$90,457.05, the leading cryptocurrency by market value, is down following the overnight Fed rate cut. The reason likely lies in the Fed’s messaging, which has made traders less excited about future easing. The Fed on Wednesday cut the benchmark interest rate by 25 basis points to 3.25% as expected and announced it will begin purchasing short-term Treasury bills to manage liquidity in the banking system. Yet, BTC traded below $90,000 at press time, representing a 2.4% decline since early Asian trading hours, according to CoinDesk data. Ether was down 4% at $3,190, with the CoinDesk 20 Index down over 4%. The risk-off action is likely due to growing signs of internal Fed divisions on balancing inflation control against employment goals, coupled with signals of a more challenging path for future rate cuts. Two members voted for no change on Wednesday, but individual forecasts revealed that six FOMC members felt that a cut wasn’t “appropriate.” Besides, the central bank suggested just one more rate cut in 2026, disappointing expectations for two to three rate cuts. “The Fed is divided, and the market has no real insight into the future path of rates from now until May 2026, when Chairman Jerome Powell will be replaced. The replacement of Powell with a Trump loyalist (who will push to lower rates aggressively) is likely the most reliable signal for rates. Until then, however, there are still 6 months to go,” Greg Magadini, director of derivatives at Amberdata, told CoinDesk. He added that the most likely occurrence as of now is a needed “deleveraging” or down-market” to convince the Fed of lower rates decidedly. Shiliang Tang, managing partner of Monarq Asset Management, said BTC is following the stock market lower. “Crypto markets initially spiked on the news but have steadily moved lower since, in conjunction with…
Share
BitcoinEthereumNews2025/12/11 17:27