> For clean Markdown of any page, append .md to the page URL.
> For a complete documentation index, see https://docs.nvidia.com/nemo/curator/llms.txt.
> For full documentation content, see https://docs.nvidia.com/nemo/curator/llms-full.txt.

> Generate image embeddings using built-in embedders for classification, filtering, and similarity search

# Image Embedding

Generate image embeddings for large-scale datasets using NeMo Curator's built-in embedders. Image embeddings enable downstream tasks such as classification, filtering, duplicate removal, and similarity search.

## How It Works

Image embedding in NeMo Curator typically follows these steps:

1. Load your dataset using `FilePartitioningStage` and `ImageReaderStage`
2. Configure the `ImageEmbeddingStage` with CLIP model settings
3. Apply the embedding stage to generate CLIP embeddings for each image
4. Continue with downstream processing stages (filtering, classification, etc.)

The embedding stage integrates seamlessly into NeMo Curator's pipeline architecture.

***

## Available Embedding Tools

<Cards>
  <Card title="ImageEmbeddingStage" href="/curate-images/process-data/embeddings/clip-embedder">
    Generate CLIP embeddings using OpenAI's ViT-L/14 model for high-quality image representations.
  </Card>
</Cards>

***