top of page

SearchAI

Semantic search for artworks using multimodal embeddings, enabling intuitive exploration beyond keywords.

Example output

Examples of artworks retrieved through semantic search queries combining visual and textual meaning.

Capture d’écran 2025-08-02 à 16.04.36.png
Capture d’écran 2025-08-02 à 16.04.21.png

What this project illustrates

  • Building end-to-end semantic search systems from data ingestion to UI

  • Leveraging multimodal embeddings to move beyond keyword-based search

  • Designing scalable search architectures for large visual datasets

  • Bridging machine learning, search infrastructure, and user experience

Context & intent

Context

Traditional search engines rely heavily on keywords and metadata, which limits discovery in large visual collections.
In art databases, users often struggle to express what they are looking for using precise terms, especially for abstract or conceptual queries.

This creates friction between user intent and available content.

​

Approach

SearchAI uses multimodal embeddings to represent both artworks and user queries in a shared vector space.
By encoding visual and textual information, the system retrieves artworks based on semantic similarity, rather than exact keyword matches.

Users can search using natural language queries such as moods, styles, or abstract concepts.

​

Value
  • Enable intuitive, natural-language search over visual content

  • Improve discovery in large-scale artwork collections

  • Retrieve conceptually related results, not just literal matches

  • Combine visual and textual understanding in a single search flow

  • Provide a foundation for scalable semantic retrieval systems

Project management & key decisions

  • Choosing a multimodal embedding approach to align text queries and images in the same representation space

  • Favoring semantic similarity search over traditional keyword indexing

  • Designing a cloud-native architecture to support scalability and deployment

  • Structuring the project end-to-end, from data scraping to UI, to reflect real product constraints

Tech stack

  • Frontend: React, Netlify

  • Backend: FastAPI, Python

  • ML: Multimodal embedding model (text + image vectors)

  • Search engine: Vector-based semantic search

  • Infrastructure: Docker, Google Cloud Run, Artifact Registry, Google Cloud Storage

  • Data source: The MET Museum Open Access API

Scientific background

SearchAI builds on research in representation learning and multimodal models, showing that text and images can be embedded into a shared semantic space.
These embeddings make it possible to measure conceptual similarity across modalities and enable retrieval of visually or semantically related content beyond surface-level descriptions.

Demo

👉 Search artworks using natural language queries combining visual and semantic information

bottom of page