<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>This volume contains the proceedings of the MMSR '24 Workshop1, a full-day workshop held in conjunction with CIKM, on October 25, 2024, in Boise, Idaho, United States. The purpose of this workshop was to explore the latest advancements, challenges, and applications in the field of multimodal search and recommendations.</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Vamsi Salaka - Head of Visual Search, Amazon Yubin Kim - Head of Engineering at Vody Chirag Shah - Professor, University of Washington</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Search and recommendation systems are nearly ubiquitous and form an integral component of modern enterprises. However, traditional search engines primarily rely on textual queries, supplemented by session and geographical data. In contrast, the advent of large language models (LLMs) like GPT-4o and Gemini has significantly enhanced the potential for multimodal search and recommendations. Multimodal systems create a shared embedding space for text, images, audio, and other modalities, enabling next-generation customer experiences. These advancements lead to more accurate and personalized recommendations, enhancing user satisfaction and engagement.</p>
      </abstract>
    </article-meta>
  </front>
  <body />
  <back>
    <ref-list />
  </back>
</article>