Paper ID: 2309.01026
Zero-Shot Recommendations with Pre-Trained Large Language Models for Multimodal Nudging
Rachel M. Harrison, Anton Dereventsov, Anton Bibin
We present a method for zero-shot recommendation of multimodal non-stationary content that leverages recent advancements in the field of generative AI. We propose rendering inputs of different modalities as textual descriptions and to utilize pre-trained LLMs to obtain their numerical representations by computing semantic embeddings. Once unified representations of all content items are obtained, the recommendation can be performed by computing an appropriate similarity metric between them without any additional learning. We demonstrate our approach on a synthetic multimodal nudging environment, where the inputs consist of tabular, textual, and visual data.
Submitted: Sep 2, 2023