Designing Multi-Modal Embedding Fusion-Based Recommender

Anna Wróblewska, Jacek Dabrowski, Michał Pastuszak, Andrzej Michałowski, Michał Daniluk, Barbara Rychalska, Mikołaj Wieczorek, Sylwia Sysko-Romańczuk

Research output: Contribution to journalArticlepeer-review

Abstract

Recommendation systems have lately been popularised globally. However, often they need to be adapted to particular data and the use case. We have developed a machine learning-based recommendation system, which can be easily applied to almost any items and/or actions domain. Contrary to existing recommendation systems, our system supports multiple types of interaction data with various modalities of metadata through a multi-modal fusion of different data representations. We deployed the system into numerous e-commerce stores, e.g., food and beverages, shoes, fashion items, and telecom operators. We present our system and its main algorithms for data representations and multi-modal fusion. We show benchmark results on open datasets that outperform the state-of-the-art prior work. We also demonstrate use cases for different e-commerce sites.

Original languageEnglish
Article number1391
JournalElectronics (Switzerland)
Volume11
Issue number9
DOIs
Publication statusPublished - 1 May 2022
Externally publishedYes

Keywords

  • data fusion
  • data representation
  • deep learning
  • embeddings
  • machine learning
  • multi-modal representation
  • recommendations

Fingerprint

Dive into the research topics of 'Designing Multi-Modal Embedding Fusion-Based Recommender'. Together they form a unique fingerprint.

Cite this