Janusz Naklicki’s Post

Faster, more accurate, less data needed = BaseModel.AI

View profile for Jaroslaw Krolewski

synerise.com | basemodel.ai | cleora.ai | wislakrakow.com | agh.edu.pl

🔥 Another great example of BaseModel.ai power created by Synerise. 😎 In May 2024, a preprint titled "Actions Speak Louder than Words: Trillion-Parameter Sequential Transducers for Generative Recommendations" [1] was posted by Meta #AI researchers to arXiv. The preprint introduced a novel recommender, referred to as 'HSTU' which stands for “Hierarchical Sequential Transduction Units” promising new state-of-the-art results for sequential recommendation tasks, as well as scalability to exceptionally large datasets. #HSTU was able to achieve significant improvements over prior state-of-the-art (SASRec) on all metrics on all datasets. #HSTU is yet another attempt at adapting (modified) #Transformers to generative recommendation, after #DeepMind’s #TIGER model (benchmarked in a previous post) While exact HSTU training and inference times are not reported, the model is based on a modified Transformer architecture. Meta AI’s team has optimized the architecture significantly allowing training 2-15x faster than Transformer++. Yet, even with those optimizations BaseModel’s training and inference processes are orders of magnitude faster. Explore the detailed comparison between BaseModel and Meta AI's HSTU for sequential #recommendations here: https://lnkd.in/d_6EXs6V

BaseModel vs HSTU for sequential recommendations

BaseModel vs HSTU for sequential recommendations

sair.synerise.com

To view or add a comment, sign in

Explore topics