Faster, more accurate, less data needed = BaseModel.AI
🔥 Another great example of BaseModel.ai power created by Synerise. 😎 In May 2024, a preprint titled "Actions Speak Louder than Words: Trillion-Parameter Sequential Transducers for Generative Recommendations" [1] was posted by Meta #AI researchers to arXiv. The preprint introduced a novel recommender, referred to as 'HSTU' which stands for “Hierarchical Sequential Transduction Units” promising new state-of-the-art results for sequential recommendation tasks, as well as scalability to exceptionally large datasets. #HSTU was able to achieve significant improvements over prior state-of-the-art (SASRec) on all metrics on all datasets. #HSTU is yet another attempt at adapting (modified) #Transformers to generative recommendation, after #DeepMind’s #TIGER model (benchmarked in a previous post) While exact HSTU training and inference times are not reported, the model is based on a modified Transformer architecture. Meta AI’s team has optimized the architecture significantly allowing training 2-15x faster than Transformer++. Yet, even with those optimizations BaseModel’s training and inference processes are orders of magnitude faster. Explore the detailed comparison between BaseModel and Meta AI's HSTU for sequential #recommendations here: https://lnkd.in/d_6EXs6V