Skip to content

Deep dive into LLM-based Recommendation Inferences #36

@udapy

Description

@udapy
  • Literature review
    • how inference have involved, which problem we have solved over years and which problem/bottleneck this adoption has created on the way?
  • Challenges:
  • "eg.": Current LLM recommendation systems suffer from massive inference overhead due to autoregressive token-by-token generation in language space. Each recommendation requires generating complete item descriptions sequentially, creating substantial latency that scales linearly with recommendation list size.

Metadata

Metadata

Assignees

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions