LLMs as Retrieval and Recommendation Engines

3 pointsposted 9 hours ago
by pongogogo

2 Comments

taintech

9 hours ago

This is a cool idea, but the cost is a killer. Normally, you would run recommendations and pre-cache them for the users or items they are associated with. Running a giant LLM for every user's recommendation is thousands of times more expensive and slower than current methods. It just doesn't seem practical for a large number of users.

pongogogo

9 hours ago

The post mentions an approach of using a large model to generate labels and then distilling this into a smaller model to lower cost (though it doesn't provide an example)