←back to thread

122 points azath92 | 2 comments | | HN request time: 0.4s | source

TLDR: Build a quick HN profile to see how little context LLMs need to personalise your feed. Rate 30 posts once, get a permanent ranked homepage you can return to.

Our goal was to build a tool that allowed us to test a range of "personal contexts" on a very focused everyday use case for us, reading HN!

We are exploring use of personal context with LLMs, specifically what kind of data, how much, and with how much additional effort on the user’s part was needed to get decent results. The test tool was a bit of fun on its own so we re-skinned it and decided to post it here.

First time posting anything on HN but folks at work encouraged me to drop a link. Keen on feedback or other interesting projects thinking about bootstrapping personal context for LLM workflows!

Show context
wickedsight ◴[] No.44455572[source]
It put this post at the top of my feed, which is cool, because it's incredibly relevant to my interests. I used to work on something similar, but way before LLMs were a thing.

Would you be willing to share some more of the architecture/tech stack?

replies(1): >>44455670 #
azath92 ◴[] No.44455670[source]
yeah we see that a bit as well, i promise its not hardcoded in haha.

On the LLM side of things we are using Gemini 2.5 flash, mostly for speed, and found it to be reasonably good quality at a vibe level compared to something heavier like claude 4, probably because we've worked hard to keep the task very simple and explicit. But in saying that there are a bunch of comments on quality that really highlight that if we want to get serious about that we should put in some user feedback loops and evals.

Its all in JS/TS, using vercel ai for the LLM calls, storage is local, but in order to really dig into quality we might start saving things, but to do that well we'd have to add auth/users etc. and we wanted to keep it light for a demo. We have been recently exploring langfuse for tracing, and are really liking that, and will probably look at using them for first pass evals when we get to it for this project.

We also talked quite a bit about non-LLM recsys and aside from time to set up and do well, something I really like is the sense of transparency and agency. you can see your profile, and edit it if you like to see the change in your results. I almost think wed lean further into that rather than folding in some trad DS or recsys stuff even if that might make the results better. Just musings at this point though.

replies(1): >>44462212 #
1. wickedsight ◴[] No.44462212[source]
Thanks for the elaborate response! I recently looked at doing something similar and ended up with the option of using vector embeddings. Is that something you've considered?
replies(1): >>44464973 #
2. azath92 ◴[] No.44464973[source]
For a first pass having some emb search would require we store and process a lot, which we dont do for this light weight demo. all the posts and recommendations are done on the fly, only looking at the current top page and re-ranking. as some others suggest, if we wanted to expand the recs more broadly than the front page then it might be a great way to run a first cut before running our rec llm on it.

For richer data to build a profile its something we look at a bunch for other projects, which could get folded in here if we decide to make it more persistent.