I used o3 to profile myself from my saved Pocket links
134 comments
·July 7, 2025saeedesmaili
juliendorra
You should be able to use Google Takeout to get all of your YouTube data, including your watch history.
This article is a nice example of someone using it:
> When I downloaded all my YouTube data, I’ve noticed an interesting file included. That file was named watch-history and it contained a list of all the videos I’ve ever watched.
https://blog.viktomas.com/posts/youtube-usage/
Of course as an European it's a legal obligation for companies to give you access, but I think Google Takeout works worldwide?
jazzyjackson
Yes I've done this in USA. pretty neat. I have it on my todo list to parse over it and find all the music videos I've watched 3 or more times to archive them.
toomuchtodo
https://archive.zhimingwang.org/blog/2014-11-05-list-youtube... might be of use along with https://github.com/yt-dlp/yt-dlp, might just grab it all and prune later due to rot and availability issues over time within YT.
tehlike
You should take this as a sign, and shoot for SWE jobs - given your interest.
What you do at work today doesn't mean you can't switch to a related ladder.
justusthane
Sometimes it’s nice for hobbies to remain hobbies
cortesoft
I believed this, which is what made me avoid computer science in college; I wanted to avoid ruining my favorite hobby.
After a few years post graduation, where I wasn't sure what I wanted to do and I floundered to find a career, I decided to give software development a try, and risk ruining my favorite hobby.
Definitely the best decision I could have made. Now people pay me a lot of money to do the thing I love to do the most... what's not to love? 20 years later, it I still my favorite hobby, and they keep paying me to do it.
formerphotoj
Exactly this. The need to make money from a thing may well eliminate the value one derives from the thing, and even add negatives such as stress, etc.
smt88
I love reading about cooking but I'd hate to become a cook
greenavocado
You need to use an iterative refinement pyramid of prompts. Use a cheap model to condense the majority of the raw data in chunks, then increasingly stronger and more expensive models over increasingly larger sets of those chunks until you are able to reach the level of summarization you desire.
datpuz
Reading 80k tokens requires more than 80k tokens due to overhead
larve
re o3: you can zip the file, upload it, and it will use python and grep and the shell to inspect it. I have yet to try using it with a sqlite db, but that's how i do things locally with agents.
saeedesmaili
Author mentions that by doing that they didn't get a high quality response. Adding the texts into model's context make all the information available for it to use.
tgtweak
I think a reasoning/thinking-heavy model would do better at piecing together the various data points than an agentic model. Would be interested to see how o3 does with the context summarized.
saeedesmaili
Agreed, that's why I used reasoning models (gemini 2.5 pro and opus 4 with extended thinking enabled).
LoganDark
> Both models thought I'm a software engineer.
You probably still are, even if that's not your career path :)
jackdawed
I've noticed a lot of people are converging on this idea of using AI to analyze your own data, the same way the companies do it to your data and serve you super targeted content.
Recently, I was inspired to do this on my entire browsing history, after reading https://labs.rs/en/browsing-histories/ I also did the same from ChatGPT/Claude conversation history. The most terrifying thing I did was having an LLM look at my Reddit comment history.
The challenges are primarily with having a context window large enough and tracking context from various data sources. One approach I am exploring is using a knowledge graph to keep track of a user's profile. You're able to compress behavioral patterns into queryable structures, though the graph construction itself becomes a computational challenge. Recently most of the AI startups I've worked with have just boiled down to "give an LLM access to a vector DB and knowledge graph constructed from a bunch of text documents". The text docs could be invoices, legal docs, tax docs, daily reports, meeting transcripts, code.
I'm hoping we see an AI personal content recommendation or profiling system pop up. The economic incentives are inverted from big tech's model. Instead of optimizing for engagement and ad revenue, these systems are optimized for user utility. During the RSS reader era, I was exposed to a lot of curated tech and design content and it helped me really develop taste and knowledge in these areas. It also helped me connect with cool, interesting people.
There's an app I like https://www.dimensional.me/ but the MBTI and personality testing approach could be more rigorous. Instead of personality testing, imagine if you could feed a system everything you consume, write, and do on digital devices, and construct a knowledge graph about yourself, constantly updating.
nottorp
> Instead of optimizing for engagement and ad revenue, these systems are optimized for user utility.
Are they, or instead they will help keeping you in your comfort cage?
Comfort cage is better than engagement cage ofc, but maybe we should step out of it once in a while.
> During the RSS reader era, I was exposed to a lot of curated tech and design content and it helped me really develop taste and knowledge in these areas.
Curated by humans with which you didn't always agree, right?
janalsncm
> Are they, or instead they will help keeping you in your comfort cage?
I’ve been paying close attention to what YouTube shorts/tiktok do. They don’t just show you the same genre or topic or even set of topics. They are constantly in an explore-exploit pattern. Constantly trying to figure out the next thing that’ll keep your attention, show you a bunch of that content, then on to the next thing. Each interest cluster builds towards a peak then tapers off.
So it’s not like if you see baking videos it’ll keep you in that comfort zone forever.
jackdawed
That's the core challenge in designing a system like this. Echo chambers and comfort cages emerge from recommendation algorithms, and before that, from lazy curation.
If you have control over the recommendation system, you could deliberately feed it contrarian and diverse sources. Or you could choose to be very constrained. Back in RSS days, if you were lazy about it, your taste/knowledge was dependent on other people's curation and biases.
Progress happens through trends anyway. Like in 2010s, there was just a lot of Rails content. Same with flat design. It wasn't really group think, it just seemed to happen out of collective focus and necessity. Everyone else was talking/doing this so if you wanted to be a participant, you have to speak the language.
My original principle when I was using Google Reader was I didn't really know enough to have strong opinions on tech or design, so I'll follow people who seem to have strong opinions. Over time I started to understand what was good design, even if it wasn't something I liked. The rate of taste development was also faster for visual design because you could just quickly scan through an image, vs with code/writing you'd have to read it.
I did something interesting with my Last.fm data once. I've been tracking my music since 2009. Instead of getting recommendations based on my preferences, I could generate a list of artists that had no or little overlap with my current library. It was pure exploration vs exploitation music recommendation. The problem was once your tastes get diverse enough, it's hard to avoid overlaps.
elcapitan
The main thing I learned from my pocket export is that 99% of the articles were "unread". Not sure if it would make sense to extrapolate something about myself other than obsessive link hording from this. :D
gavmor
For many years I've used Pocket to give myself permission to get back to work.
bryancoxwell
Well, read or not you saved those links for a reason
sandspar
Perhaps comparing your read/unread might tell something about your revealed vs stated preferences. I assume that the typical person's unread pile is mostly aspirational. I'm sure that there's lots of data on this - for example Amazon's recommendation graph may weigh our Wishlist items differently than our Purchased items.
elcapitan
I'm sure if you look long enough, you can find any pattern you want, and the opposite ;)
dankwizard
a middle aged white guy using AI, my mind is BLOWN
fudged71
I’ve been really interested in stuff like this recently. Not just Pocket saves but also meta analysis of ChatGPT/Gemini/Claude chat history.
I’ve been using an ultra-personalized RSS summary script and what I’ve discovered is that the RSS feeds that have the most items that are actually relevant to me are very different from what I actually read casually.
What I’m going to try next is to develop a generative “world model” of things that fit in my interests/relevance. And I can update/research different parts of that world model at different timescales. So “news” to me is actually a change diff of that world model from the news. And it would allow me to always have a local/offline version of my current world model, which should be useful for using local models for filtering/sorting things like my inbox/calendar/messages/tweets/etc!
asveikau
As someone with a family background of more left leaning Catholics (which I think are more common in the US northeast), it's interesting that it decided that you are conservative based on Catholicism.
cgriswald
To be fair, it actually said:
> Fiscally conservative / civil-libertarian with traditionalist social leaning
And justified it with:
> Bogleheads & MMM frugality + Catholic/First Things pieces, EFF privacy, skepticism of Big Tech censorship
First Things in its current incarnation is all about religious social conservatism. If someone is Catholic and reads First Things articles, "conservative" is a pretty safe bet.
However, I think profiling people based on what they read might be a mistake in general. I often read things I don't agree with and often seek out things I don't agree with both because I sometimes change my mind and because if I don't change my mind I want to at least know what the arguments actually are. I do wonder, though, if I tended to save such things to pocket.
kixiQu
I have a hypothes.is account where a decent amount of my annotations are little rage nits against the thing I'm reading. You'd be able to infer a ton of correct information from me if you pulled the annotations as well as the URLs, but the URLs alone could mislead.
I've had to remind myself of this pattern with some folks whose bookmarks I follow, because they'd saved some atrocious stuff – but knowing their social media, I know they don't actually believe the theses.
burnte
Born in Pittsburgh, raised Catholic, pretty darn liberal. We had alter girls in the 90s, openly gay members who had ceremonies in the church, etc. I'm not catholic now but that was a good church in the 80s and 90s.
CGMthrowaway
I would say in aggregate, both Catholics and Protestants (whichever flavor) are more likely to be liberal in the northeast / west coast and more likely to be conservative in the midwest / south. Which tells you something about the average importance of religion in 2025.
asveikau
I think it's older than 2025 and definitely has a piece of it that is specific to Catholics. I tend to think of northeastern American Catholicism from the lens of immigration. The big waves of Italians, Irish, Eastern Europeans, etc. The immigrant identity often led to left leaning economics and the parts of Christianity which are about helping the poor get emphasized.
CGMthrowaway
Idk how much experience you have with catholics outside of the northeast. I have a fair amount with all of the regions I mentioned (northeast, south, midwest, west coast). You cannot really find any American Catholic parish that is not dominated by at least one of Italians, Irish, Eastern Europeans or Hispanics. The catholic church in the US is mostly "immigrants," that is, people whose ancestors were not in the US prior to ~1850
KoolKat23
i.e. are you a charitable catholic or a prudish catholic.
nsypteras
A while back I made a little script (for fun/curiosity) that would do this for HN profiles. It’d use their submission and comment history to infer a profile including similar stuff like location, political leaning, career, age, sex, etc. Main motivation was seeing some surprising takes in various comment threads and being curious about where it might have came from. Obviously no idea how accurate the profiles were, but it was similarly an interesting experiment in the ability of LLMs to do this sort of thing.
mywittyname
I remember this. It was pretty accurate for myself, if a little saccharine (i.e., it said I was going to save the world, or some such).
nozzlegear
> Main motivation was seeing some surprising takes in various comment threads and being curious about where it might have came from.
It'd be interesting to run it on yourself, at least, to see how accurate it is.
morkalork
Someone recently did this to predict what would hit the HN front page based on article content + profiles of users.
nsypteras
That's pretty cool! Now I can imagine a tool that gives you a prediction before you even post and then offers suggestions for how to increase performance...
tencentshill
And now we see how easy it is to astroturf any given post, and that's without any budget.
threecheese
There’s no guarantee this didn’t base the results on just 1/3 of the contents of your library though, right? How can it be accurate if it’s not comprehensive, due to the widely noted issues with long context? (distraction, confusion, etc)
This is a gap I see often, and I wonder how people are solving it. I’ve seen strategies like using a “file” tool to keep a checklist of items with looping LLM calls, but haven’t applied anything like this personally.
gavmor
Maybe we need some kind of "node coverage tool" to reassure us that each node or chunk of the embedding context has been attended to.
frou_dh
Another thing one could do with a flat list of hundreds of saved links (if it's being used for "read it later", let's be honest: a dumping ground) is to have AI/NLP classify them all, to make it easy to then delete the stuff you're no longer interested in.
zkmon
What was it doing for those 13 seconds? Is it fetching content for the links? How many links could it fetch in 13 seconds? Maybe it is going by the link URLs only instead of fetching the link content?
noperator
o3 spent that time "thinking" and built the profile using only the URLs/titles, no content fetching.
GMoromisato
If you take the 13 seconds of processing time and multiply by 350 million (the rough population of the US), you get:
~144 years of GPU time.
Obviously, any AI provider can parallelize this and complete it in weeks/days, but it does highlight (for me at least) that LLMs are going to increase the power of large companies. I don't think a startup will be able to afford large-scale profiling systems.
For example, imagine Google creating a profile for every GMail account. It would end up with an invaluable dataset that cannot be easily reproduced by a competitor, even if they had all the data.
[But, of course, feel free to correct my math and assumptions.]
smokel
What will they find out? That we are humans?
Alifatisk
I did something similar, but for groupchats. You had to export a groupchat conversation into text and send it to the program. The program would then use a local llm to profile each user in the groupchat based on what they said.
Like, it built knowledge of what every user in the groupchat and noted their thought on different things or what their opinions were on something or just basic knowledge of how they are. You could also ask the llm questions about each user.
It's not perfect, sometimes the inference gets something wrong or the less precise embeddings gets picked up which creates hallucinations or just nonsense, but it works somewhat!
I would love to improve on this or hear if anyone else has done something similar
AJ007
There are other good use cases here like documenting recurring bugs or problems in software/projects.
This is a good illustration of why e2e encryption is more important than its ever been. What were innocuous and boring conversations are now very valuable when combined with phishing and voice cloning.
OpenAI is going to use all of your ChatGPT history to target ads to you, and probably will have to choice to pay for everything. Meta is trying really hard too, and already is applying generative AI extensive for advertiser's creative production.
Ultra targeted advertising where the message is crafted to perfectly fit the viewer mean devices running operating systems incapable of 100% blocking ads should be considered malware. Hopefully local LLMs will be able to do a good job with that.
After reading this I realized I also have an archive of my pocket account (4200 items), so tried the same prompt with o3, gemini 2.5 pro, and opus 4:
- chatgpt UI didn't allow me to submit the input, saying it's too large. Although it was around 80k tokens, less than o3's 200k context size.
- gemini 2.5 pro: worked fine for personality and interest related parts of the profile, but it failed the age range, job role, location, parental status with incorrect perdictions.
- opus 4: nailed it and did a more impressive job, accurately predicted my base city (amsterdam), age range, relationship status, but didn't include anything about if I'm a parent or not.
Both gemini and opus failed in predicting my role, probably understandably. Although I'm a data scientist, I read a lot about software engineering practices because I like writing software and since I don't have the opportunity at work to do this kind of work, I code for personal projects, so I need to learn a lot about system design, etc. Both models thought I'm a software engineer.
Overall it was a nice experiment. Something I noticed is both models mentioned photography as my main hobby, but if they had access to my youtube watch history, they'd confidently say it's tennis. For topics and interests that we usually watch videos rather than reading articles about, would be interesting to combine the youtube watch history with this pocket archive data (although it would be challenging to get that data).