A QR code that sends you to a different destination – lenticular and adversarial
mstdn.social
New Electrical Code Could Doom Most Common EV Charging
motortrend.com
Thank HN: My bootstrapped startup got acquired today
UI is hell: four-function calculators
lcamtuf.substack.com
Show HN: I made an open-source laptop from scratch
byran.ee
Building a full-text search engine in 150 lines of Python code (2021)
bart.degoe.de
Llama.vim – Local LLM-assisted text completion
github.com
Operator research preview
openai.com
Supercharge vector search with ColBERT rerank in PostgreSQL
blog.vectorchord.ai
Susctl CVE-2024-54507: A particularly 'sus' sysctl in the XNU kernel
jprx.io
Disabling Zen 5's Op Cache and Exploring Its Clustered Decoder
chipsandcheese.com
Surface-Stable Fractal Dithering
github.com
Building a Medieval Castle from Scratch
guedelon.fr
Psychedelic Graphics 0: Introduction
benpence.com
LoongArch64 Subjective Higlights
0x80.pl
Hacking Subaru: Tracking and controlling cars via the admin panel
samcurry.net
Show HN: I built an active community of trans people online
t4t.social
Show HN: Open-source AI video editor
github.com
Working with Files Is Hard (2019)
danluu.com
> However, generating sentence embeddings through pooling token embeddings can potentially sacrifice fine-grained details present at the token level. ColBERT overcomes this by representing text as token-level multi-vectors rather than a single, aggregated vector. This approach, leveraging contextual late interaction at the token level, allows ColBERT to retain more nuanced information and improve search accuracy compared to methods relying solely on sentence embeddings.
I don't know what it is about ColBERT that affords such opaque descriptions, but this is sadly common. I find the above explanation incredibly difficult to parse.
I have my own explanation of ColBERT here but I'm not particularly happy with that either: https://til.simonwillison.net/llms/colbert-ragatouille
If anyone wants to try explaining ColBERT without using jargon like "token-level multi-vectors" or "contextual late interaction" I'd love to see a clear description of it!