No technical report published yet, unlikely code or weights will be either given VC funding.
Chance-Device 1 days ago [-]
It’s probably something like deepseek’s native sparse attention with content based granularity. They might not be publishing anything because it’s not such a strong value proposition and doing so would lead to commentary that would tank their investment opportunities.
SilverElfin 1 days ago [-]
Or maybe because giving it away would tank their investment opportunities.
regularfry 1 days ago [-]
There's ways and means. Pushing something out in the sub-30B range would gain them mindshare and they could keep bigger models to themselves. I can't see any indication of what size their model is though.
flowerthoughts 23 hours ago [-]
For Claude Code, I feel 1M is enough. I've had a compaction once, but that was because I was forcing Claude to do something it clearly had a hard time understanding.
For general chat bots where the user doesn't understand what a context window is, what do you do about context? Latest few messages and then a memory tool? Compaction?
anon22981 20 hours ago [-]
I feel the 1M context is way too large —- the model gets ”drunk” way before it gets anywhere near 1 million. Imo the 1M context window is a huge downgrade.
vagrantJin 6 hours ago [-]
I use a tool called context-mode. Update agent to save session summary every 100k tokens
whoodle 22 hours ago [-]
This is so interesting to me, I frequently experience compaction on long running features and still find Claude is better than starting with fully fresh prompts.
Every dev seems to use these tools differently.
brianush1 23 hours ago [-]
Claude does compaction in the regular web chat interface now, too
roger_ 1 days ago [-]
Have they published?
Bombthecat 1 days ago [-]
I believe it, when I see it.
Alifatisk 21 hours ago [-]
Waiting for the paper and model card. I believe it when I see it.
t4sty 23 hours ago [-]
Feels like this is to AI what JPG is to image
dfilppi 1 days ago [-]
[dead]
Rendered at 18:03:45 GMT+0000 (Coordinated Universal Time) with Vercel.
https://news.ycombinator.com/item?id=48023079
No technical report published yet, unlikely code or weights will be either given VC funding.
For general chat bots where the user doesn't understand what a context window is, what do you do about context? Latest few messages and then a memory tool? Compaction?
Every dev seems to use these tools differently.