Hacker Newsnew | past | comments | ask | show | jobs | submit | posnet's commentslogin

I am forever reminded of this stupid cartoon: https://youtu.be/6PY8C1KmNwM?si=_WU_lstzp_5mFrxk



I feel like I'm being Poe's Law'd here.


I've definitely noticed this anecdotally.

Especially with Gemini Pro when providing long form textual references, providing many documents in a single context windows gives worse answers than having it summarize documents first, ask a question about the summary only, then provide the full text of the sub-documents on request (rag style or just simple agent loop).

Similarly I've personally noticed that Claude Code with Opus or Sonnet gets worse the more compactions happen, it's unclear to me whether it's just the summary gets worse, or if its the context window having a higher percentage of less relevant data, but even clearing the context and asking it to re-read the relevant files (even if they were mentioned and summarized in the compaction) gives better results.


Gemini loses coherence and reasoning ability well before the chat hits the context limitations, and according to this report, it is the best model on several dimensions.

Long story short: Context engineering is still king, RAG is not dead


RAG was never going away, the people who say that are the same types who say software engineers will be totally replaced with AI.

LLMs will need RAG one way or another, you can hide it from the user, but it still must be there.


Yep, it can decohere really badly with bigger context. It's not only context related though. Sometimes it can lose focus early on in a way that is impossible to get it back on track.


Yep. The easiest way to tell someone has no experience with LLMs is if they say “RAG is dead”


> someone has no experience with LLMs

Thats 99% of coders. No need to gatekeep.


Gemini loses the notion of context the longer its context is: I often ask it to provide a summary of our discussion for the outside world and it will reference ideas or documents without introducing them, via anaphore, as if the outside world had knowledge of the context.


Cursor lifted "Start a new chat" limitation on gemini and i'm actually now enjoying keeping longer sessions within one window, becuase it's still very reasonable at recall, but doesnt need to restate everything each time


Can you elaborate on how prompts enhanced with rag avoid this context pollution? I don't understand why that would be


"Compactions" are just reducing the transcript to a summary of the transcript, right? So it makes sense that it would get worse because the agent is literally losing information, but it wouldn't be due to context rot.

The thing that would signal context rot is when you approach the auto-compact threshold. Am I thinking about this right?


Yes, but on agentic workflows it's possible to do more intelligent compaction.


I feel like the optimal coding agent would do this automatically - collect and (sometimes) summarize the required parts of code, MCP responses, repo maps etc., then combine the results into a new message in a new 'chat' that would contain all the required parts and nothing else. It's basically what I already do with aider, and I feel the performance (in situations with a lot of context) is way better than any agentic / more automated workflow I've tried so far, but it is a lot of work.


Claude Code tries, and it seems to be OK at it. It's hard to tell though and it definitely feels like sometimes you absolutely have to quit out and start again.


Try using /clear instead of quitting. Doesn’t clear scrollback buffer but does clear context


Appmap's ai agent does this very well.


Have you tried NotebookLM which basically does this as an app on the bg (chunking and summarising many docs) and you can -chat- with the full corpus using RAG


This one right here Mr. Basilisk


Most exchanges do not reveal counter-party information smaller than the broker level. So you wouldn't know just from looking at market activity the same person causing the large futures move was also taking large options positions.


Doesn't matter - see a pattern, exploit it - and in doing so, make profit yourself whilst reducing the pattern.


The pattern was exploitable only on the specific days that Jane Street was allegedly manipulating. How would you have figured out, without counterparty information and before noisy sales start dragging down the index, that day X is a manipulation day?

How would you have identified that there's even such a thing as a manipulation day? Do you have a model that tells you the objectively correct number of days a non-manipulated index should be lower at close?


Simply look at the market patterns and make a profit, duh! Why doesn't everyone do that?!


Usually, everyone does do that, which is why only hard-to-detect patterns remain profitable. Not something obvious like "buy options in the morning and sell in the evening" as in this example.

But maybe Jane Street only traded like this on some days, so you would need to know whether they had done so before you could hope to exploit them.


> Simply look at the market patterns and make a profit, duh! Why doesn't everyone do that?!

I know your question was sarcastic, but not everyone is able to see patterns. And not every trader is as good as the other. Hence winners and losers.


Yes claim is price is high at open low at close. Seems pretty straightforward.


Why not reveal counter-party info?


I did this route (sans the new Laos line, which was a bus at the time) in 2014. The world really was a different place.

Shout out to 'The man in seat 61', couldn't have done it without it.

https://www.seat61.com/


Amazing! What was your budget? For how long?


I finished it in 2019~, same with skipping the Laos section that didn’t exist. I contributed a bit to Seat61 from rural local stations in Myanmar while it was still open.

https://reustle.org/rtw shows my map around the entire planet. Next time by moto!


Fwiw, https://meetup.tokyotech.com/ on https://reustle.org/ doesn't resolve. I've daydreamed "I arrive in X. Museums? Meetups!"


Median


Human intelligence roughly follows a normal distribution where the median is the same as the mean. In that sense OP was correct that half of the population are below average.


"Jumbogram", an IPv6 packet with the Jumbo Payload option set, allowing for an frame size of up to 2³²-1 bytes.

At 10Gbps it would take 3.4 seconds just to serialize the frame.


Luckily 400Gb/s nics are already on the market [1]

[1] https://docs.broadcom.com/doc/957608-PB1


Clearly all the people repeating it without truely understanding it are just simple bots with a big lookup table of canned responses.


Actually I think they're tiny homunculi, trapped in a room full of meaningless symbols but given rules on how to manipulate them.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: