{"database": "hackernews", "table": "items", "rows": [[47251900, "comment", "kamranjon", 1772649615, "Absolutely. So my codebase is huge, it&#x27;s a monolith. But my work is in very specific parts of the codebase, I don&#x27;t pull the entire code base into context (and I don&#x27;t think that is common practice even with claude) - I start at a specific point with a specific task and work with the agent to achieve something clearly defined, for example writing tests, extracting things into separate files, refactoring or even scaffolding a new feature. You have to periodically start new threads, because you&#x27;ll start hitting the limits of the context, but I max it out at over 200k because I have the memory overhead on my 128gb mbp to do that, so I can get quite a lot done.<p>I really recommend trying the Qwen models - 3 coder next is really incredible. GLM 4.7 flash is also incredibly performant on modest hardware. Important things to consider is setting the temperature and top_p and top_k values etc based on what is recommended by the provider of the model - a thing as simple as that could result in a huge difference in performance.<p>The other big leap for me was switching to Zed editor and getting its agent stuff just seamlessly integrated. If you run LM Studio on your local machine it&#x27;s super easy and even setting it up on a remote machine and calling out to LM Studio is dead simple.", 0, 47250270, null, null, null, null, null, 0, null, null, 1774132559484]], "columns": ["id", "type", "by", "time", "text", "dead", "parent", "poll", "url", "score", "title", "descendants", "deleted", "kids", "parts", "cached_at"], "primary_keys": ["id"], "primary_key_values": ["47251900"], "units": {}, "query_ms": 1.6240617260336876}