News
How sparse attention solves the memory bottleneck in long-context LLMs
1+ mon, 1+ week ago (296+ words) LLMs are getting pulled into longer and messier workflows, handling large inputs and generating longer and longer token sequences. Coding assistants need to keep track of repositories, issue threads, terminal outputs, and earlier edits. Research agents need to carry facts…...
How GhostClaw malware targets the OpenClaw AI agent boom
3+ day, 4+ hour ago (844+ words) Threat actors are exploiting the rapid adoption of AI agents by designing malware that targets the agent itself. A new malware campaign, known as GhostClaw or GhostLoader, targets AI-assisted workflows and GitHub repositories to deliver credential-stealing payloads. The campaign preys…...
Why Meta’s V-JEPA 2.1 model is a massive step forward for real-world AI
1+ week, 3+ day ago (289+ words) Experiments show that V-JEPA 2.1 yields much better and faster results in robotic grasping, autonomous navigation of the physical world, predicting object interactions, and estimating 3D depth. These are the kinds of advances that can unlock new applications for AI in the…...
How Databricks’ FlashOptim cuts LLM training memory by 50 percent
1+ mon, 8+ hour ago (762+ words) How Databricks" FlashOptim cuts LLM training memory by 50 percent'TechTalks This article is part of our coverage of the latest in AI research. Training large language models is an expensive endeavor, largely due to the massive accelerator memory required for each…...
How Sakana AI’s new technique solves the problems of long-context LLM tasks - TechTalks
2+ mon, 4+ hour ago (236+ words) A new technique developed by researchers at Sakana AI, called Context Re-Positioning (RePo), allows Large language models (LLMs) to dynamically re-organize their internal view of their input data to better handle long-context tasks." LLMs process information in a strictly linear…...
Recursive Language Models: A new framework for infinite context in LLMs
2+ mon, 1+ week ago (436+ words) Recursive Language Models (RLMs), a new framework developed by researchers at MIT CSAIL, provide a solution to the limited context window of large language models (LLMs). This approach enables models to process arbitrarily long prompts without incurring massive memory costs…...
The evolution of LLM tool-use from API calls to agentic applications
3+ mon, 4+ day ago (250+ words) This article is part of'Demystifying AI, a series of posts that (try to) disambiguate the jargon and myths surrounding AI. Tool-use has been an important part of the development of large language models (LLMs) since the release of ChatGPT in…...