LlamaCast: Daily LLM Podcast

#Reasoning
Channel
Logo of the Telegram channel LlamaCast: Daily LLM Podcast
@LlamaCastPromote
632
subscribers
9
links
Daily podcast about the published articles in the LLM field. Reach me at @ShahriarShariati
Llms Still Can't Plan; Can Lrms?
๐Ÿ“ˆ LLMs Still Can't Plan; Can LRMs?

The paper "LLMs Still Can't Plan; Can LRMs? A Preliminary Evaluation of OpenAI's o1 on PlanBench" investigates the ability of large language models (LLMs) to plan, using a benchmark called PlanBench. The authors find that while OpenAI's new "Large Reasoning Model" (LRM) o1 shows significant improvement in planning abilities, it still falls short of fully achieving the task. This research highlights the need for further investigation into the accuracy, efficiency, and guarantees associated with these advanced models.

๐Ÿ“Ž Link to paper

#Planning #Reasoning #PlanBench

@LlamaCast
On the Diagram of Thought.wav
23.7 MB
๐Ÿง  On the Diagram of Thought

This paper introduces a new framework called Diagram of Thought (DoT) that models how large language models (LLMs) reason. Unlike traditional methods that represent reasoning as linear chains or trees, DoT utilizes a directed acyclic graph (DAG) structure. This structure allows LLMs to navigate complex reasoning pathways while ensuring logical consistency. By incorporating feedback mechanisms and leveraging auto-regressive next-token prediction, DoT enables LLMs to iteratively refine their reasoning process. The authors also formalize the DoT framework using Topos Theory, providing a mathematical foundation for its logical consistency and soundness. This approach enhances both training and inference within a single LLM, eliminating the need for multiple models or external control mechanisms. DoT offers a promising framework for developing next-generation reasoning-specialized LLMs.

๐Ÿ“Ž Link to paper

#DiagramOfThought #DoT #Reasoning

@LlamaCast