PortalsOS

Related Posts

Vote to see vote counts

Podcast artwork
a16z PodcastColumbia CS Professor: Why LLM...

When solving problems, LLMs benefit from a 'chain of thought' approach. By breaking down tasks into smaller, familiar steps, they reduce prediction entropy and increase confidence in the final answer.

Vishal Misra's work on understanding LLMs is profound. He has developed models that reduce the complex, multidimensional space of LLMs into a geometric manifold, allowing us to predict where reasoning can move within that space. This approach reflects how humans simplify the complex universe into manageable forms for reasoning.

The iterative nature of science requires LLMs to engage in simulations, theoretical calculations, and experiments to discover scientific insights.

Podcast artwork
More or Less#119 OpenAI Sora vs. TikTok: C...

The debate between LLMs and other reasoning models highlights the limitations of LLMs in understanding real-world context and predicting the future.

LLMs are criticized for lacking a true world model because they predict human responses rather than actual events.

Large Language Models (LLMs) create Bayesian manifolds during training. They confidently generate coherent outputs while traversing these manifolds, but veer into 'confident nonsense' when they stray from them.

Podcast artwork
Dwarkesh PodcastSome thoughts on the Sutton in...

Current LLMs do not develop true world models; they build models of what a human would say next, relying on human-derived concepts.

Podcast artwork
Dwarkesh PodcastRichard Sutton – Father of RL ...

LLMs are criticized for lacking a true world model because they predict human-like responses rather than actual outcomes.

Podcast artwork
a16z PodcastBuilding an AI Physicist: Chat...

The integration of geometric reasoning with LLMs can enhance the representation of atoms and design geometries, benefiting scientific research.

Podcast artwork
Dwarkesh PodcastFully autonomous robots are mu...

The use of LLMs and VLMs in robotics provides a way to incorporate common sense into robotic systems, allowing them to make reasonable guesses about potential outcomes without prior experience of mistakes.