dmm: (Default)
Dataflow matrix machines (by Anhinga anhinga) ([personal profile] dmm) wrote2023-02-05 01:25 am
  • Previous Entry
  • Add Memory
  • Share This Entry
  • Next Entry
Entry tags:
  • ai safety,
  • anthropic ai,
  • artificial intelligence,
  • machine learning,
  • transformers,
  • understanding internals of ai

Technical interview with Neel Nanda

www.lesswrong.com/posts/r2yTwkGt3kbQG2mXi/axrp-episode-19-mechanistic-interpretability-with-neel-nanda

  • 3 comments
  • Post a new comment
Flat | Top-Level Comments Only
dmm: (Default)

no subject

[personal profile] dmm 2023-02-05 06:29 am (UTC)(link)
This is a Feb 3 interview.

Neel Nanda solved the mystery of "Grokking" as an independent researcher.

Before that he co-authored other cool papers about internal mechanisms of Transformer-based models while at Anthropic AI.
  • Thread
  • Reply to this
  • Thread
  • Hide 2 comments
  • Show 2 comments
dmm: (Default)

no subject

[personal profile] dmm 2023-02-05 06:30 am (UTC)(link)
See my notes at https://dmm.dreamwidth.org/64571.html

and https://github.com/anhinga/2022-notes/tree/main/Grokking-is-solved
  • Thread
  • Reply to this
  • Thread from start
  • Parent
dmm: (Default)

no subject

[personal profile] dmm 2023-02-05 06:31 am (UTC)(link)
"Hello everybody. In this episode I’ll be speaking with Neel Nanda. After graduating from Cambridge, Neel did a variety of research internships, including one with me, before joining Anthropic where he worked with Chris Olah on the Transformer Circuits agenda. As we record this, he’s pursuing independent research and producing resources to help build the field of mechanistic interpretability. But around when this episode will likely be released, he’ll be joining the language model interpretability team at DeepMind."
  • Thread
  • Reply to this
  • Thread from start
  • Parent

  • 3 comments
  • Post a new comment
Flat | Top-Level Comments Only

[ Home | Post Entry | Log in | Search | Browse Options | Site Map ]