Back
Jason Wu
Publish new episode of "Next In AI", about the foundational Google paper revealing the secret of o1
open.spotify.com/episode/3Lcu…
[Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters] #nextinai
Publish AI generated paper reading podcast "Next In AI", this episode is about "Let's verify step by step" underlying OpenAI's o1 model:
open.spotify.com/episode/0mmG…
Is this helpful?
Read [Scaling LLM Test-Time Compute Optimally can
be More Effective than Scaling Model Parameters](arxiv.org/pdf/2408.03314)
Read [Extrinsic Hallucinations in LLMs](lilianweng.github.io/posts/20…)
my impression with o1 so far: it requires more "thinking" from the human as well. To get the best result, I find myself writing a well thought through "letter" to the model vs. rapid fire "sms" to a simpler model without the reasoning capabilities. And yes, I create that "letter" with a simpler model during the initial "exploration" phase.
do you have gpt-o1 access yet? openai.com/index/introducing-…
Join LLM mooc llmagents-learning.org/f24 it starts at 3:00PM PST
And now delta has free WiFi across the board for any cabin? Nice. I’ve only been away for 6mo 😘