The subject of this post appears in the "Did you know..." section of Wikipedia's front page(archived) right now.
I'm saying "transformers" every time I am tempted to write "LLMs" because many modern LLMs also do image processing, so the term "LLM" is not quite right.
"Transformer"'s not quite right either because you can train a transformer on a narrow task. How about foundation model: "models (e.g., BERT, DALL-E, GPT-3) that are trained on broad data at scale and are adaptable to a wide range of downstream tasks".
I agree 100%. It would be interesting to explore how the term "AGI" has evolved, maybe starting with Goertzel and Pennachin 2007 who define it as:
a software program that can solve a variety of complex problems in a variety of different domains, and that controls itself autonomously, with its own thoughts, worries, feelings, strengths, weaknesses and predispositions
On the other hand, Stuart Russell testified that AGI means
machines that match or exceed human capabilities in every relevant dimension
so the experts seem to disagree. (On the other hand, Stuart & Russell's textbook cite Goertzel and Pennachin 2007 when mentioning AGI. Confusing.)
In any case, I think it's right to say that today's best language models are AGIs for any of these reasons:
In fact, GPT-2 is an AGI.
I wonder why Gemini used RLHF instead of Direct Preference Optimization (DPO). DPO was written up 6 months ago; it's simpler and apparently more compute-efficient than RLHF.
Thanks! For convex sets of distributions: If you weaken the definition of fixed point to , then the set has a least element which really is a least fixed point.
CFAR used to have an awesome class called "Be specific!" that was mostly about concreteness. Exercises included:
IIRC, Eliezer taught the class in May 2012? He talks about the relevant skills here and here. And then I ran it a few times, and then CFAR dropped it; I don't remember why.
Yep, I skimmed it by looking at the colorful plots that look like Ising models and reading the captions. Those are always fun.
No, I just took a look. The spin glass stuff looks interesting!
On 2018-04-09, OpenAI said[1]:
In contrast, in 2023, OpenAI said[2]:
Archived ↩︎
This archived snapshot is from 2023-05-17, but the document didn't get much attention until November that year. ↩︎