465 subscribers
478 photos
23 videos
36 files
942 links
Download Telegram
Forwarded from Axis of Ordinary
"Is AI task performance a type of submartingale, like a stock market index that goes up over time, but where each particular movement is intrinsically unpredictable?" https://www.lesswrong.com/posts/G993PFTwqqdQv4eTg/is-ai-progress-impossible-to-predict
Forwarded from Axis of Ordinary
"After showing a few examples, large language models can translate natural language mathematical statements into formal specifications."

[...]

"It handles the negation “there is no function” by proof-by-contradiction. It understands the phrase “into itself” and correctly formalizes the co-domain of f."

[...]

"Our finding hence shows a very surprising capability of these models. They learned very general and transferable knowledge that allows them to work with low-resource formal language."

Thread: https://threadreaderapp.com/thread/1529886847953870848.html
Forwarded from Axis of Ordinary
"Google just released a 442-author paper about a monster new test suite for evaluating Large Language Models (GPT-3 and the like), and in particular, their study of the language models' performance on their test suite as the number of parameters is scaled. As a striking example, see below for a neural net's ability to guess a movie from emojis (rather than, say, outputting random nonsense) as the number of parameters is gradually scaled from 2 million all the way to 128 billion." (via Scott Aaronson )

https://github.com/google/BIG-bench/blob/main/docs/paper/BIG-bench.pdf

This should be a little bit worrying because it makes it difficult to predict future progress. They are not just getting predictably better with more training, data, and parameters but sometimes capabilities emerge in a jumpy and unpredictable way. There can be sudden phase transitions.