#dl
https://github.com/Lightning-AI/lightning/releases/tag/2.0.0
You can compile (torch 2.0) LightningModule now.
https://github.com/Lightning-AI/lightning/releases/tag/2.0.0
You can compile (torch 2.0) LightningModule now.
import torch
import lightning as L
model = LitModel()
# This will compile forward and {training,validation,test,predict}_step
compiled_model = torch.compile(model)
trainer = L.Trainer()
trainer.fit(compiled_model)
GitHub
Release Lightning 2.0: Fast, Flexible, Stable · Lightning-AI/lightning
Lightning AI is excited to announce the release of Lightning 2.0 ⚡
Highlights
Backward Incompatible Changes
PyTorch
Fabric
Full Changelog
PyTorch
Fabric
App
Contributors
Over the last coupl...
Highlights
Backward Incompatible Changes
PyTorch
Fabric
Full Changelog
PyTorch
Fabric
App
Contributors
Over the last coupl...
#misc
This is how generative AI is changing our lives. Now thinking about it, those competitive advantages from our satisfying technical skills are fading away.
What shall we invest into for a better career? Just integrated whatever is coming into our workflow? Or fundamentally change the way we are thinking?
This is how generative AI is changing our lives. Now thinking about it, those competitive advantages from our satisfying technical skills are fading away.
What shall we invest into for a better career? Just integrated whatever is coming into our workflow? Or fundamentally change the way we are thinking?
#ml
Pérez J, Barceló P, Marinkovic J. Attention is Turing-Complete. J Mach Learn Res. 2021;22: 1–35. Available: https://jmlr.org/papers/v22/20-302.html
Pérez J, Barceló P, Marinkovic J. Attention is Turing-Complete. J Mach Learn Res. 2021;22: 1–35. Available: https://jmlr.org/papers/v22/20-302.html
#dl
I am experimenting with torch 2.0 and searching for potential training time improvements in lightning. The following article provides a very good introduction.
https://lightning.ai/pages/community/tutorial/how-to-speed-up-pytorch-model-training/
I am experimenting with torch 2.0 and searching for potential training time improvements in lightning. The following article provides a very good introduction.
https://lightning.ai/pages/community/tutorial/how-to-speed-up-pytorch-model-training/
Lightning AI
How to Speed Up PyTorch Model Training
Learn how to improve the training performance of your PyTorch model without compromising its accuracy.
#ai
A lot of big names signed it. (Not sure how they verify the signee though)
Personally, I'm not buying it.
https://futureoflife.org/open-letter/pause-giant-ai-experiments/
A lot of big names signed it. (Not sure how they verify the signee though)
Personally, I'm not buying it.
https://futureoflife.org/open-letter/pause-giant-ai-experiments/
Future of Life Institute
Pause Giant AI Experiments: An Open Letter - Future of Life Institute
We call on all AI labs to immediately pause for at least 6 months the training of AI systems more powerful than GPT-4.
#ai
The performance is not too bad. But…given this is about academic topics, it sounds terrible to have this level of hallucination.
https://bair.berkeley.edu/blog/2023/04/03/koala/
The performance is not too bad. But…given this is about academic topics, it sounds terrible to have this level of hallucination.
https://bair.berkeley.edu/blog/2023/04/03/koala/
The Berkeley Artificial Intelligence Research Blog
Koala: A Dialogue Model for Academic Research
The BAIR Blog
#data
Quite useful.
I use pyarrow a lot and also a bit of polars. Mostly because pandas is slow. With the new 2.0 release, all three libraries are seamlessly connected to each other.
https://datapythonista.me/blog/pandas-20-and-the-arrow-revolution-part-i
Quite useful.
I use pyarrow a lot and also a bit of polars. Mostly because pandas is slow. With the new 2.0 release, all three libraries are seamlessly connected to each other.
https://datapythonista.me/blog/pandas-20-and-the-arrow-revolution-part-i
datapythonista blog
pandas 2.0 and the Arrow revolution (part I)
Introduction At the time of writing this post, we are in the process of releasing pandas 2.0. The project has a large number of users,...
#ts
I love the last paragraph, especially this sentence:
> Unfortunately, I can’t continue my debate with Clive Granger. I rather hoped he would come to accept my point of view.
Rob J Hyndman - The difference between prediction intervals and confidence intervals
https://robjhyndman.com/hyndsight/intervals/
I love the last paragraph, especially this sentence:
> Unfortunately, I can’t continue my debate with Clive Granger. I rather hoped he would come to accept my point of view.
Rob J Hyndman - The difference between prediction intervals and confidence intervals
https://robjhyndman.com/hyndsight/intervals/
#code
To me, high cognitive load reduces my code quality. In thoery, there are many tricks to reduce cognitive load, e.g., better modularity. In practice, they are not always carried out. Will chatGPT help? Let’s see.
https://www.caitlinhudon.com/posts/programming-beyond-cognitive-limitations-with-ai
To me, high cognitive load reduces my code quality. In thoery, there are many tricks to reduce cognitive load, e.g., better modularity. In practice, they are not always carried out. Will chatGPT help? Let’s see.
https://www.caitlinhudon.com/posts/programming-beyond-cognitive-limitations-with-ai
Haystacks by Caitlin Hudon
Programming Beyond Cognitive Limitations with AI — Haystacks by Caitlin Hudon
Our natural processing power is limited, and leveraging AI for assistance can help us to use it more efficiently, especially when it comes to reading and understanding code. Grokking new code requires cognitive load — and can sometimes trigger cognitive…
#tool
Read on reddit about this but never really looked into details. It is actually amazing.
Just watch the video in readme.
https://github.com/Significant-Gravitas/Auto-GPT
Read on reddit about this but never really looked into details. It is actually amazing.
Just watch the video in readme.
https://github.com/Significant-Gravitas/Auto-GPT
GitHub
GitHub - Significant-Gravitas/AutoGPT: AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission…
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters. - Significant-Gravitas/AutoGPT
#academia
Data science weekly mentioned this paper.
https://arxiv.org/abs/2304.06035
Quote from the abstract:
> A growing number of AI academics can no longer find the means and resources to compete at a global scale. This is a somewhat recent phenomenon, but an accelerating one, with private actors investing enormous compute resources into cutting edge AI research.
At first, a thought it's an April fools day paper. But it seems serious.
For example, the author mentioned the strategy "Analysis Instead of Synthesis". This has already happened to many fields. Global scale and money burning experiments in physics left many teams no choice but to take other teams’ data and analyze them.
This is actually quite crazy. Thinking about how AI/ML is developing, it's almost like a paradigm shift of research.
I read about a discussion on Reddit on a similar topic. Some people are concerned that medical research is also gonna shift to the private sector because of AI, leaving many people no choice but to join these big medical corporates.
On the other hand, requirements of computing resources also made smaller companies hard to compete in some fields. We need such a guide for business too.
Data science weekly mentioned this paper.
https://arxiv.org/abs/2304.06035
Quote from the abstract:
> A growing number of AI academics can no longer find the means and resources to compete at a global scale. This is a somewhat recent phenomenon, but an accelerating one, with private actors investing enormous compute resources into cutting edge AI research.
At first, a thought it's an April fools day paper. But it seems serious.
For example, the author mentioned the strategy "Analysis Instead of Synthesis". This has already happened to many fields. Global scale and money burning experiments in physics left many teams no choice but to take other teams’ data and analyze them.
This is actually quite crazy. Thinking about how AI/ML is developing, it's almost like a paradigm shift of research.
I read about a discussion on Reddit on a similar topic. Some people are concerned that medical research is also gonna shift to the private sector because of AI, leaving many people no choice but to join these big medical corporates.
On the other hand, requirements of computing resources also made smaller companies hard to compete in some fields. We need such a guide for business too.
#ai
AI Frontiers: AI for health and the future of research with Peter Lee
https://podcasts.google.com/feed/aHR0cHM6Ly9mZWVkcy5ibHVicnJ5LmNvbS9mZWVkcy9taWNyb3NvZnRyZXNlYXJjaC54bWw/episode/aHR0cHM6Ly9ibHVicnJ5LmNvbS9taWNyb3NvZnRyZXNlYXJjaC85NTE3NTgwMC9haS1mcm9udGllcnMtYWktZm9yLWhlYWx0aC1hbmQtdGhlLWZ1dHVyZS1vZi1yZXNlYXJjaC13aXRoLXBldGVyLWxlZS8?ep=14
----
A very cool discussion on the topic of large language models.
They mentioned the early stage test of Davinci from OpenAI. The model was able to reason for AP in biology and many of the reasoning was surprising to them. Then Ashley asked the person from OpenAI why is Davinci reason like that and the person replied they don't know.
Not everyone expected that kind of reasoning in LLM. In hindsight, "It is just a language model" is a very good question. Nowadays with GPT models, it seems that this question is not a question anymore because it is becoming a fact. What is in the training texts and what is language? Karpathy even made a joke about this:
> The hottest new programming language is English
https://twitter.com/karpathy/status/1617979122625712128?lang=en
AI Frontiers: AI for health and the future of research with Peter Lee
https://podcasts.google.com/feed/aHR0cHM6Ly9mZWVkcy5ibHVicnJ5LmNvbS9mZWVkcy9taWNyb3NvZnRyZXNlYXJjaC54bWw/episode/aHR0cHM6Ly9ibHVicnJ5LmNvbS9taWNyb3NvZnRyZXNlYXJjaC85NTE3NTgwMC9haS1mcm9udGllcnMtYWktZm9yLWhlYWx0aC1hbmQtdGhlLWZ1dHVyZS1vZi1yZXNlYXJjaC13aXRoLXBldGVyLWxlZS8?ep=14
----
A very cool discussion on the topic of large language models.
They mentioned the early stage test of Davinci from OpenAI. The model was able to reason for AP in biology and many of the reasoning was surprising to them. Then Ashley asked the person from OpenAI why is Davinci reason like that and the person replied they don't know.
Not everyone expected that kind of reasoning in LLM. In hindsight, "It is just a language model" is a very good question. Nowadays with GPT models, it seems that this question is not a question anymore because it is becoming a fact. What is in the training texts and what is language? Karpathy even made a joke about this:
> The hottest new programming language is English
https://twitter.com/karpathy/status/1617979122625712128?lang=en
Google Podcasts
Microsoft Research Podcast - 137 - AI Frontiers: AI for health and the future of research with Peter Lee
Powerful new large-scale AI models like GPT-4 are showing dramatic improvements in reasoning, problem-solving, and language capabilities. This marks a phase change for artificial intelligence—and a signal of accelerating progress to come.
In this new Microsoft…
In this new Microsoft…
#misc
I was working and didn't have time watching the live streaming when they launched the starship. After some Twitter browsing, I have to say, this thing is beautiful.
https://twitter.com/nextspaceflight/status/1649052544755470338
I was working and didn't have time watching the live streaming when they launched the starship. After some Twitter browsing, I have to say, this thing is beautiful.
https://twitter.com/nextspaceflight/status/1649052544755470338
Twitter
My Autotrack software captures the moment that Starship lost control. Excitement was very much guaranteed. Great first attempt by the SpaceX team!
Tune in to hear our live reaction! @NASASpaceflight
https://t.co/uutBwWSABz
Tune in to hear our live reaction! @NASASpaceflight
https://t.co/uutBwWSABz
#coding
I had some discussions with serval people about writing good code during machine learning experimentation.
Whenever it comes to the part of writing formal code, opinions diverge. So, should we write good code that is easy to read with typing and tests, even in experiments?
The spirit of experimentation is fast and reliable. So naturally, the question comes down to what kind of coding style allows us to develop and run experiments, fast.
My experience with running experiments is that we will never run the code just once. Instead, we always come back to it and run it with different configurations or parameters. In this circumstance, how good shall my code be?
For typing and tests, I type most of my args but only write tests needed to develop and debug a function or class.
- Typing is important because people spend time figuring out what to put in there as an argument for a function. With typing, it is much faster.
- Here is an example for tests: If I need to know the shape of the tensor deep in a method of a class, I would spend some seconds writing a simple test that allows me to put breakpoints in the method to investigate inside.
But, the above is a bit trivial. How about the design of the functions and classes? I suggest taking your time writing those that are repeated in every experiment. We will hit some ceiling in development speed real quick, if we always use the first and most naive design for these. In practice, I would say, design it twice and write it once.
One such example is data preprocessing. When dealing with the same data and problems, data transformations are usually quite similar in each experiment but a bit different in details. Finding the patterns and writing some slightly generic functions would be helpful. There is always the risk of over-engineering. I prefer to improve things little by little. I might generalize a function a little bit in one experiment. And also, don't hesitate to throw away your code to rewrite. Rewriting will take little time, and it usually brings in improvements.
That's my five cents on code quality for developing and running machine learning experiments.
I had some discussions with serval people about writing good code during machine learning experimentation.
Whenever it comes to the part of writing formal code, opinions diverge. So, should we write good code that is easy to read with typing and tests, even in experiments?
The spirit of experimentation is fast and reliable. So naturally, the question comes down to what kind of coding style allows us to develop and run experiments, fast.
My experience with running experiments is that we will never run the code just once. Instead, we always come back to it and run it with different configurations or parameters. In this circumstance, how good shall my code be?
For typing and tests, I type most of my args but only write tests needed to develop and debug a function or class.
- Typing is important because people spend time figuring out what to put in there as an argument for a function. With typing, it is much faster.
- Here is an example for tests: If I need to know the shape of the tensor deep in a method of a class, I would spend some seconds writing a simple test that allows me to put breakpoints in the method to investigate inside.
But, the above is a bit trivial. How about the design of the functions and classes? I suggest taking your time writing those that are repeated in every experiment. We will hit some ceiling in development speed real quick, if we always use the first and most naive design for these. In practice, I would say, design it twice and write it once.
One such example is data preprocessing. When dealing with the same data and problems, data transformations are usually quite similar in each experiment but a bit different in details. Finding the patterns and writing some slightly generic functions would be helpful. There is always the risk of over-engineering. I prefer to improve things little by little. I might generalize a function a little bit in one experiment. And also, don't hesitate to throw away your code to rewrite. Rewriting will take little time, and it usually brings in improvements.
That's my five cents on code quality for developing and running machine learning experiments.
#misc
‘The Godfather of AI’ Quits Google and Warns of Danger Ahead - The New York Times
https://www.nytimes.com/2023/05/01/technology/ai-google-chatbot-engineer-quits-hinton.html
‘The Godfather of AI’ Quits Google and Warns of Danger Ahead - The New York Times
https://www.nytimes.com/2023/05/01/technology/ai-google-chatbot-engineer-quits-hinton.html
NY Times
‘The Godfather of A.I.’ Leaves Google and Warns of Danger Ahead
For half a century, Geoffrey Hinton nurtured the technology at the heart of chatbots like ChatGPT. Now he worries it will cause serious harm.