PDP-11π
https://www.economist.com/technology-quarterly/2020/06/11/the-cost-of-training-machines-is-becoming-a-problem The growing demand for computing power has fuelled a boom in chip design and specialised devices that can perform the calculations used in AI efficiently.β¦
Graphcore raises $222M at $2.7B valuation
https://techcrunch-com.cdn.ampproject.org/c/s/techcrunch.com/2020/12/28/ai-chipmaker-graphcore-raises-222m-at-a-2-77b-valuation-and-puts-an-ipo-in-its-sights/amp/
https://techcrunch-com.cdn.ampproject.org/c/s/techcrunch.com/2020/12/28/ai-chipmaker-graphcore-raises-222m-at-a-2-77b-valuation-and-puts-an-ipo-in-its-sights/amp/
https://www.tenstorrent.com/press/
Tenstorrent, a hardware start-up developing next generation computers, announces the addition of industry veteran Jim Keller as President, CTO, and board member.
Tenstorrent, a hardware start-up developing next generation computers, announces the addition of industry veteran Jim Keller as President, CTO, and board member.
π FPGA comes back. Titanium FPGAs from EFINIX are focused on the edge application and promises unbeatable per-Watt performance
βοΈ[pdf] Hardware Accelerator of CNN by Yann Le Cun, father of Deeo Learning revolution
πΊ [youTube] 20min introduction video from Intel about what are the FPGAs and what sort of applications can you use it for
π± [plumerAI] Yet another ML Hardware startup
keeps growing and explains why binarized neural networks do the job with less resources
π° [youTube] Bonus! The lecture from yesterday by professor Onur Mutlu, ETH, about GPU architecture.
π»π»π»
This channel is back from hibernation and more reviews will come soon :)
βοΈ[pdf] Hardware Accelerator of CNN by Yann Le Cun, father of Deeo Learning revolution
πΊ [youTube] 20min introduction video from Intel about what are the FPGAs and what sort of applications can you use it for
π± [plumerAI] Yet another ML Hardware startup
keeps growing and explains why binarized neural networks do the job with less resources
π° [youTube] Bonus! The lecture from yesterday by professor Onur Mutlu, ETH, about GPU architecture.
π»π»π»
This channel is back from hibernation and more reviews will come soon :)
PDP-11π
The latest paper by David Patterson & Google TPU team reveals details of the world most efficient and one of the most powerful supercomputers for DNN Acceleration - TPU v3. The one which was used to train BERT. We recommend that you definitely read the fullβ¦
ππΌGoogle finally released TPU v4, it will be avaliable for customers later this year.
π₯΄The previous v3 version was unveiled in 2018 and the v4 is claimed to be twice as fast.
π½TPU v4 combines in a 4096 chips sumercomputer that reaches 1 exaFLOPs (10**18) of performance
Read more on [hpcwire] and watch the video Google I/O β21
π₯΄The previous v3 version was unveiled in 2018 and the v4 is claimed to be twice as fast.
π½TPU v4 combines in a 4096 chips sumercomputer that reaches 1 exaFLOPs (10**18) of performance
Read more on [hpcwire] and watch the video Google I/O β21
π The Hardware Lottery π°
by Sarah Hooker, Google Brain [ACM]
- The very first computer hardware was extremely focused on solving one particular problem - numerical differentiation or polynomial models. In the 1960s IBM invented the concept of Instruction Set and made migration between hardware easier for software developers. Till the 2010s we have been living in the world of general-purpose hardware - CPUs.
- Computer Science Ideas win or lose not because one superior one to another, but because some of them did not have the suitable hardware to be implemented in. Back Propagation Algorithm, the key algorithm that made the deep learning revolution possible, was invented independently in 1963, 1976, 1988 and finally applied to CNN in 1989. However, it was only three decades later that deep neural networks were widely accepted as a promising research direction and the significant result was achieved with GPUs, that could run massive parallel computations.
- Today hardware pendulum is swinging back to domain-specific hardware like it was the CPU invention
- Hardware should not remain a limiting factor for the breakthrough ideas in AI research. Hardware and Software should be codesigned for the SOTA algorithms. Algorithm developers need a deeper understanding of the computer platforms.
read also here
by Sarah Hooker, Google Brain [ACM]
- The very first computer hardware was extremely focused on solving one particular problem - numerical differentiation or polynomial models. In the 1960s IBM invented the concept of Instruction Set and made migration between hardware easier for software developers. Till the 2010s we have been living in the world of general-purpose hardware - CPUs.
- Computer Science Ideas win or lose not because one superior one to another, but because some of them did not have the suitable hardware to be implemented in. Back Propagation Algorithm, the key algorithm that made the deep learning revolution possible, was invented independently in 1963, 1976, 1988 and finally applied to CNN in 1989. However, it was only three decades later that deep neural networks were widely accepted as a promising research direction and the significant result was achieved with GPUs, that could run massive parallel computations.
- Today hardware pendulum is swinging back to domain-specific hardware like it was the CPU invention
- Hardware should not remain a limiting factor for the breakthrough ideas in AI research. Hardware and Software should be codesigned for the SOTA algorithms. Algorithm developers need a deeper understanding of the computer platforms.
read also here
dl.acm.org
The hardware lottery | Communications of the ACM
After decades of incentivizing the isolation of hardware, software, and algorithm development, the catalysts for closer collaboration are changing the paradigm.
PDP-11π
π The Hardware Lottery π° by Sarah Hooker, Google Brain [ACM] - The very first computer hardware was extremely focused on solving one particular problem - numerical differentiation or polynomial models. In the 1960s IBM invented the concept of Instructionβ¦
Hey folks,
That was a long period of silence here, but I'll try to breathe a new life to this channel I'll do my best to post more frequently and not only ML hardware, but also about Zero Knowledge Proof accelerators, some areas on computer architecture, trading infrastructure and HPC
Here is a video for today β what is Zero Knowledge Proof (ZKP)
ZKP is a way for proofer to convince a verifier, who has X and Y, that given for given function F
But I find an explanation with a revealing of puffin absolutely genius in both in simplicity and clarity.
πͺ¨πͺ¨πͺ¨πͺ¨πͺ¨πͺ¨
πͺ¨πͺ¨πͺ¨π§πͺ¨πͺ¨
πͺ¨πͺ¨πͺ¨πͺ¨πͺ¨πͺ¨
Enjoy the video
Computer Scientist Explains ZKP in 5 Levels of Difficulty | WIRED
That was a long period of silence here, but I'll try to breathe a new life to this channel I'll do my best to post more frequently and not only ML hardware, but also about Zero Knowledge Proof accelerators, some areas on computer architecture, trading infrastructure and HPC
Here is a video for today β what is Zero Knowledge Proof (ZKP)
ZKP is a way for proofer to convince a verifier, who has X and Y, that given for given function F
F(x,w)=y, without revealing w to verifier. But I find an explanation with a revealing of puffin absolutely genius in both in simplicity and clarity.
πͺ¨πͺ¨πͺ¨πͺ¨πͺ¨πͺ¨
πͺ¨πͺ¨πͺ¨π§πͺ¨πͺ¨
πͺ¨πͺ¨πͺ¨πͺ¨πͺ¨πͺ¨
Enjoy the video
Computer Scientist Explains ZKP in 5 Levels of Difficulty | WIRED
YouTube
Computer Scientist Explains One Concept in 5 Levels of Difficulty | WIRED
Computer scientist Amit Sahai, PhD, is asked to explain the concept of zero-knowledge proofs to 5 different people; a child, a teen, a college student, a grad student, and an expert. Using a variety of techniques, Amit breaks down what zero-knowledge proofsβ¦
π9β€3π―1π»1π¦1
Forwarded from HN Best Comments
Re: Ask HN: How can ChatGPT serve 700M users when I can't run one GPT-4 locally?
I work at Google on these systems everyday (caveat this is my own words not my employers)). So I simultaneously can tell you that its smart people really thinking about every facet of the problem, and I can't tell you much more than that.
However I can share this written by my colleagues! You'll find great explanations about accelerator architectures and the considerations made to make things fast.
https://jax-ml.github.io/scaling-book/
In particular your questions are around inference which is the focus of this chapter
https://jax-ml.github.io/scaling-book/inference/
Edit:
Another great resource to look at is the unsloth guides. These folks are incredibly good at getting deep into various models and finding optimizations, and they're very good at writing it up. Here's the Gemma 3n guide, and you'll find others as well.
https://docs.unsloth.ai/basics/gemma-3n-how-to-run-and-fine-...
canyon289, 10 hours ago
I work at Google on these systems everyday (caveat this is my own words not my employers)). So I simultaneously can tell you that its smart people really thinking about every facet of the problem, and I can't tell you much more than that.
However I can share this written by my colleagues! You'll find great explanations about accelerator architectures and the considerations made to make things fast.
https://jax-ml.github.io/scaling-book/
In particular your questions are around inference which is the focus of this chapter
https://jax-ml.github.io/scaling-book/inference/
Edit:
Another great resource to look at is the unsloth guides. These folks are incredibly good at getting deep into various models and finding optimizations, and they're very good at writing it up. Here's the Gemma 3n guide, and you'll find others as well.
https://docs.unsloth.ai/basics/gemma-3n-how-to-run-and-fine-...
canyon289, 10 hours ago
jax-ml.github.io
How To Scale Your Model
Training LLMs often feels like alchemy, but understanding and optimizing the performance of your models doesn't have to. This book aims to demystify the science of scaling language models: how TPUs (and GPUs) work and how they communicate with each otherβ¦
π3
Hey again, folks :)
Two years have passed since my last attempt to revive this channel. In that time, I left my job in prop trading, spent a year doing the MicroMasters in Finance, and in November β24 joined a skyrocketing (according to Bloomberg) hedge fund. Now weβre driving its HFT strategies to the moon with FPGA platforms and highly optimised software.
The flip side of the coin β Iβm still far, far away from machine learning hardware. Not as far as I was when I started this channel half a decade ago, but still not quite there.
So, hereβs my proposal to make this group a bit more interactive:
I want to start a learning group on the material mentioned above β How To Scale Your Model.
Plan: 12 sections β 12 seminars. Each seminar is a 1-hour call where we read through the chapter, rephrase it, and discuss.
No fees, no recordings β pure enthusiasm and passion.
Schedule: Tuesdays and Thursdays, 7:00β8:00 BST (UTC+1).
DM me if youβre interested and ready to invest your time.π°οΈ
@vconst89
P.S. This text was grammar-checked by ChatGPT but originally typed on my phone, so any remaining quirks are 100% mine.
Two years have passed since my last attempt to revive this channel. In that time, I left my job in prop trading, spent a year doing the MicroMasters in Finance, and in November β24 joined a skyrocketing (according to Bloomberg) hedge fund. Now weβre driving its HFT strategies to the moon with FPGA platforms and highly optimised software.
The flip side of the coin β Iβm still far, far away from machine learning hardware. Not as far as I was when I started this channel half a decade ago, but still not quite there.
So, hereβs my proposal to make this group a bit more interactive:
I want to start a learning group on the material mentioned above β How To Scale Your Model.
Plan: 12 sections β 12 seminars. Each seminar is a 1-hour call where we read through the chapter, rephrase it, and discuss.
No fees, no recordings β pure enthusiasm and passion.
Schedule: Tuesdays and Thursdays, 7:00β8:00 BST (UTC+1).
DM me if youβre interested and ready to invest your time.π°οΈ
@vconst89
P.S. This text was grammar-checked by ChatGPT but originally typed on my phone, so any remaining quirks are 100% mine.
jax-ml.github.io
How To Scale Your Model
Training LLMs often feels like alchemy, but understanding and optimizing the performance of your models doesn't have to. This book aims to demystify the science of scaling language models: how TPUs (and GPUs) work and how they communicate with each otherβ¦
β4β€2π2π¨βπ»2π1πΎ1