So I randomly discovered an Ethiopian cat Instagram account with over 100 followers(same as mine lol) and casually following a bunch of other cats and dogs. And an entire childhood documented online. Honestly, I didnβt think I'd ever be jealous of a cat's social lifeπ. Now I'm thinking turning our cat into the next big influencer.
Btw this is the level of creativity LLMs won't getπ
Btw this is the level of creativity LLMs won't getπ
π25π€£4π1
We got another AI Research oriented channel by Biruk. He's a Masters student at Paris Saclay and will share cool AI research papers, lectures, codebases, experiments and opportunities from around the world. Join his channel here.
@ethio_sota
@ethio_sota
π₯16β€4π1
For anyone interested in computational neuroscience.
Apply to Simons Computational Neuroscience Imbizo which will happen in Cape Town, South Africa. They cover all your costs to attend.
https://imbizo.africa/
Apply to Simons Computational Neuroscience Imbizo which will happen in Cape Town, South Africa. They cover all your costs to attend.
https://imbizo.africa/
imbizo.africa
#Imbizo - Simons Computational Neuroscience Imbizo - #Imbizo
Simons Computational Neuroscience Imbizo summer school in Cape Town, South Africa
π₯7β€6π2
every new model is a combination of two things, training algorithms and data
open-source algorithms are up-to-date. the things that work are usually fairly simple
but the data is complex, massive, gatekept, ever-changing. here open models are way behind, and probably hopeless
Source
open-source algorithms are up-to-date. the things that work are usually fairly simple
but the data is complex, massive, gatekept, ever-changing. here open models are way behind, and probably hopeless
Source
π₯7β€3π3π―2π€1
For people who love tokens, two great papers that came out recently.
Dynamic Chunking for End-to-End Hierarchical Sequence Modeling, i saw this yesterday and they created a hierarchical network (H-Net) that learns dynamic, content-aware chunking directly from data, and it replaces tokenization, finally :)
Tokenization is NP-Complete
This one is mostly a theoretical paper and they prove that finding the optimal tokenization (either by direct vocabulary selection or merge operations) is NP-complete, shows the inherent computational difficulty of the task.
Dynamic Chunking for End-to-End Hierarchical Sequence Modeling, i saw this yesterday and they created a hierarchical network (H-Net) that learns dynamic, content-aware chunking directly from data, and it replaces tokenization, finally :)
Tokenization is NP-Complete
This one is mostly a theoretical paper and they prove that finding the optimal tokenization (either by direct vocabulary selection or merge operations) is NP-complete, shows the inherent computational difficulty of the task.
arXiv.org
Tokenisation is NP-Complete
In this work, we prove the NP-completeness of two variants of tokenisation, defined as the problem of compressing a dataset to at most $Ξ΄$ symbols by either finding a vocabulary directly...
π₯7
In the past, most ML research was driven by academia and there was a circle from people at UoToronto, UdM, Stanford, Berkeley etc. Almost everything that came out or even most pioneers were here. Now academia isn't dead but lost it's place.
Now the circle is OpenAI, Anthropic, Meta, Deepmind, XAI, Microsoft Research etc. Most of the people at the academia circle in the above are running these labs now but with less freedom to explore ideas over time.
Here comes the downfall in the second circle, in academia people spend 5 years to make some cool things and they did, most of the things were initiated in uni labs than industry. But now everyone is chasing benchmarks and 1% gain over the SOTA is almost enough to lead the race. Ofc industry made so many great things, but the same way scaling is converging, it'll be for industry labs, unless they have teams to do foundational research that might take years, e.g FAIR(Meta).
Now the circle is OpenAI, Anthropic, Meta, Deepmind, XAI, Microsoft Research etc. Most of the people at the academia circle in the above are running these labs now but with less freedom to explore ideas over time.
Here comes the downfall in the second circle, in academia people spend 5 years to make some cool things and they did, most of the things were initiated in uni labs than industry. But now everyone is chasing benchmarks and 1% gain over the SOTA is almost enough to lead the race. Ofc industry made so many great things, but the same way scaling is converging, it'll be for industry labs, unless they have teams to do foundational research that might take years, e.g FAIR(Meta).
π₯9β€2π€1
Forwarded from Beka (Beka)
rwanda and those cheerful ladies are becoming more and more attractive
π8
How many of you like or work on Computer Vision? Are there like any good Ethiopian based researchers working on CV that do some cool things, just wondering.
Currently in Kigali at ACVSS, it'll be very good to join it next year if you like Computer Vision.
Currently in Kigali at ACVSS, it'll be very good to join it next year if you like Computer Vision.
π₯16β€4
Forwarded from Beka (Beka)
This media is not supported in your browser
VIEW IN TELEGRAM
Better Auth was live in Times Square NYC last night. Crazy how far and how quick things go β€οΈ
π₯12π1
Alphafold isn't that novel work. I just learned yesterday that there were people who almost showed this 3 years prior and published a paper at a top conference, Neurips. Alphafold did almost exactly the same, used deep networks for prediction and scales it. However, they didn't even cite their paper, whyyyy sus
I learned that from Daniel Cremers yesterday and they were the ones who deserved the nobel prize.
I learned that from Daniel Cremers yesterday and they were the ones who deserved the nobel prize.
π€―7
Henok | Neural Nets
For people who love tokens, two great papers that came out recently. Dynamic Chunking for End-to-End Hierarchical Sequence Modeling, i saw this yesterday and they created a hierarchical network (H-Net) that learns dynamic, content-aware chunking directlyβ¦
So many NP-complete problems out there. What a time to be alive and see them solved/solving, proven, or just discovering them.
β€6
π₯10β€1
π19β€1