Portfolio

๐ŸŽ™๏ธ Interview with David Stutz from Google DeepMind at #HLF23
When @GoogleAI scaled machine translation to 1000 languages with monolingual data, they wrote a 57 long paper about it https://arxiv.org/abs/2205.03983. Ms. Coffee Bean summarized it ๐Ÿ˜Š: ๐Ÿ“บ https://y...
Ms. Coffee Beanโ€™s take on DALL-E 2โ€™s secret โ€œlanguageโ€. ๐Ÿ“บ https://youtu.be/MNwURQ9621k Short summary of current opinions and theories on why DALL-E 2 can create sensible pictures out of gibberish t...
Imagen from Google Brain ๐Ÿง  is competing with DALLE-2 when it comes to generating amazing images from just text! Here is an overview of Imagen and other diffusion-based text-to-image generators such...
Special AI โ˜• break edition about exciting developments in mathematical deep learning โ€“ inspired by physics. ๐Ÿ”ฅ ๐Ÿ“บ https://youtu.be/m2bXL5Z5CBM Make sure to check it out if you want a quick summary ov...
Letโ€™s VALSE ๐Ÿ’ƒ! We present our own work on the VALSE benchmark testing large pretrained vision and language models by targeting linguistic phenomena. Work accepted at ACL 2022. https://youtu.be/huml...
Itโ€™s time for a coffee break! โ˜• Weโ€™re explaining PaLM, Google AIโ€™s Pathways Language model. ๐Ÿ‘‡ Yes, itโ€™s that agglomeration of 540 billion parameters that can explain jokes. ๐Ÿ˜ ๐Ÿ“บ https://youtu.be/yi-...
The 10 billion parameter SEER model from @MetaAI is *fairer*, even though it is trained on *uncurated* data. How so? ๐Ÿ˜• Check out our take on this. ๐Ÿ‘‡ https://youtu.be/XHAoV_nKr1o
Regularization in deep learning, Lipschitz continuity, gradient regularization, adversarial defense, gradient penalty. These were topics of our daily Quiz questions! Tim Elsner (@daidailoh on Twitt...
Diffusion models beat GANs in image synthesis, GLIDE generates images from text descriptions, surpassing even DALL-E in terms of photorealism! Check out this video to learn how diffusion models wor...
We were invited to the Machine Learning Street Talk podcast! We talked about linguistics, symbolic AI and our YouTube journey. ๐Ÿ™Œ Check out the episode: ๐Ÿ‘‰ https://youtu.be/p2D2duT-R2E
Multi-head self-attention (MSA) and convolutions are complementary, or so it turns out. So what are the differences between MSA and Convs? In this video, we present the lessons of the recent "How d...
๐Ÿ“บ https://youtu.be/gzNlfvu1hqo Announcement: โ˜•โš”๏ธ๐Ÿต AMA with AI Coffee Break & Chai Time Data Science (Sanyam Bhutani) over at Weights & Biases . Happening Saturday, the 26th of Feb. at 5 PM CET / 8 ...
Can a ConvNet outperform a Vision Transformer? What kind of modifications do we have to apply to a ConvNet to make it as powerful as a Transformer? Spoiler: itโ€™s not attention. This is our animated...
What a Christmas surprise: ๐ŸŽ„ Tonight, we surpassed 10,000 subscribers on YouTube! ๐Ÿคฏ https://www.youtube.com/aicoffeebreak Ms. Coffee Bean is still drunk (with joy). Thanks to our wonderful viewers ...
Why would one build a transformer to solve linear algebra problems when there is numpy.linalg? ๐Ÿ˜ฑ ๐Ÿ“บ Check out the video to find out why F. Charton trained transformers to solve #LinearAlgebra proble...
Want to know about our conversation about #AI (mis)conceptions at @_eavi? Then check out๐Ÿ“บ the video recording: ๐Ÿ‘‰ https://youtu.be/XI5pvWzg_N4 Or read the textual summary: https://eavi.eu/ai-for-goo...
โ€œMasked Autoencoders Are Scalable Vision Learnersโ€ paper explained and animated! ๐Ÿ“บ Say goodbye to contrastive learning and say hello (again) to autoencoders in #ComputerVision! Love the simple, ele...
What could go wrong when comparing models based on the number of parameters? Size does not matter | The efficiency misnomer | What does the number of parameters mean? ๐Ÿ‘‰ https://youtu.be/l3bTCP6x-FM
Do Transformers process sequences of FIXED or of VARIABLE length? | #AICoffeeBreakQuiz question explained. ๐Ÿ‘‰ https://youtu.be/Xxts1ithupI
A confused Coffee Bean summarizes the generalization โ€“ interpolation โ€“ extrapolation debate in #MachineLearning of Yann LeCun, Gary Marcus, Francois Chollet and others. Her take on the discussion s...
SimVLM paper explained | What the paper doesn't tell you ๐Ÿ‘‰ https://youtu.be/XU4ZHUnnbpg
Ms. Coffee Bean explains, visualizes, and comments on the sobering take about ๐Ÿ“Š NLU Benchmarking by S. Bowman and G. Dahl. Check out the short, animated video: ๐Ÿ‘‡ https://youtu.be/W57u1j16iC8
๐Ÿ”ฅ NEW video! ๐Ÿ”ฅ Swin Transformer paper explained, visualized, and animated by Ms. Coffee Bean. Let us know what you think. Weโ€™re a little rusty after our vacation. ๐Ÿ“บ Check out the short video: ๐Ÿ‘‰ htt...
How to know if a face was generated by a (Style)GAN(2)? ๐Ÿ“œ We discuss โ€œEyes Tell All: Irregular Pupil Shapes Reveal GAN-generated Faces.โ€ by Guo et al. 2021. ๐Ÿ“บ If you're interested in cues for spott...
How do modern search engines work? Find out by watching this short animated video. ๐Ÿ‘‡ ๐Ÿ“บ https://youtu.be/YkK5IKgxp-c Best enjoyed during your Coffee Break! โ˜•
Sound the opinionated video alarm! ๐Ÿšจ Ms. Coffee Bean and her PhD student debate their different opinions triggered by the โ€œFoundation Modelsโ€ paper. ๐Ÿ“บ On the opportunities and risks of calling pre-...
What are tokenizers good for? Why do transformers need them? Find out in this short animated video! ๐Ÿ“บ ๐Ÿ‘‡ https://youtu.be/D8j1c4NJRfo
Our submission to the Veritasium Science Communication contest! The shortest explanation about the todayโ€™s capabilities of artificial intelligence ๐Ÿง , or rather โ€œWhen is your iPhone smarter than Ms....
How can data leakage happen during data preparation? Find out by joining Ms. Coffee Beanโ€™s visualization ๐Ÿ“บ of the answer to this question. ๐Ÿ‘‰ https://youtu.be/8_UBDTHAHqY You can find the original s...
We had a great discussion with Prof. Melanie Mitchell! For an honest, down-to-earth view on the limits of (artificial) intelligence, check out the latest Machine Learning Street Talk episode: ๐Ÿ‘‰ htt...
In the previous video, we explained the basics of positional embeddings with sines and cosines as they were introduced in the "Attention is all you need" paper. In our latest video, we dive even fu...
Positional embeddings in transformers explained and visualized! Check out the video here ๐Ÿ‘‡ https://youtu.be/1biZfFLPRSY
๐Ÿ“บ Relative position representations explained by Ms. Coffee Bean. Learn about them as they were introduced in the "Self-Attention with Relative Position Representations." paper by S. Peter, J. Uszk...