Attention as Activation
This article is an in-depth insight into the paper authored by Dai et. al., titled as "Attention as Activation".
This article is an in-depth insight into the paper authored by Dai et. al., titled as "Attention as Activation".
Aitextgen is a Python library for training text-generation models using GPT-2 and GPT-3/GPT Neo. In this tutorial you'll get aitextgen up and running quickly in a Jupyter notebook on a free GPU instance from Paperspace Gradient!
We were lucky enough to interview Demetrios Brinkmann -- the founder of the world's biggest MLOps community. He shares with us the origin story of the community as well as some observations on the state of MLOps today.
This blogpost is an in-depth discussion of the Google Brain paper titled "Searching for activation functions" which has since revived research into activation functions.
We've released a new file uploader for Gradient Notebooks. You can now upload any file to your running notebook!
What does the NVIDIA's new ARM-based CPU chip mean for the future of deep learning? We take a look at this and other questions by looking at the history of high performance computing and how HPC architectures have been adopted within deep learning.