Articles
Attention as Activation
This article is an in-depth insight into the paper authored by Dai et. al., titled as "Attention as Activation".
New ML Showcase Entry: Getting Started with aitextgen
Aitextgen is a Python library for training text-generation models using GPT-2 and GPT-3/GPT Neo. In this tutorial you'll get aitextgen up and running quickly in a Jupyter notebook on a free GPU instance from Paperspace Gradient!
Building an MLOps Community: Interview with Demetrios Brinkmann
We were lucky enough to interview Demetrios Brinkmann -- the founder of the world's biggest MLOps community. He shares with us the origin story of the community as well as some observations on the state of MLOps today.
The Swish Activation Function
This blogpost is an in-depth discussion of the Google Brain paper titled "Searching for activation functions" which has since revived research into activation functions.
File uploads arrive in Gradient Notebooks
We've released a new file uploader for Gradient Notebooks. You can now upload any file to your running notebook!
What can the history of supercomputing teach us about ARM-based deep learning architectures?
What does the NVIDIA's new ARM-based CPU chip mean for the future of deep learning? We take a look at this and other questions by looking at the history of high performance computing and how HPC architectures have been adopted within deep learning.