Neural Network
Style-based Recalibration Module (SRM) Channel Attention
In this post, we will cover a novel form of channel attention called the Style Recalibration Module (SRM), an extension of the popular TPAMI paper: Squeeze-and-Excitation Networks.
Attention Mechanisms in Recurrent Neural Networks (RNNs) With Keras
This series gives an advanced guide to different recurrent neural networks (RNNs). You will gain an understanding of the networks themselves, their architectures, their applications, and how to bring the models to life using Keras.
Accelerating Inference: Neural Network Pruning Explained
In this article, we'll discuss pruning neural networks: what it is, how it works, different pruning methods, and how to evaluate them.
Global Context Networks (GCNet) Explained
In this post, we will discuss a form of attention mechanism in computer vision known as Global Context Networks, first published at ICCV Workshops 2019.
Triplet Attention Explained (WACV 2021)
In this tutorial, we'll discuss a new form of attention mechanism in computer vision known as Triplet Attention, which was accepted to WACV 2021.
MobileNeXT (ECCV 2020)
In this article, we'll dive into an in-depth discussion of a recently proposed revamping of the popular MobileNet architecture, namely MobileNeXt, published at ECCV 2020.
Implementing Seq2Seq Models for Text Summarization With Keras
This series gives an advanced guide to different recurrent neural networks (RNNs). You will gain an understanding of the networks themselves, their architectures, applications, and how to bring them to life using Keras.