Analyzing the Power of CLIP for Image Representation in Computer Vision
In this article, we examine typical computer vision analysis techniques in comparison with the modern CLIP (Contrastive Language-Image Pre-Training) model.
In this article, we examine typical computer vision analysis techniques in comparison with the modern CLIP (Contrastive Language-Image Pre-Training) model.
In this blog post, we examine what's new in Ultralytics awesome new model, YOLOv8, take a peak under the hood at the changes to the architecture compared to YOLOv5, and then demo the new model's Python API functionality by testing it to detect on our Basketball dataset.
When it comes to image synthesis algorithms, we need a method to quantify the differences between generated images and real images in a way that corresponds with human judgment. In this article, we highlight some of these metrics that are commonly used in the field today.
In this article, we examine two features of convolutional neural networks, translation equivariance and invariance.
In this article, we examine the game theory based approach to explaining outputs of machine learning models: Shapely Additive exPlanations or SHAP. We then demo the technology using sample images in a Gradient Notebook.
In this followup article, we will be taking a look at another beneficial use of autoencoders. We explored how an autoencoder's encoder can be used as a feature extractor with the extracted features then compared using cosine similarity in order to find similar images.
In this article, we will see some GANs improvements over time, then we go through the revolutionary ProGAN paper to see how it works and understand it in depth.
In this article we took a look at one of the uses of autoencoders: image denoising. In this tutorial, we show how an autoencoder's representation learning allows it to learn mappings efficient enough to fix incorrect pixels/datapoints.
In this blogpost, we examined the architecture and capabilities of the Versatile Diffusion framework. We then demonstrated this model within a Gradient Notebook to perform txt2img, img2txt, image variation, text variation, dual-guided, and Latent Image to Text to Image synthesis.