Computer Vision Papers

Looking for a specific paper or subject?


  • Sapiens: Foundation for Human Vision Models

    Sapiens: Foundation for Human Vision Models

    Introduction In this post, we dive into a new release by Meta AI, presented in a research paper titled Sapiens: Foundation for Human Vision Models, which presents a family of models that target four fundamental human-centric tasks, which we see in the demo above. Fundamental Human-centric Tasks In the above figure from the paper, we…

  • Mixture of Nested Experts: Adaptive Processing of Visual Tokens

    Mixture of Nested Experts: Adaptive Processing of Visual Tokens

    Motivation In recent years, we use AI for more and more use cases, interacting with models that provide us with remarkable outputs. As we move forward, the models we use are getting larger and larger, and so, an important research domain is to improve the efficiency of using and training AI models. Standard MoE Is…

  • How Meta AI ‘s Human-Like V-JEPA Works?

    How Meta AI ‘s Human-Like V-JEPA Works?

    In this post, we dive into V-JEPA, which stands for Video Joint-Embedding Predicting Architecture, a new collection of vision models by Meta AI. V-JEPA is another step in Meta AI’s implementation of Yann LeCun’s vision about a more human-like AI. Several months back, we’ve already covered Meta AI’s I-JEPA model, which is a JEPA model…

  • How Do Vision Transformers Work?

    How Do Vision Transformers Work?

    Up until vision transformers were invented, the dominating model architecture in computer vision was convolutional neural network (CNN), which was invented at 1989 by famous researchers including Yann LeCun and Yoshua Bengio. At 2017, transformers were invented by Google and took the natural language processing domain by storm, but were not adapted successfully to computer…

  • From Diffusion Models to LCM-LoRA

    From Diffusion Models to LCM-LoRA

    Recently, a new research paper was released, titled: “LCM-LoRA: A Universal Stable-Diffusion Acceleration Module”, which presents a method to generate high quality images with large text-to-image generation models, specifically SDXL, but doing so dramatically faster. And not only it can run SDXL much faster, it can also do so for a fine-tuned SDXL, say for…

  • Vision Transformers Need Registers – Fixing a Bug in DINOv2?

    Vision Transformers Need Registers – Fixing a Bug in DINOv2?

    In this post we will discuss about visual transformers registers, which is a concept that was introduced in a research paper by Meta AI titled “Vision Transformers Need Registers”, which is written by authors that were part of DINOv2 release, a successful foundational computer vision model by Meta AI which we covered before in the…

  • Emu: Enhancing Image Generation Models Using Photogenic Needles in a Haystack

    Emu: Enhancing Image Generation Models Using Photogenic Needles in a Haystack

    Emu is a new text-to-image generation model by Meta AI, which was presented in a research paper titled “Emu: Enhancing Image Generation Models Using Photogenic Needles in a Haystack”. Text-to-image models are able to get a prompt as input, such as “a cat trying to catch a fish” in the example image above, and yield…

  • FACET: Fairness in Computer Vision Evaluation Benchmark

    FACET: Fairness in Computer Vision Evaluation Benchmark

    In this post we cover FACET, a new dataset created by Meta AI in order to evaluate a benchmark for fairness of computer vision models. Computer vision models are known to have biases that can impact their performance. For example, as we can see in the image below, given an image classification model, if we…

  • DINOv2 from Meta AI – Finally a Foundational Model in Computer Vision

    DINOv2 from Meta AI – Finally a Foundational Model in Computer Vision

    DINOv2 is a computer vision model from Meta AI that claims to finally provide a foundational model in computer vision, closing some of the gap from natural language processing where it is already common for a while now. In this post, we’ll explain what does it mean to be a foundational model in computer vision…

  • I-JEPA – A Human-Like Computer Vision Model

    I-JEPA – A Human-Like Computer Vision Model

    I-JEPA, Image-based Joint-Embedding Predictive Architecture, is an open-source computer vision model from Meta AI, and the first AI model based on Yann LeCun’s vision for a more human-like AI, which he presented last year in a 62 pages paper titled “A Path Towards Autonomous Machine Intelligence”.In this post we’ll dive into the research paper that…

Scroll to Top