Self-Rewarding Language Models by Meta AI
In this post we dive into the Self-Rewarding Language Models paper by Meta AI, which can possibly be a step towards open-source AGI
In this post we dive into the Self-Rewarding Language Models paper by Meta AI, which can possibly be a step towards open-source AGI
Diving into a research paper introducing an innovative method to enhance LLM inference efficiency using memory offloading
Fast Inference of Mixture-of-Experts Language Models with Offloading Read More »
In this post we dive into TinyGPT-V, a small but mighty Multimodal LLM which brings Phi-2 success to vision-language tasks
TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones Read More »
In this post we dive into LLM in a flash paper by Apple, that introduces a method to run LLMs on devices that have limited memory
LLM in a flash: Efficient Large Language Model Inference with Limited Memory Read More »
In this post we go back to the important vision transformers paper, to understand how ViT adapted transformers to computer vision
Dive into Orca 2 research paper, the second version of the successful Orca small language model from Microsoft,
Orca 2: Teaching Small Language Models How to Reason Read More »
Following LCM-LoRA release, in this post we explore the evolution of diffusion models up to latent consistency models with LoRA
In this post we dive into Microsoft’s CODEFUSION, an approach to use diffusion models for code generation that achieves remarkable results
CODEFUSION: A Pre-trained Diffusion Model for Code Generation Read More »
In this post we dive into Table-GPT, a novel research by Microsoft, that empowers LLMs to understand tabular data
In this post we explain the paper “Vision Transformers Need Registers” by Meta AI, that explains an interesting behavior in DINOv2 features
Vision Transformers Need Registers – Fixing a Bug in DINOv2? Read More »