VLM FoundationsFree

Complete VLM Interview Guide: From 1 to Infinity

Comprehensive multimodal LLM interview reference — covers vision encoders, VLM architectures, training strategies, and frontier models with update dashboard.

From 1 to Infinity: The Complete Interview Guide for Multimodal Large Model Knowledge

Multimodal large model update dashboard: Summary: Multimodal OCR Models (March 2026 Edition): OCR-VLM comparison Interviewer: How does VLM handle dynamic resolution? Project Practice: Multimodal RL Agent (DAPO Algorithm & VeRL Framework & Think with Images): Training multimodal RL agents Interviewer: Technical improvements from Qwen-VL to Qwen3.5? (February 2026 Edition): Summary of Qwen-VL series updates Qwen-3.5 Preview: Advance analysis of Qwen3.5 architecture GLM OCR: Zhipu's two-stage OCR VLM Kimi K2.5: Analysis of the 1T-parameter Kimi K2.5 model DeepSeek OCR 2: DeepSeek's latest multimodal large model STEP3-VL-10B: StepFun's open-source multimodal large model, technical report and code analysis What is Qwen3VL-Embedding's MRL Dynamic Embedding Dimension?: MRL technology analysis A Complete Guide to Qwen3VL-Embedding & Reranker: Qwen's official multimodal embedding and reranking models GLM-4.6V Preview: Analysis of Zhipu's latest native agentic multimodal large model Qwen3-VL: Complete walkthrough of the Qwen3-VL technical report [all details] HunyuanOCR: Analysis of Tencent's OCR multimodal large model; first to use GRPO for OCR training Which OCR technology leads? Comparing DeepSeek, Xiaohongshu, Baidu, and Shanghai AI Lab: OCR large model comparison AIMv2: Analysis of Apple's autoregressive ViT v2 model AIM: Analysis of Apple's autoregressive ViT v1 model The History of PatchMerger in Qwen3VL: Analysis of the development history of Qwen2VL's PatchMerger How does PatchMerger merge 2x2 Patches?: Code analysis of how Qwen2VL's PatchMerger works Interviewer: What changed between each version of Qwen-VL?: Analysis of the Qwen-VL evolution history What changed between each version of DeepSeek-VL?: Technical comparison between DeepSeek-VL and DeepSeek-VL2 DeepSeek OCR: DeepSeek OCR analysis; using the visual modality for text compression Qwen3 Omni: Analysis of the latest Qwen series model with full multimodal input and text+speech output What is DeepStack technology in Qwen3-VL?: Analysis of Qwen3-VL's new projector structure Qwen3 VL Preview: Advance analysis of Qwen3-VL model architecture! InternVL 3.5: Analysis of PJLab's latest multimodal large model Ovis 2.5: Alibaba International's latest multimodal large model Ovis 2.5 dots.ocr: Xiaohongshu's 1.7B multimodal OCR model GLM-4.5V: Zhipu's latest multimodal large model built on the GLM-4.5 text model | August 11 The History of MLP Projector in MLLMs: Review of the development history of MLP-based alignment in multimodal large models Keye-VL: Kuaishou's open-source 8B multimodal large model focused on short video understanding GLM-4.1V-Thinking: Reportedly excellent user experience. Zhipu's latest multimodal thinking model MM-R5: Improving multimodal Reranker model capability through reasoning ERNIE 4.5: ERNIE Bot 4.5 open-sourced! Technical report and code analysis VLM2Vec: Proposes the MMEB benchmark and VLM2Embedding model training framework BGE-VL: BAAI's open-source multimodal embedding model framework BGE-VL and contrastive learning data generation pipeline GME: A general-purpose multimodal embedding model trained on Qwen2-VL VRAG-RL: Qwen's open-source multimodal RAG thinking model training framework UnifiedReward-Think: Tencent Hunyuan's open-source multimodal CoT reward model, technical report analysis Seed1.5-VL: Analysis of ByteDance's latest multimodal large model technical report Project Packaging: How to Write a Multimodal RAG Resume? Multimodal RAG Evaluation: Agent Data Synthesis and Evaluation Multimodal RAG Practice - Paper Q&A mini-4o Based on Qwen2VL: Multimodal RAG code practice, retrieval-augmented generation for reading papers. ColPali: Multimodal RAG model architecture, paper interpretation. InternVL 3: Shanghai AI Lab's latest multimodal large model, code analysis. Kimi-VL-Thinking: Analysis of Moonshot AI's latest multimodal reasoning model technical report Qwen2.5-VL Supervised Fine-tuning SFT: Code practice for fine-tuning Qwen2.5-VL Qwen2.5-Omni: Alibaba's latest open-source model supporting audio, video, and text input with audio and text output. NaViT: Native dynamic-resolution ViT, the ViT training technique used in Qwen2-VL and later models. SoViT: Scaling Law in ViT architectural design. SigLIP2: Analysis of Google DeepMind's latest vision-language encoder. SigLIP: Analysis of the visual encoder used in LLaVA-Next and the DeepSeek-VL series. Development Trajectory of Multimodal Large Models: A single article to clarify the development trajectory of understanding-side multimodal large models. Quick Interview Guide: Interview question analysis and quick answers.

Ready to test yourself?

Put your knowledge to work.

Practice real interview questions on this topic. Get AI feedback on exactly what you missed.