In our latest article on To Data & Beyond, Eivind Kjosbakken discusses how you can fine-tune VLMs (visual large language models, often called VLLMs) like Qwen 2.5 VL 7B. He will first introduce you to a dataset of handwritten digits, which the base version of Qwen 2.5 VL struggles with. We will then inspect the dataset, annotate it, and use it to create a fine-tuned Qwen 2.5 VL, specialized in extracting hand-written text. 📌 The blog link is in the comment sections! ➡ If you want to master data science & AI beyond the basics, subscribe to the To Data & Beyond newsletter for in-depth tutorials and lessons: https://bit.ly/4lnAjqv #llm #opensource #largeLanguagemodels #python #datascience #machinelearning #ml #ai #artificialintelligence #genai #generativeai
Data Scientist | PhD & Generative AI Researcher | Founder @ To Data & Beyond | Opinions are my own
2moFine-Tuning VLLMs for Document Understanding: https://open.substack.com/pub/youssefh/p/fine-tuning-vllms-for-document-understanding?r=1sqbmi&utm_campaign=post&utm_medium=web&showWelcomeOnShare=false