Ollama just added support for llama3.2 vision models. This enables models to see and process images. In this video, I will walk you through a step by step process on how to set up the vision models with Ollama and use it in a RAG system for retrieval augmented generation.
LINKS:
ollama.com/blog/llama3.2-vision
Multimodal RAG: • Goodbye Text-Based RAG, Hello Vision ...
LocalGPT-Vision: github.com/PromtEngineer/localGPT-Vision
💻 RAG Beyond Basics Course:
prompt-s-site.thinkific.com/courses/rag
Let's Connect:
🦾 Discord: discord.com/invite/t4eYQRUcXB
☕ Buy me a Coffee: ko-fi.com/promptengineering
|🔴 Patreon: www.patreon.com/PromptEngineering
💼Consulting: calendly.com/engineerprompt/consulting-call
📧 Business Contact: engineerprompt@gmail.com
Become Member: tinyurl.com/y5h28s6h
💻 Pre-configured localGPT VM: bit.ly/localGPT (use Code: PromptEngineering for 50% off).
Signup for Newsletter, localgpt:
tally.so/r/3y9bb0
00:00 Introduction to OLAMA's New Vision Model Support
00:50 Setting Up OLAMA for Vision Models
01:55 Running and Testing Vision Models
03:18 Advanced Testing and Jailbreaking
06:00 Practical Use Cases and Python SDK Integration
07:38 Multimodal RAG
10:09 localGPT-Vision
All Interesting Videos:
Everything LangChain: • LangChain
Everything LLM: • Large Language Models
Everything Midjourney: • MidJourney Tutorials
AI Image Generation: • AI Image Generation Tutorials
コメント