Loading...
「ツール」は右上に移動しました。
利用したサーバー: natural-voltaic-titanium
496いいね 19836回再生

Ollama with Vision - Enabling Multimodal RAG

Ollama just added support for llama3.2 vision models. This enables models to see and process images. In this video, I will walk you through a step by step process on how to set up the vision models with Ollama and use it in a RAG system for retrieval augmented generation.


LINKS:
ollama.com/blog/llama3.2-vision
Multimodal RAG:    • Goodbye Text-Based RAG, Hello Vision ...  
LocalGPT-Vision: github.com/PromtEngineer/localGPT-Vision


💻 RAG Beyond Basics Course:
prompt-s-site.thinkific.com/courses/rag

Let's Connect:
🦾 Discord: discord.com/invite/t4eYQRUcXB
☕ Buy me a Coffee: ko-fi.com/promptengineering
|🔴 Patreon: www.patreon.com/PromptEngineering
💼Consulting: calendly.com/engineerprompt/consulting-call
📧 Business Contact: engineerprompt@gmail.com
Become Member: tinyurl.com/y5h28s6h

💻 Pre-configured localGPT VM: bit.ly/localGPT (use Code: PromptEngineering for 50% off).

Signup for Newsletter, localgpt:
tally.so/r/3y9bb0


00:00 Introduction to OLAMA's New Vision Model Support
00:50 Setting Up OLAMA for Vision Models
01:55 Running and Testing Vision Models
03:18 Advanced Testing and Jailbreaking
06:00 Practical Use Cases and Python SDK Integration
07:38 Multimodal RAG
10:09 localGPT-Vision


All Interesting Videos:
Everything LangChain:    • LangChain  

Everything LLM:    • Large Language Models  

Everything Midjourney:    • MidJourney Tutorials  

AI Image Generation:    • AI Image Generation Tutorials  

コメント