FAST Local Live LLM Preview Window - Phi-2 / Mistral 7B Uncensored
2024 ж. 20 Мам.
8 626 Рет қаралды
FAST Local Live LLM Preview Window - Phi-2 / Mistral 7B Uncensored
👊 Become a member and get access to GitHub:
/ allaboutai
🤖 AI Engineer Course:
scrimba.com/?ref=allabtai
Get a FREE 45+ ChatGPT Prompts PDF here:
📧 Join the newsletter:
www.allabtai.com/newsletter/
🌐 My website:
www.allabtai.com
In this video i create a project where we get a real time live preview of the llm output, running on local models like uncensored Phi-2 and Mistral 7B. Very fun and simple project.
00:00 Local Live LLM Preview Window Intro
00:20 Flowchart
01:32 Python Code
04:01 Live LLM Preview Test 1
04:57 Live LLM Preview Test 2
05:22 Live LLM Preview Test 3
I think that should officially replace "Hello World" from now on. 😆👍
😂 the first email was hilarious!!
Beautiful
Graet - thank you:)
🎯 Key Takeaways for quick navigation: 00:27 🐍 *Developed a local real-time LLM (Language Model) preview using Python and threading.* 01:23 🔄 *Parallel function allows capturing and processing keyboard inputs without interruption.* 01:49 📝 *Python code includes M7B function for local LM Studio, update preview function, and capture input function.* 03:27 🗣️ *System prompts for the Mistral 7B model include examples for a 4chan Reddit style and a more explicit chatbot style.* 04:09 🖥️ *Demonstrates the local real-time preview in action with user input and model responses.* 05:32 🔄 *Switches the model to Open Hermis Mistral 7P, adjusts settings, and changes the system prompt for a different tone.* 06:28 📧 *Tests the modified model by writing a short, humorous, and explicit email using keyboard input.* 08:31 💻 *Encourages support through channel membership for access to scripts and the community GitHub and Discord.*
I've had great results with the beagle models! Even the 3 bit version does better than some of the big models for things like evaluating RAG results.
How does it feel to be a demigod? Thank you for sharing your spells!
hiliarious concept
Great content. I'm LMFAO.
Can we do a real-time llm that works with ollama STT then Ollama then TTS in the terminal. Ollama is the most optimized llm solution.
check out twinny
Just use the ollama end point. All done
@Canna_Science_and_Technology with ollama, don't you need additional headers? For model and pre-prompt
@Canna_Science_and_Technology curl localhost:11434/api/generate -d '{ "model": "tinydolphin", "prompt": "As a friendly and informative assistant, provide detailed explanations. This is a test. Output:", "options": { "stop": ["Instruct:", "Output:"] }, "raw": true, "stream": true }'
@@Edward_ZS yes, host.docker.internal:11434/api
Where do I find dolphin phi 2 q5 k m
really love the web troll lol
you obviously need a debounce on that input
I feel like this would have been much better in streamlit