Stop paying for ChatGPT with these two tools | LMStudio x AnythingLLM

2024 ж. 15 Мам.
215 206 Рет қаралды

In this video, we are installing two user-friendly tools that make downloading, running, and managing a powerful local LLM to replace ChatGPT. Seriously.
Today, with only a desktop computer with a retail GPU, and two free applications you will get a fully private Local LLM + RAG chatbot running in less than 5 minutes!
This is no joke - the teams at LM Studio and AnythingLLM are now fully integrated for your convenience. Run models like LLama-2, Mistral, CodeLLama, and more to make your dreams a reality at no expense of privacy.
LMStudio: lmstudio.ai
AnythingLLM: useanything.com/download
Chapters:
0:00 Introduction to LMStudio x AnythingLLM
0:57 What is AnythingLLM?
1:20 Installing LMStudio
1:53 Installing AnythingLLM
2:10 LMStudio Basic use tutorial
4:28 Testing out our model
5:32 How to level up your LLM chat abilities
6:00 Connecting LMStudio to AnythingLLM
7:53 Send a basic chat on AnythingLLM to our custom model
8:26 Adding knowledge to our LMStudio model
10:08 What the future of chat with local LLMs is going to look like

Пікірлер
  • Please do a dedicated video on training minimal base models for specific purposes. You're a legend. Also a video on commercial use and licensing would be immensely valuable and greatly appreciated.

    @codygaudet8071@codygaudet80712 ай бұрын
    • +1

      @Al-Storm@Al-Storm2 ай бұрын
    • Where to start with in the path of learning AI (llm, rag, generative Ai..,)

      @akram5960@akram59602 ай бұрын
    • +1

      @fxstation1329@fxstation1329Ай бұрын
    • Yes!

      @vulcan4d@vulcan4dАй бұрын
    • Very nice question i am waiting for the same. Wish Tim make that video soon

      @nasirkhansafi8634@nasirkhansafi863411 күн бұрын
  • Thanks mate. Had them up and running in a few minutes.

    @alanmcoll101@alanmcoll1012 ай бұрын
  • Fantastic! I've been waiting for someone to make RAG smooth and easy :) Thank you for the video!

    @autonomousreviews2521@autonomousreviews25212 ай бұрын
  • thanks for the tutorial, everything works great and surprisingly fast on M2 Mac Studio, cheers!

    @TazzSmk@TazzSmk2 ай бұрын
  • Wonderful. Thanks Tim. You have solved a world of problems by publishing this video. I am sure AnythingLLM will keep adding more supported file formats such as excel, csv etc.

    @dipteshbosedb@dipteshbosedb2 ай бұрын
    • waiting for good cdv support

      @JonathanDeCollibus@JonathanDeCollibusАй бұрын
    • CSV

      @JonathanDeCollibus@JonathanDeCollibusАй бұрын
  • I’m just about to dive into LM Studio and AnythingLM Desktop, and let me tell you, I’m super pumped! 🚀 The potential when these two join forces is just out of this world!

    @PCFix41@PCFix41Ай бұрын
  • Excellent tutorial. Thanks a bunch😊

    @olivierstephane9232@olivierstephane92322 ай бұрын
  • Very nice tutorial! Thanks Tim,

    @claudiantenegri2612@claudiantenegri26122 ай бұрын
  • Thank you, I've been struggling for so long with problematic things like privateGPT etc. which gave me headaches. I love how easy it is to download models and add embeddings! Again thank you. I'm very eager to learn more about AI, but I'm absolute beginner. Maybe video on how would you learn from the beginning?

    @bradcasper4823@bradcasper48232 ай бұрын
  • Great stuff,this way you can run a good smaller conversational model like 13b or even 7b,like Laser Mistral. Main problem with this smaller LLM are massive holes in some topics,or informations about events,celebs or other stuff,this way you can make your own database about stuff you wanna chat. Amazing.

    @user-ld8sy9xu2v@user-ld8sy9xu2vАй бұрын
  • Awesome man. Hope to see more video with AnythingLL!

    @williamsoo8500@williamsoo850023 күн бұрын
  • I’d love to hear more about your product roadmap - specifically with how it relates to the RAG system you have implemented . I’ve been experimenting a lot with Flowise and the new LlamaIndex integration is fantastic - especially the various text summarisation and content refinement methods available with a LlamaIndex based RAG. Are you planning to enhance the RAG implementation in AnythingLLM?

    @sitedev@sitedev2 ай бұрын
  • Thank you for making this video. This helped me a lot.

    @cee7004@cee7004Ай бұрын
  • Thank you so much for your generosity. I wish the very best for your enterprise . God Bless!

    @AC-go1tp@AC-go1tp29 күн бұрын
  • Thank you! Very useful info. Subbed.

    @BudoReflex@BudoReflexАй бұрын
  • Thanks a ton ...you are giving us power on working with our local documents... its blazingly fast to embed the docs, super fast responses and all in all i am very happy.

    @vivekkarumudi@vivekkarumudi2 ай бұрын
    • thats liberating ! i was really concerned about privacy especially when coding or working on refining internal proposals> Now I know what to do

      @ashleymusihiwa@ashleymusihiwa2 ай бұрын
    • What type of processor/GPU/model are you using? I'm using version 5 of Mistral and it is super slow to respond. i7 and an Nvidia RTX 3060ti GPU.

      @BarryFence@BarryFenceАй бұрын
  • Thanks for building this.

    @rowbradley@rowbradleyАй бұрын
  • Thanks a lot , very informative , you are my new mentor :)

    @aiforall217@aiforall2172 ай бұрын
  • It's very helpful. Thank you!

    @wingwing2683@wingwing26832 ай бұрын
  • changing the embedding model would be a good tutorial! For examle how to use a multi langual model!

    @MCSchuscha@MCSchuscha2 ай бұрын
  • Thanks for this, about to try it to query legislation and case law for a specific area of UK law to see if it effective in returning references to relevant sections and key case law. Interested in building a private LLM to assist with specific repetitive tasks. Thanks for the video.

    @fieldpictures1306@fieldpictures1306Ай бұрын
  • Thanks a lot! This tutorial is a gem!

    @craftedbysrs@craftedbysrs12 күн бұрын
  • That's really amazing 🤩, I will definitely be using this for BIM and Python

    @thualfiqar87@thualfiqar87Ай бұрын
  • Wow, great information. I have a huge amount of documents and everytime I search for something it's getting such a difficult task to fulfill

    @Helios1st@Helios1st16 күн бұрын
  • This is exactly what I've been looking for. Now, I'm not sure if this is already implemented, but if the chat bot can use EVERYTHING from all previous chats within the workspace for context and reference... My god that will change everything for me.

    @kylequinn1963@kylequinn19632 ай бұрын
    • It does use the history for context and reference! History, system prompt, and context - all at the same time and we manage the context window for you on the backend

      @TimCarambat@TimCarambat2 ай бұрын
    • @@TimCarambatbut isn’t history actually constrained by active model’s context size?

      @IrakliKavtaradzepsyche@IrakliKavtaradzepsyche2 ай бұрын
    • @@IrakliKavtaradzepsyche yes, but we manage the overflow automatically so you at least don't crash from token overflow. This is common for LLMs, to truncate or manipulate the history for long running sessions

      @TimCarambat@TimCarambat2 ай бұрын
  • Thanks dude! Great video

    @Babble_Gum@Babble_GumАй бұрын
  • Brilliant. Thank you.

    @moreloveandjoy@moreloveandjoy2 ай бұрын
  • IMO anythingLLM is much userfriendly and really has big potential. thanks Tim!

    @stanTrX@stanTrX19 күн бұрын
  • AnythingLLM looks super awesome, cant wait to setup with ollama and give it a spin. tried chat with rtx but the youtube upload option didnt install for me and that was all i wanted it for

    @BotchedGod@BotchedGodАй бұрын
  • Just got this running and it's fantastic. Just a note that LM Studio uses the API key "lm-studio" when connecting using Local AI Chat Settings.

    @jimg8296@jimg8296Ай бұрын
    • does it provide script for youtube?

      @thegoat10.7@thegoat10.7Ай бұрын
  • Great overview!

    @boomerstrikeforce@boomerstrikeforce2 ай бұрын
  • Great work Tim, I'm hoping I can introduce this or anything AI into our company

    @drew5834@drew5834Ай бұрын
  • Also, how is this different from implementing RAG on a base foundation model and chunking our documents and loading it into a vector db like pinecone? Is the main point here that everything is locally run on our laptop? Would it work without internet access?

    @continuouslearner@continuouslearner2 ай бұрын
  • Excellent work. Please make a video on text to sql and excel csv sql support for llms and chatbot. Thank you so much ♥️

    @shabbirug@shabbirugАй бұрын
  • This video changed everything for me. Insane how easy to do all this now!

    @karlwireless@karlwirelessАй бұрын
  • Love you for your explain Love from INDIA 😊

    @Namogadget_@Namogadget_Ай бұрын
  • I notice some of the models are 25GB+.. BLOOM, Meta's Llama 2, Guanaco 65B and 33B, dolphin-2.5-mixtral-8x7b etc Do these models require training? If not, but you wanted to train it with custom data, does the size of the model grow, or does it just change and stay the same size? Aside from LMStudio , AnythingLLM , any thoughts on other tools that attempt to make it simpler to get started, like Oobabooga , gpt4all io , Google Colab , llamafile , Pinokio ?

    @bennguyen1313@bennguyen13132 ай бұрын
  • looks soo good! I have a question : is there some way to add chat diagram like voiceflow or botpress ? For example, guiding the discussion for an ecommerce chatbot and give multiple choice when ask questions ?

    @immersift7856@immersift78562 ай бұрын
    • I think this could be done with just some clever prompt engineering. You can modify the system prompt to behave in this way. However, there is no voiceflow-like experience built-in for that. That is a clever solution though.

      @TimCarambat@TimCarambat2 ай бұрын
  • Absolutely stellar video, Tim! 🌌 Your walkthrough on setting up a locally run LLM for free using LM Studio and Anything LLM Desktop was not just informative but truly inspiring. It's incredible to see how accessible and powerful these tools can make LLM chat experiences, all from our own digital space stations. I'm particularly excited about the privacy aspect and the ability to contribute to the open-source community. You've opened up a whole new universe of possibilities for us explorers. Can't wait to give it a try myself and dive into the world of private, powerful LLM interactions. Thank you for sharing this cosmic knowledge! 🚀👩‍🚀

    @cosmochatterbot@cosmochatterbot2 ай бұрын
  • @Tim, this episode is brilliant! Let me ask you one thing. Do you have any ways to force this LLM model to return the response in a specific form, e.g. JSON with specific keys?

    @TheDroppersBeats@TheDroppersBeatsАй бұрын
  • I mean this is pretty useful already, is there plans to increase the capabilities to include other formats of documents, images, etc?

    @WestW3st@WestW3st2 ай бұрын
  • I had a spare 6800xt sitting around that had been retired due to overheating for no apparent reason, as well as a semi-retired ryzen 2700x , and i found 32 gigs of ram sitting around for the box. Just going to say flat out that it is shockingly fast. I actually think running Rocm to enable gpu acceleration for lm studio is running llm's better than my 3080ti in my main system, or at the very least, so similar i cant perceive a difference

    @djmccullough9233@djmccullough9233Ай бұрын
  • So if in case we need to programmatically use this, does anythingllm itself offer a ‘run locally on server’ option to get an API endpoint that we could call from a local website for example? i.e. local website -> post request -> anythingllm (local server + PDFs)-> LMstudio (local server - foundation model)

    @continuouslearner@continuouslearner2 ай бұрын
    • Did you get an answer?

      @clinbrokers@clinbrokers2 ай бұрын
  • software engineer and AI knowledge? You got my sub.

    @xevenau@xevenau2 ай бұрын
  • LM Studios TOS paragraph: "Updates. You understand that Company Properties are evolving. As a result, Company may require you to accept updates to Company Properties that you have installed on your computer or mobile device. You acknowledge and agree that Company may update Company Properties with or WITHOUT notifying you. You may need to update third-party software from time to time in order to use Company Properties. Company MAY, but is not obligated to, monitor or review Company Properties at any time. Although Company does not generally monitor user activity occurring in connection with Company Properties, if Company becomes aware of any possible violations by you of any provision of the Agreement, Company reserves the right to investigate such violations, and Company may, at its sole discretion, immediately terminate your license to use Company Properties, without prior notice to you." Several posts on LLM Reddit groups with people not happy about it. NOTE: I'm not one of the posters, read-only, I'm just curious what others think.

    @LiebsterFeind@LiebsterFeind2 ай бұрын
    • Wait so their TOS basically says they may or may not monitor your chats in case you are up to no good with no notification? okay. I see why people are pissed about that. I dont like that either unless they can verifiable prove the "danger assessment" is done on device because otherwise this is no better than just cloud hosting but paying for it with your resources

      @TimCarambat@TimCarambat2 ай бұрын
    • Thanks for bringing this to my attention btw. I know _why_ they have it in the ToS, but I cannot imagine how they think that will go over.

      @TimCarambat@TimCarambat2 ай бұрын
    • Ancient idea clash between wanting to be a good "software citizen" and the unfortunate fact that their intent is still to "monitor" your activities. As you said in your second reply to me, "monitoring" does not go over well with some and the consideration of the intent for doing so, even if potentially justified, is a subsequent thought they will refuse to entertain. @@TimCarambat

      @LiebsterFeind@LiebsterFeind2 ай бұрын
    • ​@@TimCarambatLet's say there is a monitoring background behind, what if we setup a vm that did not allow to connect to the internet, does that will make our data safe ?

      @alternate_fantasy@alternate_fantasyАй бұрын
    • @@alternate_fantasy it would prevent phone homes, sure, so yes. That being said I have Wiresharkd lmstudio while running and did not see anything sent outbound that would indicate they can view anything like that. I think that's just their lawyers being lawyers

      @TimCarambat@TimCarambatАй бұрын
  • This is an amazing tutorial. Didn't know there were that many models out there. Thank you for clearing the fog. I have one question though, how do I find out what number to put into "Token context window"? Thanks for your time!

    @djkrazay7791@djkrazay7791Ай бұрын
    • Once pulling into LMStudio, its in the sidebar once the model is selected. Its a tiny little section on the right sidebar that say "n_ctxt" or something similar to that. Youll then see it will explain further how many tokens your model can handle at max, RAM permitting.

      @TimCarambat@TimCarambatАй бұрын
    • @@TimCarambat your the best... thanks... 🍻

      @djkrazay7791@djkrazay7791Ай бұрын
  • Thank you so much for the concise tutorial. Can we use ollama and LM studio as well with AnythingLLM. It only takes either of them. I have some models in ollama, and some in LM. Would love to have them both in AnythingLLM. I don't know if this is possible though. Thanks!

    @fxstation1329@fxstation1329Ай бұрын
  • Nice one Tim. It’s been on my list to get a private LLM set up. You’re guide is just what I needed. I know Mistral is popular. Are those models listed on capabilities, top being most efficient? I’m wondering how to choose the best model for my needs.

    @Chris.888@Chris.8882 ай бұрын
    • Those models are curated by thr lmstudio team. Imo they are based on popularity. However, if you aren't sure what model to chose, go for Llama2 or Mistral, can't go wrong with those models as they are all around capable

      @TimCarambat@TimCarambat2 ай бұрын
    • Thanks Tim, much appreciated.

      @Chris.888@Chris.8882 ай бұрын
  • Very useful video!! Thanks for the work. I kept a doubt about the chats that take place, there is any registration of the conversations? For commercial purposes it will be nice to generate leads with the own chat!

    @HugoRomero-mq7om@HugoRomero-mq7om22 күн бұрын
    • Absolutely, while you can "clear" a chat window you can always view all chats sent as a system admin and even export them for manual analysis or fine-tuning.

      @TimCarambat@TimCarambat21 күн бұрын
  • Can't wait to try this. I've watched a dozen other tutorials that were too complicated for someone like me without basic coding skills. What are the pros/cons of setting this up with LMStudio vs. Ollama?

    @MrAmirhk@MrAmirhk19 күн бұрын
    • If you don't like to code, you will find the UI of lmstudio much more approachable, but it can be an information overload. Lmstudio has every model on huggingface. Ollama is only accessible via terminal and has limited model support but is dead simple. This video was made before we launched the desktop app. Our desktop comes with ollama pre-installed and gives you a UI to pick a model and start chatting with docs privately. That might be a better option since that is one app, no setup, no cli or extra application

      @TimCarambat@TimCarambat19 күн бұрын
  • Great content❤

    @atharvadeshpande6647@atharvadeshpande66472 ай бұрын
  • thanks, i love it. beautiful work

    @lalamax3d@lalamax3dАй бұрын
  • I get this response every time: "I am unable to access external sources or provide information beyond the context you have provided, so I cannot answer this question". Mac mini M2 Pro Cores:10 (6 performance and 4 efficiency) Memory:16 GB

    @jakajak1991@jakajak199110 күн бұрын
  • Hello Tim, you can make a video connecting Ollama with AnythingLLM?

    @milorad9301@milorad93012 ай бұрын
  • Can you do more of these demonstrations or vidoeos, is anythingLLM capable of generating visual conent like a dalle3 or video, assuming using a capable open sourse modell is there a limitation other then local memory as to the size of the vector databases created? this is amazing ;) Thanks for this video truly appreciated man. Liked and subscrided to support you.

    @FisVii77@FisVii772 ай бұрын
  • Im on a Linux machine, and want to set up some hardware ... recommended GPU (or even point me to the direction for good information?) or better yet can an old bitcon rig do the job somehow seeing as though theyre useless for bitcoin these days?! Great tutorial too mate, really appreciate you taking the time!

    @CaptZenPetabyte@CaptZenPetabyteАй бұрын
  • Thanks for the insights. What's the best alternative for a person who doesn't want to run locally yet he wants to use opensource LLMs for interacting with documents and webscraping for research.

    @NaveenKumar-vj9sc@NaveenKumar-vj9sc2 ай бұрын
    • OpenRouter has a ton of hosted open-source LLMs you can use. I think a majority of them are free and you just need an API key.

      @TimCarambat@TimCarambat2 ай бұрын
  • Very cool, I'll check it out. Is there a way to not install this on your OS drive?

    @Al-Storm@Al-Storm2 ай бұрын
  • Thank you!

    @YannMetalhead@YannMetalheadАй бұрын
  • Thanks, Tim, for the good video. Unfortunately I do not get good results for uploaded content. I'm from Germany, so could it be a language problem, cause the uploaded content is german text? I'm using the same mistral model from your video and added 2 web pages to anythingLLMs workspace. But I'm not sure if the tools are using this content for building the answer. In the LM studio log I can see a very small chunk of one of the uploaded web pages. But in total, the result is wrong. To get good embeddings values I downloaded nomic-embed-text-v1.5.Q8_0.gguf and use it for the Embedding Model Settings in LM Studio which might be not necessary, cause you didn't mentioned such steps in your video. I would appreciate any further hints. Thanks a lot in advance.

    @uwegenosdude@uwegenosdude29 күн бұрын
  • This is an amazing video and exactly what Ineeded. Thank you! I really appreciate it. Now the one thing,how do I find the token context window for the different models? I'm trying out gemma?

    @Jascensionvoid@Jascensionvoid2 ай бұрын
    • up to 8,000 (depends on VRAM available - 4096 is safe if you want best performance). I wish they had it on the model card on HuggingFace, but in reality it just is better to google it sometimes :)

      @TimCarambat@TimCarambat2 ай бұрын
    • I gotcha. So for the most part, just use the recommended one. I got everything working. But I uploaded a PDF and it keeps saying I am unable to provide a response to your question as I am unable to access external sources or provide a detailed analysis of the conversation. But the book was loaded and moved to workspace and save and embed? @@TimCarambat

      @Jascensionvoid@Jascensionvoid2 ай бұрын
    • For what its worth in LM Studio, on the sidebar there is a `n_cntxt` param that shows the maximum you can run. Performance will degrade if your GPU is not capable though to run the max token context.

      @TimCarambat@TimCarambat2 ай бұрын
  • Hi Tim, Fantastic. Is it possible to use anythingllm with gpt4 directly, for local use? like the example you demonstrated above.

    @Djk0t@Djk0t2 ай бұрын
    • Can't imagine that's possible with GPT4. The VRAM requires for that model would be in the hundreds of GB.

      @thedeathcake@thedeathcake2 ай бұрын
  • Very helpful video. I'd love to be able scrape an entire website in Anything LLM. Is there a way to do that? Is there a website where I can ask help questions about Anything LLM?

    @stevekirsch8284@stevekirsch828416 күн бұрын
  • Does AnythingLLM have its own version? How I recognize since time how relevant my installed version is or I should update to receive some new features...

    @iiiBog@iiiBog2 ай бұрын
  • thats great, I was getting tired of the restrictions in the common AI platforms

    @CrusaderGeneral@CrusaderGeneralАй бұрын
  • Use with source code? This might be ideal, can this understand my VS database and ETL projects? Asking questions of source would be my #1 usecase. I've installed, and 2 things...I don't see a way to point it at a folder - or do I have to injest each file individually, and second when navigating into the code folder, the SQL, dtsx, ps1 type files are filtered out. I'm assuming that means it won't work with those filetypes - which would rule it out for a code chatbot...

    @TrevorMatthews@TrevorMatthews2 ай бұрын
  • Hi Tim, What GPU have you got ?

    @temp911Luke@temp911Luke2 ай бұрын
  • Very nice. Will definitely try it. Is or will there be an option to integrate a anything LLM workspace in a python code to automate task via API?

    @MrNatzu@MrNatzu2 ай бұрын
    • Yes, but the api is only in the docker version currently since that can be run locally and on cloud so an API makes more sense for that medium

      @TimCarambat@TimCarambat2 ай бұрын
  • You are a superhero!

    @FlutterXP@FlutterXP23 күн бұрын
  • To add PDFs in the chat and make a Pools of knowledge to select from would be great.

    @nightmisterio@nightmisterioАй бұрын
  • Hi Tim I am farley new to this. But going to as a silly question. will this method have up to date information knowledge similar to GPT4 using bing etc.... ? Thans this is a great video!

    @gigawerx@gigawerx2 ай бұрын
    • The model cutoff dates vary, so no. however! We are going to be adding live search via various means (from free to connecting to an external paid services). Live web browsing is the term for this. Some models are even "online" and have this baked in, but they are not private or something you can run yourself. See perplexity AI for that kind of functionality. We want to unlock this for local LLMs for you though

      @TimCarambat@TimCarambat2 ай бұрын
  • How well does it perform on large documents. Is it prone to lost in the middle phenomena?

    @Augmented_AI@Augmented_AIАй бұрын
    • That is more of a "model behavior" and not something we can control.

      @TimCarambat@TimCarambatАй бұрын
  • This is superb. Would it be possible to provide Web interface to small number of people? Protected by username and password? Do you provide paid webinars? Amazing stuff.

    @maximmono1@maximmono12 ай бұрын
  • Looks really clean thank you ! quick question wanted to test with a 50mb txt log file but after sometime embedding got an error cannot create a string longer than 0x1 and didn't catch the rest any thoughts on how I could add big log files ? Used the default embedder and vector store with ollama codellama 7b

    @mohamedemarah9379@mohamedemarah9379Ай бұрын
    • That is a bizzare error i have never seen. What operating system?

      @TimCarambat@TimCarambatАй бұрын
  • Downloaded. I'm excited to install on my 3 GPU rig.

    @s.c.r.e.a.m.forp.e.a.c.e.3541@s.c.r.e.a.m.forp.e.a.c.e.3541Ай бұрын
  • I want to try it in a Linux VM, but from what I see you can only make this work on a laptop with a desktop OS. It would be even better if both LMstudio and AnythingLLM could run in one or two separate containers with a web UI

    @Equality-and-Liberty@Equality-and-LibertyАй бұрын
  • Awesome!

    @dadlord689@dadlord6892 ай бұрын
  • works for understand code projects?

    @adriAM55@adriAM552 ай бұрын
  • Thanks for this video. Does AnythingLLM a good job in terms of writing SEO optimised blog articles?

    @MarsX69@MarsX692 ай бұрын
    • It is a model that you choose in ML Studio that affects the output.

      @IrakliKavtaradzepsyche@IrakliKavtaradzepsyche2 ай бұрын
  • Thanks for the great tutorial! I was hoping I would be able to dump a github repo in and talk to it about entity framework relationships. Saw .cs and other files are not valid. I was going to iteratively go through my repo and dump all the source into a single .txt file, then put that in the vectordb. Have you a better solution?

    @brettgarnier107@brettgarnier1072 ай бұрын
    • We have a Github repo connector that can pull an entire repo!

      @TimCarambat@TimCarambat2 ай бұрын
    • I spent a while getting that to work with my private repo. I made the temp token and gave it read access to my files. I paste my url, and access token, "Could not prepare github repo for lading! Check URL" is what I get every time, tried for an hour @@TimCarambat

      @brettgarnier107@brettgarnier1072 ай бұрын
  • Thank you

    @valenciawalker6498@valenciawalker64982 ай бұрын
  • Hey, Great video. For some reason I don't have LMStudio as an optional provider in AnythingLLM, any thoughts? thanks.

    @JasonStorey@JasonStorey20 күн бұрын
    • That certainly isn't right... Where are you in the UI where you do not see lmstudio?

      @TimCarambat@TimCarambat20 күн бұрын
  • which local llm is the best and have largest context token window? Mine only shows around 2k context token window

    @zeroxinfinity5816@zeroxinfinity5816Ай бұрын
  • Hmm using it on a Mac, on ollama it seems faster than LM Studio, but the main issue seems that besides the AnythingLLM question it seems not to remember the content in a reliable way. It seems to not involve local docs all the time. I did threw away all the settings an started over from scratch … it’s more a miss than a hit?!

    @yvesvandenbroek6055@yvesvandenbroek60552 ай бұрын
  • you are amazing 🎉🎉🎉

    @apricotcomputers3943@apricotcomputers3943Ай бұрын
  • Thanks a lot for this amazing tutorial. Unluckily, I've tried on my pc(no gpu) and it takes more than 30 seconds to generate the first token of the 'how are you' query. Do you think the wait is justified by the fact that using coding specific models, I can just make progress faster? I'm asking this because even when using gpt-4, often I have to correct the work it does several times? Thanks again for the creation of AnythingLLM

    @zacca03@zacca032 ай бұрын
    • It's not atypical. My MacBook (Intel cpu) takes about that long for even a single token query. It's rough. More token = longer wait. Model size does impact the time to load but smaller models can often perform work. Unfortunately with a great gpu you may be limited to cloud based models if you want really concise response that are accurate. That's just the state of things currently

      @TimCarambat@TimCarambat2 ай бұрын
    • ​@@TimCarambatI'm studying to become a mechanical engineer, so it's not really my field. But something that comes to my mind is the following: if you cannot exploit the GPU, why not exaggerate on RAM or SWAP usage? I have 16gb of RAM with 512gb of NVME storage, there's plenty to be given to AnythingLLM but mostly LMStudio. Seeing your brilliance developing AnythingLLM, I'm sure you will figure out something. For me, having a laptop and being a heavy gpt user, it's not cost effective to actually get another PC for this usage. You have also to account for probably expensive energy bills. Still, the project has a LOT of potential! Keep it up with your work mate!

      @zacca03@zacca032 ай бұрын
  • Thanks for the video! I did it as you said and got the model working (same as you picked). It ran faster than i expected and I was impressed with the quality of the text and the general understanding of the model. However when i uploaded some documents [in total just 150 kb of downloaded HTML from a wiki] it gave very wrong answers [overwhelmingly incorrect]. What can i do to improve this?

    @Mursaat100@Mursaat100Ай бұрын
    • two things help by far the most! 1. Changing the "Similarity Threshold" in the workspace settings to be "No Restriction". This basically allows the vector database to return all remotely similar results and no filtering is applied. This is based purely on the vector database distance of your query and the "score" filtered on, depending on documents, query, embedder, and more variables - a relevant text snippet can be marked as "irrelevant". Changing this setting usually fixes this with no performance decrease. 2. Document pinning (thumbtack icon in UI once doc is embedded). This does a full-text insertion of the document into the prompt. The context window is managed in case it overflows the model, however this can slow your response time by a good factor, however coherence will be extremely high.

      @TimCarambat@TimCarambatАй бұрын
    • Thank you! But i dont understand what you mean with "Thumbtack icon in UI once doc is embedded". Could you please clarify?@@TimCarambat

      @Mursaat100@Mursaat100Ай бұрын
  • Nice work! Unfortunately, after installing anythingllm on windows 10, and running it, its stuck at "loading workspaces".

    @reinerheiner1148@reinerheiner11482 ай бұрын
  • This video is gold. Push this to the top people.

    @codygaudet8071@codygaudet80712 ай бұрын
  • What is best for translating from english to another languages?

    @anysobyspbyany7299@anysobyspbyany72992 ай бұрын
  • hey tim im on debian, lm studio runs and wonderfully well too, however im having a small issue, on the sidebar the only icon that isnt a weird square is the local server icons... what icon pack or font do i need from the repo?

    @jongdonglu@jongdongluАй бұрын
    • phosphor icons

      @TimCarambat@TimCarambatАй бұрын
  • Cool! Is there any AI tool that can learn something when being fed with some websites, then can answer my questions about stuff mentioned on those websites? For example, I want to learn Rust (a programming language). I can give that AI tool websites about the language, the libraries in that language ... Then the AI tool must be able to write some Rust applications when being given enough details. Is it feasible now or we need to wait for a few more years/decades?

    @avalagum7957@avalagum7957Ай бұрын
    • You could do that exact thing now. There is a website scraper in the app currently

      @TimCarambat@TimCarambatАй бұрын
  • Great stuff, my man! If you don't mind sharing, what is your business model with AnythingLLM?

    @MoeShlomo@MoeShlomo2 ай бұрын
    • We offer a hosted multi-user version of AnythingLLM that is managed on our cloud account. We additionally offer a services arm for doing on-premise installation of our product alongside setting up a local (or remote) LLM. The desktop app as it stands will be free. It costs us nothing to run desktop apps as it runs privately on your machine. If there is ever anything we need to host for functionality - that may be paid. However, I am very adamant having a kickass desktop LLM experience should be accessible and free since its using your compute resources and can maintain your level of privacy.

      @TimCarambat@TimCarambat2 ай бұрын
    • To also be clear. If there is some external functionality for the desktop app that we have to host and is paid it would not be core to the functionality of the app and would be a plug-in or opt-in thing. The other paid segments will hopefully prove to be enough for a sustainable model without getting in the way of our core offering so everyone can be happy.

      @TimCarambat@TimCarambat2 ай бұрын
  • Can any of these locally installed llm's use the internet to search or in any other way?

    @MikeG-js1jt@MikeG-js1jt10 күн бұрын
  • Can this use plugins and such like oobaboga's web ui?

    @SK-gc7xv@SK-gc7xv2 ай бұрын
  • unable to run you module on LMStudio with my 12thgen I3 adn 8 gb ram..whats the least required configuration?

    @krunalvyas3141@krunalvyas3141Ай бұрын
  • How do I go in an change the LS-Studio base model URL and Token context window in AnythingLLM after install.

    @mstew8386@mstew8386Ай бұрын
  • That's a great one .. Just got stucked in one scenario after some time of use on asking any questions giving respone- could not respond to messages. Request failed with status code 400. Please help!

    @bhushan80b@bhushan80b27 күн бұрын
  • I am trying to access pdf and documentation present on a website I have given AnythingLLM, but it seems not to be working. Is it possible to do so, or do I need to manually download them from the website and attach them in AnythingLLM?

    @avantigaming1627@avantigaming16272 ай бұрын
  • hello i was wondering if there is away to use the new lm studio muli mode . i tried to request and its saying that it needs the model specified is there any way to put this parameter

    @deandresnago2796@deandresnago2796Ай бұрын
KZhead