Create fine-tuned models with NO-CODE for Ollama & LMStudio!

  Рет қаралды 42,826

Tim Carambat

Tim Carambat

Күн бұрын

Пікірлер: 110
@richardallison1576
@richardallison1576 4 ай бұрын
This is amazingly simple! Great job! But I have the horse power so PLEASE get the locally run version out soon!! 👍🏻
@okra3000
@okra3000 Ай бұрын
You're an excellent teacher. Thanks
@revalenz-wissenswerkstatt
@revalenz-wissenswerkstatt 4 ай бұрын
great🔥fantastic what the Anythingllm team develops and makes available open-source 👍now I understand what the thumbs up button is for, that already exists in the workspace chats for the language model answers
@vulcan4d
@vulcan4d 3 ай бұрын
Very nice, though I would love to see more local functionality since that is really the goal of running something local. If I wanted cloud I would stick with Claude or OpenAI and fund their empire ;)
@coltnavydaddy4261
@coltnavydaddy4261 Күн бұрын
hello tim. i am pretty interested in the possibilities anythingLLM and Ollama presents, but the question for you (as a person with lots of experience in my book): if I want to make an agent, who would analyse some twitter threads / post in a theme (let's say, nature, or sports), and then write / reply based on what was read. what would be the tool or a service to make such a beast? would it be done greatly in a no-code pipeline (many of them, possibly), or there would be an easier way around?
@PiyapongBoonrit
@PiyapongBoonrit 25 күн бұрын
So good!! Thank You Man.
@joe_dearman
@joe_dearman 4 ай бұрын
Sounds powerful. Looking forward to checking this out.
@drp111
@drp111 3 ай бұрын
This is an interesting approach to get local llms be of higher usefulness. The 250 for the service sound reasonable if the accessibility for lesser technical users or user with limited time is given.
@TimCarambat
@TimCarambat 3 ай бұрын
@@drp111 that is great feedback, as that really is my core intention with that. I shopped around myself and was quoted as low as 1k to much much more and sometimes I never even get the model back. The local training stuff is in the works but I already foresee someone getting issues thinking a 10 year old gpu should be able to do this, which is why we offered cloud first since it for sure works for everyone
@theuniversityofthemind6347
@theuniversityofthemind6347 7 күн бұрын
Hey Tim, great video! I have a question: I write short 10-minute stories, all based on a five-step storytelling structure. Is it possible to fine-tune a model so that, whenever I ask it to write a story on a given topic, the model consistently follows this five-step process?
@mad7505
@mad7505 3 ай бұрын
Cant wait to be able to train my models locally through AnthingLLM
@drpchankh
@drpchankh 4 ай бұрын
Thanks for sharing, please do help zoom into the screen text as you elaborate. The font size is really too small and the screen estate is not well used. Appreciate the sharing.
@beissel_glitch
@beissel_glitch 3 ай бұрын
Can you make a complete video on the installation of Docker Anything LLM please ??
@TimCarambat
@TimCarambat Ай бұрын
Its just a single command! If you have docker in your terminal just run this command: docs.anythingllm.com/installation-docker/local-docker#recommend-way-to-run-dockerized-anythingllm
@j0hnc0nn0r-sec
@j0hnc0nn0r-sec 4 ай бұрын
9:53 I have a large obsidian notebook (.Md files) of about 10,000 notes. Can I use these notes instead of the “chats” when doing the fine tune?
@TimCarambat
@TimCarambat 4 ай бұрын
The next version of this fine-tuning will indeed enable "Raw-document" tuning so you don't need to chat. Totally understand making chats seems annoying and direct content tuning would be much faster! One step at a time :)
@j0hnc0nn0r-sec
@j0hnc0nn0r-sec 4 ай бұрын
@@TimCarambat hell yeah
@j0hnc0nn0r-sec
@j0hnc0nn0r-sec 4 ай бұрын
@@TimCarambat I suppose it would be a good idea if I exported my chats from OpenAI and Anthropic and used those chats for fine tuning.
@TimCarambat
@TimCarambat 4 ай бұрын
@@j0hnc0nn0r-sec Soon you should be able to "port" those messages in as well so you can continue chatting in AnythingLLM as well. But yes, we could tune off that file directly as well
@atanudasgupta
@atanudasgupta 3 ай бұрын
this is a great tool indeed, installed it and works great !. Could you help me understand how you got Ollama models to have function call to work, I find it errors out when I use with langchain. Appreciate your help. It seems to work ok in Anything LLM, at least better than using langchain
@EchoesOfTomorrowFilms
@EchoesOfTomorrowFilms 3 ай бұрын
love it!!
@alandelagarza463
@alandelagarza463 13 минут бұрын
Thanks.
@mehdihachana_3138
@mehdihachana_3138 4 ай бұрын
This is awesome keep it up!
@RadioDJDude
@RadioDJDude 20 күн бұрын
Great content! Thanks for "birthing" this cool tech. I'm a documentarian with a library of interview transcripts. I'm trying to find the best local method to create a RAG to search and interact with them. Currently, they're PDFs and CSV files. I installed AnythingLLM and have been getting some lackluster results. Probably my fault! What's the best set-up for my use case re. LLM, embedder, vector db, etc? Currently, I'm using all the native options and I find the RAG is hallucinating a lot and/or not showing all results of a search term. Any tips would be appreciated! Thank you!
@TimCarambat
@TimCarambat 19 күн бұрын
Sure! So PDFs are pretty cut and dry and the defaults work. CSVs and other tabular data though are HARD. The nature of CSV's often requires full-document comprehension which basically means you need a massive model (Google Gemini, Anthropic) to digest them. The alternative is to load them into a proper database so at least relationships exist. CSV's cannot be effectively used in RAG since there is no "semantic" relationships among datapoints for the most part. This makes "chunking" effectively worthless. Checkout Document pinning here for solving CSV issues, you may have to use a cloud model for those since they are probably thousands of lines long docs.anythingllm.com/llm-not-using-my-docs
@RadioDJDude
@RadioDJDude 19 күн бұрын
@@TimCarambat Thanks so much for your reply, Tim! I first started with PDFs, but because the timecode info wasn't consistent across all PDFs, I converted them into 3 column CSVs. In my system prompt, I explain how the CSVs are structured and where to look for the data. But, you're saying, even with my instructions, CSVs are still difficult to work with? Side note...I'm about to start creating content for my filmmaker's channel. I know transcript wrangling is a popular topic/pain point. I'm sure you're super busy, but if you'd be interested in doing a quick interview, perhaps we can shine the light on a local llm solution for filmmakers who typically wouldn't consider it. Let me know!
@aauu6896
@aauu6896 3 ай бұрын
The training dataset that is sent, consist of only chat record or including reference dataset too? Why fine tuned LLM size doubled? If the fine tune data is increased causing fine tuned LLM increasing too? Thank you.
@matthewmuller
@matthewmuller 3 ай бұрын
Is there a way to include the knowledge from the RAG assets into fine tuning? Can you fine tune with multi modal assets?
@muraliytm3316
@muraliytm3316 3 ай бұрын
You should add an check for updates feature, it is difficult to uninstall and download everytime
@jordyvu999
@jordyvu999 2 ай бұрын
thank your awesome post, just clarify: though with just 14 dataset, but during your training process, you use RAG so that it can expand knowledge in 14 data and can respond with longer text, am I right?
@TimCarambat
@TimCarambat Ай бұрын
if i understand correctly, yes
@Gravitized
@Gravitized 3 ай бұрын
Thank you!
@nobady6022
@nobady6022 4 ай бұрын
I would like to do this locally
@igorshingelevich7627
@igorshingelevich7627 4 ай бұрын
Do yousee his face expression when he explained why you do need to pay for remote processing. 😂
@CenterCred
@CenterCred 4 ай бұрын
what time
@Automan-AI
@Automan-AI 2 ай бұрын
bravo
@shawngoodin
@shawngoodin 4 ай бұрын
This is great will this extend into creating LoRAs?
@TimCarambat
@TimCarambat 4 ай бұрын
It is a LoRA! If you need the .safetensors we can also provider those for export as well since they tend to be more portable than a hefty GGUF. I just thought most people would get more confused with all the files for a LoRA and unsure how to use them
@DeeKrypted
@DeeKrypted 4 ай бұрын
How did you get the localhost:3000? Love the app and the content, will test this later !!
@TimCarambat
@TimCarambat 4 ай бұрын
Using our docker image! hub.docker.com/r/mintplexlabs/anythingllm
@airforce__ro
@airforce__ro Ай бұрын
One point if is possible . I'm not coder but i try to create a python script to store my "personal behavior" in a plain text/texts in a folder to be loaded automatic by model, like that every time i load the model, he will know my old conversations, know my plans , my direction for a specific field . Of course , in time the model should try to sort the data from files in a manner that made him more useful in responses. Obviously didn't work well 😐
@greensynergy
@greensynergy 3 ай бұрын
waiting for local install video
@jwickerszh
@jwickerszh 4 ай бұрын
Any thoughts on using LORA adapters?
@shubhamkhichi_cyberagi
@shubhamkhichi_cyberagi 4 ай бұрын
@Tim Does the fine-tune pipeline take the documents and websites you vectorized into consideration or just the chat? If its the former and latter then its super powerful, if its just the latter then I have to generate thousands of chats to just get close to an 800 page book that I vectorized in my knowledge base.
@shubhamkhichi_cyberagi
@shubhamkhichi_cyberagi 4 ай бұрын
Also second question: Can I import chats from my chatGPT export where I have been chatting about technical knowledge for a while now?
@TimCarambat
@TimCarambat 4 ай бұрын
@@shubhamkhichi_cyberagi for now it's the chats WITH citations but we are working on a way to generate data direct from documents. Believe me I know it's tedious to generate the chats. Next iteration is raw docs + chats!
@TimCarambat
@TimCarambat 4 ай бұрын
@@shubhamkhichi_cyberagi I didn't even know this was a thing. You can export data chats from chat gpt?
@shubhamkhichi_cyberagi
@shubhamkhichi_cyberagi 4 ай бұрын
@@TimCarambat Yup click on your profile -> Settings -> Data Controls -> Export Data. You get multiple json as well as an HTML file I have chat dating back 2023 March which have gold in it. I need it out and imported to AnythingLLM
@shubhamkhichi_cyberagi
@shubhamkhichi_cyberagi 4 ай бұрын
@@TimCarambat Yes its under Profile -> Settings -> Data control -> Export Chat
@roygatz
@roygatz 4 ай бұрын
One technical quesion, I learned that for training, it's better to use the higher precision version if not full prevision for better quality. Is it because the 8b model is so small, a quantized version does not make any meaningful difference in terms of quality?
@TimCarambat
@TimCarambat 4 ай бұрын
Correct, with smaller param models (
@zbyszeklupikaszapl3280
@zbyszeklupikaszapl3280 4 ай бұрын
Can I train that on pictures and video data?
@PredictAnythingSoftware
@PredictAnythingSoftware 4 ай бұрын
Its a language model, so the answer is no. Unless you use a vision model like llava, but it uses different dataset format structure..
@monbeauparfum1452
@monbeauparfum1452 Ай бұрын
Can I finetune models from the cloud version? I have absolutely no coding skills.
@TimCarambat
@TimCarambat 24 күн бұрын
Yes, but i would recommend still using the desktop app so you can easily load that model in locally once its ready
@SiliconSouthShow
@SiliconSouthShow 4 ай бұрын
Now that the agent is working with ollama i dont think ill log out of anythingllm, lol
@TimCarambat
@TimCarambat 4 ай бұрын
@@SiliconSouthShow we are working on custom agents right now so that you are not limited to what we provide out of the box. let me know if there is a tool that would be more useful if there is one top of mind
@SiliconSouthShow
@SiliconSouthShow 4 ай бұрын
​@@TimCarambat First, thank you so much. I am thrilled to see the progress with AnythingLLM. I've spent a lot of time building agents from scratch in Python using Ollama, and while it's been a tremendous learning experience, having robust tools at my disposal would be a game-changer. I would love to see the addition of a comprehensive Tools Library. A moonshot feature could be a Memorize Tool for unsupervised data collection and learning. A tool for handling webhooks, dialers, and callers, akin to those available with LangChain, would also be fantastic. However, if those are outside the immediate scope, enhancing the current web search and web scraping tools would be invaluable. I advocate for AnythingLLM passionately. I’ve introduced over 100 people to it, often speaking about it in live Zoom sessions. The platform stands out because it’s accessible enough for anyone to start using immediately while still being powerful. It’s well-designed, user-friendly, and out-of-the-box ready. Given the chance, I'd love to run a channel teaching others how to leverage AnythingLLM for various applications, from work to play. I’m particularly excited about potential memory features that would allow for advanced projects like multiplayer RPGs. AnythingLLM is in a class of its own. Unlike other tools that are merely interfaces for other functionalities, AnythingLLM is a powerhouse. It’s a unique tool that truly delivers, and I can’t praise it enough. My wife even jokes that I should be a spokesperson for AnythingLLM because of my enthusiasm. Looking forward to more great features and continuing to support this amazing platform! (I"m such a loser, I spent 2 hrs oen night teaching, talking and complain aLLM when the last update came out and ollama was missing from the agent support system, but I am a huge advocate and fan of aLLM, period, I love it.) PPS: I could see aLLM with a multiAgent system in it, I mean, I see it clearly, doing it all.
@advanced_code_made_easy
@advanced_code_made_easy 3 ай бұрын
You should really release this for people to do locally (because obviously you have the code for it lol), and then have the finetuning service for bigger companies... Someone else is going to soon if you don't
@TimCarambat
@TimCarambat 3 ай бұрын
You can already fine-tune locally via tons of services you have to write the glue code yourself
@RahulGupta1981
@RahulGupta1981 4 ай бұрын
is there a portable version available? Like download and run in place of installation?
@TimCarambat
@TimCarambat 4 ай бұрын
For the fine-tuning or AnythingLLM? For the fine-tuning we give you a model GGUF file you can take anywhere and AnythingLLM has a desktop version.
@RahulGupta1981
@RahulGupta1981 4 ай бұрын
@@TimCarambat For AnythingLLM, I actually have no option to use docker and installed application are tracked/listed on server. So I mostly find simple ready to run apps which I don't have to install. :)
@DanielSchweinert
@DanielSchweinert 4 ай бұрын
Wow that is huge! Thank you! Can I import *.gguf file directly into AnythingLLM like the downloaded system models? Couldn't find any answer to that simple task. Don't want to install LM Studio or Ollama.
@TimCarambat
@TimCarambat 4 ай бұрын
We will allow that on the desktop app - since the docker version does not have an LLM inside of it. We _do_ have llama cpp in the docker image but we will be removing it soon due to the complexity of maintaining it - which is why standalone llm runners tools like Ollama or LMStudio exist. It's a project in and of itself to maintain. Can i ask what you have against for installing LMStudio or Ollama?
@DanielSchweinert
@DanielSchweinert 4 ай бұрын
​@@TimCarambat just showed AnythingLLM to some ordinary people and they were amazed what can be done locally and offline but when I told them they need to install also LM Studio and start server etc. than it was already too much work. LOL. And I have to agree with them. It would be really great if everything was in a desktop app. I know this is not for the masses they just use chatgpt and be done with it.
@TimCarambat
@TimCarambat 4 ай бұрын
@@DanielSchweinert Well then the desktop app fits that. It has an LLM inside of it. Only for the multi-user docker version do you need to have some other external LLM. By default the desktop app has an LLM built into it which makes requirement to install and external runner extraneous. From how it sounds they should use the desktop app. At the start of the video I mentioned I was using the multi-user browser based version because that is just where this feature is live now, that's all
@DanielSchweinert
@DanielSchweinert 4 ай бұрын
@@TimCarambat Thank you! I know there are system llm inside of AnythingLLM like Meta Llama3 etc. for download but I really don't know how I can import other LLM's into it like "dolphin Llama3". Where is the location or path where to put those other LLM's without using LM Studio. Btw Im on a mac.
@boardsontt1756
@boardsontt1756 3 ай бұрын
What’s the difference between this route and creating a custom GPT
@TimCarambat
@TimCarambat 3 ай бұрын
@@boardsontt1756 custom GPTs are just OpenAI models with a system prompt + rag and sometimes basic tools. A fine tune is basically a custom llm where it already knows your documents inherently and does not need rag. Nor does it need a system prompt to behave in a specific manner. Lastly, it can run fully offline and have additional rag and agent tooling on top of all that.
@NLPprompter
@NLPprompter 4 ай бұрын
can i do this in future= run llama3 400 billion (i believe frontier model) make it chat with all kind of complicated information (with frontier model it might able answer them) then use those chat into... json? to fine tune llama3 8b?
@TimCarambat
@TimCarambat 4 ай бұрын
@@NLPprompter that is exactly what you can do, I just used OpenAI here because I can. Same principle though of using a more powerful model to fine tune a smaller one. Also we should have Llama 3.1 live soon for tuning as well so best of both worlds
@NLPprompter
@NLPprompter 4 ай бұрын
@@TimCarambat I'm sorry to ask here, but I'll ask any way maybe you know something, there was a paper by OpenAI about Grokking, it's about... when a model were in fine tune phase and got overfitting continuously then at some point it become able to generalize have you seen such phenomenon in your system? if yes then if got time... i would like to hear more.
@asimabusallam3147
@asimabusallam3147 4 ай бұрын
@mohammedsaleh-ck8jf
@mohammedsaleh-ck8jf 4 ай бұрын
🙏
@musicg6847
@musicg6847 4 ай бұрын
How can we also fine-tune whisper?
@TimCarambat
@TimCarambat 4 ай бұрын
That is an STT model, does not work the same as text text models and is a very different set of data to train on
@hasangh4678
@hasangh4678 4 ай бұрын
When it will be released for windows?
@TimCarambat
@TimCarambat 4 ай бұрын
@@hasangh4678 should be able to get the desktop app updated with this by the end of the week
@alvintoms1806
@alvintoms1806 4 ай бұрын
@@TimCarambat❤❤
@AlehandroDAO
@AlehandroDAO 3 ай бұрын
Just FYI it is against OpenAI usage policy to use their models to create content for any model training
@TimCarambat
@TimCarambat 3 ай бұрын
This is a common misunderstanding when it comes to fine-tuning specifically. The generation of a fine-tuned model from API output (NOT ChatGPT) is not the generation of a new full-weight competing model with respect to their terms. If we used the output to generate a brand new foundational model - like LLama 3.2 or something, that would be a violation as it is a new-weight full-parameter model that would compete with OpenAI. Creating a fine tune from any foundational model, using responses from their API, is permissible within those terms. References from their TOS ----- Use Output to develop models that compete with OpenAI. ----- Ownership of content. As between you and OpenAI, and to the extent permitted by applicable law, you (a) retain your ownership rights in Input and (b) own the Output. We hereby assign to you all our right, title, and interest, if any, in and to Output. ------ Source:openai.com/policies/terms-of-use/
@atanudasgupta
@atanudasgupta 3 ай бұрын
also how can I add some custom tools?
@TimCarambat
@TimCarambat Ай бұрын
Like this: docs.anythingllm.com/agent/custom/introduction
@AnOldMansView
@AnOldMansView Ай бұрын
How do you delete a workspace?
@TimCarambat
@TimCarambat Ай бұрын
Click on the "Gear" icon on a workspace. On the "General settings" tab its a big red button
@trentadams5582
@trentadams5582 3 ай бұрын
FOR ANYONE THATS CURIOUS UNSLOTH DOES THIS FREE
@TimCarambat
@TimCarambat 3 ай бұрын
I don't know how many times I said this in the video, but this is not for people who know how to use libraries like Unsloth. Unsloth is amazing, but you still have to know how to code AND have a GPU you can even use. Funny enough, for the promised local version I am using Unsloth because it's so simple. Sure you can run their colab example but that isn't even close to what everyday people need. The issue is not gatekeeping, it's making it easy and accessible to those who probably dont even have a GPU they can use to fine-tune. Hell, even if you go to unsloths website they dont offer this for free. The LIBRARY for custom code is free, their hosted version is still paid and is not even open to the public.
@thunderwh
@thunderwh 4 ай бұрын
Looks awesome but for some reason the way you talk is making me want to check my wallet is still in my pocket.
@TimCarambat
@TimCarambat 4 ай бұрын
I took your Blockbuster card
@louisduplessis5167
@louisduplessis5167 23 күн бұрын
50usd pm min ?!
@TimCarambat
@TimCarambat 22 күн бұрын
@@louisduplessis5167 you can run it free for desktop. You can self host at your cost, and if you can't do any of that, yeah you can pay us to host it for you.
Unlimited AI Agents running locally with Ollama & AnythingLLM
15:21
Tim Carambat
Рет қаралды 155 М.
EASIEST Way to Fine-Tune LLAMA-3.2 and Run it in Ollama
17:36
Prompt Engineering
Рет қаралды 46 М.
МЕНЯ УКУСИЛ ПАУК #shorts
00:23
Паша Осадчий
Рет қаралды 5 МЛН
Don't underestimate anyone
00:47
奇軒Tricking
Рет қаралды 16 МЛН
Why no RONALDO?! 🤔⚽️
00:28
Celine Dept
Рет қаралды 61 МЛН
RAG vs. Fine Tuning
8:57
IBM Technology
Рет қаралды 62 М.
Stop paying for ChatGPT with these two tools | LMStudio x AnythingLLM
11:13
Qwen Just Casually Started the Local AI Revolution
16:05
Cole Medin
Рет қаралды 84 М.
"okay, but I want Llama 3 for my specific use case" - Here's how
24:20
Unleash the power of Local LLM's with Ollama x AnythingLLM
10:15
Tim Carambat
Рет қаралды 126 М.
Run ALL Your AI Locally in Minutes (LLMs, RAG, and more)
20:19
Cole Medin
Рет қаралды 250 М.
Fine-tuning Large Language Models (LLMs) | w/ Example Code
28:18
Shaw Talebi
Рет қаралды 358 М.
МЕНЯ УКУСИЛ ПАУК #shorts
00:23
Паша Осадчий
Рет қаралды 5 МЛН