Please make a video showing the best role playing models to use for sillytavern. Thanks! Maybe like a top 10 or top 5 or something. Which roleplay models are best?
@TheYashakamiСағат бұрын
This is also my primary goal for AI. The censorship absolutely ruins the experience. I remember one time where GPT-4 told me it was unethical to beat up a wolf to make sure it doesn't wake up and attack me in my sleep.
@Kutsushita_yukino3 сағат бұрын
what do you think is the best model for RP’s as of right now though? and how does it compare to a model made from scratch tailored for RP like character Ai. CAI has a natural human like creative quality to it that even claude opus can’t replicate(rightfully so because it’s made for general purposes) so whenever a new big model comes out i compare between the two.
@SandTiger428 сағат бұрын
Just came across this video/channel. I was heavy into RP a bit ago, but got burnt out and haven't touched it for months. I think I'm going to get AMD's new video card when it releases, and I want to play around with local LLM's this time, instead of using API keys in SillyTavern. I normally say that I want fully quanted model at 120B or higher to have a good experience, 70B minimum. So I'm wondering if I'll become majorly frustrated by the limitation of 48gigs of ram. Now I just need to look for a good tutorial on how to locally set up everything. (I tried looking in your channel, but it's full of meme's with a scattering of helpful guides. heh)
@ChameleonAI7 сағат бұрын
I strongly suggest investing in as much CPU RAM as you can. Yes, VRAM is important, but now that we can split between CPU and GPU, RAM has become way more important than just a year ago. If you really want to use large LLMs you'll want like 64GB+. With 48GB I don't think you're going to enjoy 70B models very much and 120B will be pretty slow if you can host them at all. On my machine that has 32GB RAM and 12GB VRAM, I feel that models in the 20B range are about as high as I can go without them being too slow for my taste. I didn't mention this in the video but I actually have two machines, the smaller one I mentioned and the bigger one with 64GB RAM and 16GB VRAM. Even the bigger machine doesn't like 120B models very well and 70B is about as big as I can stand on it. I don't have a tutorial for setting up local. I've considered it, and now that I've become familiar with a few different LLM backends I feel a bit more confident discussing the topic. You're not the first to ask for a guide on hosting LLMs and I feel a noob guide is long overdue. If you're interested in just my tutorials, use my tutorial playlist. My channel is a mix of things but I've split them into categories for people interested in just the technical discussions (which I completely understand).
@МануелСтоев15 сағат бұрын
Thank you for this video! I learned a lot!
@StevenHuxley20 сағат бұрын
Could you make a video about the latest Mistral Small? I have high expectations, but there must be some issues as it seems incredibly worse than the previous Mistral small
@ChameleonAI12 сағат бұрын
It would take a long time to make a full video about it, but I did try it out today. I went back and forth between 2501 and 2409 for a while to try and get a direct comparison, and honestly I didn't see a huge difference. I felt like 2501 was sometimes a bit more descriptive. According to the UGI leaderboard, it's all a little better and it has a higher willingness score. I'm a bit surprised that you found it to be way worse. I think if I did a blind comparison I'd have a hard time picking out which one is which. Any issues I had like speaking out of turn or not adhering to the prompt were present in both models. Maybe some of your settings are making it act weird or there's a difference in the quantization quality? If the settings are the same between the two, I'd expect them to act similarly. I specifically tried the bartowski Mistral-Small-24B-Instruct-2501-Q6_K_L.gguf quant and the equivalent for 2409 if that matters.
@richterbelmont791422 сағат бұрын
Thank to your video, I finally stop try too force Stable Diffusion v1.5 from Hugging Face to work and let it back from it retirement. XD
@f4tureКүн бұрын
what is the reason you dont use ur real voice instead of this A.I TTS?
@ChameleonAIКүн бұрын
I don't want my real voice out there, and I wouldn't even have this youtube channel if it weren't for voice cloning. It's me speaking into the microphone and transforming my voice with speech to speech, I've only used TTS quite recently in a couple videos, and for the most part I prefer to still narrate things myself.
@mudamudamudamudamudaa50802 күн бұрын
I realy need this video. Thanks for your work.
@laden66752 күн бұрын
Great video, but I wish you tried the full model as hosted by DeepSeek. The distills are not great
@laden66752 күн бұрын
lol I skipped over the clip where you mention that the first time I watched the video. Still, you should mention it's a distill either in title or description
@ChameleonAI2 күн бұрын
I can update the description to be more clear about it being a distillation. Running the full model is a little out of reach for my machine, unfortunately, and I focus on models I can host myself.
@martytraceytravel2 күн бұрын
for the life of me I can't figure out how to utilize this...though I would love to be able to have this style for my D&D game prologue!
@ChameleonAI2 күн бұрын
Traditionally, you can feed it into RVC and train a voice model. More recently, there are utilities that require just a 15 second sample and can do instant voice cloning. I have a couple videos that talk about this, but MaskGCT and Amphion Vevo are two examples.
@martytraceytravel2 күн бұрын
@@ChameleonAI many thx for getting back to me...i'll give it a shot!
@8bitstargazer2 күн бұрын
Unrelated but what is your goto RP llm for silly tavern atm? I feel like no matter what progress is made i end up going back to mistral large / miqu / wizard. I do enjoy the unlsop models though. I enjoy the style of qwen the most but the censorship is just too heavy handed
@ChameleonAI2 күн бұрын
I've been out of the game a little bit. My old favorite was Chronomaid Storytelling 13B, but it's a bit old. I'm still evaluating Cydonia-v1.3-Magnum-v4-22B but I like it so far. If you like qwen or anything else that's pretty censored, check out a concept called abliteration. It's a process of uncensoring a model. I've not personally messed with abliterated versions but it seems to be pretty popular. One good way of looking for models is the UGI leaderboard, that's how I find out about most of them.
@owhite24452 күн бұрын
stable diffusion 1.5 rules. all downhill from there.
@johnny58052 күн бұрын
Why do you keep going on about disgusting incel stuff like 'waifus' and 'furries' ?? They are irrelevant to AI, just as they irrelevant to society.
@ChameleonAI2 күн бұрын
Those are the people who trained the models, like it or not.
@LostLibrarian4 күн бұрын
... PERFECT! JUST PERFECT!!!
@NinuRenee5 күн бұрын
okay
@vichikaru7 күн бұрын
Sd1.5+flux do all the thing, skip others XD
@Bruno_B_RJ7 күн бұрын
This video really gives a great overview of the whole thing. Thank you so much! I've been using MidJourney for years, but I can't afford the subscription anymore. Based on your experience, are any of these tools on the same level or even better than MidJourney?
@ChameleonAI7 күн бұрын
I'd say the latest XL models are pretty competitive with midjourney, and XL has fairly modest system requirements. SD3.5 is the best and biggest but it's lacking the user support and history of XL.
@EkShunya8 күн бұрын
great summary
@ling67019 күн бұрын
Thanks, amazing info summary. Subscribed
@DoubleBob10 күн бұрын
You are really good at making some, objectively speaking, boring topic quite interesting to listen to. I hope you do a follow up when history has progressed further.
@ChameleonAI10 күн бұрын
Thanks for the kind words. 2025 should be interesting, I'm keeping my eye on SD 3.5 and alternatives like Flux but still waiting on stuff to shake out.
@Archydoge10 күн бұрын
Loved this video thank. Great seeing the evolution of the branches and what they are based on. Would great if you can do periodic update videos with. As i have veen learning about ai recently im seeing noobAI and illutrious and seee where these are evolving would ve good.
@tucos20411 күн бұрын
Very helpful video in understanding the models and the history. Please make an annual update to this and give us a followup video that includes the different models as well, hunyuan etc.
@ChameleonAI11 күн бұрын
I do want to cover 3.5 at some point and get into the non SD models as well, so yeah. I do have to do my homework on all that stuff though and see what shakes out in 2025.
@JarppaGuru12 күн бұрын
if know trained image then know what images it do. it wont do nothing thin air
@starcrafter13terran12 күн бұрын
Spot on
@dantartable12 күн бұрын
i love you so much !!!!!!!!!!!!!!!
@MsNemaki12 күн бұрын
Thank you for this awesome video!!!
@LuizCarlosTomaz13 күн бұрын
This video is amazing!🥰
@Sl1555515 күн бұрын
good video
@sdfrasvxc291318 күн бұрын
5:45 am i just stupid or didnt pay attention or how did those prompts appear for you in the CMD bar?? were they automatically created? if so how do you do so? i tried to rewatch these couple minutes multiple times but i still couldn't figure out how did those get created. was it the style preset or something in the prompt templates? Cause when i tried to do the same with SD forge and used /sd you , the ai just completely copy pasted something random from the template and create an entirely different image that's not even close to the character and blud just created a blurry blue vomit , i need help💀
@ChameleonAI18 күн бұрын
The prompts are automatically created. I tried explaining it at 4:15 but it's difficult to convey. What it generates entirely depends on the LLM you're using, so that's why it's important to look at it in the cmd window to see what it's really outputting. The style prefix is only some text that's tacked on to the beginning of whatever the LLM outputs, so that's not very important. The important part is the prompt template, that's how SillyTavern asks for a prompt from the LLM. Basically it's the same as if you manually used the prompt template as your text input, then copy-pasted its response into SD. Now if it's creating blue vomit, that's probably a SD setting or something to do with the model (missing vae, wrong sampler, not enough steps, model incompatibility with SD forge, it could be a lot of different things).
@sdfrasvxc291317 күн бұрын
@@ChameleonAI oooh i get it now, you need an LLM like oobabooga? i think that's where it went wrong for me cause i dont have that. I tried to make this work solely by using sillytavern, stable diffusion forge and koboldcpp unofficial google colabs (to run cloudflare kobold api's) so thats why i was confused why the prompts for the SDforge appeared simply from the templates passed raw lol
@shailendrarathore44519 күн бұрын
Provide a text to song as necer loose the original voice and pitch
@ChameleonAI18 күн бұрын
Amphion is working on Text to Music but it's not ready yet. I'm not aware of any other open source solutions but if I find one I'll try it out.
@LoveMapV420 күн бұрын
Please make an in-depth comparison between MaskGCT and F5 TTS. Some people say F5 is better, but in my opinion, MaskGCT gives better results. The only downside of MaskGCT is that it is currently unable to generate very long prompts. Also, please provide the entire project as a zip file, as I was unable to make it work. It says I need espeak, and I have already installed it, but the problem persists. So, I cannot test MaskGCT personally.
@ChameleonAI20 күн бұрын
I have another video where I compare a few different TTS solutions including F5 but I wouldn't say it's in-depth. Interestingly enough, I have the opposite problem as you where I haven't had a lot of success running F5 locally but I can run MaskGCT. I believe MaskGCT sounds generally better but takes longer and uses much more VRAM. I think most fans of F5 think that the quality is acceptable for the speed and resources that it uses. As for providing the project as a zip, it's not my project, it's from Amphion, and zipping the fully installed project with dependencies wouldn't work anyway due to cross-platform issues and how virtual environments work (I've tried. Maybe a docker container or something would work but you can't just zip it up). You're not the only one experiencing the espeak problem, though, and there are several open issues on the Amphion github about it, for example: github.com/open-mmlab/Amphion/issues/323
@LoveMapV420 күн бұрын
@@ChameleonAI Unfortunately, my problem isn't related to the RuntimeError: failed to load voice 'ja'. In my case, it says that espeak was not installed on the system, even though I already installed it. I tried both the 32-bit and 64-bit versions, but it's still not working. I also added all possible paths to the environment variables, but nothing seems to work. Maybe there's something wrong with my PC, not the code of the program itself. I don't know. PS: I am using another AI project that requires espeak to be installed on my system. I was able to use it without any issues, so I believe the problem isn't related to not installing espeak properly.
@ChameleonAI20 күн бұрын
@@LoveMapV4 Yeah, looking around on the issues, some people are saying that the espeak errors aren't really about espeak so there's something to that. Unfortunately since it's not an error I can replicate on my system, the only advice I have is to check out the github issues page or simply wait and hope Amphion fixes it.
@kowy_m21 күн бұрын
it's confusing detecting u'r voice(does it exits !?), place a beep or other to know that's narrative
@ChameleonAI20 күн бұрын
You mean my "real" voice? I always use a voice changer, so in a sense my real voice doesn't exist.
@mariogan22 күн бұрын
this is the best tutorial I had ever seen, nice vocal in the end
@ChameleonAI22 күн бұрын
Good to know that a tutorial made in the SD 1.5 days is still relevant today.
@kwikb21 күн бұрын
i agree Well done
@jess_o24 күн бұрын
Great video, thanks for the recap! I've recently returned to this hobby, and I'm really, really liking Illustrious
@deviantai_com25 күн бұрын
Great ramble, learned some stuff but most of all it cleared up a lot of maybe stuff for me , thanks.
@Kinos14126 күн бұрын
Personally, I'm still on 1.5. It's giving me exactly what I want
@numberonedad26 күн бұрын
that american guy is missing an alveolar flap in "water" americans say wadder not waTer
@ChameleonAI26 күн бұрын
Probably over-annunciating because he was doing a reading. I do that myself sometimes.
@Maxume26 күн бұрын
Could you maybe explain what benefit this offers over F5-TTS? The quality doesn't seem to be better but the VRAM requirement is double or more than that of F5.
@ChameleonAI26 күн бұрын
There is a TTS option but this is primarily voice-to-voice, so it's a different application. Unless I missed something, F5 doesn't have an sts mode.
@Maxume26 күн бұрын
@@ChameleonAI Oh, sorry. I thought it was TTS. It makes sense now. Thanks
@JuicyBenji26 күн бұрын
Nice, is this real time as well or only recordings?
@ChameleonAI26 күн бұрын
Only recordings. I didn't see a way of feeding in a stream real-time with the code they have right now, though it may be theoretically possible because it takes less time to run the conversion than the duration of the clip (on my machine anyway).
@markus_knoedel27 күн бұрын
Thank you. So much wisdom in here.
@troexnevelin9927 күн бұрын
That's a good content don't let the number of views fool you! Really good analogy of what happening today back to it's roots seeded hundred years ago in first movie, I guess digging deeper into literature we could find even more examples. Good job!
@blacout128 күн бұрын
bro first of all hats-off to u .. your explanation was thorough, and your voice is amazing and perfect pace that helps me digest the topic.. thank you for this.. i watched whole video
@endymionm137428 күн бұрын
Great video as I'm new to this software and it was good history as well as concept lesson on what the different models do and how they function
@mgga2429 күн бұрын
sent a discord dm 😶🌫
@ParvathyKapoor29 күн бұрын
pinokio install?
@ChameleonAI29 күн бұрын
Great idea. I just published github.com/chameleon-ai/vevo-pinokio
@ParvathyKapoor29 күн бұрын
@ChameleonAI thanks boss 🙏
@guitarmeetsscience29 күн бұрын
You deserve a bigger following - Love these videos!
@brootalbap29 күн бұрын
Well the video was informative in two ways. 1) I know more alternatives to SoViTTS 2) I know why i won't buy an AMD GPU. The current AI landscape revolves around CUDA by nvidia. AMD sleeps on AI big time.
@ChameleonAI29 күн бұрын
Life is pain, I wasn't into AI stuff when I bought my rig. So much for python being cross-platform. lmao.
@brootalbap29 күн бұрын
My comment regarding SD3: woman lying on grass. people saw the SD1.5 untuned quality of SD3 (PLUS the licensing scheme!) and people were just done with SD3. SD3 was even banned for a while on civitai because stability AI tried to sneak in legal terms that would have forced people to delete their images and fine-tune models if they ever quit their subscription to stability ai. What they tried to do was insane.
@ChameleonAI29 күн бұрын
Yeah licensing for SD3 was a big part of why it died on the vine and why people were eager to adopt Flux. Though they have since updated their terms and now I'm seeing some early fine-tunes for 3.5 that look promising.
@XDM_StudiosАй бұрын
what do you think might happen to stable diffusion as a whole now that 4 is not in the works? will someone be able to step up the plate and take over or will other models just emerge? i see flux is giving it a go so maybe more like that will emerge.
@ChameleonAI29 күн бұрын
Since I made this video, SD 3.5 came out and is pretty competitive with flux. Stability AI also has partnered with Amazon, so maybe they've gotten past their financial troubles. As of now, there is still no SD4 announced, but I guess we'll see what they do in 2025.