"not everyone is crazy enough to have a server rack in their garage" yeah i got mine in my bedroom LMAO
@bb2ridder7578 ай бұрын
yeah got mine in the attic
@blackryan52918 ай бұрын
bad_dragon...I'm freaking dead bro 😂
@marcogenovesi85708 ай бұрын
not everyone is deaf and can do that
@TopHatProductions1158 ай бұрын
lol same 😂
@SvRider5128 ай бұрын
Mine is in my room too!
@iraqigeek83638 ай бұрын
You can get P100 for $100 only since at least last October if you "Make an Offer" Not all sellers will accept it, but a few will. I bought 4 of them at $100 last year
@CraftComputing8 ай бұрын
That's very good to know, as I might be picking a couple more of these up shortly...
@theangelofspace1558 ай бұрын
@@CraftComputingwell, maybe after thos video that wont be a thing anymore 😢
@KiraSlith8 ай бұрын
@@CraftComputing Every time a medium-large KZbinr makes a video, prices spike. I doubt they'll remain that accessible for too long now you've published this vid. :P
@milescarter78038 ай бұрын
I got the 12GB for 90, d'oh
@stefanl51838 ай бұрын
Make sure you get the PCIe version and not the SXM2, unless you have an adapter or a server with SXM2 sockets. The SXM2 versions are cheap because of this.
@lewzealand47178 ай бұрын
7:57 Oops, you compared Time Spy on the P100 VM to Time Spy *Extreme* on the Radeon 7600. The 7600 gets ~11,000 in Time Spy, or about twice the single-VM score shown.
@Chad_at_Big_CAT_Networking8 ай бұрын
The cloud gaming aspect initially got my attention, but I think a lot of us are going to be more curious how they perform running Ollama at home. Looking forward to more of this series.
@fujinshu8 ай бұрын
Not all that great, considering they lack the Tensor cores that have since appeared on newer GPUs since Volta and Turing, which are kinda the reason there’s not a lot of support for Pascal and older GPUs.
@yzeitlin8 ай бұрын
We are still patiently awaiting the Hyper-V homelab video you mentioned on talking heads! love the content
@xpatrikpvp8 ай бұрын
One thing to mention would be that latest supported vgpu driver for the P100 (also other pascal gpus P4/P40) is version 16.4 They dropped support in the latest 17.0/17.1
@TheInternalNet8 ай бұрын
This is really really exciting. Thank you for never giving up on this project. This is the exact card I am considering for ML/AI to run in my R720xd.
@SvRider5128 ай бұрын
I have a Tesla P4 in my 720xd
@richardsontm8 ай бұрын
Would be great to see a P4 v P40 v P100 head to head. Having a blend of Cloud Gaming and Ollama performance would be interesting for those looking for a homelab/homegamer/AI tinkerer all-rounder too 👍
@CraftComputing8 ай бұрын
P40 just arrived :-) In a couple weeks, I'm going to be testing out every GPU I have on hand for performance and power draw. Stay tuned...
@conscience_cat11468 ай бұрын
@@CraftComputingI've heard that the P100 doesn't have H.265 support, and only includes a H.264 encoder. If that is the case, then theoretically the P40 should look alot better with Sunshine and Moonlight. Can you test this out and possibly confirm in your next video? This info will make or break which card I end up getting.
@richardsontm8 ай бұрын
@@CraftComputing We look forward to it - thank you for the fun content, it's always an interesting watch @ Craft Computing
@d0hanzibi8 ай бұрын
Yeah, comparison in dimensions of gaming, general workstation tasks and LLMs would be really awesome
@keylanoslokj18068 ай бұрын
What is cloud gaming
@th3r3v928 ай бұрын
I've been using my Tesla P4 with an Erying 12800H ITX board as a home server for almost a year now, and I absolutely love it. I have a Win10 vGPU VM running on it, primarily used by my girlfriend, but it's also great when friends come over for a quick LAN session. I was really disappointed when I found out a few weeks ago that NVIDIA dropped Pascal support from the GRID driver.
@d3yuen8 ай бұрын
We (my company) owns 5 Dell PowerEdge 740s with two P100-16G-HBM each, but we used VMware vSphere as our hypervisor. Going on 4+ years, they continue to be excellent and reliable cards - still in active service today for VDI. With Dell Premier Enterprise pricing, we got them at considerably less than MSRP. It's the ongoing support and maintenance paid periodically to Dell, VMware and NVidia that's the killer. Pro tip: it's important that you line up the driver versions from the hypervisor down to your guests. That is, the driver version on your guest must be supported by the driver running in the hypervisor. 😅
@joshklein8 ай бұрын
Any videos with proxmox and gpus i love to watch! Keep them coming!
@TheXDS8 ай бұрын
I love your vGPU content Jeff! You actually inspired me to build a small homelab with a P100, though the CPUs on my server are somewhat older than I would like (a pair E5-2690 v4 CPUs)
@joseph31648 ай бұрын
Great video, always love to see the enterprise hardware for home server use. What are you using for licensing servers for these cards? Are you just using the 90 day trial from nvidia, or are you using some type of fastapi-dls server?
@mikboy0188 ай бұрын
Just one of those could power an Unreal Tournament 2004 LAN party... Of note, there is also a 12GB P100 -- they don't perform terribly, either.
@b127_18 ай бұрын
3:10 Both GP100 and GP102 have 3840 cores. However, GP100 has only ever been sold with 3584 cores active, while you can buy versions of GP102 with the full die, like p10, p40, p6000 and titan Xp (but not titan X (pascal), that has 3584, just like the 1080ti).
@ProjectPhysX8 ай бұрын
GP100 is a very different microarchitecture despite the same "Pascal" name. It has a 1:2 FP64:FP32 ratio, in contrast to all other Pascal GPUs which have 1:32. FP64 is only relevant for certain scientific workloads. Today there is only very few GPUs that can till do FP64 with 1:2 ratio, and .Ost are super expensive: P100, V100, A100, H100, B100, MI50, MI60, Radeon VII (Pro), MI100, MI210.
@smalle8 ай бұрын
These are pretty dang impressive. This might be the coolest/most approachable card you’ve shown so far!
@cabbose25528 ай бұрын
the pcie power connector can usually deliver 200+ watts thanks to over spec'd cables but the standard only requires 150
@ATechGuy-mp6hn8 ай бұрын
I've seen the sponsor spot for the cloud gaming machine from maximum settings on this channel before, but its kinda ironic that you are setting up your own gaming server afterwards
@MeilleureVie20248 ай бұрын
All P100 Ebay listings went up 50$ since you posted this video hahaha
@chromerims8 ай бұрын
Love the vGPU cloud gaming content 👍
@j_ferguson8 ай бұрын
I absolutely need those BEER coasters. Also, I was very lucky to go to Block 15 during the last solar eclipse and had Hypnosis a cognac barrel aged barleywine. Their Nebula and Super Nebula stouts are way more common and still delicious though.
@laberneth8 ай бұрын
I'm impressed what is possible today. I was a Administrator of a community school in germany 25 years ago. Time is running so fast. Informative Video. Subscribed!
@fruitcrepes48758 ай бұрын
Good to know all these Liquidations we've been doing for the P100s at the datacenter are going to good use! I've boxed up thousands of these bad boys and off they go
@MyAeroMove8 ай бұрын
Best series! Love to watch such kind of tinkering!
@calypsoraz43188 ай бұрын
As a fellow Oregonian, how do you mitigate the humidity in your garage? Or is it not bad enough to affect the rack?
@greenprotag8 ай бұрын
Time spy (Standard) vs Time spy (extreme) results? I suspect you are closer to a standard run on Ryzen 3600 + GTX 1060 @ 4 693 (Graphics Score4 464 CPU Score6 625) but at that point I am splitting hairs. Your result is 2x playable gaming experiences on a single $150-180 enterprise GPU WITH a nice automated script for set up. This is a nice alternative to the P4 especially if the user has only 1 x16 slot.
@brahyamalmonteruiz99848 ай бұрын
Jeff, how did you cooled the P100? I've seen your video on cooling the Tesla GPUs but which option did you used in the video ?
@nexusyang48328 ай бұрын
Just looked up the spec sheet for that P100 and saw that the 16GB memory is what they called “Chip-on-Wafer-on-Substrate.” Very cool.
@LetsChess18 ай бұрын
Have you don’t anything with the Tesla p40? I was wondering how different performance is between the p40 and the p100.
@subven18 ай бұрын
I can't wait for SR-IOV to be available on Intel Arc! This would open up a more modern and potentially even cost-effective approach to cloud gaming and VDI solutions. Unfortunately, Arc SR-IOV support is currently only available in the out-of-tree driver.
@d1m188 ай бұрын
Look forward to the next video that compares full single 11:56 performance
@EyesOfByes8 ай бұрын
So...Are there any specific games that benifit from increased memory bandwith
@hazerdoescrap8 ай бұрын
Just wondering if you could "expand" on the use case testing when you build out the Master List? For example I have a GPU stuffed in a system that I'm going to be using as an Encoder Rig for 264/265 Encoding (cause nv charges too much for AV1 right now) and I'm wondering how that would effect the GPU performance? Or if I were executing LLM testing via oLlama at the time someone were running a game.....
@DanteTheWhite8 ай бұрын
When it comes to testing ollama, you need the amount of VRam that will hold whole model at once otherwise some model layers are delegated to cpu which hinders performance considerably.
@hazerdoescrap8 ай бұрын
Yeah, I fired it up against a 32 core Epyc server.... it was not pretty.... Would be interesting to see how the GPU balancing handles the RAM juggling for that kind of load when split with other non-LLM functions....
@ewenchan12398 ай бұрын
Two questions: 1) Are you able to play Halo Infinite with this setup? 2) What client are you using to connect to your system remotely? I am asking because I tried Parsec, and even with an actual, real monitor connected to my 3090, it still stuttered a lot. Thank you.
@rudypieplenbosch67528 ай бұрын
wow this is great info, I just finished my Epyc Genoa build and was looking for a proper ways to get graphical performant VMs, amazing 👏. Does this also work for Linux VMs?
@blakecasimir8 ай бұрын
The P100s and P40s are very commonplace from China, and inexpensive. They are both decent for running large language models as well. But it's still Geforce 10 era performance, so don't expect wonders.
@insu_na8 ай бұрын
I have a P100 in my old HPE server, but I've stopped using it because the P100 doesn't have p-states, meaning it can't do any power-saving mode. Now in practice when no load is applied this means the GPU idles at 30-40W which still isn't awful, but when you compare it with other GPUs even from the same generation (such as the P40) which can idle at 5W-8W it's quite the difference (I live in Germany and electricity costs actual money here). That's on top of my servers' already high idle power. My EPYC GPU Server ***idles*** at 200W without any GPUs installed, so that's a thing..
@czarnicholas2k6988 ай бұрын
I guess I leave this comment as a heads up that PVE 8.2 (Kernel 6.8) breaks nvidia driver compile, and I'll add that I'd love to see a video about how to fix it.
@tylereyman52907 ай бұрын
has there been any word on how to work around this?
@badharrow8 ай бұрын
Has anyone else encountered the issue when proxmos is installed in a raid that you dont enable IOMMU within Grub?
@cberthe0677 ай бұрын
Does this procedure continue to work with PVE 8.2 (Kernel 6.8) update ? I have some doubt ... seems people are getting error compiling some drivers (vg_unlock) ?
@JoshuaBoyd8 ай бұрын
While you are working on those benchmarks, I'd love to see something done with LLMS, say a quick test or two using Mistral and tiny llama on each card?
@xmine088 ай бұрын
That ultra fast memory makes it interesting for LLMs, could you try that? It's only 16GiB, but really fast at that and cheap so might be a solution for some!
@ezforsaken8 ай бұрын
question! Are there any modern non-nvidia options for doing 'vgpu' (shared pcie graphics card)? Amd? does intel allow srv-io on their stuff for sharing them? just asking!
@nyavana8 ай бұрын
Really hope unraid gets proper vGPU support. I have a p100 but since there is no easy way to get vGPU working, I can only use it for stuff like transcode
@mechanizedMytha8 ай бұрын
i wonder if it would be worth it to use this to upgrade an aging pc... most definitely gonna be hard powering it... how did you get around to solving power delivery, in fact?
@user-hv5jv9gb6c8 ай бұрын
Would a P100 have any advantage over a RTX 3060 or 4060 in stable diffusion?
@AlexKidd4Fun8 ай бұрын
The more modern consumer cards (30xx or 40xx) would be way better for AI than what is being discussed here.
@ProjectPhysX8 ай бұрын
There is no advantage for AI; AI needs low/mixed precision matrix acceleration which P100 lacks. The P100 smokes the newer cards in scientific FP64 workloads though. It's 3.7x faster in FP64 than even the RTX 4090.
@user-hv5jv9gb6c8 ай бұрын
@@ProjectPhysX Ah ok, I understand now. Thank you....
@victorabra8 ай бұрын
For gaming i recommend to use a Tesla P40 it have GDDR5X ram but it better GPU frequency and 24 GB good for AI too.
@tylereyman52907 ай бұрын
How did you get around the degradation that occurs after 20 minutes of use without a purchased license?
@yellowood2525 ай бұрын
Would it be possible to use two P 100's for a single VM? And would it actually reflect in higher performance (reasonably reliable) gaming?
@CraftComputing5 ай бұрын
Nope.
@Jeremyx96x8 ай бұрын
Can a standard CPU be usable for "cloud" gaming? I have my old 2600x and was wondering if I can get a m40 or similar to pair with it.
@bradnoyes79558 ай бұрын
How would the P100 perform doing AI video upscaling (such as Video 2X)? I've got several DVD rips (even a couple VHS 'rips') I'd like to try to AI upscale for my Plex server; so if I can throw one of these in my Proxmox box (an HPe DL360p Gen8 SFF) and setup a batch to churn through them without hogging up my gaming machine, that would be nice.
@rtu_karaidel1154 ай бұрын
I am playing on my P100 games such as Rainbow Six Siege , Mortal Kombat II and etc , and it is awesome!!! Moreover with Moonlight + Wireguard + Dualshock 4 , i am now able to play all PC Games on my phone😛 This is crazy , unbelivable and weird feelings income ... still can't belive it is possible!
@d.73038 ай бұрын
Would a amd Mi25 be just as good? Seem pretty similar besides the memory bus width I picked one up for $60 off of ebay and havent decided what i wanted to do with it lol
@CraftComputing8 ай бұрын
I have a couple Mi25s. Unfortunately AMD has never publicly released their MxGPU drivers for them.
@arjanscheper7 ай бұрын
any chance we can see an updated vgpu tutorial? using the proxmox-vgpu-v3 script.. but cannot seem to get the p4 nor any gpu working with plex hw transcoding in a proxmox ubuntu vm. I can see them in nvidia-smi in the vm but hw transcoding keeps on failing. and the old guide just spews out errors etc.. as most guides on yt are 3/4 years old already. Just want to setup a fileserver > plex and then optional windows gaming vm or homeassistant
@anzekejzar32338 ай бұрын
Hi there, I have been thinking about setting up the system like this myself. Do you by any chance know if there is some weird Nvidia licensing thing, or could these VMs run Linux?
@wayland71508 ай бұрын
This is cool and I have a question. How does this differ in practice from using two dedicated GPUs? For example I have a Proxmox machine with two A380 cards eached dedicated to an auto starting Windows 10 VM. Yes two physical Windows PCs in one SFF case. Each VM has to have it's own hardware dedicated to it. Yes another VM could use the same hardware but the first one would have to close first. In the P100 setup do you have to dedicate a particular slice of the P100 to the particular VM like I'm doing with my two PCIe GPUs? Following on from that would it be possible to have a number of non-running VMs that could be started in any order?
@lilsammywasapunkrock8 ай бұрын
You are talking about vgpu. These video accelerators can be split up evenly, meaning, you can allocate resources to vm's in multiples if two, and the VM will think it has its own graphics card. Meaning 16gb total memory, split once would be 2 gb "video cards" or 4 4gb cards ECT. The vram will be reserved for each vm, but the host computer will allocate resources for each vm. Meaning, if you only have one VM running, it will see slightly lower then 100% processing utilization. If you have a second VM running, it does not split it an even 50/50 unless both vm's are asking for 100% usage. Meaning one could just be playing a KZbin video for example and won't need more then 10%, but the other could be playing a game using 90%. I encourage you to watch Jeff's other videos.
@wayland71508 ай бұрын
@@lilsammywasapunkrock yes, I comprehend what you've said but I'm still confused by the splitting up. For instance I had a Proxmox with an HD 5950, this card was actually two GPUs, I split that up with one VM dedicated to the first GPU and the other to the second GPU. Never the twain shall meet. If for instance I had another VM that could use GPU1 I could only run that when GPU1 was not being used by the first mentioned VM. So with this P100 is one slice of it dedicated to the particular VM I set it up with? It will it pick what it needs dynamically like it does with system RAM. When setting up a VM I don't pass through a particular RAM chip.
@tanmaypanadi14148 ай бұрын
@@wayland7150The vram is not dynamically allocated. it is predetermined but the actual Cuda cores running the workload will work in parallel across the VMs as long as the vram is sufficient.
@jcrus20018 ай бұрын
Jeff; I was excited to used the new script to setup GPU passthrough for my Tesla M-40. I ran into an issue, I am wondering if anyone else ran into this issue. Before going through this I went through the intial setup following your Telsa M40 video and Windows 10 could see the M40. After using the script, M40 show with the varous MDEV types after the install . I selected one of the Q-types. When I went into my Windows the Tesla M40 did not show correctly, I installed the lastest NVida drivers and Nvidia drivers could not find the card. I re-read the installation website and notice they had used a older Nvida driver, I tried to install this driver and got the same results. I went back and removed the MDev type and now the Nvida driver recognised the card was present in Windows. Has any one else ran into this issue. Jeff - did I miss a step ? -- thank you
@sisakamence8 ай бұрын
what do you mean when you say it doesn't find card under windows? have you installed the grid drivers for windows 10 ? windows should see like a quadro q6000
@jcrus20018 ай бұрын
@@sisakamence yes, loading the Nvida grid drivers fail to find the card, even though windows shows it in the device manager
@mastermoarman5 ай бұрын
im debating the p40, p100 and quadro a4000 for a transcode and codeproject ai for security camera image recognition
@ryancholmes8 ай бұрын
Anyone know how this card would do with transcoding for Plex? 4K down to 1080p
@pops718 ай бұрын
Can this gpu be used in a standard gaming PC or only a server rack. It seems for the price it could be a powerful addition to any gaming rig
@SoraIroNaKISEKI8 ай бұрын
Does it work with games that have anti-cheat and banning VMs? I tried a while ago and none of my game works, even single player games (cuz they have cloud saves?)
@joshuastelly42568 ай бұрын
If you're using Proxmox, populate the manufacturer field in the bios section of the VM config. EAC is that easy to bypass for VM detection.
@wayland71508 ай бұрын
I want to build a Clown Gaming Server, I should be able to get a heck of a lot of Clowns in a Small Form Factor case.
@techpchouse8 ай бұрын
Great 💪🏽 thinking about an upgrade from the P4 to 10/100
@thebrentfamily76027 ай бұрын
Any recommendations when you don't have extra EPS power connectors but plenty of pcie connectors?
@CraftComputing7 ай бұрын
You can get adapters for 2x8-pin PCIe to 8-pin EPS. Current is the same (300W) for them.
@hamzahassan67263 ай бұрын
is this setup limited to p100 pcie? or can it work with sxm2 as well?
@DarrylAdams8 ай бұрын
So when will NVIDIA release Grace Hopper vGPU? The upscaling alone will be worth it
@CraftComputing8 ай бұрын
Never. Hopper doesn't support graphics APIs.
@Zozzle8 ай бұрын
Have you seen that 12gb tesla m40s are like $50 on ebay?
@CraftComputing8 ай бұрын
Yes I have! Tesla M60s too! I'm going to be re-reviewing those cards shortly.
@w3isserwolf8 ай бұрын
Does someone know what has changed on proxmox? I cant enable iommu in the config file Like in crafts Tutorial because there is on file? Did they changed the bootloader/kernel? I want to pass a asmedia satacontroller to an truenas vm
@CraftComputing8 ай бұрын
drive.google.com/file/d/1rPTKi_b7EFqKTMylH64b3Dg9W0N_XIhO/view?usp=drivesdk This is my tutorial for IOMMU, upgrades for v7 and v8. They did change from grub to cmdline for UEFI boot systems for enabling IOMMU.
@w3isserwolf8 ай бұрын
@@CraftComputing thanks. Will give it a try
@w3isserwolf8 ай бұрын
@@CraftComputing do you know why i only see 4 PCI IOMMU Groups? There are more than 1 Devices in each group. I want to go back to grub :( The Videotutorial was so helpfull to get it to work but now with cmdline im screwed. Im Using an asrock b450m pro4 r2.0 with an Ryzen 5 5600G but cant get IOMMU working.
@bkims8 ай бұрын
I believe the tesla P40 is actually the same silicon as the titan xp which would be somewhat faster than the P100 for gaming id expect. To the best of my knowledge the P100's superior memory performance is really only meaningful for things like AI inferencing workloads. Either way both cards are around the same price these days which is pretty cool. edit: though perhaps the extra BW is beneficial for cloud gaming, I don't have any experience there.
@ProjectPhysX8 ай бұрын
Yes P40 and Titan Xp are identical, except doubled 24GB capacity on P40. The P100 is a very different microarchitecture from the rest of Pascal: it supports FP64 with a 1:2 ratio. All other Pascal cards only do 1:32 ratio.
@ngrok_urlurl8 ай бұрын
but can we use those to run moder pytorch for nn training?
@MP-ul8 ай бұрын
Are the Kepler tesla GPU's still okay for this type of thing? I want to buy few K80s to make a remote workstation for few people but they do not need a ton of gpu horse power for what they do.
@CraftComputing8 ай бұрын
I'm actually going to be re-reviewing the K80, M40S and M60 in a video shortly. Stay tuned.
@MP-ul8 ай бұрын
@@CraftComputing Looking forward to it. Good luck.
@computersales8 ай бұрын
I never have understood the preference towards the P100 of the P40. My only assumption is the higher memory bandwidth of the P100 is beneficial for AI workloads.
@VinnyG9198 ай бұрын
got 3 of them how do you get these drivers are they paid? do you really need windows vm for games or would a bare metal headless linux server do?
@CraftComputing8 ай бұрын
The script, uh, *ahem* installs the drivers for you.
@michaelamos758 ай бұрын
Will this work to stream oculus/steam to a VR headset like Oculus 2?
@CraftComputing8 ай бұрын
Yes actually!! Though for latency, you need to be on the same LAN. Works great though through Virtual Desktop or Oculus wireless link.
@chriskarnes34108 ай бұрын
How do you power one of these? None of my cables seem to match the connector
@stefanl51838 ай бұрын
You didn't watch the video or didn't pay close enough attention. It's EPS 12v, not 8 pin PCIe. Many data center GPUs use EPS 12v instead of 8 pin PCIe. Anyway, you need an 8 pin EPS 12v connector on your power supply or and adapter that takes 2 PCIe connectors and powers feeds an EPS 12v connector. EPS 12v was used because it can deliver more power.
@itzsnyder72718 ай бұрын
How much is the power consumption?
@CraftComputing8 ай бұрын
Heaven benchmark consumed around 100W (60 FPS cap). Both VMs utilized 100% of the GPU, and power use was between 180 and 200W. The card technically has a 250W TDP.
@itzsnyder72718 ай бұрын
@@CraftComputing that’s completely reasonable! Wow! Considering to build the same server with this wattage usage. Can you say something about the idle consumption?
@CraftComputing8 ай бұрын
Idle was around 25W IIRC.
@BrandonTahedl8 ай бұрын
Does anyone have any idea of how well the P40 / P100 would play Star Citizen? I know that's a bit of a complicated question, but wondering if anyone that has one has tried.
@SaifBinAdhed8 ай бұрын
I was on this all day, I still have no output when I do mdevctl types... I don't want to give up yet, im very new to proxmox, I am doing it on a fresh install, intel i7 7700K and an RXT 2060 Super which supposed to be compatible, I run the script, all goes well, but no mdev
@curtiscrowethephotopro24538 ай бұрын
I have a similar setup but with a P4, I can't get the resolution to go above 1360x768. I have tried reinstalling the video driver, and I have reinstalled the virtual display driver and its still giving me trouble. I was thinking about doing a fresh install of Windows but if anyone wants to give their thoughts I would be happy to listen.
@saifemran45288 ай бұрын
Thanks for the video! can this gpu work with esxi?
@xerox94268 ай бұрын
Its actually not correct that the vGPU drivers are limited to 60 FPS. You can disable the Frame Rate Limiter (FRL) by changing the scheduling from best-effort to another profile via nvidia-smi on your host. I tested this myself. Have a look at the docs. This has impact on general performance within a VM as it is a fixed or equal share. This means 2 VMs have fixed 50%. 4 have 25%.
@hugevibez8 ай бұрын
Tbh this mostly verifies that having a dedicated GPU per VM might still be the best option if they are all going to be gaming. Sure this is a much more cost effective option than 4 1080s or whatever, but move up on the scale having 4 RTX 4090's is much more cost effective than an L40/RTX 6000 Ada (with the sweetspot being somewhere along the way) and you gain a lot of performance. I do wish it was easier to have a GPU serve both a VM and your containers at the same time on modern consumer GPUs.
@EvertG80868 ай бұрын
I’m using an Intel ARC 750 for mine. Works really well.
@NecroFlex8 ай бұрын
I was lucky about 2 years ago, snagged a Tesla M40 24GB and Tesla P100 for around 90€ both, seller had them as unknown condition. Got both of them running just fine, M40 with modified registry to see it as a high performance GPU, the P100 with a bit more driver fuckery to get it to work. Have been thinking of flashing the P100 to a Quadro GP100 to see if i can use it like that with just a reg edit aswell, but no luck so far.
@ericspecullaas28418 ай бұрын
I have 2 of them. I was thinking of getting 2 more a second cpu and a new motherboard. If you want to run a localhost LLM 2 of them work really freaking well. When i was running a local LLM the response time was insane. Im talking within 1 second before it generated a response. Yeah i have 256gb of ram so that helps.
@bigTBoss8 ай бұрын
Have you fone vr gaming with your cloud gaming setup?
@remcool12588 ай бұрын
How a V100 would compare with a P100 or a P40 ?
@SvRider5128 ай бұрын
I'd like to see Jeff get a hold of some of those Intel flex GPUs.
@CraftComputing8 ай бұрын
Me too fam. Me too.
@liamgibbins4 ай бұрын
mines not in the garage but next to the tv in the lounge on my r730, a tad loud but it heats my home ok
@titaniummechanism32148 ай бұрын
2:19 Well, the pcie 8 Pin is RATED for 150 Watts, but I think it's widely accepted that it is in fact much more capable than that.
@JPDuffy8 ай бұрын
I picked up a cheap Tesla P4 a while back to play with headless gaming. That never worked out too well as streaming seemed to cut too deeply into the tiny 75W max TDP. Instead I have it in an old Haswell Xeon with Windows 11 and with the clocks unlocked it's fantastic. I'm getting well over 60FPS in everything I play, often 90+ at high/med settings. I'd have gotten a P100, but after trying to cool an M40 and failing a couple years ago I decided to keep it simple. Take the shroud off the P4, strap on a fan and you're done.
@BrunodeSouzaLino8 ай бұрын
The EPS connector is so better at power that the RTX A6000 only needs a single connector instead of multiple 8 pin connectors or that abomination that is the 12VHPWR connector. And its TDP is only 100W less than the 4090.
@SpreadingKnowledgeVlogs8 ай бұрын
amazon luna uses the tesla cards for their cloud gaming service. I believe its the t4 but i cant remember which one exactly but can double check again the way you find out is doing a bench mark in some games it will list the specs. They use crap xeon cpus though.
@MrButuz8 ай бұрын
It would be great if you could do a chart with bare metal score in the game then 1 virtual with the game then 2 virtual one game one heaven benchmark. That would let us know what we're in for in all scenarios.
@realiesOne8 ай бұрын
can the 3090 be patched for vgpu use?
@remcool12588 ай бұрын
I would love to see how it compares with a p40
@recoveryguru8 ай бұрын
That computer at the beginning has got to be from about 2001, doubt it's going to work with Maximum Settings. 🤣
@markthompson42258 ай бұрын
You forgot about video rendering in jellyfin, pled and emby...
@lil.shaman63848 ай бұрын
I dont have friends, can you finally make a video about using multiple vGPUs for one host or a cluster of proxmox nodes automatically scaling on a single vm load?
@ProjectPhysX8 ай бұрын
The P100 is a beast in FP64 compute, it smokes all of the newer cards there, 3.7x faster than even an RTX 4090. P100 is a very different microarchitecture from the rest of Pascal cards, with 1:2 FP64:FP32 ratio. Today this is only seen in A100/H100 data-center cards which cost upwards of $10k. FP64 is required for certain scientific workloads like orbit calculation for satellites.