A better example than a sausage factory, is a film production, all the scenes are filmed out of order, based on locations, crew availabilty etc.. But as long as the editing is ordering the scenes in the script's order (program) the result for the viewer will be acceptable
@MattGodbolt7 ай бұрын
Nice analogy!
@johngaughan93997 ай бұрын
Let's throw branch prediction into the mix. You have two units, each filming different scenes that take place at different times in the movie. Director/writer/producer make a script change, now the second unit has to re-record a new scene and disregard what they already filmed.
@klaxoncow7 ай бұрын
Ah, we'll just fix it in post, right?
@borisdorofeev56022 ай бұрын
Fantastic analogy
@gojohnniegogo7 ай бұрын
Bonus points for a Winamp reference. Topical too since the source code is being released in September!
@mgancarzjr7 ай бұрын
It really whips the llama's....
@MAGNETO-i1i7 ай бұрын
What really? Why they release it?
@user-qf6yt3id3w7 ай бұрын
It's crazy how Tomasulo's algorithm for out of order execution was was first used in the IBM System/360 Model 91 released in 1967, predating the integrated circuit.
@logantcooper67 ай бұрын
We stand on the shoulders of giants.
@MattGodbolt7 ай бұрын
I see folks are speculating on the topic already ;)
@MattGodbolt7 ай бұрын
And the people posting "first" ? " second" are displaying...out of order ;D
@giveaway40027 ай бұрын
hiii Mat, thanks for Compiler Explorer... i really love it. u r my inspiration
@MenaceInc7 ай бұрын
Your puns, much like nearly all of your talks and tools, are great 👏
7 ай бұрын
The whole time you were talking about "as long as nobody notices", my brain was going "Spectre! Spectre! Spectre!" So, nice that you did indeed point it out at the end. I find it hard to explain to people who are technical enough to understand a little bit, but not deeply technical, how it is possible that the same bug existed in essentially every mainstream CPU on the planet, including both RISC and CISC ones, both big (e.g. POWER) and small (e.g. Cortex) ones, in Intel and AMD and Apple silicon, among others. And the answer is that it exploits one of the fundamental "tricks" that make *every* CPU fast, which is basically the trick you are explaining here. Of course, this is only the beginning as the new Go.Fetch vulnerability has shown which exploits data-dependent prefetching.
@comradepeter877 ай бұрын
WAIT WHAT THE FUDGE YOU ARE THEEE MATT GODBOLT!! WE HAVE A CELEBRITY OVER HERE OMG
@frankbucciantini3887 ай бұрын
Years later someone finally explaining the Spectre / Meltdown bug properly. This is why software mitigations make the CPUs affected much slower: they limit if not disable entirely this kind of mechanism.
@DubioserKerl7 ай бұрын
Hang on, I know that logo on the shirt.... is that...? Yes, it IS Godbolt!
@linuxguy11997 ай бұрын
Men of culture, we meet again.
@bloated_complacency4 ай бұрын
I have taken uni courses here in the statesbthat are not as clear and concise as this video has been. Thank you so much for this breakdown, it has added some much needed perspective into my computer science endeavors; keep up the great work!
@jaffarbh7 ай бұрын
Simply brilliant. I would highly suggest covering Intel's "Itanium" CPU, and why it went awfully wrong.
@RKelleyCook7 ай бұрын
Love me the robots where W == waiting and W == working.
@MattGodbolt7 ай бұрын
Yeah... I didn't think that through properly did I. Last minute changes never a good idea. Hopefully you get the idea though?
@chitlitlah7 ай бұрын
I was waiting for P to go from processing to paused.
@IceMetalPunk7 ай бұрын
@@MattGodbolt Always allocate at least two bytes to your status strings 😉
@klaxoncow7 ай бұрын
@@IceMetalPunk Strings? How wasteful. What you need is an enum. 0 = empty, 1 = ready, 2 = waiting, 3 = working, 4 = completed, 5 = retired. Only 3 bits needed (with a couple of statuses to spare - I already used one to represent "empty", as a null reference to say "ignore this row, as we've not filled it out with data yet". And if we need a human-readable string then we can have an array of strings: ['empty', 'ready', 'waiting', 'working', 'completed', 'retired'], using the status as the index. And you don't need the English language to start all the words you want to use with different letters at any point.
@golangismyjam6 ай бұрын
@@klaxoncowthis is why the high level language Devs don't get much respect. They are severely lacking in fundamentals because they generally just know how to make a website.
@marcwinner5677 ай бұрын
Thanks so much for these videos. Matt is truly a great teacher of these concepts!
@nicksamek127 ай бұрын
This series has been fantastic so far!
@SimGunther7 ай бұрын
Modern x86 CPUs will have a RISC kind of pipeline in the microcode level despite the base ISA not implying such pipelining. So to answer the main question, technically the CPU is doing (N core * pipeline phases / average cycles per instruction) things all at once, but that's not a neat answer since hyperthreading is possible and not all instructions run equally as latency from prediction rollbacks, cache locality/coherency, and write interference need to be considered, among many other things.
@surters7 ай бұрын
N cores * issue width * average micro ops per instruction (max execution units) etc. etc. hundreds of limitations.
@monad_tcp7 ай бұрын
This concept of RISC/CISC is an outdated idea. All modern CPUs are implemented using microcode, which is technically RISC and on top of it we have a CISC ISA, they're all hybrid . Yes, even the simpler ones like those who use RISC-V end up having even smaller microcode for implementation, its just convenient.
@trevinbeattie48887 ай бұрын
@@monad_tcpI wouldn’t consider microcode a type of RISC since that layer typically isn’t exposed at the programmable instruction level and only microchip designers who are working with something like Verilog actually use that level of coding.
@ArneChristianRosenfeldt7 ай бұрын
MIPS exposes the pipeline as branch delay slot. RISCV kicked it out again. We are back at 8086 like ISA. MIPS already hid any data hazards of the pipeline and would rather wait for the cache or main memory. Ah, you mean that a load instruction is kinda the start of a pipeline. Next comes a compute instruction. Yeah.
@thewhitefalcon85397 ай бұрын
There isn't a fixed number of pipeline phases.
@gekko4346 ай бұрын
Thanks Computerphile, I've been loving these series on how CPUs work. Absolutely fascinating, even for a layperson like me
@unvergebeneid7 ай бұрын
14:31 "Honey, why are you shouting'Pee! Pee!' into your phone?" "No, mum! I'm shouting 'p' _at_ my phone." "Well whatever it is you do with pee and your phone, it is rather odd, isn't it, and I'd prefer for you to do it in your room."
@giveaway40027 ай бұрын
please, please more cpu videos....
@godnyx1177 ай бұрын
Agree! Low level, hardware related videos are awesome and very valuable!
@unvergebeneid7 ай бұрын
The times where a sqrt would take 100 cycles are long gone. These days it's more on the order of 16-20 cycles. Still slow compared to a multiply or add but in the ballpark of a division.
@JonBrase7 ай бұрын
20 cycles is still enough time for ~100 instructions to pile into the ROB behind the sqrt at typical pipeline widths (plus whatever was already behind the sqrt when it started executing).
@yashsinghal10238 күн бұрын
Correction Its true that multiple operations can happen in the same clock cycle , its still largely dependpended on the instructions . The floating point operations are lot different than integer point opearations and depending upon the type, the functional units that work on it may not be empty at which it would still be stalled
@JobvanderZwan5 ай бұрын
The timing of this video is interesting: Lynn Conway, the computer scientist who invented some of the fundamental techniques for out-of-order execution in the 1960s, passed away at the beginning of June.
@IceMetalPunk7 ай бұрын
As a web dev, this reminds me a ton of database transactions. You can update many columns in many rows in a database table at once, but if one fails, the transaction fails as a whole; it's "undone" because the changes never get committed to the database in the first place.
@BarafuAlbino7 ай бұрын
What is above, so is below. What is below, so is above.
@EvanTech-v3q7 ай бұрын
"Well I feel like this is a i don't know if you've ever watched Qi where the big bell goes-" *"BRRRRRRRR"*
@RecycleBin07 ай бұрын
lego can do almost anything
@EvanTech-v3q7 ай бұрын
@@RecycleBin0 It's GR8BRIK's old logo
@quitethecontrary18465 күн бұрын
best spectre explanation ever...thank you
@Darkknight5127 ай бұрын
This really kicks the llamas ass.
@loudej7 ай бұрын
Well and truly kicked indeed
@Omnifarious07 ай бұрын
5:00 - I think the first to do "pseudo-assembly" was Knuth with MIX. Though, his "pseudo-assembly" was perhaps more rigorous than what you intended.
@zwanz0r7 ай бұрын
Very nice episode! Great explanation with the todo board 😊. A great follow-up would be how modern CPUs prevent specter-like attacks, because i assumed the cache would also be reset after branch prediction failed.
@jeromethiel43237 ай бұрын
Winamp, it whips the llamas ass! ^-^ What a blast from the past.
@chaoslab7 ай бұрын
Coding assembly on the first few Archimedes machines was very interesting, multiple logical options with instructions (like conditional execution and barrel shifting).
@goshisanniichi7 ай бұрын
I remember from a computer engineering course long ago, that there was a branch "prediction" scheme with superscalar processors where the prediction part was skipped altogether. Because the processor was capable of doing multiple things at once, it could just process both branches simultaneously and throw out the one that wasn't needed at the point when the branch was finally processed. I don't think it was actually ever used and if it was then not much because actual attempts at prediction are still better in most cases.
@techmage897 ай бұрын
Some GPUs actually do this! They run hundreds or thousands of threads over the same code, but execute groups of threads in lockstep, so if a group of threads encounters a branch and they don't all take the same path, they will all run both branches and then mask out the results from the wrong branch. It's basically a way of fitting branching into a SIMD pipeline.
@masterdjon5 ай бұрын
Really great video and really great series. I would like to submit a suggestion: when talking about parallelism, I think you would have been a bit easier to follow with tick/frame number. Still following you from years. Continue your great work!
@veers0r7 ай бұрын
Love the compiler explorer. :)
@jeromethiel43237 ай бұрын
I just learned not that long ago, that Cray computers invented a lot of what you were talking about in this video. Multiple streams, pipelining (not the idea, making it better), out of order execution, et al. It's why Cray was able to hold on to the title of supercomputer for so long. They were doing things nobody else at the time could. And now it's part of pretty much every modern processor. It was a heady time.
@JonnyPowell7 ай бұрын
it was a Crayzy time
@johngaughan93997 ай бұрын
For much of computing history, microcomputers (i.e. Intel+AMD) followed the lead of mainframes (i.e. Cray+IBM). Modern NASes allow the user to hot-swap hard drives. Back in the 1970s, mainframes allowed the user to hot swap CPUs and RAM. I still can't do that in an AMD64 system in 2024.
@R.B.7 ай бұрын
I can see how Cray multicore systems could schedule a process across multiple cores, but I don't think it is the same as multicore CPUs of today. Memory access and caching would be significant in an SMP way. At the processor level you can take advantage of shared L1 cache on a single thread. I think Cray was more about distrubuted threads, but each thread wasn't context switching. What's interesting to me about this is that it suggests that the core of a CPU could get into a deadlock lower than potential software dreadlocks between threads.
@bemk7 ай бұрын
This something you don't generally need, but when you do need it it's very useful to have. Had a bug the other day in an embedded system where the file system chip would put data on the bus despite its driver never asking for it, because an RSA calculation somehow triggered a speculative read in the chip's DMA region. Debugging was infuriating, 'cause every time you'd change some code to do an inspection, or would even interfere with a debugger, you'd interfere with the pipeline causing the bug to disappear. Only reason we ended up finding the solution was due to some very kind and skilled people in some forums all over the internet and an erratum for the chip
@andrewharrison84367 ай бұрын
Parallel processing can be "fun" to debug. I once had an exit process that started failing. It ran perfectly in debug because waiting for me to respond to the prompt gave the parallel clean up process time to complete. My fix was a bodge: I shuffled the order of my code till it worked. My best guess as to the cause of the failures was that a release of the compiler or operating system had done something to the timings.
@bemk7 ай бұрын
@@andrewharrison8436 been there, done that, joined the club, got the t-shirt. At least it's a problem you can solve with some semaphores though... That said, the feeling you get when you actually solve an issue like this. Very satisfying
@FutureAIDev20157 ай бұрын
Jump instructions have entered the chat
@esra_erimez7 ай бұрын
😂
@ScottLovenberg7 ай бұрын
I know Goto when I see their alt!
@scaredyfish7 ай бұрын
I like the desktop background - where's it from?
@adfaklsdjf7 ай бұрын
thanks for mentioning spectre at the end :)
@morwar_5 ай бұрын
One time I asked a question on Matt's youtube channel and it was answered.
@MissNorington7 ай бұрын
CPUs are actually much faster than in this example. They can see ahead of time that you are squaring the numbers, so there is no need to square root at the end. The robots inside the CPU are probably trained with captcha as well, which is still legal when this comment was written
@Kalernor7 ай бұрын
Which area of computer science is this? Where can I learn more about it? Also, what is this topic called? Out of order operations? Again, where can I learn more about it?
@axelanderson20307 ай бұрын
Google and literature piracy
@borchen07 ай бұрын
Do modern compilers produce machinecode that help the CPU with this process? A sort of preprocess, so less stuff has to be undone?
@marsovac5 ай бұрын
A compiler does not have the runtime values that trigger branching. Also there is no mechanism in the ISA to tell to the cpu: "you're most likely to go this way" even if you knew it. But the cpu does keep track of it in repeated branches by itself, it is what branch prediction does.
@MartinLindsay7 ай бұрын
One nitpick on the description of the "retirement" stage you described, and I apologize as I realize it may be a detail intentionally not visited here. While it is simple to say that it goes down in order committing things back out of the cpu, it only prevents committing if there is an unresolved conditional branch before the instruction. Things otherwise can commit back to main memory or cache out of order, and is one reason that atomics use acquire/release semantics to assert control over the ordering.
@stephenwhite5067 ай бұрын
In a single cycle, modern CPUs are performing thousands of XORs in parallel for every cache line check and can do this in parallel for each of the instruction, data or address translation caches. So, the answer should be thousands.
@TymexComputing7 ай бұрын
No Byte Bug puzzle here? Jane Street has addressed today's Numberphile BugByte puzzle 1-24 as Computerphile puzzle.
Obviously when the branch prediction fails, a lot of these instructions need to be thrown out and redone, but is there anything stopping a CPU core from taking instructions from both branches, adding it to this out of order table, and letting both get executed until it knows which branch is taken and can evict all of the instructions from the wrong branch?
@nayjames1237 ай бұрын
Given you'd have to create a new instruction stream for every branch you see, and how frequently you come across branches. You'd end up needing a huge amount of instruction streams that are nearly all going to be discarded. The cost of adding all these streams is also gonna be very high and cause bottlenecks. To help the throughput when there are branch mispredictions CPU vendors created things like hyper threaded cores, where if one stream of instructions has a mis predict, the other stream will have less contention for execution units so likely be able to run a little quicker
@jeromethiel43237 ай бұрын
I thought this a while ago, and realized that it would fail pretty quickly, just based on how often branches happened. The real issue is that the processor is already so much faster than the memory, that the memory is the weakest link here, not the processor. If you tried to fill both branches of the pipeline, that's twice the memory bandwidth. If it's all in cache, you might be okay, but can you guarantee that? You cannot. Especially when your pipeline is as deep as it is in modern CPU's.
@Revoker12217 ай бұрын
What you're describing is predicated execution which tend to happen a lot in vectorised computing and pipelines, think SIMD or GPUs, where both sides of a branch are executed but only one side of the branch is kept at a time. An advantage of this approach is that you get to keep the speed boost of vectorised computing, but at the cost of reduced throughput as extra code is ran. If you're interested in learning more, you can read "Conversion of control dependence to data dependence" for one of the first places where this idea was introduced and fleshed out, or for something more modern, give "ISPC: A SPMD compiler for high performance CPU programming" a look over (or any other modern day SIMD tutorial. Most SIMD these days come with instructions purpose built for this kind of predicated filtering)
@ArneChristianRosenfeldt7 ай бұрын
@@jeromethiel4323branch prediction also fails pretty fast based on how often branches happen. Or is there some “bunching” of successful predictions?
@henryprickett58997 ай бұрын
Usually successful predictions happen in a row. 99% hit rate looks more like 1 million hits followed by 10k misses than 100 hits followed by 1 miss.
@Ascended_BUPАй бұрын
How do the operation get decoded faster than the robots are carrying them out, i thought all operations in a computer happened more or less at the same speed as dictated by the clock cycle (with the exception of things like a square root for example which is actually many operations one after the other)? I havent watched the other videos in this series so apologies if my question was already answered in one of them
@eliasross45767 ай бұрын
Interestingly for concurrent programming, out of order operations like stores can be visible to other threads since the pipeline executor won’t have visibility into what another pipeline executor is doing.
@MattGodbolt7 ай бұрын
It very much depends on your CPU architecture. X86 makes some pretty strong guarantees (for normal load and stores). And speculative stores shouldn't be visible to other threads under any circumstance that I'm aware of.
@axelBr17 ай бұрын
It's amazing what can be implemented using logic gates etched into silicon.
@stevefan82837 ай бұрын
7:12 but can't this example be implemented with a fused-multiply add and a hardware square root instead...?
@trevinbeattie48887 ай бұрын
It’s only a model
@kevinscales7 ай бұрын
The CPU he is emulating doesn't have that, no
@solhsa7 ай бұрын
you might as well have a single instruction that calculates distance. It's just an example.
@davidhand97217 ай бұрын
How is it writing values to that table, then checking those values to see if it's ready, without going through at least one clock cycle? How are they managing synchronization?
@The_Pariah7 ай бұрын
To this day, I STILL use WinAmp. Best music player ever.
@rogerlevasseur3977 ай бұрын
Let's dive into Very Long Instruction Set computer architecture (Intel's Itanium and Multiflow's minisuper) where the compiler from it's analysis determines which instructions run in parallel together. simplifies the pipeline.
@YaofuZhou7 ай бұрын
I guess if you throw security considerations into the mix, things quickly become super complicated ;)
@Tahgtahv7 ай бұрын
So, mentioned at the end of the video were issues that sounded like they might affect timing. (eg, something unexpected was in a cache, or got bumped out of a cache.) However, how do you account for actual memory side effects? For example, there are some chips that automatically advance an address on reads/writes. How do you ensure that access to such a chip is strictly in order, with no speculative/out of order access?
@henryprickett58997 ай бұрын
You only write at the retire stage (writes are in order) and you cache "writes" to another buffer so that out of order reads get the right result at the right time. Beyond that, cache coherency protocols keep it safe between processors.
@henryprickett58997 ай бұрын
That is to say, you cache out of order writes before retirement locally, so that out of order reads see the correct values and don't need to wait for writes to retire.
@Tony_Indiana6 ай бұрын
Who/Where do we send out money to so we keep on getting more videos? According to my recent polling myself and others would be most interested in LLM. Though anything amazing is ok. We just need to know where to send the cash/xmr/"coin"
@discoisdead85047 ай бұрын
Nice vid about cpu internals 👍
@unvergebeneid7 ай бұрын
You just know he chose a whiteboard because he was under the illusion that he'd just erase the old values in the cells instead of making a complete mess by repeatedly crossing things out 😄
@zxuiji7 ай бұрын
7:07 For the game example you gave this would actually be way more than you need. Just multiply √0.5 against t1 + t3 and you'll get a close enough (if not exact) result. √0.5 just needs to be done once at the start of the game and stored for later use.
@TiT88517 ай бұрын
Great, but what happen to the pipeline when an interrupt occur (like syscall or time slice expired from OS)?
@trevinbeattie48887 ай бұрын
An interrupt is basically a jump with some extra steps to save the current CPU state and switch to a new context, so I imagine it has to go through the pipeline just like any other part of the program. Whatever is currently in the pipeline may either finish going through or be discarding depending on how interrupt handling is implemented.
@TiT88517 ай бұрын
Thank you.
@VandalIO7 ай бұрын
Is there a video of how alu does barrel shifting ?
@AnotherPointOfView9447 ай бұрын
several ways, depending on the number of bits/word. Easiest way is lookup table.
@user-qf6yt3id3w7 ай бұрын
If you have a bunch of registers of the input shifting 0 bits, 1, bit, up to 32 or 64 bits and then just select from the one you need you've got a barrel shifter. It's simple but it obviously takes up a lot of area.
@VandalIO7 ай бұрын
@@user-qf6yt3id3w I know how that works ! But how’d you implement that in pure logic ? , just try designing an 8 bit barely shifter with logic gates and multiplexors, it gets crazy complicated
@VandalIO7 ай бұрын
@@AnotherPointOfView944 lookup table is cheating 😂
@user-qf6yt3id3w7 ай бұрын
@@VandalIO On an IC the wiring is not too bad because you can run the bit lines diagonally. Obviously you can't get around the fact that an n bit barrel shifter needs an n*n array of cells. Or even 2n*n if you want expanding shifts which is handy in graphics.On the other hand the cell doesn't need to be a flip flop. Actually if you read the Wiki article there's a better way to do it which is better for wider devices" "The very fastest shifters are implemented as full crossbars, in a manner similar to the 4-bit shifter depicted above, only larger. These incur the least delay, with the output always a single gate delay behind the input to be shifted (after allowing the small time needed for the shift count decoder to settle; this penalty, however, is only incurred when the shift count changes). These crossbar shifters require however n2 gates for n-bit shifts. Because of this, the barrel shifter is often implemented as a cascade of parallel 2×1 multiplexers instead, which allows a large reduction in gate count, now growing only with n x log n; the propagation delay is however larger, growing with log n (instead of being constant as with the crossbar shifter). For an 8-bit barrel shifter, two intermediate signals are used which shifts by four and two bits, or passes the same data, based on the value of S[2] and S[1]. This signal is then shifted by another multiplexer, which is controlled by S[0]"
@Samhain__UK2 ай бұрын
I was hoping for more in this series. Anything in the (OoO) pipeline?
@user-dv5gm2gc3u7 ай бұрын
once heard jim keller talk about this in an interview. Kinda insane what's happening in a cpu nowadays.
@ScottLovenberg7 ай бұрын
"I'll give myself infinite registers." - hold every possible outcome and just keep paying the memory wall tax. Now it's a physics problem.
@ScottLovenberg7 ай бұрын
Wait.... We can't pack them close enough to use this solution. Nevermind. I just moved the memory wall to everywhere and introduced a singularity to the least efficient solution. My bad, guys.
@solhsa7 ай бұрын
@@ScottLovenberg See also: SSA form, example: spir-v.
@RealCadde7 ай бұрын
The answer to the question changes based on definition. If you subscribe to the notion that computers can do as many things as there are cores in the system, then computers can to thousands of things every clock cycle. Assuming the computer has a graphics card with that many cores. If you also subscribe to the notion that each core does many things at the same time on every clock cycle, well then you have many thousands of things happening every clock cycle. And if you then look at a "super computer" (a rack of computers all tied together) well now they do millions to billions of things, billions of times per second. But realistically, computers (doesn't matter how many cores they have, how many memories they have, how their pipelines are set up) can only really do ONE thing per clock cycle. At the end of the day, there NORMALLY is a single INPUT and a single OUTPUT. You can do one INPUT every clock cycle and you can do one OUTPUT every clock cycle. Yes, there are many cases where inputs and outputs are completely separated from each other. They don't cross communicate in ANY way, they could just as well be separate computers doing separate things. But then i'd argue that you should have asked "how many things can N number of computers do" and the answer would be N things. So WHY can't many inputs and outputs be handled by a single computer at the same time, all the time? Because the inputs and outputs, somewhere in the process, needs to "talk" to each other or the operating system or a shared storage or the network. Once they do, they are bottlenecked by that interaction. Just because a modern computer can work on many different tasks at the same time, it still isn't truly completely parallel. The universe is truly parallel, there are no bottlenecks in that sense. So are, i would argue at least, quantum computers. Digital computers are only multitasking MID process, but at the end of the day everything comes together at a single input and output point.
@bity-bite7 ай бұрын
If a CPU had two cores, it means it has four threads, which means it can do 4 things at the same time, right? Would it mean the CPU accepts 4 inputs and with 4 outputs?
@landsgevaer7 ай бұрын
A computer can simultaneously display a cursor, rotate its fan, make a mouse click noise, stimulate my retina, heat the room, annoy the cat, support a geranium, etc...
@RealCadde7 ай бұрын
@@landsgevaer In a time slice...
@RealCadde7 ай бұрын
@@bity-bite A CPU can have 1, 2, 3, 4... cores. Doesn't matter though because it still depends on definition. I see each core as a separate computer.
@trevinbeattie48887 ай бұрын
If you go to the microarchitecture level (as suggested in the video), a single CPU core can do as many operations as it has processing units for. If you go down further to the transistor level, a CPU can do millions of gate operations at once. It all depends on what “things” you’re counting.
@froop23937 ай бұрын
Where can I get such a cool compiler explorer hoodie? btw: great tool, really love it!
@dakiloth7 ай бұрын
Is it possible for a processor to calculate both branches of a branch?
@solhsa7 ай бұрын
yes, and I think I've read some cpu archs doing that... might have been itanium, but I'm not sure. It's wasteful, though.
@deltamico7 ай бұрын
If this paralelism is integrated, does the bend language achieve anything?
@solhsa7 ай бұрын
Compilers can do larger scale things, and discard operations altogether, which is way harder on this level.
@NatePerdomo7 ай бұрын
What is his purple buffalo desktop wallpaper image?
@MattGodbolt7 ай бұрын
It's one of the default Ubuntu ones; a Minotaur I think?
@74_Green7 ай бұрын
@@MattGodbolt Mantic Minotaur 23.10
@3rdalbum7 ай бұрын
He'll want to upgrade that before too long, doesn't it go end of life in a few days?
@Razzbow5 ай бұрын
Yeah a CPU can only do one thing at a time. That's why modern CPUs are composed of many seperate processors which themselves break tasks down into primitive blocks which can be parsed and dispatched in paralel.
@henrycobb7 ай бұрын
When a machine is out of order it needs to get fixed. C.f. the Spectre and Meltdown bugs.
@kayakMike10007 ай бұрын
Depends on how many hardware thread cores it has. RISC-V calls these harts.
@EdgyNumber17 ай бұрын
Remember the early days when XBox programmers slated PS3 CELL because it was 'too slow?' It took a while to shake off the old way of doing things and really jump on board the idea of parallel computing. Are modern x86 processors able to handle true parallelism these days or have they simply had to bodge something extra onto what is ancient architecture?
@Lion_McLionhead7 ай бұрын
So what's the Apple M3 & raspberry pi 5's biggest weapon?
@seanharricharan76026 ай бұрын
What does he actually mean when he says he has 4 CPUs on his laptop with 10 Units each?
@ukyoize7 ай бұрын
I wonder why there isn't "Don't predict this branch, or predict it THIS way" for security.
@JohnSmith-op7ls7 ай бұрын
The correct answer is one. Just because multi-core CPUs can do more than one at a time doesn’t change that because they’re really just multiple CPUs with some shared resources, on the same wafer. It’s a more compact version of multi-processor motherboards. Each core is a CPU and doing one thing at a time. We’ve bastardized the meaning of CPU for simplicity when it comes to multi-core chips, but logically they are not a single unit, they are multiple units working in tandem.
@professorx40477 ай бұрын
Well he said computer not cpu
@ArneChristianRosenfeldt7 ай бұрын
Isn’t it interesting how you can lock() an object in Java on one core, and the other cores respect this despite the widespread use of cache? How did that work with CPUs, like the two SH2 in SEGA 32x? Jaguar and PS3 cell only have scratchpad memory. JRISC in the Jaguar can access shared memory, but lacks instructions for locks. I think that it was not allowed for two of the 3 (4) processors to write to the same object . Only producer, consumer pipeline was allowed. So, in a queue each party only writes to their pointer: read or write. Each party reads the other pointer to check if the queue is full or empty.
@nathanbolstad95677 ай бұрын
You can only write to a RAM cell from one source. The bus gets locked. CPUs have had a test and set instruction basically forever. The instruction locks the bus until complete. Locks in high level languages mark that team cel (logically) as volatile which forces the view on the CPU to skip the cache and go out to the memory every time to ensure the lock can be acquired. As a result, locking has greater overhead and has greater affect than code that doesn't require locking/semaphores. Locking hurts parallelism too. As a result, in recent years, there has been a fair amount of work on basic data structures and the like that can be implemented without locks to address the pain of locking and scalability.
@ArneChristianRosenfeldt7 ай бұрын
@@nathanbolstad9567 volatile in C just prevents that the value is „cached“ in a register. A binary compiled for my 386sx doesn’t know about cache. Only thing is that writes to video ram or sound queues are so large that it flushes any write back buffer. X86 has ports to signal that the value has to go out. Quake light maps were calculated on a 4 CPU Pentium board. No test and set instruction. So like on game consoles of the time lockless data structures were common. Amiga broke the 68k TAS instruction because the blitter and the CPU were never supposed to write into the same data structure.
@illford6 ай бұрын
weird quotation marks, what language uses them? Real question not anything malicious
@NatePerdomo7 ай бұрын
I'm usually lost about 45 seconds into these videos. Still love 'em, though.
@surters7 ай бұрын
Hey, just read the entire "Computer Architecture - a Quantitative approach" and you will understand!!!
@phasm427 ай бұрын
Interesting that Itanium tried to move this complexity out of the CPU and into the compiler, but ultimately keeping this stuff in the hardware (x86) won.
@joshd797 ай бұрын
Winamp!
@BurgerKingHarkinian7 ай бұрын
It really whips the lamas ass!
@nathanbolstad95677 ай бұрын
And it's now going "open sauce" (sic).
@christopherlawley18427 ай бұрын
still using it
@alex846327 ай бұрын
Why don't compilers do this with cpu cores to automatically make every program multi-threaded? Too much overhead?
@solhsa7 ай бұрын
There is actually some research into "multi-core single threading", but as far as I know it never got anywhere.
@kazedcat7 ай бұрын
The problem is branch prediction. Unless you are running and calculating real values you cannot do branch prediction at compiler time.
@ProjectPhysX7 ай бұрын
My computer does 3584 instructions at once. GPU SIMT magic.
@gregorymorse84237 ай бұрын
Square root is a one tick unary operation as fast as an add. Common misconception. Trig instructions are slowest on modern CPUs
@martinbakker76157 ай бұрын
Hate these kind of questions because they are never specific enough. Eg, he says in the 80 computers did one thing at a time. Reading bits out of memory still was 8 at a time.
@werds13927 ай бұрын
He’s talking about one clock cycle
@ArneChristianRosenfeldt7 ай бұрын
Bit logic is a Vector Computer Right there. Archimedes had a 32bit ALU in 1987. And 32 bit memory bus. 68k had 16 bit in the 70s, but needs 4 cycles for everything.
@toby99997 ай бұрын
Depends on how you define one thing... 8bits OR is it one byte...
@dipi717 ай бұрын
21:34 Of course we notice - branch prediction uses more power and generates more waste heat. It can also rip nice holes into your cyber security. It angers me to state all that in the first place.
@floppy85686 ай бұрын
This guy has 4 10-core CPUs in his laptop
@PhilipMurphy8Extra7 ай бұрын
Computerphile would never air on UK TV, it's way too smart for the TV broadcasters these days
@humansizedaperture6 ай бұрын
❤
@drooplug7 ай бұрын
RIP Winamp.
@nathanbolstad95677 ай бұрын
Nope. It's being released as open source (recent news!) Expect it to reappear soon!
@nathanbolstad95677 ай бұрын
24 Sep 2024 is the announce release of the source!
@M0rn1n6St4r6 ай бұрын
4 weeks since Computerphile's last video? What's going on? Are they done making videos? ¯\_(ツ)_/¯
@VaughanMcAlley7 ай бұрын
I wonder how much higher level scripting languages benefit from this. Presumably a fair bit as the interpreters are mostly just C programs…
@badcrab74947 ай бұрын
Before watching, Speculative computation?
@MrJleonp7 ай бұрын
That's an acceleration method, they will probably talk about how CPU aren't really multitasking, just assigning cru cycles to different tasks.
@TheEulerID6 ай бұрын
It was most certainly not true in the 1980s that computers could do only one thing at a time. There were mainstream SMP computers around from the early 1960s.
@phill6859Ай бұрын
The question is too vague. What do you mean by "computer" and what do you mean by "things"? Do you just mean the cpu and running instructions? Because that isn't what you asked.
@ElvenSpellmaker7 ай бұрын
Spectre and Meltdown want to fight!
@ve3uom7 ай бұрын
OMG I suddenly miss Winamp!
@snowballeffect78127 ай бұрын
love the QI reference lol
@esra_erimez7 ай бұрын
26th?
@vadrif-draco6 ай бұрын
Damn, winamp
@love_exegence7 ай бұрын
He’s cute
@var677 ай бұрын
Has a whiteboard which is easy to wipe, still crosses out all the old words...........
@syjwg7 ай бұрын
Instead of two inputs, I'm sure a computer in the future will handle three inputs. Something like 1 + 2 + 3, That would be awesome.
@pudy24877 ай бұрын
Three-operand instructions are as old as x86 is with LEA being the earliest and most important among them. Most vector instructions nowadays take 3 operands as well, so that a separate destination register may be specified.
@Disguised_Hawk7 ай бұрын
Please use the correct terms for things you are explaining. Like, don't call a processor that has more cores than one a computer with multiple CPUs. The computer still has one CPU that contains multiple cores, which individually are hyperthreading the instructions. Not naming things correctly or at all makes the thing you explain very hard to relate to other information that one might have on this topic. Therefore people will learn less because they can't connect it. Naming cores CPUs makes people who already what you are talking about either confused or suspicious of your research. PLEASE just name things the CORRECT way. Otherwise the main part of this video is better than the beginning, thankfully
@KingJellyfishII7 ай бұрын
the problem is that "CPU" is quite an ill-defined term. Most people use it to refer to the chip or even whole unit that contains one or more cores, but by a computer science definition each core is in fact its own CPU. It's simply an inconsistency in naming between the "layperson" and the "academic", for lack of better words.
@Roomsaver5 ай бұрын
@@KingJellyfishIIWhy would the whole unit not be the CPU? It’s the central processing UNIT, after all
@KingJellyfishII5 ай бұрын
@@Roomsaver CPU isn't defined by being a central processing unit (what does that even mean, anyway?). that's just a quick description. the reason is, in the design of one multi core processor, each core is its own CPU. it could operate separately from the other cores as its own standalone CPU. This is done commercially to some extent, when producing a silicon die for a processor sometimes there's a defect in one or more cores, so they're simply switched off and the die can still be used albeit with a lower core count.
@savagesarethebest72515 ай бұрын
Hyperthreads is an Intel terminology and it doesn't really corresponds to physical cores on the CPU chip.
@juliethecyborgКүн бұрын
@@KingJellyfishIIit's more complicated than that -- they can share higher level caches, clock circuitry, and so on