Great TCP "dynamic window" explanation with a history lesson and explanation not only how it works but why we needed it in the first place. Much, much better than in CCNA books.
@EmptyGlass992 жыл бұрын
I love the way Dr Clegg explains this and interacts with his audience. Great teaching.
@I_leave_mean_comments2 жыл бұрын
I need to listen to him at 1.5 speed minimum... or else his awkward pausing drives me insane.
@zelllers2 жыл бұрын
TCP congestion control is one of my favorite topics in the entire world. Even today we're still trying to modernize it, there's been some great work with BBR
@chrisknestrick3742 жыл бұрын
I remember reading - and re-reading - that paper in grad school. Truly a seminal paper and an EXCELLENT job of describing it!
@MalongaModeste Жыл бұрын
What paper is it?
@dreamzens2 жыл бұрын
I've been learning about this in class this quarter, it's such a cool concept. Currently, our class is working on a project to make UDP reliable with congestion control too. So awesome to see a Computerphile video on it as well! Thank you Dr. Clegg!
@richardclegg80272 жыл бұрын
You might want to look up QUIC if you did not already.
@RussellTeapot2 жыл бұрын
Wait what? Don't take this as an harsh comment, I'm genuinely interested and don't know much about communication protocols, so bear with me: isn't this kind of "reinventing the wheel"? As far as I know, UDP by design doesn't care about reliability and congestion, just "multiplexing", in the sense that using the concept of ports multiple applications can communicate through the same connection. By adding the other features (typical of TCP) don't you make it more heavy and defeat is purpose?
@richardclegg80272 жыл бұрын
@@RussellTeapot so Google's QUIC basically uses UDP and implements reliability features at the application layer. They can use some tricks to then get improved performance versus TCP. When I type this comment it is most likely being sent over that style of connection. A high proportion of traffic to/from Google owned services does this.
@dreamzens2 жыл бұрын
@@RussellTeapot It's just a class-assigned project for learning purposes to understand the underlying mechanisms of TCP, so the intention is to "reinvent the wheel", so to speak. You are right, though. It's for the sake of pedagogy, I suppose.
@dreamzens2 жыл бұрын
@@richardclegg8027 We briefly discussed QUIC with HTTP/2.0 in class and it seemed quite interesting. I will do more research on it, thank you!
@Phlarx2 жыл бұрын
I had the opportunity to work with Mike Karels (one of the coauthors of that paper) a few years back. He's a great role model... humble and eager to help out the less-experienced programmers (like me). He'd probably be happy to speak with the Computerphile team, if they wanted to connect.
@artiem52622 жыл бұрын
Thank you for discussing this classic and important paper -- we need periodic reminders on these issues, as the problems keep re-appearing when the long-ago solutions are forgotten. Another I'd like to see (from the dark ages) is Denning's Working Set Model for Program Behaviour from the late 60's, when computers were still huge beasts that remembered using little donuts made of rust.
@oresteszoupanos2 жыл бұрын
These days, with 32 kbps you can have a very fine phone call, but that's how far lossy audio compression has come. Opus codec for the win :-)
@13cbt132 жыл бұрын
I didn't believe you and I did some research. Listened to a conversation at 16kbps and it was crystal clear. Opus codec for the win indeed.
@JoshWalker12 жыл бұрын
indeed
@BinaryCounter2 жыл бұрын
Opus is amazing. You can get very intelligible speech down to 6kbits. I used this to listen to podcasts over 2G internet. I had a server that I would connect to through SSH and then instruct to download videos or podcast from various sites, convert them to Opus 16kbits and put them on a webserver. I would then use VLC on my phone to stream those over a 32kbit 2G connection and listen to my favorite creators that way. Desperate times hehe.
@kvatikoss17302 жыл бұрын
@@BinaryCounter I've been thinking of making a custom Spotify like this
@unfa002 жыл бұрын
Opus is absolutely amazing. It's hard to imagine there will ever be anything better invented for lossy digital audio compression. It can also use extremely small buffers for minimal encode/decode latency and change all of it's properties smoothly mid-stream - it's perfect for voice chat, especially in multi-year games where you don't want to use too much bandwidth, or it could delay the most important game update packets.
@Lumaraf12 жыл бұрын
Thanks for the nice explanation. Could you also do a video on quic and how that handles congestion differently?
@richardclegg80272 жыл бұрын
QUIC is super interesting for sure. I thought this background was necessary first though - otherwise I need to explain TCP and QUIC.:)
@SakarPudasaini102 жыл бұрын
@@richardclegg8027 Please do make a video on QUIC. How acknowledgements, packet loss, flow control and congestion control works with QUIC.
@SakarPudasaini102 жыл бұрын
++Global Synchronization
@aande12 жыл бұрын
@@richardclegg8027 Yes, please make a video on QUIC. I'd be very interested as well!
@mrxmry32642 жыл бұрын
thing is, i hadn't even heard about quic until i started using syncthing, which explains why i don't know anything about it.
@bluegizmo19832 жыл бұрын
The early days of the internet were so fascinating! I was born in '83 and got my first computer around 1995, it had a Pentium CPU clocked at a whooping 133 MHz. I remember dialing into BBS boards and accessing the early internet via AOL and getting kicked off the internet because someone picked up the phone and buying my first 1GB 5.25" hard drive thinking there's no way I'd ever be able to fill that much space 😂
@ingolfstraube84332 жыл бұрын
Wonderful explanation. I've wondered in the early dayes, how they cept IT working
@kesslerdupont60232 жыл бұрын
Thanks for the Greta video! I have seen my network bandwidth throttle in task manager in a similar way to the curve described and I always wondered why.
@kentw.england23052 жыл бұрын
Van had to have his arm twisted to publish because journals didn't want articles that cited an email. Now it is one of the most cited articles in internet research.
@richardclegg80272 жыл бұрын
Yes. I only met him once but he seems a very modest guy.
@RonJohn632 жыл бұрын
I remember Van Jacobson header compression in PPP configurations.
@johseh53122 жыл бұрын
I like this guy, he's got a jazzy manner of communicating. There's pep to it.
@MrKristian2522 жыл бұрын
I really love how he talks and explains
@thomasbonse2 жыл бұрын
I've actually had worse speed tests when my only ISP option was Comcast. I would generally only able to get 37bps with 75% packet loss on top of that. And that was in 2012-2013 in the DC metro area!
@JorgetePanete2 жыл бұрын
rip
@laurendoe1682 жыл бұрын
What I find interesting is that in addition to a missing packet, sometimes there are errors in the received packet. Sometimes these errors can be corrected, and sometimes they cannot. When they cannot, the packet needs to be resent. This factor was not mentioned.
@autohmae2 жыл бұрын
yes, a lot of the time in those days not whole packets were lost, but bits of packets
@richardclegg80272 жыл бұрын
Yes --- there are a lot of mechanisms I didn't really have time to go through including things like duplicate ACKs (which is the mechanism for what you mention, a packet that fails to checksum at the receiver). When I teach the topic throughly it's about 10 hours of lectures.
@laurendoe1682 жыл бұрын
@@richardclegg8027 Thank you for your reply. Eagerly awaiting the videos! (Although I'd love them, I do realize how much work it would be so I kid instead).
@richardclegg80272 жыл бұрын
@@laurendoe168 Happy to get the feedback. There's a lot to say about TCP for sure.
@kvetter2 жыл бұрын
My understanding is that there is a problem with slow start vis-a-vis web browsing. Slow start works great for connections which are open for a long time--the delay caused by slow start isn't noticeable. But web browsing usually involves lots of short connections for which slow start is a significant problem.
@richardclegg80272 жыл бұрын
So there's lots of work on this, some approaches include TCP "Fast open" and "QUIC".
@squelchedotter2 жыл бұрын
Would love to learn a bit about BBR, a proposed replacement for this algorithm!
@richardclegg80272 жыл бұрын
BBR is a TCP flavour - so it uses these mechanisms but in a clever way. At heart though it is still a window increasing and decreasing with congestion.
@perrylund39952 жыл бұрын
Will be using in my college Data Communications class as supplement.
@SeeonX2 жыл бұрын
Can you please do a video about when Covid started a lot of areas ISPS in US and EU region hit prime time issues. Like between 5PM to 11PM areas bandwidth would drop in download speed. This was a huge issue but no one really talked about it.
@syntaxerorr2 жыл бұрын
Great video. Thanks.
@s_s_s_s_s_s_s_s_s_2 жыл бұрын
i think brady has osmosis'd a computer science degree at this point, his suggestions are almost always on point
@kevintedder42022 жыл бұрын
Can you do a further video on the unintended effects of this congestion avoidance, such as TCP synchronisation, and how this can be avoided?
@BobFrTube2 жыл бұрын
NIce - I had assumed that back off was in the original implementation. Have you done a video on the modern version of this problem -- buffer bloat?
@richardclegg80272 жыл бұрын
Backoff is also part of the solution and backoff techniques are in the paper too. I don't think there is a computerphile on bufferbloat -- would be an interesting one.
@mrxmry32642 жыл бұрын
8:07 that is how the old xmodem protocol worked, isn't it? yes, i did some BBS stuff back in those days, first using an acoustic coupler (300 bps, yawn), later using a series of increasingly fast modems... man, this brings back memories...
@schifoso2 жыл бұрын
XMODEM was very basic and just sent a packet after every ACK. ZMODEM was probably the best as it allowed for streaming, file names and sizes, batch sends, packet size adjustments due to poor line quality, etc.
@mrxmry32642 жыл бұрын
@@schifoso i don't remember exactly how they all worked, but i do remember that ymodem was better (faster) than xmodem, and zmodem was better than ymodem. and then there was the hardware or software handshake between the modem and the computer...
@adfaklsdjf2 жыл бұрын
@@mrxmry3264 If only the alphabet provided letters faster than Z :(
@jwydubak96732 жыл бұрын
It is not a problem for xmodem because the link between two machines is direct and it is impossible there to be more than one packet in-flight at any given time. Whereas in a network there may be roughly as many packets in-flight as there is routers between two machines.
@rfvtgbzhn11 ай бұрын
@@adfaklsdjf they could have used the first version AMODEM instead of XMODEM, then they would have had enough letters left.
@kufena2 жыл бұрын
In my bit of Norfolk, sometimes, 40bps would seem good.
@gabrieldesantanalacerda2 жыл бұрын
That was so good! I have a question: does these tricks are today implemented automatically by a network protocol or are those a set of practices that the ISP needs to configure manually?
@danielhartley132 жыл бұрын
TCP congestion control is implemented by the protocol itself. Its end-to-end driven, rather than network assisted. This is because internet protocol doesn't pass congestion information to TCP, so TCP uses its own end-to-end driven method where it inferes the network congestion state itself using backoff timers, or 'probing' as the guy in the video puts it.
@richardclegg80272 жыл бұрын
TCP is implemented at the end computers (sender and receiver). It is part of your operating system. Depending on your OS you can tweak exact details.
@shez6662 жыл бұрын
I'm glad he did correctly point out that the TCP window is based on bytes not packets as he kept saying but disappointed it took 15 minutes
@MalongaModeste Жыл бұрын
Can you do a video on how TCP/IP work please Dr.
@phutureproof2 жыл бұрын
Im so glad you cut away to those shots of yourself drinking tea and nodding /s
@andljoy2 жыл бұрын
I have a question for Dr Richard Clegg. What would cause packet loss on a network that increases and decreases in a almost prefect sign wave with peaks every 10 mins ( around 7% loss ). I believe its a buffer that is filling up on the main router causing it. The clients in turn reduces the TCP frame size to a point where the speed on the network drops and the router can then handle it. The frame size then slowly increases again and the problem repeats.
@richardclegg80272 жыл бұрын
Hmm... Honestly that periodicity is very slow. Most things operate on a quicker time frame than that. (Excuse me for asking could there be a milisecond microsecond confusion - if the timescale was much quicker the behaviour would be more normal). I guess you have grabbed a pcap and looked at it through wireshark to determine this behaviour?
@framegrace12 жыл бұрын
Industrial environment? Check for interferences....
@jwydubak96732 жыл бұрын
Can we have a description of ECN (Explicit Congestion Notification) protocol in one of future videos?
@elraviv2 жыл бұрын
It's a very good video. Do a follow up with "silly window syndrome" that will be fun.
@calxier2 жыл бұрын
This seems to rely on cooperation of all endpoints on the network to implement the same multiplicative decrease. Are there additional built in protections against a "bad actor" who tries to keep sending packets as close to the bandwidth limit as possible?
@ГеоргиГеоргиев-с3г2 жыл бұрын
I don't know how the protocol works, my hunch says the problem would be easy to spot, if it is a direct connection, it would be obvious you are not using the protocol and could get cut(if the middle node gets bothered with you, or just drops every however many of your packets to go up to that amount, after all internet uses a handshake, you can't just opt out of an agreement without the other side noticing), otherwise it would be out of your control so you get what you get even if you request for more, so it will be only a one layer deep problem, if i have to guess.
@TomStorey962 жыл бұрын
The routers in the network can also implement various "Quality of Service" schemes, some of them very complex, which can take into account an individual hosts traffic patterns and limit their overall bandwidth. But the more complex things get the more memory and processor intensive it gets on the router, and there are limits to how much complexity or how many hosts you can manage. Out of the box I seem to recall that Cisco routers, at least at one stage, would have several queues per interface that packets would be placed into based on a hash, and then those queues were processed in a round robin fashion, such that one busy host does not deny other hosts some packet forwarding time. It's not perfect because one host can still affect others that have the same hash, but overall the effect is minimised for the greater majority.
@richardclegg80272 жыл бұрын
UDP can do exactly this. If you wished to configure your machine to just push out as many packets as you can nothing stops you.
@kentw.england23052 жыл бұрын
This requires the routers to get involved. Read up on random early drop.
@richardclegg80272 жыл бұрын
@@kentw.england2305 sure RED BLUE etc are designed to send loss based signals to well behaved TCP senders to make them back off. But a poorly behaved sender just keeps sending ignoring the drops.
@LimitedWard2 жыл бұрын
I'm wondering, why couldn't the intermediate nodes on the network look at the window header, compare against the current state of its own receiver buffer, and then modify the window to be the lesser of the two values? This assumes the bandwidth is limited by the overall bandwidth of all the receiver buffers in the network. That seems like a more deterministic approach than this kind of guess and check. Edit: I realize why this wouldn't work. The traffic speed would then be limited to the max speed of the slowest intermediate node.
@DrRasputin20122 жыл бұрын
I guess that this only works if every TCP implementation honours it - what if one implementation chooses to be greedy? How do you protect against that?
@kentw.england23052 жыл бұрын
The routers get involved. It's called queue management.
@richardclegg80272 жыл бұрын
Really there is no protection. This is called the "TCP fairness" problem. You could in theory tweak your own settings to be more greedy (back off slowly push others out).
@aazimmermann2 жыл бұрын
Thanks for the video! Just wondering if collision avoidance is still in use today and if congestion is still a common problem which results in packet loss?
@richardclegg80272 жыл бұрын
Sure -- Carrier Sense Multiple Access/Collision Avoidance is basic to how your WiFi works. Try: "WiFi's Hidden ____ Problem - Computerphile" on youtube. Steve Bagley has a great explanation. Congestion can still occur though. Modern WiFi recovers loss on the local link so the loss is typically not there. However, modern WiFi is high bandwidth so the congestion is elsewhere on the network. You can get (say) 500Mb/s through your WiFi but somewhere on path will be a slower link (say at a congested router mid network). That is where the congestion happens and packets are lost.
@I.____.....__...__2 жыл бұрын
1:04 Nobody would _want_ to make a phone-call through the Internet in 1995, they'd just use a normal phone. The Internet was still for transferring files, which were still relatively small at the time (though I distinctly remember staying up all night a few times, trying to download files 😕).
@andrewharrison84362 жыл бұрын
Yesterday my internet just worked - now whatever I do I will know that the packet rate is modulated by congestion control. Well, I think that's worth knowing.
@network_king2 жыл бұрын
Interesting, should do one on like CSMA/CD, CSMA/CA. I would love to see one on Radia perlamand and spanning tree.
@martingriffiths98512 жыл бұрын
Love these vids and ths one inparticular. Try running it at 1.25 times speed for optimal effect !!! ;)
@adfaklsdjf2 жыл бұрын
So when I'm downloading a large file, the steady state is actually the sender is constantly ramping up then halving its send rate, "bouncing off the limiter" as it were? The rate looks so much more steady on my end..
@metaMichi2 жыл бұрын
The queues (the buffers) gets drained in the meantime (after the multiplicative decrease of the CWND). If the buffer is large enough (> 1 BDP) it can keep the delivery rate up.
@ciarfah2 жыл бұрын
This reminds me of an 'issue' I had transferring files to a slow USB drive on a Linux machine. The file transfer would show 100% then hang because the file would be fully buffered on the machine, but not yet transfered to the flash drive over the slow bus.
@kipp142 жыл бұрын
I now wonder if this might be the fatal problem with the current philosophy that us ISPs have with their cost benefit analysis where the packet drops are more to do with bad route choice on a given bundle for a certain bandwidth and less to do with the total amount of bandwidth available. I feel like at some point that the faster you confirm receipt the less congestion you have and the more reliable the connections become
@bs_blackscout2 жыл бұрын
please more networking videos!!
@jaybrooks10982 жыл бұрын
That iis bug around 2004 degraded the internet a lot too
@zxuiji2 жыл бұрын
Why not just start with all packets and move straight into the halving after... assuming there was missing acknowledgements that is, as for the data itself rather than (if I remember rightly that is) resending everything in the event of a fail, just send those that were missed, it's easy to setup after all as you can just do something like this with linux based server: Step 1: launch a dedicated process for the connection Step 2: process creates folder ~/connections// Step 3: process create all packets expected to be sent in said folder naming them .packet (since process memory might not be enough) Step 4: process sets up an internal buffer of values to hold the acknowledgement state of each packet Step 5: process creates a pool of threads for as many packets as can be handled Step 6: each thread sends it's own packet & waits on acknowledgement, setting it's acknowledgement value to 1 if it does get acknowledged, leaving it if it doesn't Step 7: main thread determines if any packets need to be resent or still need to be sent and re-uses the pooled threads for those packets while doing the mentioned congestion control (in other words loop back to step 6 until all have been acknowledged or some abandon condition is triggered) Step 8: empty the folder of every *.packet file and exit Step 9: server re-uses the POOL_ID for the next connection it is handling With the above it is not necessary to re-generate the packets, just send what is not acknowledged over and over until the acknowledgement is retrieved or abandon it midway for whatever reason, it's the clients job to put them in the right order using the sequences after all
@richardclegg80272 жыл бұрын
If you start by sending fast each new connection causes problems. Also worth noting writing to files can be pretty slow so you want to avoid any file writes here.
@zxuiji2 жыл бұрын
@@richardclegg8027 Well the files could be avoided with a memory check via attempted realloc but as far as the "each new connection causes problems" it's gonna do that anyways when incrementing so why not just get it over with from the start and see whether those problems actually occur, if they don't then great, if they do then if we're lucky at least some got through and that will reduce the number of packets we next need to send as it's the servers in between that will lose the packets after all
@richardclegg80272 жыл бұрын
@@zxuiji there is a huge amount of research into the ideal starting window size. The problems caused by slightly going over bandwidth with a modest increase in windowsize are much less than the problems caused by going hugely over bandwidth by starting too aggressively. Basically did you dump 100 packets the network could not handle or just one. (You'll also cause knock on problems for other traffic on your own network.) If you want to know current state of the art TCP CUBIC or TCP BBR are where to look. It also depends on setting (data centre vs general internet).
@zxuiji2 жыл бұрын
@@richardclegg8027 You're halving at each failure anyways so why not just try all out at the start and see if you need to halve the amount you send?
@richardclegg80272 жыл бұрын
@@zxuiji that was part of the original problem solved by this paper. Lots of new connections happen all the time on a well functioning network. If they all start high the result is congestion collapse. But don't take my word for it. Read the paper.
@rolandtennapel50582 жыл бұрын
The internet is represented as a cloud because it is nebulous; You know somewhat what's going on in there, but it's such a jumble it's next to impossible to map it. 😉 You can pull a cable out of a server, but the 'traffic controllers' won't know immediately that that server is no longer available, and that principle applies to those controllers between themselves as well, of course. So in a very real sense it's like looking into a cloud or mist, but the several waves to indicate a mist could be confused as 'somewhere between; 1~2', an unary programming symbol, a flow or a volume. Different eyes look differently at such symbols so a cloud makes the most sense.
@LeeSmith-cf1vo2 жыл бұрын
Is this window size calculated per remote host:port? If not, wouldn't firewalls that drop packets play havoc with the algorithm?
2 жыл бұрын
It is done for each TCP connection individually. A TCP connection is defined by its source and destination IP address and port numbers. The algorithms used will account for most kinds of bottlenecks in the network, firewalls being just one of them.
@richardclegg80272 жыл бұрын
Each TCP sender tracks its own as Orjan says. If a firewall drops your packets you slow down. (But it would be unusual for a firewall to drop middle packets for a flow - usually they block a flow or do not, it does not usually make sense for a firewall to kill only some of a connections packets).
@kelvinluk91212 жыл бұрын
So does it mean if I wanna enjoy the bandwidth I was promised by ISP, I should go shut down neighbors' network for a larger congestion windows right? :)
@richardclegg80272 жыл бұрын
Well it could work (if they are on the same ISP). Your neighbours might not enjoy it.
@adfaklsdjf2 жыл бұрын
Technically yes. There are problems with this approach but they aren't technical.
@carvoloco42292 жыл бұрын
That graph with steady ups and sharp downs reminds me of the bitrate graph I get when I copy a large number of files from one location to another, especially if the destination lies on an external drive. Is it possible that the OS is using somewhat similar congestion contention algorithms to allow the use of shared IO resources by many processes in parallel at the maximum possible rate?
@richardclegg80272 жыл бұрын
It will be using the TCP algorithm almost certainly.
@carvoloco42292 жыл бұрын
@@richardclegg8027 Well, I doubt so. But I've seen nothing in the congestion contention algorithm described that requires it to be applied exclusively on TCP connections; it seems to me that it could be applied whenever an uncoordinated set of information providers all try to push their data through the same shared channels expecting some sort of acknowledgements in return. There are plenty of shared information channels within a computer, both physical (e.g. the bus) and logical (e.g. a memory buffer held by a disk driver). And multiple processes running simultaneously, many of which may try to make use of the same channels at the same time, knowing nothing about each other. Data fragments cannot be stuck in a router inside a PC, but they can certainly be stuck in queues, so algorithms that allow the OS to maximize the throughput are certainly desirable. Now, the graph with steady ups and sharp downs shown on the video does remind me of the throughput graph that my computer displays when operating with large numbers of files, which makes me wonder if the lessons learned in the eighties were useful not only to solve the network congestion problem but also other congestion problems in different situations.
@richardclegg80272 жыл бұрын
I am describing TCP in the video which covers about 85% of internet traffic. Until recently it was almost the only choice for reliable data transfer. A few years back Google found a way to run TCP like algorithms over QUIC doing the congestion control in the browser. That is mainly applied between some web browsers and Google owned web sites.
@richardclegg80272 жыл бұрын
What I think you are talking about here is internal buses in PCs. They don't really need this same kind of protocol as they (a) work on a known fixed bandwidth and (b) can be coordinated in other ways.
@And_Rec2 жыл бұрын
what if it’s an ack that get lost? did i miss it?
@matthewparker92762 жыл бұрын
I've seen my own internet drop as low as 40 bps, for hours at a time. But that's Australian broadband for you.
@asandax62 жыл бұрын
I came to this video because I've been researching ways to bring down data usage. In my country South Africa data is really expensive and the infrastructure is really poor.
@lucidmoses2 жыл бұрын
Doesn't the "Cloud" come from the military pre-electronics. When a battle started, amongst all the smoke from poor gun powder a message was sent via carrier pigeon.
@adfaklsdjf2 жыл бұрын
Not sure, but I'm confident that it pre-dates wide use of the term "cloud computing".. I've been drawing the internet as a cloud on diagrams for more than 20 years..
@melanierhianna2 жыл бұрын
Is that why there is an RFC for IP via carrier pigeon, so it can properly use the cloud.
@lucidmoses2 жыл бұрын
@@melanierhianna Yes, my brothers internet feels like that out on the farm.
@olivier25532 жыл бұрын
Bandwith of the classic phone was about 3KHz, that is it needed 3kbps. 32 k would allow 10 phones calls simultaneously. That is why ADSL was working: phone only needed a very small portion of the bandwidth that could be carried on a copper twisted pair.
@kesslerrb2 жыл бұрын
What codec gets a phone call down to 3Kbps? G.729 compresses calls down to ~6Kbps but I haven’t seen any that can reliably support calls with less bandwidth
@olivier25532 жыл бұрын
@@kesslerrb No codec, just normal analog phone, like it has been used for 100 years :) Phone has been designed to use only 3KHz bandwidth because it covers most of the spectrum of human voice.
@Phroggster2 жыл бұрын
@@kesslerrb The voiceband of traditional plain old telephone service (POTS) was roughly 300-3,300 Hz bidirectional. The signals were confined to this range via analog filtering, no codecs required. Just a very restrictive bandpass filter at both ends, and several more along the connection path.
@WobblycogsUk2 жыл бұрын
So much confusion here. The bandwidth of a standard voice only phone line is about 3kHz and operates at the regular voice frequencies. The data rate possible on that line is given by the Shannon-Hartley theorem which is dependent on the signal to noise ratio of the line. It was initially assumed that phone likes would be about to achieve data rates of about 33kbps but development of the phone network resulted in better signal to noise ratios which allowed for higher speeds such as 56kbps. In theory they could probably have gone a little faster but the world switch to ADSL. ADSL uses basically the same ideas but it operates at higher frequencies which, in turn, give it more bandwidth to play and so it achieves better data rates.
@olivier25532 жыл бұрын
@@WobblycogsUk Yet, you don't need a minimum of 32Kbps to transmit voice only.
@cadekachelmeier72512 жыл бұрын
Ack
@uplink-on-yt2 жыл бұрын
Good thing all the implementations out there adhere to this, right? This pretty much relies on the cooperation of the network users. Luckily, I guess, if there’s only a small number of jerks on the network, things will continue to work.
@jca1112 жыл бұрын
I watched at 125%
@GilesBathgate2 жыл бұрын
He talks with an additive increase and multiplicative decreace in speed 😂
@dp1212732 жыл бұрын
Don't look at my left ear the whole time 🤣
@andredejager36372 жыл бұрын
love it 😀
@wahyu94202 жыл бұрын
My internet speed right now isn't better than internet speed 40 years ago, suck.
@NickNorton2 жыл бұрын
It's faster than rfc 1149
@idontknowreally2 жыл бұрын
Damn you, this is basically the rick roll of RFCs... 😁
@richardclegg80272 жыл бұрын
Without looking I guess "avian carriers"?
@nikkiofthevalley2 жыл бұрын
@@richardclegg8027 Yep, avian carriers. I'm less of a networking guy, so I tend to use HTTP response code 418 - "I'm a teapot" for these sorts of jokes.
@kesslerrb2 жыл бұрын
Wasn’t there an extension to RFC1149 that strapped usb thumb drives to the pigeons? 🤣
@adfaklsdjf2 жыл бұрын
@@kesslerrb probably.. I heard someone looked into it and found you could achieve decent throughput with usb sticks, if you didn't care about latency.
@AcornElectron2 жыл бұрын
My internet feels like it collapses daily
@barebears2892 жыл бұрын
Who's here from Hussein Nasser's channel?
@der.Schtefan2 жыл бұрын
First time ever I used the 1.5x speed button on KZbin. He speaks so sloooooooooooooooooooooooooooooooow and seems to micro-nap between sentences losing orientation. His bandwidth certainly is below 40bps.
@DumbledoreMcCracken2 жыл бұрын
This is a low bandwidth presentation
@imamalox2 жыл бұрын
The person who edited this video could've maybe put in some effort to remove the annoying background noise 😅
@lenorkhide28732 жыл бұрын
You can see that graph in action on steam during the initial release of an AAA game
@three-card-dead2 жыл бұрын
really not happy with the quality of this video -- any chance we can get another pass on this?
@colonelhacker36612 жыл бұрын
FECN and BECN.
@Yupppi2 жыл бұрын
At this point I'm starting to feel like most of the data on internet transfer is all kinds of protocols, keys, encryption and notes about the package rather than the actual data. Bernoulli's law would imply that the packages would flow faster in the bottleneck!
@DFPercush2 жыл бұрын
You should write an angry letter to Al Gore. :P Although if you think about the "bottleneck" being the fiber optic cables that carry hundreds of gigabits of bandwidth, it kind of does.
@adfaklsdjf2 жыл бұрын
You're not wrong about protocol overhead. It varies as a percentage of total traffic, with higher-data applications generally having a much lower protocol overhead as a % of total traffic... for watching a youtube video, it'll be a fraction of a percent.. for a short email, it's basically all overhead.
@hrford2 жыл бұрын
I'd tell you a TCP joke, I'm sure you'll get it. Are you ready to hear it?
@kevintedder42022 жыл бұрын
I don't think anyone got it. Try sending it again, slowly. 😀
@sjatkins2 жыл бұрын
40 bps? Wow. I got way better than that on FidoNet years earlier over phone modems.
@sandordugalin89512 жыл бұрын
Briliiant
@polares81872 жыл бұрын
faster details pleease
@leviath0n2 жыл бұрын
Best at 1.25 speed.
@AcornElectron2 жыл бұрын
Yeah?
@Ur112 жыл бұрын
"fork handles"🕯🕯🕯🕯?
@alexguillen24932 жыл бұрын
Syn Ack!!!!!
@romandobra31512 жыл бұрын
What?
@skytech25012 жыл бұрын
wornderful vid, if you want to use your full brain bandwidth play the video at 1.75 :)
@AmanSharma-fr4uu2 жыл бұрын
Increase it exponentially and decrease when there's overflow
@DarthAthar2 жыл бұрын
why not directly 2?
@skytech25012 жыл бұрын
@@DarthAthar to avoid congestion
@allesarfint2 жыл бұрын
Had to reduce it to 0.5, the buffer has not much space
@akashpawar90582 жыл бұрын
akash
@Schwuuuuup2 жыл бұрын
Not an uninteresting video but for my taste well too long winded.. basicly 16 min of pramble in a 20 minute video....
@vicentelhs2 жыл бұрын
those laptops looks SUS
@skorp56772 жыл бұрын
1024th
@du42bz2 жыл бұрын
2048th
@dakoderii42212 жыл бұрын
@@du42bz 4096th
@pacifico49992 жыл бұрын
@jewtube Overflow, back to 4096
@NoEgg4u2 жыл бұрын
CNN is reporting that Al Gore wrote the paper on TCP/IP and congestion.
@domtorque2 жыл бұрын
25th
@cubeboi70702 жыл бұрын
sixty ninth
@tubeWyrme2 жыл бұрын
This could have been better explained in 5 mins with no protocol details
@adfaklsdjf2 жыл бұрын
I think you're looking for another channel
@MichaelKingsfordGray2 жыл бұрын
A functionally illiterate Doctor working at a University! How standards have avalanched. Sad.
@sam-you-is2 жыл бұрын
sixth
@jmshrv2 жыл бұрын
Second
@anon33082 жыл бұрын
Fifth
@eekee60342 жыл бұрын
@Yummy Spaghetti Noodles That's a result of using more than one server to handle comments. Tom Scott has a clear and concise video on why the number of likes goes up & down, (something like, "Why KZbin lies about likes") and if you think about it, you can see how it applies to comments too.