Blake [he/him]

  • 2 Posts
  • 244 Comments
Joined 1 year ago
cake
Cake day: June 12th, 2023

help-circle

    • 7zip (archive tool) People still use Winrar for whatever reason.
    • Firefox (web browser) - It’s well known but still under-used. Stop using chrome. Chrome will be the end of the free internet.
    • windirstat (visualise used disk space) - I use this to clear free space
    • musicbee (music player) - replaced VLC with this because VLC has had a bug that makes it not playback FLAC files properly for like… A year?
    • Screen2Gif (gif recorder & editor)
    • ShareX (screenshot/video tool)
    • Open Video Downloader (GUI tool to rip YouTube etc.)
    • uBlock Origin (ad-blocker) addon
    • Sponsorblock (sponsorship skipper) addon









  • As SI prefixes, they’re all multiples of ten, technically speaking. So referring to 1,024 bytes as a kilobyte is incorrect, it’s 1.024 kilobytes or 1 kibibyte. Microsoft deciding to ignore industry and international standards is the reason for the confusion.

    But either way, hard drive manufacturers will sell a 1TB drive, and Windows will see that as a 935GB drive - that’s basically the difference between 2^40 bytes vs. 10^12 bytes



  • You can’t really compare small-scale clusters of highly available services with the scale of the entire Internet, it’s just an entirely different ballgame. Though even in small scale setups, there is always a sweet spot between too many paths and not enough paths - VRRP (which is the protocol usually used for high availability) actually has quite a big overhead, you can’t have too many connections on the same network or it causes lots of problems.

    Internet scale routing usually uses BGP, which also has quite a heavy overhead.

    I guess all you need to understand is that routing isn’t free, and the more routes, the more overhead. So there’s always going to be a point where adding more routes just makes things slower rather than faster. And BGP… is just a bit of a mess, right now, honestly. The BGP table has grown so big that a lot of older devices can’t keep it in fast memory anymore, so they either have to be replaced with newer hardware or use slow memory (and therefore slow processing of packets). So it’s not really in everyone’s best interests to just keep adding more routes. It’s harder and harder to justify.

    why there are so many more connections in the north east and west coast if more connections slows the whole system down

    I’m not from the US, so at best it would be an educated guess.

    Firstly, it’s not as simple as just “more connections is more slow”, it means there’s a greater overhead. If the improvement from adding another line is greater than the overhead, then it can be worthwhile. For example, imagine a simple network with three routers, A, B and C, where A is connected only to B, and C is connected only to B, meaning that B is connected to both A and C. If there is a large amount of traffic between A and C, it may be worth adding a direct connection between them. If there isn’t, then it’s probably not worth doing.

    I guess it’s a bit like adding a new road between two existing roads. Is it worth adding a junction and a set of traffic lights to some existing roads, or would that slow down traffic enough not to be worth doing?

    Maybe, since you work with software more, it would make sense to put it this way: why don’t you create an index for every single possible column and table in SQL?

    Or just look at it like premature optimisation. There’s a saying about premature optimisation in software engineering! ;-)

    Another thing to keep in mind though is that there’s definitely still quite a few bad decisions still kicking around from when the internet was new. It takes time and effort to get rid of the legacy junk, same as in programming.


  • The problem that I’m having (and why I asked that) is because I was assuming that you would have some knowledge which you don’t seem to have with a lot of my comments. I’m really not trying to be rude, but it makes it a lot more difficult to explain the flaws in your reasoning when you’re talking about topics that are beyond your knowledge as if you know them well.

    I have explained the realities of the situation to you, if you don’t want to accept them, that’s fine, but you’re basically arguing with an expert about something you don’t really understand very well. I’m happy to explain stuff but you should just ask rather than assume you know better because it makes it much more difficult for me to understand the gaps in your understanding/knowledge.

    So ultimately, for routers, we have a number of limited resources. Firstly, yes, interfaces, but also the usual stuff - CPU, RAM, etc.

    Now, I mentioned before that routing protocols are very complex - they have many metrics which are taken into account to determine what path is ultimately best for each packet. This is a process which can be quite intensive on CPU and RAM - because the router needs to “remember” all of the possible routes/destinations a packet can travel, as well as all of the metrics for each destination - distance, delays, administrative distance, TTL, dropped packets, etc. and then make a decision about processing it. And it needs to make these decisions billions of times a second. Slowing it down, even a tiny bit, can hugely impact the total throughout of the router.

    When you add another connection to a router, you’re not just increasing the load for that one router, but for the routers which connect to the routers which connect to those routers which route to the routers that route to that router… you get the idea. It increases the number of options available, and so it places additional burden on memory and processing. When the ultimate difference in distance even an extra 100 miles, that’s less than a millisecond of travelling time. It’s not worth the added complexity.

    That’s what I meant when I said that an extra hop isn’t worth worrying about, but adding additional connections is inefficient.





  • Yep, I mean, the comment you’re replying to literally contains the phrase, “the biggest issues are interference…” haha

    Likewise, it’s something that’s likely to improve as we tend to move away from the 2.4GHz band.

    Dropping packets is definitely more of a problem for streaming in particular, rather than anything else, because like you said, if you drop packets you’re going to get degraded quality video. If you were gaming locally, it wouldn’t really affect you as much. Online games have extremely good, well designed methods of compensating for dropped packets in a way that streaming will never be able to match.


  • There’s absolutely nothing wrong with that topology - the fact that you seem to think that the design is a bad thing really demonstrates your lack of understanding here.

    For example, have you never wondered why we don’t just connect every device in a network all together like a big daisy chain? Or why we don’t use a mesh network? There is a large number of reasons why we don’t really use those topologies anymore.

    I don’t want to get into the specifics, but in general, the more networks a router is connected to, the less efficient it is overall.

    The propagation delay is pretty insignificant for most routers. Carrier grade routers like those at the core of the internet can handle up to 43 billion packets per second, another hop is absolutely nothing in terms of delay.


  • Sorry, what I wrote here was unclear, I wrote it needs less boosting in another comment, but re-reading this one, it does sound like I’m claiming it needs no boosting over any distance - that’s not what I meant though! I just meant that you can run an equivalent link without any boosting further than you could with copper.

    Interference isn’t actually that big of a deal for Ethernet over copper, unless the installer does something silly like run UTP alongside high power electrical lines, or next to a diesel generator, or something. Between shielding, the use of balanced signals, and the design of twisted pair, most interference is eliminated.