i love reading about technology

  • 0 Posts
  • 3 Comments
Joined 8 months ago
cake
Cake day: March 8th, 2024

help-circle

  • Ryuzo@lemmy.worldtoUsenet Invites@lemmy.world[W] DrunkenSlug
    link
    fedilink
    English
    arrow-up
    3
    ·
    8 months ago

    Usenet (or nntp) relies on physical servers at set locations around the world, these normally reside in large datacenters. these physical servers are owned by or maintained by the “backbones” so for example, Giganews owns and maintains its own servers, and sells access to them as either giganews, or supernews.com. They also sell access to 3rd party resellers who brand the access as their own and sell it onto you giganews resellers are RhinoNewsgroups, Usenet.net, Powerusenet - these may use their own cache for recent content, but once past that they all share the same content.

    When a “linux distro” is uploaded to Usenet, it is uploaded to one of these physical servers, either directly through a provider like giganews, or through a reseller, this doesn’t matter as these physical servers are all connected via certain propagation agreements with all of the other backbones and caches be it xsnews, searchtech limited, this process is called propagation, and it can take up to 40 minutes and even longer for these physical servers to become synced.

    Here’s a little diagram: https://lemmy.world/pictrs/image/d04f5e90-1187-477c-a7a8-4638303299e5.png

    This is as you can see a cyclical process - the posters chosen provider gets the article, and then other servers go “oh what do you have that we dont? lets grab that” this process occurs until the servers are >=99% complete, when a poster uploads they normally upload their release with par files to account for 10% of the overall file - so if 5% of the release is damaged due to poor transfer or such - your download client can grab these par files and fix the missing sections - kinda cool!


  • Ryuzo@lemmy.worldtoUsenet Invites@lemmy.world[W] DrunkenSlug
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    8 months ago

    Despite Torrents benefiting from decentralized file distribution, its efficacy remains governed by seeding availability. Usenet, with centralized storage, boasts impressive retention periods, often spanning several years. This is why it does not require seeding unlike torrents.