this post was submitted on 17 Sep 2023
669 points (92.2% liked)
Games
31990 readers
4 users here now
Welcome to the largest gaming community on Lemmy! Discussion for all kinds of games. Video games, tabletop games, card games etc.
Weekly Threads:
Rules:
-
Submissions have to be related to games
-
No bigotry or harassment, be civil
-
No excessive self-promotion
-
Stay on-topic; no memes, funny videos, giveaways, reposts, or low-effort posts
-
Mark Spoilers and NSFW
-
No linking to piracy
More information about the community rules can be found here.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
It's the most obvious example of a digital media problem. Computers might be able to account for every bit with the use of parity files and backups with frequent parity checks, but the fact is most people aren't running a server with 4 separately powered and monitored drives as their home computer, and even the most complex system of data storage can fail or degrade eventually.
We live in a world of problems, like the YouTube problem, compression problems, encoding problems, etc. We do because we chose efficiency and ease of use over permanency.
Yes, and this can be done through mostly automatic or distributed processes.
I wouldn't describe it as complex, just the bare minimum of what is required to actually preserve data with no loss. All physical mediums may degrade through physical processes, but redundant systems can do better.
It isn't hard to seed a torrent. If a group of people want to preserve a file, they can do it this way, perfectly, forever, so long as there remain people willing to devote space and bandwidth.
All of these problems boil down to intent. Do people intend to preserve a file, do they not care, do they actively favor degradation? In the case of the OP game, it seems that the latter must be the case. Same with Youtube, same with all those media companies removing shows and movies entirely from all public availability, same with a lot of companies. If someone wants to preserve something, they choose the correct algorithms, simple as that. There isn't necessarily much of a tradeoff for efficiency and ease of use in doing so, disk space is cheap, bandwidth is cheap, the technology is mature and not complicated to use. Long term physical storage can be a part of that, but it isn't a replacement for intent or process.
I think you didn't read correctly on the statement about the most complex system failing. I'm not saying that is the most complex system, I am saying the most complex system will fail.
LMAO at the idea of comparing every bit of every portion of every seeder's copy with each other simultaneously and then cross referencing every parity file to be doubly safe, and then failing to see the chance of loss of parity during transmission of said files even after that. I will admit it would take a lot longer for a torrented file to degrade than some other forms of file distribution, but it's not going to last for a thousand years, mate.
And I am saying complexity has little to do with it and also that a system can exist that will not fail.
Specifically why not? What is unrealistic about this scenario, assuming enough people care to continue with the preservation effort? All nodes must fail simultaneously for any data to be lost. The probability of any given node failing at any given time is a finite probability, independent event. The probability of N nodes failing simultaneously is P^N. That is exponential scaling. Very quickly you reach astronomically low probabilities, 1000 years is nothing and could be safely accomplished with a relatively low number of peers. Maybe there are external factors that would make that less realistic, like whether new generations will even care about preserving the data, but considering only the system itself it is entirely realistic.
Read the above conversation to find out.