I know we can’t do this with any copyrighted materials. But a lot of books, music, art, knowledge is in the creative commons. Is it possible to create one massive torrent that includes all that can be legally included and then have people only download what they actually want to enjoy?
So like… a meta-torrent that is a torrent of all the other torrents?
Edit: or we could just create a website that had a list of all the torrents. And give it a clever name that reminds us of the fact that we’re pirating things. Oh… wait
With the way that the BitTorrent v2 protocol works, each file of the original, underlying torrents wouldn’t have to be re-seeded, but rather would reuse each file’s individual hash and thus incorporate those files into the meta torrent without necessarily having to download or even upload any part of the meta torrent.
That said, the .bittorrent file would be massive and might run up against certain limits in the current protocol.
But don’t lose the list. Losing the list would be bad. We’ll need to keep the list in a safe place.
I mean… That pretty much describes torrents period… What is the functional difference between hosting a single torrent with everything, and hosting a torrent per item?
If the expectation is that you only include files you want when downloading the torrent, you’re only going to be seeding that portion.
Seems like it would just make the search function harder, and make it harder to determine the “health” of individual items…
I don’t understand the benefit…
For example zlibrary is 220TB of books and scientific articles, that included in the torrent would be great along with all the stuff that is arts and music.
Basically it would be a way to combat media vanishing off of the net over time. Basically a noah’s arche for all of mankinds knowledge.
It would be great to have everything in one single spot to make it easier to contribute and get stuff. We’d also be more easily capable of combining our forces to maintain/create the thing.
You’re describing leeching from something like Anna’s Archive datasets.
Not quite what you’re getting at, but the entirety of Wikipedia without images is available as a 20-30GB download: https://en.wikipedia.org/wiki/Wikipedia:Database_download
Pretty sure this stuff already exists in some form. /r/datahoarder people would probably be able to steer you in the right direction though you may need to lay out several thousand for enough HDDs to hold it.
No
maybe
Scihub + libgen archives (torrent list) come pretty close, even that it’s not what you’re asking about (100TB? total)
“All” is impossible. You’re going to miss something. And it’s a lot of work. Maybe have a look at the datasets people/researchers use to train Artificial Intelligence. I think some people put in the effort to compile large datasets with just freely licensed data.
it’s a lot of work
so per your suggestion using for example the zlibrary book/paper repo and training sets of openai as starting point one could maybe get around the brunt of the work.
You could seed the Torrents by the internet archive.
Yes