this post was submitted on 14 Jan 2024
200 points (94.6% liked)

Piracy: ꜱᴀɪʟ ᴛʜᴇ ʜɪɢʜ ꜱᴇᴀꜱ

54758 readers
364 users here now

⚓ Dedicated to the discussion of digital piracy, including ethical problems and legal advancements.

Rules • Full Version

1. Posts must be related to the discussion of digital piracy

2. Don't request invites, trade, sell, or self-promote

3. Don't request or link to specific pirated titles, including DMs

4. Don't submit low-quality posts, be entitled, or harass others



Loot, Pillage, & Plunder

📜 c/Piracy Wiki (Community Edition):


💰 Please help cover server costs.

Ko-Fi Liberapay
Ko-fi Liberapay

founded 1 year ago
MODERATORS
 

For instance, say I search for "The Dark Knight" on my Usenet indexer. It returns to me a list of uploads and where to get them via my Usenet provider. I can then download them, stitch them together, and verify that it is, indeed, The Dark Knight. All of this costs only a few dollars a month for me.

My question is, why can't copyright holders do this as well? They could follow the same process, and then send takedown requests for each individual article which comprises the movie. We already know they try to catch people torrenting so why don't they do this as well?

I can think of a few reasons, but they all seem pretty shaky.

  1. The content is hosted in countries where they don't have to comply with takedown requests.

It seems unlikely to me that literally all of it is hosted in places like this. Plus, the providers wouldn't be able to operate at all in countries like the US without facing legal repercussions.

  1. The copyright holders feel the upfront cost of indexer and provider access is greater than the cost of people pirating their content.

This also seems fishy. It's cheap enough for me as an individual to do this, and if Usenet weren't an option, I'd have to pay for 3+ streaming services to be able to watch everything I do currently. They'd literally break even with this scheme if they could only remove access to me.

  1. They do actually do this, but it's on a scale small enough for me not to care.

The whole point of doing this would be to make Usenet a non-viable option for piracy. If I don't care about it because it happens so rarely, then what's the point of doing it at all?

you are viewing a single comment's thread
view the rest of the comments
[–] ShitpostCentral@lemmy.world 5 points 10 months ago* (last edited 10 months ago) (2 children)

That makes some amount of sense. I'm not sure exactly how each article is stitched together to create the full file. Do you happen to know if it's just put together sequentially or if there's XORing or more complex algorithm going on there? If it's only the former, they would still be hosting copyrighted content, just a bit less of it.

EDIT:
https://sabnzbd.org/wiki/extra/nzb-spec
This implies that they are just individually decoded and stitched together.

[–] Shadow@lemmy.ca 11 points 10 months ago

Compressed into a set of small archives, then each one is posted.

Usually par files are included so you can regenerate a few missing archives. https://en.m.wikipedia.org/wiki/Parchive

[–] grue@lemmy.world 8 points 10 months ago* (last edited 10 months ago)

Do you happen to know if it’s just put together sequentially or if there’s XORing or more complex algorithm going on there? If it’s only the former, they would still be hosting copyrighted content, just a bit less of it.

Copyright is a legal construct, not a technological one. Shuffling the file contents around doesn't make the slightest bit of legal difference, as long as the intent is to reconstruct it back into the copyrighted work.

(Conversely, if the intent was to, say, print out the file in hexadecimal and wallpaper your house with it, that wouldn't be copyright infringement even if you didn't rearrange it at all because the use was transformative. Unless the file in question was a JPEG of hex-digit wallpaper, of course.)