this post was submitted on 02 Apr 2025
701 points (99.2% liked)

Technology

68305 readers
5592 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 121 points 2 days ago (2 children)

Laws should be passed in all countries that AI crawlers should request permission before crawling whatever target site. I haver no pity to AI "thiefs" that get their models poisoned. F...ing plague, wasn't enough the adware and spyware...

[–] [email protected] 20 points 1 day ago (1 children)

i doubt the recent uptick in traffic is from “stealing data” for training but rather from agents scraping them for context, eg Edge Copilot, Google’s AI search, SearchGPT, etc.

poisoning the data will likely not help in this situation since there’s a human on the other side that will just do the same search again given unsatisfactory results. like how retries and timeouts can cause huge outages for web scale companies, poisoning search results will likely cause this type of traffic to increase and further increase the chances of DoS and higher bandwidth usage.

[–] [email protected] 7 points 1 day ago (1 children)

So? Break context scrapers till they give up, on your site or completely.

[–] [email protected] 2 points 1 day ago
[–] [email protected] 21 points 2 days ago (3 children)

An HTTP request is a request. Servers are free to rate limit or deny access

[–] [email protected] 1 points 23 hours ago (1 children)

Bots lie about who they are, ignore robots.txt, and come from a gazillion different IPs.

[–] [email protected] 1 points 23 hours ago

That's what ddos protection is for.

[–] taladar 14 points 1 day ago

Rate limiting in itself requires resources that are not always available. For one thing you can only rate limit individuals you can identify so you need to keep data about past requests in memory and attach counters to them and even then that won't help if the requests come from IPs that are easily changed.

[–] [email protected] 19 points 2 days ago (1 children)

And Wikimedia, in particular, is all about publishing data under open licenses. They want the data to be downloaded and used by others. That's what it's for.

[–] [email protected] 6 points 1 day ago (1 children)

Even so I think it would be totally reasonable for them to block web scrapers, as they provide better ways to download all their data.

[–] [email protected] 9 points 1 day ago

At the root of this comment chain is a proposal to have laws passed about this.

People can set up their web servers however they like. It's on them to do that, it's their web servers. I don't think there should be legislation about whether you're allowed to issue perfectly ordinary HTTP requests to a public server, let the server decide how to respond to them.