Codeberg: army of AI crawlers are extremely slowing us; AI crawlers learned how to solve the Anubis challenges.
-
It matters a lot though what kind of goal the system incentivises. Imagine if it was people's happiness and freedom instead of quarterly profits.
That’s the part people never really seem to understand. It makes sense though because we’re subjected to the system from birth and it’s all a lot of people know so they can’t grasp the idea of a world outside of that so it can sometimes be difficult to get through to people on that
-
The crawlers are likely not AI though, but yes OCR could be done effectively without AI anyways. This idea ultimately boils down to the same hope Anubis had of making the processing costs large enough to not be worth it.
OCR could be done effectively without AI
OCR has been neural nets even before convolutional networks emerged in the 2010s
-
cross-posted from: https://programming.dev/post/35852706
reminder to donate to codeberg and forgejo
-
Something needs to route the PC to the server
I don't want client-server model. I want sharing model. Like with Briar.
The only kind of "servers" might be relays, like in NOSTR, or machines running 24/7 like Briar mailbox.
IP. How would I go about replacing it? I don't know, I think Yggdrasil authors have written something about their routing model, but 1) it's represented as ipv6, so IP, 2) it's far over my head, 3) read the previous, I don't really want to replace it as much as not to make it the main common layer.
client-server model. I want sharing model. Like with Briar
Guess what
Briar itself, and every pure P2P decentralized network where all nodes are identical... are built on Internet Sockets which inherently require one party ("server") to start listening on a port, and another party ("client") to start the conversation.
Briar uses TCP/IP, but it uses Tor routing, which is IMO a smart thing to do
-
If it becomes popular enough that it's used by a lot of people then the bots will move over there too.
They are after data, so they will go where it is.
One of the reasons that all of the bots are suddenly interested in this site is that everyone's moving away from GitHub, suddenly there's lots of appealing tasty data for them to gobble up.
This is how you get bots, Lana
Yes, I know. But, while trying to find a way to bomb the AI datacenters (/s, hopefully it doesn't come to this), we can stall their attacks.
-
Could you give some examples?
Socialized healthcare
-
You want some bipcoin to buy weed drugs on the slip road? Don’t bother figuring out how to set up that wallet shit, come to our nifty token exchange where you can buy and sell all kinds of bipcoins
Maybe I'm slow today, but what is this referencing? Most dark web sites use Monero. Is there some centralized token that people used instead?
Edit: Oh, I guess you're referring to Mt.Gox? I mean yeah, people were pretty stupid for keeping their bitcoin in exchange wallets (and sending it right to the drug dealers directly from there? Real dumb). That's always a bad idea. I don't think they transferred it there instead of something else, they just never took custody of the coins after buying them on the exchange.
Monero
Satoshi was right and Crypto absolutely has valid use cases. What if your government doesn't want you accessing meds you need at prices you can afford? What if your government doesn't like your sexual orientation, but you want a subscription to a dating site? What if your government throws up unjust export controls or tariffs that suddenly make you and your business impossible?
Crypto's best killer use case is uncensorable, untraceable money
Bitcoin is neither of those things. There is a reason people buy heroin with Monero. It actually does what crypto is supposed to do, which means it could safeguard your Grindr XTRA subscription.
-
Yeah but there's would be scrappers if the robots file just pointed to a dump file.
Then the scraper could just do a spot check a few dozen random page and check the dump is actually up to date and complete and then they'd know they don't need to waste any time there and move on.
Given that they already ignore robots.txt I don't think we can assume any sort of good manners on their part. These AI crawlers are like locusts, scouring and eating everything in their path,
-
I feel like half of the blame capitalism gets is valid, but the other half is just society. I don't care what kind of system you're under, you're going to have to deal with other people.
Oh, and if you try the system where you don't have to deal with people, that just means other people end up handling you.
I would give this reddit gold
Instant easy complaints help-i'm-oppressed-by-Capitalism today sound an awful lot like the instant easy complaints help-i'm-oppressed-by-Communism I used to hear from rednecks
Ask someone who starved & died under either system how obviously superior it is, you will find millions on either side
-
It matters a lot though what kind of goal the system incentivises. Imagine if it was people's happiness and freedom instead of quarterly profits.
Imagine if it was people’s happiness and freedom instead of quarterly profits
- Whose happiness and freedom?
- How is it to be measured?
- Capitalists honestly believe that free trade is the best albeit flawed way to do both of the above
It's definitely valid to disagree about point #3, but then you need to give a better model for #1 and #2
-
Can there be a challenge that actually does some maliciously useful compute? Like make their crawlers mine bitcoin or something.
The Monero community spent a long time trying to find a "useful PoW" function. The problem is that most computations that are useful are not also easy to verify as correct. javascript optimization was one direction that got pursued pretty far.
But at the end of the day, a crypto that actually intends to withstand attacks from major governments requires a system that is decentralized, trustless, and verifiable, and the only solutions that have been found to date involve algorithms for which a GPU or even custom ASIC confers no significant advantage over a consumer-grade CPU.
-
Given that they already ignore robots.txt I don't think we can assume any sort of good manners on their part. These AI crawlers are like locusts, scouring and eating everything in their path,
Crawlers are expensive and annoying to run, not to mention unreliable and produce low quality data.
If there really were a site dump available, I don't see why it would make sense to crawl the website, except to spot check the dump is actually complete.
This used to be standard and it came with open API access for all before the silicon valley royals put the screws on everyone -
Socialized healthcare
And in what way does capitalism socialize healthcare in the United States?
-
And in what way does capitalism socialize healthcare in the United States?
Socialized healthcare exists there -- at least until the current administration finishes ripping it away
-
Socialized healthcare exists there -- at least until the current administration finishes ripping it away
The US uses a welfare state, that isn’t socialism and no one is praising it
-
The US uses a welfare state, that isn’t socialism and no one is praising it
Socialized healthcare isn't socialism...
Interesting.
How about Canada?
-
Socialized healthcare isn't socialism...
Interesting.
How about Canada?
The US does not have socialized healthcare. It uses a welfare state. What about Canada?
-
The US does not have socialized healthcare. It uses a welfare state. What about Canada?
Does Canada have socialized healthcare?
Does anywhere?
-
OCR could be done effectively without AI
OCR has been neural nets even before convolutional networks emerged in the 2010s
Yeah you're right, I was using AI in the colloquial modern sense. My mistake. It actually drives me nuts when people do that. I should have said "without compute-heavy AI".
-
Does Canada have socialized healthcare?
Does anywhere?
Yeah they do! A few places do. I’m just enjoying seeing you prove your own sarcasm as actual truth because these things formed in spite of capitalism