Monday, 2026-03-16

ykarelHi releases.openstack.org is down, is that known issue?04:11
ykarelmost of the jobs red due to this, example failure build https://zuul.opendev.org/t/openstack/build/88e88f89dddb41888430f6ad7be7673904:13
ykarelalso sent over openstack-discuss to hold recheck https://lists.openstack.org/archives/list/openstack-discuss@lists.openstack.org/thread/X4YBRF6LOPJTPHRRLRYOQ6W7Y5RWKYNI/05:03
ykarelsame is with docs.openstack.org, also down06:15
fricklerykarel: yes, there was an issue again like on Friday, but mnasiadka already restarted it08:58
tkajinamsent out a reply to broadcast that :-)09:40
ykarelthx frickler tkajinam mnasiadka 09:49
mnasiadkaThanks tkajinam, was planning to do that now :)09:50
tkajinammnasiadka, thank you for fixing it :-)09:52
*** haleyb|out is now known as haleyb13:02
JayFdocs.openstack is hurting again :( sometimes I wonder if I'm getting caught in glue traps being setup for AI bots20:34
fungiJayF: i think the server is caught in a trap set by ai bots20:36
fungiwe're continuing to try to tune apache and waf rules to block the worst offenders20:36
mnaserthis is probably not going to be the hottest take, but what about setting up cloudflare for these things, takes the headache out of handrolling all of this stuff20:37
fungii'm open to the possibility, but we will be accepting that it's currently impossible to run a website on the internet today without handing over all our user data to a compamny who wants to resell it to the highest bidder20:39
fungialso admitting that you can no longer operate a website with open source software20:39
mnaseri think the space is moving so fast and things are hurting a bit, it could be something to give us some breathing space as the oss world figures things out20:40
mnaseri think there's alot of things we can explore such as nginx for example which historically has been far more powerful and resource efficent than apache to serve static content, etc20:40
JayFmnaser++++ took the words outta my mouth20:40
JayFI'd rather have things working and you all free to do things other than tune apache20:41
JayFthan stick to an ideal that clearly we're not able to fund20:41
fungiat the moment, it appears that some organized crime outfit in control of hundreds of thousands of backdoored mobile devices has decided to flood every website with random url guesses to see if they can find anything that's not already been indexed normally and fed into existing llm models20:41
mnaserafaik there are things like https://github.com/TecharoHQ/anubis for example, but again deploying this takes time20:42
mnaserheck, they even say you should get cloudflare and this is only if you really dont want to use it =)20:42
JayFfungi: How trust-based is the infra team? We have infra resources at GR-OSS I could potentially ask if they could help e.g. deploy anubis, but I think it'd be a ... project-based interaction, not a long-term-ongoing devops commitment 20:43
fungiour systems administration is done entirely in the open through code review, so there's no trust concerns, they wouldn't need the ability to ssh into servers20:45
fungijust pushing (mostly ansible) changes into gerrit20:45
JayFack20:45
JayFno promises but I can ask20:45
fungiand we're not exactly rolling our own solutions to this stuff, it's just that e.g. traditional waf approaches like apache mod_security aren't entirely suited to the latest wave we've seen over the past week-ish20:46
fungipreviously it was crawlers following links from pages to other pages, but now most of the requests the server is handling are for randomly-guessed nonsexitent urls, and not even repeated ones, so things like 404 caching don't help because the server has never seen the request and has to check anyway20:48
fungiurls like20:49
fungihttps:///developer/ironic/webapi/contributor/contributor/contributor/contributor/support/admin/drivers/install/admin/cli/contributor/support/contributor/admin/user/contributor/cli/contributor/admin/admin/install/refarch/contributor/api/contributor/admin/contributor/contributor/contributor/contributor/admin/drivers/contributor/admin/admin/install/configuration/policy.html20:49
fungier,20:49
fungihttps://docs.openstack.org/developer/ironic/webapi/contributor/contributor/contributor/contributor/support/admin/drivers/install/admin/cli/contributor/support/contributor/admin/user/contributor/cli/contributor/admin/admin/install/refarch/contributor/api/contributor/admin/contributor/contributor/contributor/contributor/admin/drivers/contributor/admin/admin/install/configuratio20:49
fungin/policy.html20:49
fungiwhatever bot this is seems to take the path components of existing nearby urls and recompose/repeat them in every permutation20:50
fungii can only guess it's in hopes of finding pages that are otherwise unknown20:50
clarkbI think it is important to call out that all of this is publicly managed and tested pre merge. If people feel strongly that nginx would help (I'm not convinced since it is a ddos effectively) then you can write that change and push it up and start the conversation20:52
clarkbwe have also been fairly candid about the mitigation steps we have already taken and those that we think we could use to further alleviate the issue in our matrixroom20:52
fungi"our" being opendev's not the tact sig's20:52
clarkbif people are interested in this stuff then please help otu and get involved. But fungi and I are basically already pretty focused on this and have been since last week and even prior to that (the waf stuff we're doing is preexisting work that started over the holidays)20:52
clarkbits really easy to show up and say You should just use cloudflare or that we're using the wrong webserver20:53
clarkbits actually really not easy to keep all of these services up and running under an onslaught of poorly behavior bots (that know they are poorly behaved and spoof their details)20:53
clarkball while trying to avoid making the problem worse by over correcting and blocking legit traffic (my phone currently can't talk to github for a concrete exampel of this problem. I think because I have js disabled?)20:54
clarkbthis is one of my major concerns with anubis20:54
clarkbparticularly for static content that runs zero to minimal js20:55
clarkb(which is the bulk of the content on this particular server)20:55
stmcginnisLong time now see! .o/20:57
clarkbftr I don't want to use cloudflare. I think that their business practices around this particular issue are bit scammy (they just added "crawl this site" endpoint). So now rather than being crawled by random people on the internet you're getting crawled by the people you are paying to protect you.20:57
stmcginnisWanted to make sure someone here saw that http redirects to https are no longer working: https://www.reddit.com/r/openstack/comments/1ruupac/opesntack_docs_down/20:58
fungiit's not the redirects, it's just that apache is taking a while to process the request regardless of whether it's going to redirect or not20:59
fungiredirects will take longer though, because the client then has to wait for the redirect before requesting the actual page it's being redirected to21:00
clarkbright the problem is a ddos. It isn't specific to things like redirects21:01
fungiclarkb: yes, and probably the biggest upshot of all of this is that even if i was mildly curious about trying out some "ai" based tools before, i have absolutely zero interest in supporting anything to do with the obnoxious cesspool that the ai craze has turned into, burning the internet to the ground in order to try to scam a few more bucks out of people21:02
fungillms are a menace, plain and simple. their very existence has created perverse incentives to tragedy-of-the-commons everything in sight with absolutely no concern for what collateral damage the do along the way21:03
fungiit's like we gave an nuclear arsenal to a bunch of children and then asked them to play nicely21:05
clarkbit is worth noting that many of the big players seem to identify themselves correctly and don't create these situations. Unfortunately, other crawlers spoof user agents, come from botnet looking diverse ip ranges, request content that doesn't exist, and so on. Its these that are particularly problematic. Blocking google, openai, anthropic, meta, et al is unlikely to help much21:05
clarkbthis is where tools like anubis may help by forcing the bots to do proof of work before they can fetch the data. Unfortunately, they require your client to do the same21:06
clarkbif someone wants to deploy anubis and we see how bad it is for users and whether or not it helps against these particular botnets I am not opposed21:06
fungisadly, the "legitimate" llm operators have also been caught buying training data from dubious sources, perpetuating this stripmine-the-planet goldrush21:06
clarkbit is sad that non js content would require js just to filter out the bots though21:06
fungialso tomorrow's llm training crawlers will just throw in a client-side js processor to work around anubis, but hey maybe it buys us a few days before we have to replace it with yet-something-else21:08
clarkbyes, I'm honestly surprised they haven't done this yet. I Think tehy can share the cookie even so you don't have to calculate it over and over again21:08
fungimy guess is that the only reason they haven't so far is that so little of value to them on the internet puts a solution like that in place, so it's cheaper to just skip/ignore whatever's there21:09
fungibut as it becomes more popular, there will be an increasing incentive to work around21:09
fungiand by putting our eggs in that basket, we're increasing that incentive21:10
clarkbthinking out loud about the patterns we've seen for docs particularly I wonder if some sort of rate limiting mehcanism would be a good idea21:17
clarkbthough possibly easily defeated by the many many ip addresses21:17
fungiyes, i'll run some numbers, but probably better to discuss in #opendev:opendev.org matrix21:19
clarkb++ moving there21:19
JayFMy big concern is just we have hundreds of people who are slowed down by these issues; they aren't the fault of OpenDev but we need to shine more light on them so we can resolve them. It's going to drive away users and contributors if it hasn't already. 21:27
JayFI am not trying to proscribe any solution, just indicating that no solution is existentially threatening to openstack21:27
fungihopefully we can get the word out that trying to operate services supporting multi-million dollar companies on a semi-volunteer skeleton crew is a risk, and that some additional investment in the way of more actual sysadmins would go a long way to mitigating it. not when things are broken and falling down around us and we have to time to train anyone up, but during the calm21:29
fungitimes when it's running smoothly21:29
clarkband it goes beyond just opendev. If you're talking existential threats to openstack I could probably write up a list a mile long. But starting where there is acute pain is probably a good starting point21:31
fungiwe have a couple of close-to-full-time people, a couple of part-time people, and another part-time person being trained up to help21:31
clarkbre nginx it was pointed out that openstac documentation currently relies heavily on htaccess files for redirects22:00
clarkbso that won't work wtihout some conversion to nginx redirects or similar22:00
fungiyes, more to the point it's not just that it relies on .htaccess, but has a bunch of bespoke tooling and associated test framework for generating templated .htaccess rules and making sure they're not broken22:02
mnaserclarkb: You can just opt out of it. At some point we need to stabilize the infra here..22:08
clarkbsure that is an option too22:09
clarkbI just want to call it out as "this will berak and opesntadck will be broken in a different way unless this is handled too" for the nginx idea22:09
clarkbI don't thinka nyone is arguing about that. But what is missed is that fungi and I have been actively working on this specific problem since last week. And actively dealing with the dramatic change in internet traffic for months22:10
clarkbtoday everyone notices because the mitigations are not sufficient. But no one is aware of the other ddos storms that we've managed to deal with22:10
clarkbthis isn't new or foregin to us and it has been something that has eaten signficiant amount of time. Taking time away from other important tasks. This is why more help would be great22:11
fungiyes, the pattern of abuse has changed significantly this time, and the various solutions we've had in place weren't designed for this particular case so it's taking some new engineering22:11
clarkbit doesn't help that the openstack docs team which would be responsible for helping us disolved years ago and the TC which took on those duties has basically ignored them22:12
fungithe docs team used to develop their own site management tooling, even22:12
clarkbwe're what is left and we're doing our best. If that isn't good enough we'd really appreciate some help22:13
fungibut as people leave, responsibility gets increasingly concentrated on the ever fewer who stick around, until it burns them out too22:13
fungithese tasks have not traditionally been well-staffed because this is a volunteer project and services are maintained on a best-effort basis, but sometimes our best effort isn't enough to avoid a bit of pain22:14
fungiwe make up for and paper over a lot of it through automation, so it often goes unnoticed how small of a group of people we really have keeping it all running22:15
JayFclarkb: I told fungi this in DM, I'll say it here so you can hear: part of why I'm raising the alarm is it's *obvious* you all have been doing heroic things to keep it online. That is extremely appreciated ... and not sustainable. There need to be solutions that don't involve the infra team burning themselves out :( (I hope they exist)22:42

Generated by irclog2html.py 4.1.0 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!