[01:15:43] FIRING: [4x] HaproxyKafkaSocketDroppedMessages: Sustained high rate of dropped messages from HaproxyKafka - https://wikitech.wikimedia.org/wiki/HAProxyKafka#HaproxyKafkaSocketDroppedMessages - https://alerts.wikimedia.org/?q=alertname%3DHaproxyKafkaSocketDroppedMessages [01:20:43] FIRING: [9x] HaproxyKafkaSocketDroppedMessages: Sustained high rate of dropped messages from HaproxyKafka - https://wikitech.wikimedia.org/wiki/HAProxyKafka#HaproxyKafkaSocketDroppedMessages - https://alerts.wikimedia.org/?q=alertname%3DHaproxyKafkaSocketDroppedMessages [01:30:43] RESOLVED: [9x] HaproxyKafkaSocketDroppedMessages: Sustained high rate of dropped messages from HaproxyKafka - https://wikitech.wikimedia.org/wiki/HAProxyKafka#HaproxyKafkaSocketDroppedMessages - https://alerts.wikimedia.org/?q=alertname%3DHaproxyKafkaSocketDroppedMessages [03:50:43] FIRING: [8x] HaproxyKafkaSocketDroppedMessages: Sustained high rate of dropped messages from HaproxyKafka - https://wikitech.wikimedia.org/wiki/HAProxyKafka#HaproxyKafkaSocketDroppedMessages - https://alerts.wikimedia.org/?q=alertname%3DHaproxyKafkaSocketDroppedMessages [03:55:43] FIRING: [9x] HaproxyKafkaSocketDroppedMessages: Sustained high rate of dropped messages from HaproxyKafka - https://wikitech.wikimedia.org/wiki/HAProxyKafka#HaproxyKafkaSocketDroppedMessages - https://alerts.wikimedia.org/?q=alertname%3DHaproxyKafkaSocketDroppedMessages [04:00:43] FIRING: [9x] HaproxyKafkaSocketDroppedMessages: Sustained high rate of dropped messages from HaproxyKafka - https://wikitech.wikimedia.org/wiki/HAProxyKafka#HaproxyKafkaSocketDroppedMessages - https://alerts.wikimedia.org/?q=alertname%3DHaproxyKafkaSocketDroppedMessages [04:05:43] RESOLVED: [9x] HaproxyKafkaSocketDroppedMessages: Sustained high rate of dropped messages from HaproxyKafka - https://wikitech.wikimedia.org/wiki/HAProxyKafka#HaproxyKafkaSocketDroppedMessages - https://alerts.wikimedia.org/?q=alertname%3DHaproxyKafkaSocketDroppedMessages [13:30:00] I'm having another look at T406222. I'll start following https://wikitech.wikimedia.org/wiki/LVS#Add_a_new_load_balanced_service in a few minutes unless someone tells me to stop. [13:30:01] T406222: Add druid coordinator service to LVS for the druid_public cluster. - https://phabricator.wikimedia.org/T406222 [13:33:17] if someone could review https://gerrit.wikimedia.org/r/c/operations/puppet/+/1216793 [13:47:51] 06Traffic, 06Data-Platform-SRE (2025.11.07 - 2025.11.28), 07Essential-Work, 13Patch-For-Review: Add druid coordinator service to LVS for the druid_public cluster. - https://phabricator.wikimedia.org/T406222#11443942 (10Gehel) deploying following instructions in https://wikitech.wikimedia.org/wiki/LVS#Confi... [13:48:02] lgfm too! [13:58:35] fabfur: thx [13:58:48] 06Traffic, 06Data-Platform-SRE (2025.11.07 - 2025.11.28), 07Essential-Work, 13Patch-For-Review: Add druid coordinator service to LVS for the druid_public cluster. - https://phabricator.wikimedia.org/T406222#11443987 (10Gehel) This seems to be working, sending an HTTP 307 redirect to one of the druid node:... [14:09:19] fabfur: if you have a minute to check on https://gerrit.wikimedia.org/r/c/operations/puppet/+/1216797 and the prerequisite for merging it are fullfilled. [14:10:52] have all previous steps been successful ? [14:11:05] as far as I know, yes. [14:11:18] successful test on https://phabricator.wikimedia.org/T406222#11443987 [14:12:09] ok so puppet ran on all lvs hosts and pybal has been restarted too I suppose [14:13:45] I completed steps 4 and 7 with no error! [14:14:40] lgtm! [14:14:45] thx! [14:17:11] 06Traffic, 06Data-Platform-SRE (2025.11.07 - 2025.11.28), 07Essential-Work, 13Patch-For-Review: Add druid coordinator service to LVS for the druid_public cluster. - https://phabricator.wikimedia.org/T406222#11444052 (10Gehel) 05Open→03Resolved [14:22:04] 06Traffic, 06Data-Platform-SRE (2025.11.07 - 2025.11.28), 07Essential-Work, 13Patch-For-Review: Add druid coordinator service to LVS for the druid_public cluster. - https://phabricator.wikimedia.org/T406222#11444059 (10Gehel) HTTP calls to `druid-public-coordinator.svc.eqiad.wmnet:8081` result in an HT... [15:53:10] 06Traffic, 06Security-Team, 10WMF-General-or-Unknown, 07ContentSecurityPolicy, 13Patch-For-Review: Add restrictive CSP to upload.wikimedia.org - https://phabricator.wikimedia.org/T117618#11444306 (10ssingh) Thanks for the update @sbassett. Noting that Traffic requested the move to January. [17:35:46] 10netops, 06Infrastructure-Foundations, 10Observability-Logging: ~5k/logs/sec from netdev - https://phabricator.wikimedia.org/T412143 (10colewhite) 03NEW [17:36:17] 10netops, 06Infrastructure-Foundations, 10Observability-Logging: ~5k/logs/sec from netdev - https://phabricator.wikimedia.org/T412143#11444783 (10colewhite) [17:41:26] 06Traffic, 06SRE: Getting forbidden from public CI runners on forgejo with opendatasync - https://phabricator.wikimedia.org/T412142#11444796 (10Reedy) [18:02:32] 06Traffic, 06SRE: Getting forbidden from public CI runners on forgejo with opendatasync - https://phabricator.wikimedia.org/T412142#11444919 (10taavi) Please provide the full HTTP response body, which either details why the request is blocked or contains an internal identifier for us to locate the relevant WAF... [18:21:37] 10netops, 06Infrastructure-Foundations, 10Observability-Logging: ~5k/logs/sec from netdev - https://phabricator.wikimedia.org/T412143#11444984 (10ayounsi) Looks like a repeat of {T398433} but for a different switch. As it's not in a VXLAN fabric, we should look at upgrading that one switch. [18:32:44] 10netops, 06Infrastructure-Foundations, 06SRE: Nokia SR-Linux ARP resolution bug on v24.10.x+ - https://phabricator.wikimedia.org/T409178#11445009 (10ops-monitoring-bot) Icinga downtime and Alertmanager silence (ID=2a98251c-6798-469c-a3de-57fcfb13969f) set by cmooney@cumin1003 for 2:00:00 on 17 host(s) and t... [18:36:43] 06Traffic, 06SRE: Getting forbidden from public CI runners on forgejo with opendatasync - https://phabricator.wikimedia.org/T412142#11445020 (10EvanCarroll) `

Our servers are currently under maintenance or experiencing a technical issue