[04:50:31] 10DBA, 10Operations, 10ops-eqiad, 10Patch-For-Review: Physically move es1017 from D to C row - https://phabricator.wikimedia.org/T197072 (10Marostegui) Anything left after repooling the host? [04:54:14] 10DBA: switchover es1014 to es1017 - https://phabricator.wikimedia.org/T197073 (10jcrespo) [04:54:18] 10DBA, 10Operations, 10ops-eqiad, 10Patch-For-Review: Physically move es1017 from D to C row - https://phabricator.wikimedia.org/T197072 (10jcrespo) 05Open>03Resolved a:03Cmjohnson I don't think so. [09:10:28] marostegui: I got a spike of errors on s6, probably replication [09:10:45] or could be just loadbalancer as usual [09:10:47] in codfw [09:10:49] I guess? [09:10:53] nope [09:11:11] That is strange, I haven't done anything that could cause lag there [09:11:15] all hosts? [09:11:20] 8:52, eqiad [09:11:37] db1085 I think [09:12:13] Interesting…that was the last host I did, not sure why it would have spiked [09:12:16] I will check [09:12:28] something related to ROW format, maybe? [09:12:44] did as in, at that time, or yesterday? [09:13:16] No, today, can't say it was around that time yeah [09:13:27] It is not even on graphs, so it must have been something fast no? [09:13:32] Could be related to ROW [09:13:35] Somehow [09:13:48] https://logstash.wikimedia.org/goto/ca6ac1c091fb6c426202766d7faf1157 [09:13:54] it seems for a single second [09:14:05] but who knows with the LB in the state it is- not reliable [09:14:30] also server db1093 is not replicating? [09:14:40] at the same time [09:15:06] That is weird, because I am touching one at the time [09:15:20] and db1093 was done long before db1085 [09:15:21] I am not too worried, I just ping when I see something weird just in case [09:16:06] yeah yeah [09:16:10] thanks for doing so [09:16:24] I am monitoring it actually, but 1-2 seconds can easily slip indeed [09:19:27] I also check more throwly [09:19:30] when I deploy [09:19:33] that is why [09:23:04] I wonder if I could build a better kibana dashboard with all errors without the known issues but only for eqiad [10:39:56] 10DBA: Check recentchanges table and query errors on wikis other than commonswiki and ruwiki - https://phabricator.wikimedia.org/T178290 (10Marostegui) [12:20:03] 10DBA, 10Schema-change: Schema change: Add unique index on archive.ar_rev_id - https://phabricator.wikimedia.org/T196379 (10Marostegui) a:03Marostegui [12:20:26] 10DBA, 10Schema-change: Schema change to drop default from externallinks.el_index_60 - https://phabricator.wikimedia.org/T197891 (10Marostegui) a:03Marostegui [12:45:51] only 3 servers left with jessie on codfw [12:45:58] (not counting parsercaches) [12:52:08] \o/ [12:52:29] how many do we have in eqiad (not counting masters)? [12:55:35] let me see [12:56:28] I think 7 [12:57:49] 7 masters eqiad + 7 replicas eqiad + 3 servers codfw + 6 parsercaches [12:58:00] 23 [12:58:42] marostegui: do you consider standalone ones as masters? [12:59:00] 4 es hosts [12:59:06] 1 dbstore1002 [12:59:23] and 5 labs related [12:59:28] we don't have much say on [13:01:00] Nice, not many then! [13:01:26] I may finish the codfw ones this week [13:01:35] * marostegui hugs jynus [13:25:42] I have pending repooling