[01:03:39] 10DBA, 06Labs, 10Tool-Labs, 13Patch-For-Review: Tool Labs queries die - https://phabricator.wikimedia.org/T127266#2730527 (10Dispenser) Queries are being killed after 2 hours since June, leaving data for tools stale. [06:20:43] 10DBA, 13Patch-For-Review: Unify commonswiki.revision - https://phabricator.wikimedia.org/T147305#2730655 (10Marostegui) db1069's ALTER is now done so I believe we can close this ticket: ``` MariaDB MARIADB db1069 commonswiki > show create table revision\G *************************** 1. row ******************... [06:29:12] 10DBA, 06Operations, 10ops-codfw: db2037: Disk in predictive failure - https://phabricator.wikimedia.org/T148373#2730656 (10Marostegui) And everything looks good now, thanks a lot @Papaul ``` => ctrl slot=0 physicaldrive all show Smart Array P420i in Slot 0 (Embedded) array A physicaldrive 1I:... [06:29:25] 10DBA, 06Operations, 10ops-codfw: db2037: Disk in predictive failure - https://phabricator.wikimedia.org/T148373#2730657 (10Marostegui) 05Open>03Resolved [07:31:19] 10DBA, 06Labs, 10Labs-Infrastructure, 06Operations: Create maintain-views user for labsdb1001 and labsdb1003 - https://phabricator.wikimedia.org/T148560#2730717 (10jcrespo) I suggest we should avoid providing wildcard grants. Those have created issues in the past, and we want to move away from them. When g... [08:08:20] 10DBA, 13Patch-For-Review: Reimage dbstore2001 as jessie - https://phabricator.wikimedia.org/T146261#2730743 (10Marostegui) The import finished. One table crashed though, however everything looks fine. This is the timeline: `transcode`table (which as a PK) crashed while importing. Started with innodb_force_re... [08:09:28] jynus ^ I have edited the comment (I hit enter by mistake) just in case you only got to "S4", so just refresh :) [08:10:33] I assume every time things crases, no replication was running, or was it? [08:10:46] No, it wasn't running no [08:11:01] configuring and starting replication was the last step :) [08:11:15] but maybe running for other shards? [08:11:41] for other shards yes, and I left it running because I wanted to see if we can import data without stopping replication from other sources [08:11:44] Which should be that way [08:11:49] Otherwise it is a bit crappy [08:11:52] (I think) [08:11:55] and gtid on or off? [08:12:05] For S3 ON [08:12:14] For S1 I will enable it now [08:12:16] the one that was running? [08:12:23] yep, S3 was the one that was running [08:12:28] S1 was the one being imported [08:12:32] then it should be secure [08:12:46] gtid in mariadb does provide innodb transactional control [08:13:25] transcode, what about this table? [08:13:39] no PK like the others, or any other strange characteristic? [08:13:41] Don't know really, first time it crashes during these imports [08:13:46] No no, it does have a PK [08:14:07] we should do more tests on eqiad [08:14:28] probably there is some kind of ancient row format issue with certain tables [08:14:38] I am going to add to the documentation to the well known issues the PK thingy as it is clearly an issue (or can be) [08:14:41] and all of codfw probably comes from the same original server [08:14:54] I do not know [08:14:58] Could be, but this table never crashed before while change_tag crashed everytime [08:15:01] it is strange [08:15:22] what I am a bit disappointed is with importing tablespaces [08:15:32] which doesn't seem so great for us [08:15:35] Yes, it doesn't look super reliable [08:15:41] slow, unreliable, yes [08:15:47] Although, this time it worked, we will see what happens with S4 [08:15:57] Maybe it was just that problematic table: change_tags [08:16:04] it would be nice to see how it works in reverse [08:16:13] in reverse? [08:16:20] because if xtrabackup creates problems with many tables [08:16:47] these servers as the source of copies seems not a great idea [08:17:34] and then we have compression [08:17:43] which I do not think will make things more stable... [08:17:49] indeed [08:18:04] everything you are doing is great [08:18:12] Thanks for your support :) [08:18:17] I am just thinking the best way to move forward [08:18:28] At some point, maybe it is even something we need to consider to create a source host built from a logical backup [08:18:39] I think we can enable compression in S1 now [08:18:40] oh, yes [08:18:51] we just never find the time [08:19:03] And see how it works with tables that have been copied via moving .ibd files [08:19:07] As I am a bit scared [08:19:07] takes 1-2 days [08:19:22] Because if that fails, then that is a blocker [08:19:38] I think it is better to try that before importing S4 [08:19:54] So we can have a full picture of how transportable spaces + compression works [08:19:54] note we could have been carrying 4.x unsupported formats for a long time [08:20:21] or maybe our versions has bugs on features like this, which are not so popular [08:20:30] could be [08:20:43] In that case, the source host needs to be built from logical backups [08:20:49] Which is the safest thing probably [08:20:54] also [08:21:07] I would love to have per-db tablespaces for s3 [08:21:25] but that is 5.7-only [08:21:54] we are getting better, at least we will not have 40K files on a single dir [08:21:57] as with tokudb [08:22:01] yeah XDDD [08:22:07] by the way ,db1069 finished the alter [08:22:16] saw it [08:22:20] resolve [08:22:27] TokuDB didn't play trick on us this time XD [08:22:33] ha [08:22:39] was it online? [08:22:46] No XD [08:22:56] But I was expecting it to crash or soemthing as we were manipulating PKs [08:23:03] look, when even innodb is more online [08:23:12] My faith on tokudb is quite down now :( [08:23:12] than toku, which is one of its selling points [08:23:46] it is not that the balance is weighted, it is that there is no reason to use it [08:23:51] 10DBA: Rampant differences in indexes on enwiki.revision across the DB cluster - https://phabricator.wikimedia.org/T132416#2730764 (10Marostegui) [08:23:54] 10DBA, 13Patch-For-Review: Unify commonswiki.revision - https://phabricator.wikimedia.org/T147305#2730763 (10Marostegui) 05Open>03Resolved [08:24:17] for innodb compression [08:25:25] there was a task [08:27:15] https://phabricator.wikimedia.org/T139055 [08:27:16] this? [08:27:23] yes [08:27:29] it didn't show on searches [08:27:40] Nop, it didn't I had to go tot he DBA dashboard :| [08:28:52] so apparently I tested it on production on db1073 [08:28:56] before: https://phabricator.wikimedia.org/T139055#2418115 [08:29:14] ALTER TABLE revision row_format=compressed key_block_size=8\G [08:29:28] 16 hours nice [08:29:45] (13 hours 10 min 8.65 sec) for pagelinks [08:29:49] I am so non confident about transportable spaces + compression... [08:29:54] I hope MariDB proves me wrong [08:30:02] ha [08:30:39] Probably I will stop slave; to alter them [08:30:48] db2034 was cloned from db1073 [08:32:24] https://grafana.wikimedia.org/dashboard/db/server-board?from=1464602836457&to=1468447311389&var-server=db1073&var-network=eth0&panelId=17&fullscreen [08:32:35] Note the increase in size before the reduction [08:33:00] Why is that? [08:33:38] 10DBA, 07Schema-change, 07Tracking: Schema changes for Wikimedia wikis (tracking) - https://phabricator.wikimedia.org/T51188#2730768 (10hoo) [08:33:39] is that the table rebuilds during the alters? [08:36:00] yes [08:36:15] so it is dangerous to do it with low space [08:36:25] yeah [08:36:39] for dbstores we should be safe, but it is indeed something to keep in mind [08:36:42] how much space do you have now used/free there? [08:37:04] A lot, 3.2T available but I have around 600G of the snapshot [08:37:19] So 3.8T [08:37:22] ok, then used by the fs? [08:37:32] 3.8 by 3 shards without compression? [08:37:38] is that right^? [08:38:06] 2 shards (S1, S3) without compression are 2.9T [08:38:16] And the snapshot of those two compressed is 600G [08:38:24] ah, ok [08:38:29] root@dbstore2001:/srv# df -hT /srv/ [08:38:29] Filesystem Type Size Used Avail Use% Mounted on [08:38:29] /dev/mapper/tank-data xfs 6.6T 3.4T 3.2T 53% /srv [08:38:49] So we should be fine to compress S1 [08:43:01] 10DBA: Test InnoDB compression - https://phabricator.wikimedia.org/T139055#2730773 (10jcrespo) db1073 as of now: ``` +----------+------------------+--------+------------+----------+--------+--------+---------+---------+ | DATABASE | TABLE | ENGINE | ROW_FORMAT | ROWS | DATA | IDX | TOTAL |... [08:43:11] see my update here^ [08:43:35] it seems I only compressed the top 14 tables [08:43:43] to get a 50% reduction [08:45:22] clearly, just doing *links tables + revision is a huge help [08:45:52] because the amount of text they have [08:46:17] so do you have more or less a plan? [08:46:38] whatever you learn about compression specifically, add it to that ticket, too [08:51:38] Yeah, I do have a plan [08:51:50] So I am going to wait for S1 to catchup, create a snapshot and start compressing S1 [08:51:54] I will add it to the ticket [08:53:18] revision | InnoDB | Compressed | 679.52M [08:53:24] Almost 700M rows, amazing [08:54:15] uf, mechanical disks + snapshot ? [08:54:43] Creating the tar with pigz last time took around 2 hours or so [08:54:58] nothing to object, I just will hope the alter will finish at some point [08:55:02] :-) [08:55:11] XDDDD [08:55:24] and that you will not run out of snapshot space? [08:55:42] I will delete the previous snapshot [08:55:55] So we would have around 3.6T to start doing the alters (one by one) [08:56:15] I mean, automatically but on a sequential way I think [08:57:00] es2015 is not ok :-/ [08:57:19] raid again? :( [08:57:19] "WARNING: could not read sysctl settings" [08:57:30] :( [08:57:37] that should be purely software [08:57:47] kernel, or whatever [08:57:56] Interesting sysctl -a does work [08:58:05] maybe it is icinga itself? [08:58:11] but it happend last time when there were hw issues [08:58:19] Yeah, it happened to db1082 too [08:58:22] maybe I am just being irrational here [08:58:36] because that does not make sense [08:58:39] No, that did happen to db1082 when it was having its crisis [10:01:23] jynus: you joining me in the call? :) [10:01:29] yes [10:01:33] :) [10:01:46] I just see to events [10:01:49] 2 [10:01:51] ah [10:01:55] the DBA triage one [10:02:00] We are both there [10:02:15] I think itis the same hangout [10:02:25] should be [10:09:22] 07Blocked-on-schema-change, 10DBA, 10Wikidata, 07Schema-change: Drop eu_touched in production - https://phabricator.wikimedia.org/T144010#2730915 (10hoo) [10:45:22] 10DBA, 07Epic, 07Tracking: Database tables to be dropped on Wikimedia wikis and other WMF databases (tracking) - https://phabricator.wikimedia.org/T54921#2730965 (10mark) [10:54:42] 07Blocked-on-schema-change, 10DBA, 10Expiring-Watchlist-Items, 10MediaWiki-Watchlist, and 3 others: Add wl_id to watchlist tables on production dbs - https://phabricator.wikimedia.org/T130067#2730969 (10mark) [11:54:29] jynus: I lost you [11:54:38] let me try to reconnect [11:54:45] ok [12:48:42] Progress: 31.151 Time: 10411 / Progress: 15.940 Time: 10411 [12:51:45] Not bad [12:51:53] https://grafana-admin.wikimedia.org/dashboard/db/server-board?panelId=17&fullscreen&var-server=db1053&var-network=bond0&from=now-7d&to=now-4m [12:53:46] We will see the final graph.. [13:29:26] 10DBA, 06Operations, 10ops-eqiad: Degraded RAID on db1046 - https://phabricator.wikimedia.org/T148633#2731366 (10elukey) p:05Triage>03High [13:52:14] 10DBA, 10MediaWiki-Database, 07Wikimedia-log-errors: Special:Block request makes the database timeout on db1055 - https://phabricator.wikimedia.org/T140650#2731432 (10Johan) [14:18:01] 10DBA, 06Operations, 10ops-eqiad: Degraded RAID on db1046 - https://phabricator.wikimedia.org/T148633#2731507 (10Cmjohnson) 05Open>03Resolved a:03Cmjohnson The disk was replaced yesterday. All systems go root@db1046:~# megacli -PDList -aALL |grep "Firmware state" Firmware state: Online, Spun Up Firmw... [14:44:20] do you know what is the deal with s3 on dbstore2001? [14:45:19] jynus: I am updating the ticket right now :_( [14:45:21] But this is the deal [14:45:24] All of a sudden [14:45:32] 161020 14:36:56 [ERROR] Master 's3': Slave I/O: Got fatal error 1236 from master when reading data from binary log: 'Error: connecting slave requested to start from GTID 0-171974683-4736235774, which is not in the master's binlog', Internal MariaDB error code: 1236 [14:45:36] 161020 14:36:56 [Note] Master 's3': Slave I/O thread exiting, read up to log 'db2018-bin.002452', position 4; GTID position 0-171974683-4736235774 [14:45:41] just like that [14:45:47] And s3 has been working fine for days [14:46:12] what if you disable GTID? [14:49:38] I was first checking the binlogs in the master [14:49:43] to see if that position exists [14:54:06] I guess I can do a change master to and use db2018-bin.002452', position 4; [14:54:14] But I am pretty sure that is wrong and won't work [14:54:43] don't [14:55:01] first, is id 171974683 the server id of the master? [14:56:29] No, the master's id: 180359174 (db2018) [14:56:32] second, try to identify the last executed event of s3 from the binlogs of the dbstore [14:56:43] ah, then it may be something else [14:56:51] like a local write [14:57:09] I think disabling GTID will work [14:57:26] but try first to check the above^ to be sure [14:59:51] I am still exploring the master binlog to see if any of them contains 171974683 [15:01:29] I think it will be on the slave only, by experience [15:01:48] Yeah, it is there indeed [15:01:50] in the salve [15:01:51] slave [15:02:00] so that's it [15:02:09] https://phabricator.wikimedia.org/P4275 [15:02:12] that is in the slave [15:12:44] It is interesting that that query isn't even on the master's binlogs [15:22:17] jynus: What do you suggest? Trying to do a change master to master_user_gtid=slave_pos ? [15:22:37] to none [15:22:46] or whatever is disabling it [15:23:00] Ah, yes: "no" [15:23:05] That is the one to disable it I believe [15:23:32] I remember somthing similar happening to a dbstore [15:23:50] all of sudden like this? [15:24:04] not all of a sudden [15:25:22] As expected, duplicate entries are now showing up [15:40:52] strange [15:42:13] Yeah, it is very strange that it crashed like that [15:42:26] there is something writing where it shouldn't [15:42:44] we need to know what and where [15:44:19] I am connecting dots now [15:44:28] Because S1 got a duplicate entry as well there, pretty much at the same time [15:44:43] (I am still updating the ticket) but I am giving you a break news :p [15:57:12] 10DBA, 13Patch-For-Review: Reimage dbstore2001 as jessie - https://phabricator.wikimedia.org/T146261#2731817 (10Marostegui) Looks like after the crash today, S3 replication thread got broken - reminder: s3 has been untouched for almost a week after it was copied over: ``` 161020 14:23:02 [ERROR] Master 's3':... [15:58:42] if it is after the crash, it is not somthing [15:59:04] if a server crashes and does not use gtid, it immediately gets corrupted [15:59:16] jynus: s3 was using gtid :) [15:59:18] something external [15:59:44] then we have a configuration problem [16:00:18] It is weird that after a week replicating just fine, all of a sudden it gets that error [16:00:26] mmmm [16:00:27] (there have been many crashes) [16:00:35] what is the default database for that change? [16:00:44] the one on the ticket? [16:00:53] what do you mean? [16:01:13] I am thinking of a filtering issue [16:01:37] filtering as in mysql replication filters? [16:01:51] yes [16:02:03] Interesting…let me know what you are thinking :) [16:02:33] This is what S3 has: Replicate_Wild_Do_Table: %wik%.%,heartbeat.% [16:02:50] yes, that is why I mentioned [16:02:51] it [16:02:57] But you think that issue can happen after a week replicating finely? [16:03:21] do you know which server has server id 171974683 ? [16:03:53] No idea, but I can try to check :) [16:06:24] db1057.eqiad.wmnet [16:06:25] server_id 171974683 [16:06:42] that is the master [16:07:45] mmm now that I think about it, I checked the master's (codfw) binlog, but not the upstream one [16:08:00] It was not appearing in codfw one (which is weird) but it might appear in the primary one [16:08:09] Which still doesn't make sense to me [16:08:10] what is the intermediate master? [16:08:13] 2016? [16:08:37] for s3 db2018 [16:08:49] s3 is the one that had the weird issue [16:21:53] too much table corruption [16:22:05] I do not even think this is a wise method to continue [16:23:03] I really cannot believe this transportable spaces is that bad :( [16:23:15] try 5.7 [16:23:47] But we'd need another host to be 5.7 to use as source [16:24:18] we will just try with 10 -> 5.7 [16:24:26] do you have a better suggestion? [16:24:41] No, we can try that [16:24:48] because my other suggestion is to do a logical import [16:24:57] and you will like that less [16:24:58] Let's try 10.0 -> 5.7 [16:25:00] as that can be faster [16:25:01] XDDD [16:25:04] :-/ [16:25:17] I will go for 5.7 in dbstore2001 then [16:25:23] go for a smaller sample [16:25:30] Yeah [16:25:41] if a single table doesn't work, it is not worth to continue [16:25:48] indeed [16:26:02] Well, for S1 for instance, it failed 1 table out of a lot [16:26:04] what script are you using anyway? [16:26:35] For moving the tablespaces? [16:28:11] for whatever you are doing [16:28:39] Ah, for the normal copies (not moving ibd) just a single nc and for moving the ibd files I am doing it manually too, as I do not want to add more variables [16:28:43] ie: the script might be broken [16:28:48] no [16:28:53] So I am executing the MySQL commands with a bash oneliner [16:28:57] but how are you automatizing [16:29:00] each table [16:29:05] Ah [16:29:09] bash so far [16:29:11] I suppose you do not run 40000 commands? [16:29:16] haha no XDDD [16:29:22] is it on the documentation? [16:29:23] Just simple bash oneliners [16:29:26] I think so [16:29:39] yep: • for i in `mysql --skip-ssl enwiki -e "show tables;" -B`; do echo $i; mysql --skip-ssl enwiki -e "flush table $i for export;";done [16:31:56] I am going to do some groceries - will talk to you tomorrow!! [16:31:58] wait [16:32:01] what [16:32:16] I think there are things there that should not work [16:32:28] like? [16:33:09] go, I will have a closer look at it [16:33:18] sure! [16:33:26] let me know if you see something strange [16:33:30] thanks! [16:37:41] if you are doing as documented, I would not know how it can even work? [16:38:09] flush table $i for export; should be done within a session [16:38:24] if you exist mysql, the locks should be dropped [16:38:34] and mysql cannot be stopped [16:38:49] plus you are coping ibdata1 over the other? [16:39:09] it will crash, and when it restarts, it will not likely work again [17:03:05] I think the last part you aren't, it is just a typo of the documentation [17:03:20] but let's talk tomorrow about improving the method :-) [17:04:42] check before I connect https://www.percona.com/blog/2014/12/09/mysql-5-6-transportable-tablespaces-best-practices/ [17:13:57] 10DBA, 06Labs, 10Labs-Infrastructure, 06Operations: Create maintain-views user for labsdb1001 and labsdb1003 - https://phabricator.wikimedia.org/T148560#2732127 (10AlexMonk-WMF) >>! In T148560#2730717, @jcrespo wrote: > Right now we maintain a blacklist on realm.pp. We should transform that into a white li... [17:21:42] 10DBA, 06Labs, 10Labs-Infrastructure, 06Operations: Create maintain-views user for labsdb1001 and labsdb1003 - https://phabricator.wikimedia.org/T148560#2732172 (10AlexMonk-WMF) I suppose we go from one list that needs updating when private wikis are created to one that needs updating when non-private (the... [17:23:19] 10DBA, 06Community-Tech, 10MediaWiki-extensions-CentralAuth: Add indices for local_user_id and global_user_id in production - https://phabricator.wikimedia.org/T148243#2718099 (10kaldari) [18:12:22] 10DBA, 06Operations, 10ops-codfw, 13Patch-For-Review: es2015 crashed with no os logs (kernel logs or other software ones) - it shuddenly went down - https://phabricator.wikimedia.org/T147769#2732383 (10Papaul) p:05High>03Normal [20:04:53] jynus: Yeah - I was not copying the ibdata1 :-) [20:05:05] But what you said about closing the session can be the key. However, S1 is working fine :| [20:05:14] Let's discuss that tomorrow yeah [20:05:19] Thanks for the feedback though! [20:20:52] 10DBA: apply events killing long running queries to db1070; any other production server - https://phabricator.wikimedia.org/T148790#2732802 (10jcrespo) [20:21:20] 10DBA: apply events killing long running queries to db1070; any other production server - https://phabricator.wikimedia.org/T148790#2732815 (10jcrespo) p:05Triage>03Unbreak! [22:04:50] 10DBA, 10CirrusSearch, 06Discovery, 06Discovery-Search (Current work), and 2 others: CirrusSearch SQL query for locating pages for reindex performs poorly - https://phabricator.wikimedia.org/T147957#2733044 (10Deskana) [22:05:41] 10DBA, 10CirrusSearch, 06Discovery, 06Discovery-Search (Current work), and 2 others: CirrusSearch SQL query for locating pages for reindex performs poorly - https://phabricator.wikimedia.org/T147957#2710120 (10Deskana) The above is a hack that unblocks @smalyshev's work temporarily, but we are awaiting a p...