[00:01:43] PROBLEM Free ram is now: WARNING on bots-4.pmtpa.wmflabs 10.4.0.64 output: Warning: 17% free memory [00:02:38] Ryan_Lane: would be nice if you can find time to "clone" the openid-wiki to an openid-wiki2 . I just "pulled" the recent core and E:OpenID. [00:03:03] and I am logged out from there [00:03:24] I mean I am not logged in to openid-wiki [00:04:49] I can't right now [00:04:54] I'll try to get this done tonight [00:05:18] crunch crunch crunch. Gawd. I haven't seen a package this long to build since Qt! [00:05:27] ok, that'll be fine ! [00:06:29] "see" you tomorrow, good night (Berlin local time: 01:6) [00:07:46] andrewbogott: can you add me as a reviewer on those changes? [00:07:53] they aren't in my list [00:10:09] Ryan_Lane: OK, done. [00:10:13] cool [00:13:07] * Coren wishes he had OGS installed so he could use qmake to build OGS. [00:29:22] RECOVERY Total processes is now: OK on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS OK: 146 processes [00:37:12] RECOVERY Total processes is now: OK on bastion1.pmtpa.wmflabs 10.4.0.54 output: PROCS OK: 148 processes [00:37:12] RECOVERY Free ram is now: OK on sube.pmtpa.wmflabs 10.4.0.245 output: OK: 22% free memory [00:37:32] RECOVERY Free ram is now: OK on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: OK: 23% free memory [00:39:52] RECOVERY Free ram is now: OK on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: OK: 22% free memory [00:41:52] RECOVERY Free ram is now: OK on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: OK: 22% free memory [00:47:23] PROBLEM Total processes is now: WARNING on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS WARNING: 154 processes [00:49:29] andrewbogott: reviewed one [00:49:35] I'm reviewing the other [00:49:46] it would likely be good to stop unused volumes [00:49:51] err [00:49:53] unloved [00:50:24] unused volumes are more difficult because it could lead to flapping, if a user create/deletes/creates/deletes a single instance [00:50:54] for that situation we could have a cache file that lets us know when a volume was last stopped [00:51:04] if it was stopped within an hour, it would not stop it [00:52:52] PROBLEM Free ram is now: WARNING on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: Warning: 18% free memory [00:56:45] OK, I'll look at adding stopping. Getting things to start again will be a bit messy. [01:00:32] PROBLEM Free ram is now: WARNING on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: Warning: 17% free memory [01:02:22] RECOVERY Total processes is now: OK on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS OK: 143 processes [01:03:26] Ryan_Lane, are you starting/stopping/messing with gluster on labstore2? [01:05:11] hm, maybe not so messy [01:05:57] andrewbogott: I disabled the script temporarily [01:06:17] OK. 'gluster volume info' was telling me that gluster was down [01:06:19] But no longer [01:09:54] PROBLEM Free ram is now: WARNING on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: Warning: 19% free memory [01:10:13] PROBLEM Free ram is now: WARNING on sube.pmtpa.wmflabs 10.4.0.245 output: Warning: 14% free memory [01:14:53] PROBLEM Current Load is now: WARNING on parsoid-roundtrip6-8core.pmtpa.wmflabs 10.4.0.222 output: WARNING - load average: 8.50, 7.23, 5.92 [01:20:24] PROBLEM Current Load is now: WARNING on parsoid-roundtrip3.pmtpa.wmflabs 10.4.0.62 output: WARNING - load average: 8.64, 7.74, 5.70 [01:24:54] PROBLEM Current Load is now: WARNING on ve-roundtrip2.pmtpa.wmflabs 10.4.0.162 output: WARNING - load average: 5.98, 8.00, 6.02 [01:34:30] Ryan_Lane: I'd like to add the starting/stopping as a separate patch. So go ahead and submit comments on the patch in gerrit as it is. [01:35:15] ok [01:46:42] PROBLEM Free ram is now: CRITICAL on bots-4.pmtpa.wmflabs 10.4.0.64 output: Critical: 3% free memory [01:47:20] !log webtools installed libssl0.9.8 (dependency for OGS) [01:47:22] Logged the message, Master [01:51:43] PROBLEM Free ram is now: WARNING on bots-4.pmtpa.wmflabs 10.4.0.64 output: Warning: 6% free memory [02:00:23] PROBLEM Total processes is now: WARNING on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS WARNING: 163 processes [02:01:43] RECOVERY Free ram is now: OK on bots-4.pmtpa.wmflabs 10.4.0.64 output: OK: 78% free memory [02:09:52] RECOVERY Current Load is now: OK on ve-roundtrip2.pmtpa.wmflabs 10.4.0.162 output: OK - load average: 3.87, 4.20, 4.97 [02:11:20] Ryan_Lane, do you know how to make 'gluster volume stop' skip the 'Do you want to continue?' prompt? [02:11:24] Oddly, 'force' does not do that. [02:11:26] yep [02:13:50] Ryan_Lane: So… how? [02:13:57] one sec [02:17:53] PROBLEM Current Load is now: WARNING on ve-roundtrip2.pmtpa.wmflabs 10.4.0.162 output: WARNING - load average: 4.98, 4.93, 5.07 [02:21:10] andrewbogott: --mode=script [02:23:16] works! [02:25:23] RECOVERY Total processes is now: OK on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS OK: 147 processes [02:44:52] RECOVERY Current Load is now: OK on parsoid-roundtrip6-8core.pmtpa.wmflabs 10.4.0.222 output: OK - load average: 3.47, 3.87, 4.87 [02:49:22] PROBLEM Total processes is now: WARNING on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS WARNING: 156 processes [02:55:32] RECOVERY Current Load is now: OK on parsoid-roundtrip3.pmtpa.wmflabs 10.4.0.62 output: OK - load average: 4.47, 4.35, 4.96 [03:03:42] Ryan_Lane: I'm noticing that right after I start and stop a volume, subsequent 'gluster volume' commands fail for a few seconds. [03:03:49] yep [03:03:53] it's really annoying [03:04:24] This is suspiciously similar to the behavior in the last version when I manually restarted the service after a start or stop... [03:04:28] we should likely change the cron to run every 5 minutes [03:04:32] Suppose they fixed the memory leak by having the service restart itself? [03:04:39] heh [03:05:12] Anyway… perhaps this script should just exit after it either starts or stops a volume. Since it won't be getting any more work done after that... [03:05:30] yeah. maybe so [04:37:52] RECOVERY Free ram is now: OK on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: OK: 22% free memory [04:39:52] RECOVERY Free ram is now: OK on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: OK: 27% free memory [04:40:17] RECOVERY Free ram is now: OK on sube.pmtpa.wmflabs 10.4.0.245 output: OK: 22% free memory [04:40:33] RECOVERY Free ram is now: OK on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: OK: 22% free memory [04:43:43] PROBLEM Free ram is now: WARNING on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: Warning: 14% free memory [04:53:44] PROBLEM Free ram is now: CRITICAL on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: Critical: 5% free memory [04:58:32] PROBLEM Free ram is now: WARNING on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: Warning: 17% free memory [04:58:42] PROBLEM Free ram is now: WARNING on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: Warning: 6% free memory [05:02:52] PROBLEM Free ram is now: WARNING on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: Warning: 13% free memory [05:03:12] PROBLEM Free ram is now: WARNING on sube.pmtpa.wmflabs 10.4.0.245 output: Warning: 14% free memory [05:05:52] PROBLEM Free ram is now: WARNING on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: Warning: 17% free memory [06:30:52] PROBLEM Total processes is now: WARNING on parsoid-roundtrip5-8core.pmtpa.wmflabs 10.4.0.125 output: PROCS WARNING: 151 processes [06:35:52] RECOVERY Total processes is now: OK on parsoid-roundtrip5-8core.pmtpa.wmflabs 10.4.0.125 output: PROCS OK: 150 processes [07:24:13] PROBLEM Free ram is now: WARNING on bots-cb.pmtpa.wmflabs 10.4.0.44 output: Warning: 18% free memory [07:29:12] RECOVERY Free ram is now: OK on bots-cb.pmtpa.wmflabs 10.4.0.44 output: OK: 34% free memory [08:37:52] RECOVERY Free ram is now: OK on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: OK: 22% free memory [08:38:12] RECOVERY Free ram is now: OK on sube.pmtpa.wmflabs 10.4.0.245 output: OK: 22% free memory [08:38:32] RECOVERY Free ram is now: OK on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: OK: 23% free memory [08:38:42] RECOVERY Free ram is now: OK on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: OK: 27% free memory [08:40:52] RECOVERY Free ram is now: OK on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: OK: 21% free memory [08:46:42] PROBLEM Free ram is now: WARNING on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: Warning: 19% free memory [08:51:32] PROBLEM Free ram is now: WARNING on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: Warning: 19% free memory [08:56:33] RECOVERY Free ram is now: OK on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: OK: 20% free memory [09:10:52] PROBLEM Free ram is now: WARNING on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: Warning: 14% free memory [09:11:12] PROBLEM Free ram is now: WARNING on sube.pmtpa.wmflabs 10.4.0.245 output: Warning: 14% free memory [09:13:52] PROBLEM Free ram is now: WARNING on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: Warning: 17% free memory [09:18:32] PROBLEM Current Load is now: WARNING on parsoid-roundtrip3.pmtpa.wmflabs 10.4.0.62 output: WARNING - load average: 6.62, 6.58, 5.59 [09:21:42] RECOVERY Free ram is now: OK on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: OK: 20% free memory [09:55:32] PROBLEM Free ram is now: WARNING on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: Warning: 19% free memory [09:58:32] RECOVERY Current Load is now: OK on parsoid-roundtrip3.pmtpa.wmflabs 10.4.0.62 output: OK - load average: 1.55, 2.67, 4.59 [10:37:44] PROBLEM Free ram is now: WARNING on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: Warning: 19% free memory [12:02:22] PROBLEM Total processes is now: WARNING on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS WARNING: 151 processes [12:07:25] RECOVERY Total processes is now: OK on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS OK: 147 processes [12:07:42] PROBLEM Disk Space is now: WARNING on wikidata-dev-9.pmtpa.wmflabs 10.4.1.41 output: DISK WARNING - free space: / 521 MB (5% inode=71%): [12:37:42] RECOVERY Free ram is now: OK on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: OK: 27% free memory [12:38:52] RECOVERY Free ram is now: OK on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: OK: 22% free memory [12:40:32] RECOVERY Free ram is now: OK on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: OK: 24% free memory [12:40:52] RECOVERY Free ram is now: OK on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: OK: 27% free memory [12:41:12] RECOVERY Free ram is now: OK on sube.pmtpa.wmflabs 10.4.0.245 output: OK: 22% free memory [12:50:42] PROBLEM Free ram is now: WARNING on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: Warning: 19% free memory [12:56:52] PROBLEM Free ram is now: WARNING on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: Warning: 17% free memory [12:58:52] PROBLEM Free ram is now: WARNING on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: Warning: 13% free memory [13:00:16] Change on 12mediawiki a page Wikimedia Labs/Tools Lab was modified, changed by Hashar link https://www.mediawiki.org/w/index.php?diff=649504 edit summary: [+7] pretty table: border="1" --> class="wikitable" [13:04:13] PROBLEM Free ram is now: WARNING on sube.pmtpa.wmflabs 10.4.0.245 output: Warning: 14% free memory [13:13:33] PROBLEM Free ram is now: WARNING on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: Warning: 19% free memory [13:17:27] Hi! At the moment, where does mail for (instance) users get sent to? I. e., output of user crontabs, at jobs, etc.? [13:30:22] PROBLEM Free ram is now: WARNING on bots-liwa.pmtpa.wmflabs 10.4.1.65 output: Warning: 18% free memory [13:40:37] scfc_de`: to the wikimedia root i think [13:42:15] scfc_de`: in a crontab you can use the MAILTO environment variable [13:45:22] RECOVERY Free ram is now: OK on bots-liwa.pmtpa.wmflabs 10.4.1.65 output: OK: 20% free memory [13:59:08] hashar: Thanks. Do you know a solution for at jobs? [14:00:20] scfc_de`: This needs a more permanent solution, clearly. [14:01:27] Coren: On Toolserver, mail is sent to the address set in LDAP (overwritable in ~/.forward with an ugly hack, "aliasd"). [14:02:48] scfc_de`: Yeah, because of the role accounts on the Tool Labs especially I'm looking at a number of possible alternatives. I'll certainly talk with CT and Geoff about the possibility of having local mailboxen and/or permit procmail. [14:03:22] PROBLEM Free ram is now: WARNING on bots-liwa.pmtpa.wmflabs 10.4.1.65 output: Warning: 19% free memory [14:04:00] scfc_de`: So that things like cron reports can be stuffed in a mbox instead of being emailed. [14:05:57] Coren: Well, whether I read it via "ssh $HOST mail" or it is delivered to my inbox doesn't matter to me, but I think the legal implications are the same :-). [14:07:21] scfc_de`: They may or may not be; I could see reasons why Geoff might want to control and/or add rules to the idea of store-and-forward; the question needs to be posed. [14:09:44] !log wikidata-dev wikidata-dev-9: Started to play with WikibaseSolr extension. Solr is in /opt, solarium inside the extension's directory, it's not working. [14:09:46] Logged the message, Master [14:13:37] Coren: Well, as I said: "I think" :-). [14:14:33] scfc_de`: Well, Geoff is paid the Big Bucks to think of such things. i'll just dump that on his desk and think about technology instead. :-) [14:16:53] Coren: If at the end some form of "mail" is implemented, I'm (well, somewhat) fine with that :-). [14:28:52] PROBLEM Current Load is now: CRITICAL on deployment-cache-upload-test.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [14:29:39] PROBLEM Disk Space is now: CRITICAL on deployment-cache-upload-test.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [14:30:02] scfc_de`: Hm, I don't expect that using command-line mail is a reasonable solution for anyone involved. :-) If there are local mailboxen, there would be some more better way to get at them. :-) [14:30:12] PROBLEM Free ram is now: CRITICAL on deployment-cache-upload-test.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [14:30:32] scfc_de`: I'm thinking procmail to make logs of automated email, the rest forwarded. :-) [14:31:42] PROBLEM Total processes is now: CRITICAL on deployment-cache-upload-test.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [14:32:22] PROBLEM dpkg-check is now: CRITICAL on deployment-cache-upload-test.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [14:34:55] Coren: What automated email do you think of? [14:35:48] scfc_de`: Cron, for one. Possibly things like (hey, your job just went BOOM!) also. [14:38:51] Coren: That's one of the things I definitely want mail for :-). I absolutely hate about SGE that the job output is saved on disk. The nicety of cron is that the output is delivered "to your doorstep", so you don't have to regularly check some log. [14:42:05] Coren: BTW, did the OGS build work? [14:43:12] scfc_de`: Some breakage caused by braindead dependencies. I'm currently porting some debian patches to gridengine. [14:48:13] do we have some documentation about puppetizing stuff? [14:48:17] Coren: gridengine = Ubuntu's gridengine-* package of SGE?! (Even though I have no clue how it is packaged ATM on Ubuntu or Debian.) [14:48:44] scfc_de`: Well, Debian's but yeah. [14:49:12] Platonides: https://labsconsole.wikimedia.org/wiki/Help:Self-hosted_puppetmaster [14:50:07] wtf, it creates a puppet class automatically? [14:50:14] how can it do so? [14:50:27] What do you mean? [14:51:17] (I haven't actually tried puppetmaster::self myself.) [14:51:21] oh, "You now have a local puppet repo" [14:51:37] I understood that as you now have a copy of your system in puppet [14:51:50] That would be nice :-). [14:52:13] that "Force another puppet run, and behold! " didn't imply that it did everything [14:52:51] (What also would be nice: A parameter to puppetmaster::self that is a link to a Git repository/branch so that the back and forth between instance and repo is easier.) [14:54:32] PROBLEM Free ram is now: CRITICAL on puppet-dev.pmtpa.wmflabs 10.4.1.85 output: Connection refused by host [14:54:42] PROBLEM dpkg-check is now: CRITICAL on puppet-dev.pmtpa.wmflabs 10.4.1.85 output: Connection refused by host [14:55:12] so I guess I should create modules/webtools, with subfolders and manifests and files [14:55:52] PROBLEM Current Load is now: CRITICAL on puppet-dev.pmtpa.wmflabs 10.4.1.85 output: Connection refused by host [14:56:02] PROBLEM Disk Space is now: CRITICAL on puppet-dev.pmtpa.wmflabs 10.4.1.85 output: Connection refused by host [14:56:02] PROBLEM Total processes is now: CRITICAL on puppet-dev.pmtpa.wmflabs 10.4.1.85 output: Connection refused by host [14:56:11] Yep, that was my idea. It's just 0 and 1s, so later we could move it to "tools" if that becomes necessary. [15:02:12] what's the difference between require and include in puppet? [15:05:06] Feb 20 14:38:25 deployment-cache-upload-test puppet-agent[891]: Did not receive certificate [15:05:12] any idea what "did not receive certificate" is ? [15:05:43] perhaps it is connectinbg through https? [15:05:48] I guess my lab is broken again [15:05:52] that is a fresh instance though [15:05:56] hashar I believe it does not receive the desired certificate :) [15:06:11] you need to show it more love [15:07:22] every time I create an instance something screw up :( [15:08:12] * hashar gives up [15:08:17] can't create any instance [15:08:18] pff [15:08:52] PROBLEM Current Load is now: CRITICAL on puppet1.pmtpa.wmflabs 10.4.0.251 output: Connection refused by host [15:09:22] hashar: Could that ^^^ be connected? :-) [15:09:28] someone deleted the 000-default symlink from /etc/apache2/sites-enabled :S [15:09:32] PROBLEM Disk Space is now: CRITICAL on puppet1.pmtpa.wmflabs 10.4.0.251 output: Connection refused by host [15:09:40] no wonder why it stopped working [15:10:12] PROBLEM Free ram is now: CRITICAL on puppet1.pmtpa.wmflabs 10.4.0.251 output: Connection refused by host [15:10:27] scfc_de`: maybe [15:10:57] !log bastion puppet1.pmtpa.wmflabs is dead. [15:10:58] Logged the message, Master [15:11:11] I give up for today [15:11:42] PROBLEM Total processes is now: CRITICAL on puppet1.pmtpa.wmflabs 10.4.0.251 output: Connection refused by host [15:12:22] PROBLEM dpkg-check is now: CRITICAL on puppet1.pmtpa.wmflabs 10.4.0.251 output: Connection refused by host [15:28:52] PROBLEM Current Load is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused by host [15:29:32] PROBLEM Disk Space is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused by host [15:30:12] PROBLEM Free ram is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused by host [15:31:42] PROBLEM Total processes is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused by host [15:32:22] PROBLEM dpkg-check is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused by host [15:35:36] hashar, maybe you want o take a look if my manifest looks right? [15:36:15] I could :-D [15:36:24] though it is probably better to have it reviewed directly by someone from ops hehe [15:37:11] I don't see any of them here :) [15:37:21] and you have *some* idea of puppet :P [15:40:05] does the name of the files where you define the classes matter? [15:40:39] Platonides: in modules yes :-] [15:40:53] /modules/manifests/.pp [15:40:53] or : [15:41:03] /modules/manifests//.pp [15:41:18] but can one file contain several classes ? [15:41:30] aka apache::website would be in /modules/apache/manifests/website.pp [15:41:38] grmblbl [15:41:50] my templates above are wrong, but that last example is probably fine [15:42:03] the "apache" class would be : /modules/apache/manifests/init.pp [15:42:25] in a module context, I have no idea whether you can have sub classes in the same file [15:42:32] I guess that might work if you only include it locally [15:42:58] I think I will make another file and play safe for now [15:43:17] Platonides: Do you have a public repo to look? [15:43:35] what do you mean? [15:45:40] So that not only hashar can view your manifests :-). [15:45:41] it's not hidden [15:45:41] https://gerrit.wikimedia.org/r/50011 [15:45:42] what is the point of asking for review when the manifest does not even lint? :-] [15:45:45] https://integration.mediawiki.org/ci/job/operations-puppet-validate/1729/console [15:46:51] why doesn't it validate? [15:47:02] ah, syntax error [15:47:17] but it's a file [15:47:23] why is it trying to lint it? [15:47:34] if you install puppet on your machine (something like "gem install puppet") , you should be able to lint locally [15:47:38] ah, no [15:48:10] it's the manifest what files [15:48:15] but I have no idea why [15:48:38] which only shows my ignorance [15:49:42] hmm [15:49:46] ok, a ; at the end of each file [15:50:32] yeah [15:50:33] hmm [15:50:41] Platonides: just put each file in its own {} statement [15:50:54] and there is really no point in enabling the default apache :] [15:51:09] I copied that from changeset 49270 [15:51:16] maybe [15:52:06] Platonides: Is this just testing Puppet? I don't think UserDir will work with this config. [15:52:07] and you probably don't want to use /data/project :D [15:52:18] i am off [15:52:24] * hashar waves [15:52:59] hashar: Au revoir! [15:53:04] bye hashar [15:54:05] ;;-) [15:56:09] scfc_de`, it's not exactly UserDir [15:56:20] but it works so far [15:56:34] in its simplified form [16:01:05] And the tools are executed as the user's account? [16:02:13] not yet [16:02:51] Ah, okay. [16:11:42] RECOVERY Total processes is now: OK on puppet1.pmtpa.wmflabs 10.4.0.251 output: PROCS OK: 84 processes [16:12:22] RECOVERY dpkg-check is now: OK on puppet1.pmtpa.wmflabs 10.4.0.251 output: All packages OK [16:13:52] RECOVERY Current Load is now: OK on puppet1.pmtpa.wmflabs 10.4.0.251 output: OK - load average: 0.23, 0.67, 0.52 [16:14:32] RECOVERY Disk Space is now: OK on puppet1.pmtpa.wmflabs 10.4.0.251 output: DISK OK [16:15:13] RECOVERY Free ram is now: OK on puppet1.pmtpa.wmflabs 10.4.0.251 output: OK: 81% free memory [16:22:52] PROBLEM Current Load is now: WARNING on parsoid-roundtrip6-8core.pmtpa.wmflabs 10.4.0.222 output: WARNING - load average: 7.08, 7.35, 5.66 [16:23:32] PROBLEM Current Load is now: WARNING on parsoid-roundtrip3.pmtpa.wmflabs 10.4.0.62 output: WARNING - load average: 7.66, 7.78, 6.22 [16:35:53] PROBLEM Current Load is now: WARNING on ve-roundtrip2.pmtpa.wmflabs 10.4.0.162 output: WARNING - load average: 9.56, 7.55, 6.00 [16:38:33] RECOVERY Free ram is now: OK on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: OK: 24% free memory [16:38:53] RECOVERY Free ram is now: OK on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: OK: 22% free memory [16:39:13] RECOVERY Free ram is now: OK on sube.pmtpa.wmflabs 10.4.0.245 output: OK: 22% free memory [16:40:42] RECOVERY Free ram is now: OK on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: OK: 27% free memory [16:41:52] RECOVERY Free ram is now: OK on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: OK: 22% free memory [16:48:43] PROBLEM Free ram is now: WARNING on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: Warning: 19% free memory [16:56:23] PROBLEM dpkg-check is now: CRITICAL on puppet1.pmtpa.wmflabs 10.4.0.251 output: DPKG CRITICAL dpkg reports broken packages [16:57:13] PROBLEM Free ram is now: WARNING on sube.pmtpa.wmflabs 10.4.0.245 output: Warning: 14% free memory [17:06:22] RECOVERY dpkg-check is now: OK on puppet1.pmtpa.wmflabs 10.4.0.251 output: All packages OK [17:06:53] PROBLEM Free ram is now: WARNING on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: Warning: 13% free memory [17:09:52] PROBLEM Free ram is now: WARNING on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: Warning: 17% free memory [17:11:33] PROBLEM Free ram is now: WARNING on nova-precise2.pmtpa.wmflabs 10.4.1.57 output: Warning: 19% free memory [17:29:42] heya, so, the page to sign up for developer access isn't working, it gives me "Incorrect or missing confirmation code. " (This page: https://labsconsole.wikimedia.org/wiki/Special:UserLogin/signup ) [17:35:56] greg-g, have you talked to Ryan Lane? [17:36:18] He doesn't seem to be on IRC at the moment :/ [17:36:35] Krenair: what's his nick? [17:36:46] "Confirmation code"? The one you should receive in your mail? Did you try several times? [17:37:15] I tried a few times, yes :) (thought it might have been captcha confirmation code related, or something of the sort) [17:37:39] Ah yes, there is a captcha too :/ [17:37:40] and I havne't received any email about it, just working through the new tech employee wiki page [17:38:04] you get to that signup page from: https://www.mediawiki.org/wiki/Developer_access [17:42:35] I guess you'll have to wait for someone knowledgeable about this. [17:47:35] Darkdadaah: no worries, no rush [17:47:53] RECOVERY Current Load is now: OK on parsoid-roundtrip6-8core.pmtpa.wmflabs 10.4.0.222 output: OK - load average: 3.53, 3.34, 4.67 [17:52:22] greg-g, still having login trouble? [17:55:51] andrewbogott: yeah, haven't tried anything new since I first said anything in here :) [17:56:39] greg-g, can you catch me up with what's happened so far? [17:57:39] andrewbogott: yeah, so, I'm just reading the new tech employee docs: https://office.wikimedia.org/wiki/New_tech_employee_orientation [17:57:51] andrewbogott: one of the steps is to request a Developer Account [17:58:23] RECOVERY Current Load is now: OK on parsoid-roundtrip3.pmtpa.wmflabs 10.4.0.62 output: OK - load average: 1.99, 2.89, 4.71 [17:59:29] andrewbogott: and this page: https://www.mediawiki.org/wiki/Developer_access is where I'm brought to the labconsole account creation page, and there, after filling out, gives me the "Incorrect or missing confirmation code" error [18:00:09] greg-g: OK, so, backing up… on that page, under 'Accounts', we're talking about the top two bullet points right? [18:00:50] andrewbogott: right, the first one really (since I don't think I'll need the ops stuff) [18:00:52] RECOVERY Current Load is now: OK on ve-roundtrip2.pmtpa.wmflabs 10.4.0.162 output: OK - load average: 1.65, 2.47, 4.78 [18:01:19] yep, ok. I'm clarifying because the process used to invole 'requesting' but now it's self-serve. In theory. [18:01:28] ah, gotcha, yay theories [18:01:30] OK, so you visited the labsconsole account creation page and filled out the form and... [18:01:47] and then the error message [18:02:14] Ah, ok. Any chance you just got an extra-difficult capcha? [18:02:37] well, I've tried about 3 times, I hope I'm really a human [18:02:53] lemme try again, for completeness's sake [18:03:21] [bz] (8UNCONFIRMED - created by: 2greg.chaki, priority: 4Unprioritized - 6normal) [Bug 45202] Personal information posted that needs to be removed - https://bugzilla.wikimedia.org/show_bug.cgi?id=45202 [18:04:08] andrewbogott: same :/ [18:04:37] ok, nothing interesting in /this/ log, lemme look in another [18:04:48] k [18:04:55] thanks, andrewbogott [18:05:53] greg-g, are you in the office perchance? [18:07:16] hm, confirmed in the source that that is indeed a capcha rejection. [18:07:41] (Which is not to say that you aren't human; likely something is going wrong with the validation) [18:07:59] Re: bz, I really wonder what "business disclosure agreements" might have been compromised by that post. Ts, ts, ts. [18:09:23] <^demon|lunch> We don't remove e-mails just because a name appears really. [18:09:32] <^demon|lunch> Plus, there's zero we can do about 3rd party archives. [18:10:29] Hah [18:10:36] Also, the guy's name is in his e-mail address [18:10:42] There is zero expectation of privacy there [18:11:21] Is his email address 'jowblow@Iusedtobeinthemafia.com'? [18:11:49] <^demon|lunch> He already e-mailed me as list admin, I just hadn't responded yet. [18:12:10] andrewbogott: I am, over by robla's desk [18:12:27] <^demon|lunch> andrewbogott: And no, his e-mail is fname.lname@gmail. [18:12:30] andrewbogott: firstname.lastname@foobar.com is upset that we revealed his name is Firstname Lastname [18:14:26] RoanKattouw: Nitpicking: The post doesn't even say this, and if, it wouldn't have been us, but him! :-) [18:15:39] But I try to get my head around why he is trying to purge this message 4.5 years after the fact ... [18:16:04] Yeah... [18:17:07] Perhaps trying to start a new business. Well, whatever. [18:19:32] <^demon|lunch> Marked WONTFIX. [18:19:53] <^demon|lunch> Basically, unless someone included their home address or a phone number, we don't do this. It's hugely disruptive. [18:20:59] andrewbogott: thanks for the help [18:21:46] [bz] (8NEW - created by: 2Andrew Bogott, priority: 4Unprioritized - 6normal) [Bug 45203] Dumb error message on account creation shell name collision - https://bugzilla.wikimedia.org/show_bug.cgi?id=45203 [18:21:59] Argh! [18:22:04] Man that yellow hurts my eyes [18:23:55] <^demon|lunch> Yellow? [18:34:20] ^demon: I assume RoanKattouw means the "NEW" in the bug announce, does your irc client show colors? [18:34:43] <^demon> Nope, and I forget that sometimes :p [18:50:26] Coren: why a database to back the info? [18:50:48] what kind of stuff will go into it? [18:52:01] That's my usual MO. Have management tools used to edit a db (often just a flatfile), and use that db as source to generate config from. The idea is idempotence; if an operation fails just doing the 'regenerate stuff' operation restores to a known state (as opposed to, say, account creation failing to create /some/ things and needing partial fixes) [18:52:58] In this case, the db should just contain: "tool id" "list of maintainers" "other-config-stuff-I-dunno-yet", and be the authoritative source to generate users, groups, wiki pages, directories, etc. [18:53:20] generate users? [18:53:30] we'll manage the users and groups in ldap [18:53:47] the wiki pages themselves will be generated when the users are created [18:53:53] I know, but there would be a script that takes the list of tools, makes a list of users, and syncs up LDAP [18:54:10] RoanKattouw_away so change that [18:54:30] what kind of programmers are you guys, if you can't change the config of a simple bot [18:54:46] This way, the management tools don't need to know how the users are setup on the backend [18:54:57] Makes things MUCH less brittle. [18:55:38] New tool -> just add it to the db. The rest is done backend. [18:56:09] syncs up ldap? [18:56:47] Yep, and would create the directory in /data/projects if it doesn't exist (possibly copying a skeleton), and any other things we add to this in time for the "new project" task. [18:57:12] Like: stuff.d/00sync-users stuff.d/10create-project-dir and so on [18:57:41] Each of those scripts being written to be idempotent. [18:58:13] This is why I like puppet. I'm a big fan of state definitions as opposed to "list of stuff to do" :-) [18:58:50] s/"new project"/"new tool"/ [18:59:23] I'm not sure why we'd write into a database, then into ldap [18:59:24] If one of the steps break, fix the breaking and just rerun the "sync everything up" tool. Partial failures are thus gracefully handled. [18:59:47] all of the necessary info will be in ldap and the wiki immediately [18:59:54] I don't see the need for a separate database at all [18:59:59] it's just another thing to fail [19:00:22] the scripts to manage the /data/project directories can pull from ldap [19:00:23] Hm. How flexible is out LDAP schema then? It could very well be "the" database [19:00:35] what information is really needed? [19:00:41] a tool is a user/group [19:00:57] the sudo policy can be injected into ldap immediately when the user/group is created [19:01:18] the tool information can be created immediately when the user/group is created too [19:01:21] At this time? I don't know. A tool is a user/group, but it might also be needed resources, or requirements, or even things like "only instance x" [19:01:23] as a wiki page [19:01:42] Sure, but you don't want to use the wiki page as a source for configuration. [19:02:39] To me, a DB for the tool info is necessary for proper management. Can LDAP be used for this? Sure. It might even be the best possible place depending on a few things. [19:02:50] I don't see why it is [19:03:00] if we're already using puppet, we should use it for this [19:03:18] those requirements you're mentioning can be puppet vars or classes [19:03:47] Sure, but where do you pull the puppet config /from/? Certainly, you don't intend to have a tool accessible from end-users being able to directly write to puppet config? [19:04:02] why not? [19:04:18] we already do ;) [19:04:28] the puppet config is stored in ldap [19:04:31] ... o_O [19:05:16] hm. maybe we should write an ENC [19:05:30] Wait, wait. I mean, from the labsconsole, project admin clearly can fiddle with the puppet config -- but that's not endusers. [19:05:38] end users can doo [19:05:39] *too [19:05:54] on the manage instances page, there's a "configure" action [19:06:10] also, managing what's available can be done per-project by project admins [19:06:22] PROBLEM Total processes is now: WARNING on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS WARNING: 155 processes [19:06:23] hm. this wouldn't really be doable per-tool, though [19:06:42] and the puppet config is per instance, not global [19:06:59] of course, there's nothing stopping us from changing how that works with an ENC [19:06:59] Right, and I would have expected we'd be a bit more... miserly with root on instances. [19:07:17] we can control how access works to the puppet config [19:07:45] Allright, I'm open to alternatives; there's much I don't know about the internal SOPs and MOs yet. :-) [19:07:54] for instance, we could let a tool author manage config per tool and only allow them to modify prefix'd variables [19:08:15] let's keep things simple right now [19:08:26] But I do have two objectives: (a) rock solid configuration, (b) proper compartmentalization. :-) [19:08:34] and assume that tools don't need to be configured [19:08:43] or if they do, we can handle it inside of the project [19:09:01] for instance, we could have flat files in /data/project that are config files [19:09:18] then the file/directory permissions would serve well enough [19:09:25] and we wouldn't need a custom interface [19:09:52] creation of tools can honestly be self-service [19:09:58] there's no reason it can't [19:10:03] Oh, I agree. [19:10:09] we can have an interface for that [19:10:13] That's what I am gunning for. [19:10:20] that will create the ldap user/group [19:10:30] and will also create a documentation page [19:10:34] on the wiki [19:10:49] But you'd have that tool create the ldap stuff and the documentation directly? [19:10:54] <^demon> Ryan_Lane: https://gerrit.wikimedia.org/r/#/c/49987/, https://gerrit.wikimedia.org/r/#/c/50029/ and https://gerrit.wikimedia.org/r/#/c/50016/ are all ready to go in. [19:10:58] a script can run that will create the directory in /data/project by pulling the info from ldap [19:11:05] Coren: yes [19:11:18] the script can also put in a basic config file for the tool [19:11:25] See? You have /almost/ the same approach than I do, except you put different things in different places. [19:11:26] it can be yaml [19:11:34] Whereas I say "have exactly one mechanism" [19:11:37] yeah, you're approach isn't bad :) [19:11:54] but it's doing it differently from how we're already doing things [19:11:57] The only difference is that I'd have that tool ONLY put the stuff in LDAP, which would be authoritative [19:12:13] And the doc and directory would be created /from/ the info in the LDAP instead. [19:12:47] that's fine as well [19:12:55] the documentation can be updated from the script [19:13:09] it can also document how the tool is configured by reading the config file [19:13:19] hello Ryan_Lane :-] [19:13:22] I'm a fan of YAML for configuration that needs to be editable [19:13:30] and of python for the script ;) [19:13:45] and it's also nice if the script is run as a daemon with an init script, rather than via cron [19:14:07] hashar: howdy [19:14:08] I have discovered YAML with OpenStack python tools. That is a nice file format. [19:14:16] <3 yaml [19:14:18] let you do more than .ini [19:14:32] and is way easier to edit than json (*cough* trailing comma, no comments) [19:14:56] so anyway, I had a puppet issue this afternoon on a fresh labs instance [19:15:03] some certificate is not received : http://pastebin.com/3iX8PvW5 [19:15:15] instance is deployment-cache-upload-test2 i-000005e9 [19:15:22] https://labsconsole.wikimedia.org/w/index.php?title=Special:NovaInstance&action=consoleoutput&project=deployment-prep&instanceid=ea0947ba-3b5b-44ad-b3c8-357fddcbaa78®ion=pmtpa [19:15:43] I wanted to reboot it but then though that you might be interested in investigating that issue :-] [19:16:00] hashar: that's a newly created instance? [19:16:03] yup [19:16:13] did it have the same name as another instance? [19:16:15] though I might have rebooted it once [19:16:22] I don't think [19:16:24] ooohhhh [19:16:26] I know the issue [19:16:38] anyone against me submitting my last patchset of E:RSS Version 2.18 ?? https://gerrit.wikimedia.org/r/#/c/3925/ Code Review was done for previous patchsets [19:16:57] Wikinaut: #mediawiki / #wikimedia-dev :-] [19:17:01] by CSteipp & Ori.Livneh [19:17:24] crossposted FROM #mediawiki [19:17:54] hashar: i dont know the EXACT policy for Extension authors and maintainer (like me) [19:18:02] Let me know this... [19:18:04] hashar: fixed [19:18:19] puppet isn't running properly on virt0 [19:18:27] and it didn't update a changed password [19:18:33] ohh [19:18:42] bye [19:19:02] also I was wondering if we could have a puppet run to be forced on instance startup [19:19:27] (ideally a "Run puppet" button in labsconsole would be ideal hehe) [19:19:49] I sometime find myself breaking an instance because of LDAP configuration being overridden / some DNS conf mistake and such [19:20:17] we have plans on adding a "run puppet" button when we have salt-api up and running [19:20:19] so I end up waiting for puppet to run again [19:20:27] great :-] [19:20:48] are you adding the node entry to site.pp on puppetmaster::self instances? [19:20:53] that'll break ldap [19:20:57] if done improperly [19:21:07] Ryan_Lane: pls. (if you have time), don't forget to setup "my" openid-wiki2 [19:21:18] :-) [19:21:20] Wikinaut: it'll be way quicker if you do it [19:21:28] uh [19:21:28] Ryan_Lane: hop I never do that. I add the classes I want in labsconsole then check them in the instance configuration [19:21:32] @search mediawiki [19:21:32] Results (Found 13): morebots, labs-home-wm, labs-nagios-wm, labs-morebots, gerrit-wm, extension, revision, info, bots, labs-project, openstack-manager, wl, deployment-prep, [19:21:32] how? [19:21:47] I can never find this damn documentation [19:21:58] try google [19:21:58] from the bot anyway [19:22:15] Wikinaut: https://labsconsole.wikimedia.org/wiki/Help:Single_Node_MediaWiki [19:23:02] Ryan_Lane: ty. Hopefully, I can do this w/o further assistance... [19:23:09] let me know if you run into any issues [19:23:27] and you answer ??? [19:23:29] ;-) [19:24:36] yes [19:24:47] Coren: so, let's make a project plan for this [19:24:55] Coren: so that I know we're on the same page for design [19:25:12] RECOVERY Free ram is now: OK on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: OK: 92% free memory [19:25:25] ^^^^^^ good nagios ! [19:26:42] RECOVERY Total processes is now: OK on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: PROCS OK: 90 processes [19:27:03] hashar: Question: given, that a patchset x is codreviewed, and I add further ps x+1, x+2 ... which do not contain substantial changes... [19:27:04] I think, I can submit. Can you confirm? [19:27:22] RECOVERY dpkg-check is now: OK on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: All packages OK [19:27:44] Wikinaut: depends ojn the not substantial changes hehe :-] [19:27:57] hehehe :-) [19:28:00] ty [19:28:13] csteipp answered me the same on the other channel [19:28:17] basically the same [19:28:25] Wikinaut: as I understand it your RSS patch has been reviewed by Ori + Chris [19:28:34] Wikinaut: so that is most probably fine. Make sure to ping sam about it [19:28:36] Yes [19:28:42] Wikinaut: (if that is deployed on wikimedia cluster) [19:28:46] Sam - who? [19:28:52] RECOVERY Current Load is now: OK on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: OK - load average: 0.09, 0.46, 0.37 [19:28:53] Reedy <-- that Sam [19:29:06] Wikinaut: if anything screw up, at worse we will revert the patch in production [19:29:31] Wikinaut: or send some patches to fix it up if it is not too badly broken :-] [19:29:32] RECOVERY Disk Space is now: OK on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: DISK OK [19:29:57] hashar: I tested it over and over, I have this in production, so I am convinced it will not break (unless I and others have overlooked something...) [19:30:08] but one can never be sure [19:30:16] I will ping Reedy [19:31:27] Ryan_Lane: ok that fixed the issue. Thank you! [19:32:56] hashar: great [19:34:27] hashar: FYI, I added Reedy to the reviewers, removed the "previous" reviewers and wrote him a nice comment. [19:34:39] So let's see, what he is answering. [19:34:43] Wikinaut: that would be fine :-] [19:35:07] hashar: I am happy to have learned finally git. I use it for ALL my private projects, locally, and on all machines [19:35:26] nice is the portable windows version of git [19:35:30] for special cases.... [19:35:34] Wikinaut: I should have taught that during the Berlin hackaton. I was a bit sick though :/ [19:35:37] pls. dont comment on windows [19:35:52] I don't care of your OS / tools. as long as they are used to produce open source software! [19:35:53] hashar: de rien [19:36:01] * hashar uses a Mac [19:36:02] yep! [19:36:26] uses raspberry, Windows, Linux SLES, openSuse, Debian [19:36:31] but nobody cares [19:48:21] Ryan_Lane: That, design docs are my first order of business [19:48:49] Ryan_Lane: But, between you and me, I have much hatred for yaml. I'll cope. :-) [19:48:54] hahaha [19:48:57] what do you prefer? [19:49:10] For human editable? [19:49:10] it's easy from a user editing perspective [19:49:22] json is terrible for it [19:49:27] ini files are annoying [19:49:31] (and limiting) [19:49:37] json is teh suxx0rs for humans. [19:49:53] yes [19:50:32] I like perl-as-config, since I tend to do perl for scripting. Since you guys play with python, I'll suffer it and deal with yaml. :-) [19:51:58] I never experimented with using python for python config, though. [19:52:11] awwwwwwww perl [19:52:14] err [19:52:15] ewwwww [19:52:26] python for python config works [19:52:31] but yaml is better [19:52:40] Perl, done right, is most excellent. The problem with perl is that it's soooo easy to do it /wrong/. [19:52:52] it's good to use something non-language specific for config [19:52:57] yep [19:54:00] The advantage of using code as config is ... well, you can use code as config (computed values, stuff nabbed from the environment, or from an external datasource) without having to make the tool itself need to know about all of the edge cases. [19:55:00] So your do_stuff.pl doesn't need to be messed with if you've got that one use case where the config value should come from a mysql or something. [19:56:07] But for the simple case, "$foo = [ 'bar', 'baz' ]" is no worse than "foo:\n - bar\n - baz" [19:56:47] Meh. Dif'rent strokes and all that. [19:57:28] Ryan_Lane: I got a few input/output errors on my cvn vm [19:57:28] * Coren has an intense dislike of the concept of significant whitespace, which influences his opinion there. [19:57:36] rm: cannot remove `.bash_history': Input/output error [19:57:36] rm: cannot remove `.bashrc': Input/output error [19:57:44] $ source .bashrc [19:57:44] -bash: .bashrc: Input/output error [19:57:50] ls [19:57:50] ?????????? ? ? ? ? ? .bash_history [19:58:05] Sounds like ldap shiz [19:58:12] You created these files when I was in the office 2 weeks ago, remember? It was right after the gluster issue [19:58:46] It's on cvn-apache2 in /home/kirnkle [19:58:51] /home/krinkle * [20:00:43] Damianz: likely a split brain [20:00:50] Ryan_Lane: ? [20:00:54] gluster [20:01:19] Krinkle: I'll fix it for you [20:01:21] ohh I see - sorry that comment was in the wrong window... trying to fix exchange D: [20:01:22] gimme a little bit [20:01:56] Ryan_Lane: Don't worry about recovering the files (in case you are), I can re-create them in a snap if you can get them to go away. [20:02:13] Krinkle: if that's the case, then I think rm'ing them will work [20:02:39] that's what I tried at first [20:02:41] ^^ [20:02:42] ah [20:02:45] ok. one sec [20:03:05] neither open, edit or rm worked [20:08:06] Hm, small question : are we supposed to update Ubuntu ourself in the instances ? (I don't mean upgrade) [20:09:08] Krinkle: ok. it should be fixed [20:09:17] Thx [20:16:24] RECOVERY Total processes is now: OK on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS OK: 149 processes [20:27:00] Ryan_Lane: I think, you meant https://labsconsole.wikimedia.org/wiki/Help:Instances (not: Help:Single_node ) can you confirm ? [20:27:24] I followed that instructions to https://labsconsole.wikimedia.org/wiki/Special:NovaInstance .. [20:27:33] .. but I am not allowed to add a new one [20:27:58] I only see three etherpad [20:27:59] bastion [20:28:02] openid and a submit button [20:28:09] lack of rights ? [20:28:53] Ryan_Lane: change request: pls. set permissions to me that I can add a new instance (if this correct) [20:32:34] RECOVERY Free ram is now: OK on aggregator2.pmtpa.wmflabs 10.4.0.193 output: OK: 92% free memory [20:37:13] RECOVERY Free ram is now: OK on sube.pmtpa.wmflabs 10.4.0.245 output: OK: 22% free memory [20:38:43] RECOVERY Free ram is now: OK on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: OK: 27% free memory [20:39:53] RECOVERY Free ram is now: OK on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: OK: 21% free memory [20:41:53] RECOVERY Free ram is now: OK on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: OK: 27% free memory [20:50:12] PROBLEM Free ram is now: WARNING on sube.pmtpa.wmflabs 10.4.0.245 output: Warning: 14% free memory [20:55:22] PROBLEM Total processes is now: WARNING on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS WARNING: 153 processes [20:59:40] Wikinaut: Isn't there a listing beneath openid mentioning pmtpa? [21:04:56] PROBLEM Free ram is now: WARNING on conventionextension-trial.pmtpa.wmflabs 10.4.0.165 output: Warning: 13% free memory [21:05:42] [bz] (8NEW - created by: 2Marc A. Pelletier, priority: 4Highest - 6enhancement) [Bug 45119] Add per-project service/role user accounts and groups - https://bugzilla.wikimedia.org/show_bug.cgi?id=45119 [21:07:03] scfc_de`: what do you mean exactly ? [21:11:42] PROBLEM Free ram is now: WARNING on mediawiki-bugfix-kozuch.pmtpa.wmflabs 10.4.0.26 output: Warning: 19% free memory [21:13:01] Wikinaut: you have permissions.... [21:13:38] you should see: pmtpa [Toggle, https://labsconsole.wikimedia.org/w/index.php?title=Special:NovaInstance&action=create&project=openid®ion=pmtpa] [21:13:53] underneath openid [21:13:57] in manage instances [21:22:52] PROBLEM Free ram is now: WARNING on integration-jobbuilder.pmtpa.wmflabs 10.4.0.21 output: Warning: 17% free memory [21:25:22] RECOVERY Total processes is now: OK on bots-bnr1.pmtpa.wmflabs 10.4.1.68 output: PROCS OK: 148 processes [21:25:35] Ryan_Lane: just as info, I do _not_ see it underneath OpenID in manage instances which is https://labsconsole.wikimedia.org/wiki/Special:NovaInstance [21:25:44] but I use your link [21:26:50] and will use the default settings [21:26:56] Wikinaut: you need to modify your project filter [21:27:00] to add the openid project in [21:27:06] ah, moment [21:27:38] ok! I now do understand. [21:27:50] it was not apparent, that the "add" link is shown then [21:28:00] well, let me now add openid-wiki2 [21:28:04] ... [21:31:25] https://commons.wikimedia.org/wiki/File:Pouso_da_Apollo_11_na_Lua.ogg [21:34:32] PROBLEM Disk Space is now: CRITICAL on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [21:35:13] PROBLEM Free ram is now: CRITICAL on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [21:35:53] PROBLEM Current Load is now: CRITICAL on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [21:36:43] PROBLEM Total processes is now: CRITICAL on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [21:37:23] PROBLEM dpkg-check is now: CRITICAL on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: Connection refused by host [21:37:38] Ryan_Lane: my PuTTY says Server sent disconnect message .Type 2 Auth failure. Too many authentication failures for user Wikinaut" Have you seen this before ? Do I have to set up my key somewhere ? [21:37:50] on the new instance ? [21:37:51] Wikinaut: it's probably not finished building [21:37:55] ok [21:37:56] active is misleading [21:38:10] the virtual machine is active, but until puppet finishes running it isn't done [21:38:15] (I thought this too) [21:38:47] (Wikinaut to Wikinaut: Try to be patient. Try harder.) [21:39:24] Ryan_Lane: So I could also set up an instance for testing the bleeding edge E:RSS ? [21:40:02] why not use the current instances for it? [21:40:12] RECOVERY Free ram is now: OK on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: OK: 90% free memory [21:40:16] they are development instances, that's kind of the idea :) [21:40:52] RECOVERY Current Load is now: OK on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: OK - load average: 0.81, 0.97, 0.55 [21:40:53] Last week you told me, not to set up Etherpad-Lite [21:41:05] yes, because it's an openid project [21:41:08] but I also wanted to test my E:EtherpadLite [21:41:11] oh [21:41:13] you said rss [21:41:19] I read that incorrectly [21:41:36] and Etherpad -- separately., or together. As you like [21:41:42] RECOVERY Total processes is now: OK on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: PROCS OK: 84 processes [21:41:52] From my view ONE instance is enough [21:41:56] Wikinaut: you should be able to log into that instance now [21:41:57] for: [21:42:12] E:RSS + E:EtherpadLite + E:OpenID [21:42:22] RECOVERY dpkg-check is now: OK on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: All packages OK [21:44:32] RECOVERY Disk Space is now: OK on openid-wiki2.pmtpa.wmflabs 10.4.1.80 output: DISK OK [21:54:14] :-( Authenticating with public key -- Putty closes. [21:54:39] key works on the other instance (openid-wiki) [21:55:35] Wikinaut: Look up; is there a "can't create /home/you" message /above/ the motd? [21:59:51] Wikinaut: And what does the console log under https://labsconsole.wikimedia.org/wiki/Special:NovaInstance say? Has the setup of the instance including keys finished? [22:24:43] doomed again :-] [22:25:11] scfc_de`: it looks so [22:25:37] could use a fix to i-000005e9 instance. It has invalid devices in /etc/fstab ( sda3 / sdb3 ) :( [22:25:41] I'll try a reboot [22:25:49] oops I stop [22:26:03] PROBLEM SSH is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused [22:26:20] hashar: your message ^ was not related to my problem. Or ? [22:26:23] PROBLEM dpkg-check is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused by host [22:26:35] Wikinaut: unrelated [22:26:38] ok [22:26:44] I#ll try a reboot [22:27:00] basically need someone with console to fix up my instance :( [22:27:00] PROBLEM Current Load is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused by host [22:27:02] or I can be lazy and create a new one [22:27:33] PROBLEM Disk Space is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused by host [22:27:35] bah can't even create an instance [22:28:11] hashar, do you happen to have PHPUnit/Framework.php in your phpunit install? [22:28:23] I suspect this file was deleted in newer versions [22:28:23] PROBLEM Free ram is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: CHECK_NRPE: Socket timeout after 10 seconds. [22:28:28] Platonides: maybe :-] [22:29:04] Platonides: that is a directory with v3.7.10 [22:29:43] PROBLEM Total processes is now: CRITICAL on deployment-cache-upload-test2.pmtpa.wmflabs 10.4.1.55 output: Connection refused or timed out [22:30:12] PROBLEM host: deployment-cache-upload-test2.pmtpa.wmflabs is DOWN address: 10.4.1.55 CRITICAL - Host Unreachable (10.4.1.55) [22:30:22] look at the top of tests/selenium/installer/MediaWikiInstallerTestSuite.php: [22:30:26] require_once 'PHPUnit/Framework.php'; [22:30:26] require_once 'PHPUnit/Framework/TestSuite.php'; [22:30:40] cannot login to my new instance :-( [22:30:40] in a conf call can't really look at that [22:30:46] I suspect we should remove the first one [22:30:47] AFAIK nobody uses the selenium files [22:31:09] so we might want to remove them all, I can't remember though why they are still there. [22:31:45] the file was deleted in 2010 [22:35:30] hashar: for my new isntance, do I have to add something ? my key ? or is this done automatically ? [22:35:56] Wikinaut: should be automatically done by puppet [22:36:03] though sometime that does not work :-] [22:36:22] you can look at the instance console to find out what it is doin [22:36:23] g [22:41:43] Wikinaut: ah. I know the issue [22:42:01] I stopped the script that allows filesystem access [22:42:21] I do know the issue. Wiat [22:42:23] wait [22:42:26] wait [22:42:28] wait [22:43:50] no. It is YOUR turn,. Can send you a screenshot, but do not want to upload to imgur [22:44:02] eh? [22:44:02] Creating directory '/home/wikinaut'. [22:44:04] Unable to create and initialize directory '/home/wikinaut'. [22:44:07] right [22:44:07] Wikinaut: send it to commons ? :D [22:44:18] no need for anyone to send any screenshots [22:44:20] Ryan_Lane: I CAN login to bastion [22:44:31] as I just said, I know what the issue is [22:44:34] When I then login to instance, this works but then the error comes [22:44:34] and the cause [22:44:35] yes [22:44:37] yes [22:44:39] yes [22:44:40] and I'm fixing it [22:44:51] "yes, please" [22:45:48] hashar: Ryan_Lane: the other reason (why my PuTTY) closes down is on my side: the very first connect is prompting for the server key fingerprint (you know this) [22:46:03] and one has to confirm the fingerprint once manually [22:46:10] Ryan_Lane: whenever you can, I could use some cleanup of /etc/fstab on deployment-cache-upload-test i-000005e9 . Some manifest added /dev/sda3 /dev/sdb3 in it :-] [22:46:15] (I need to document this. my job) [22:50:33] [16:55:35] Wikinaut: Look up; is there a "can't create /home/you" message /above/ the motd? <-- told ya. :-) [22:50:50] Wikinaut: it's working now [22:50:53] What is motd [22:51:06] Coren: I do not know anything... [22:51:07] I just reenabled the script that managed access [22:51:15] of the day [22:51:18] m ? [22:51:20] Message of the Day. The text blurb that reads, right now, "Problems with access..." [22:51:21] motto ? [22:51:23] ah [22:51:26] Now I know [22:51:31] okoko [22:51:54] "The Eagle has landed" [22:52:03] thanks [22:52:23] Ryan_Lane: Will you fix this home-directory creation problem, or shall I file a bug ? [22:52:30] I mean, in general [22:52:33] dude. I just said it's fixed [22:52:38] when one creates an instance [22:53:05] the script manages this [22:53:07] so your answer is: "Do not file a bug." [22:53:07] it's enabled [22:53:09] no more problem [22:53:10] yes [22:53:14] bravo [22:54:12] (going back to my new server...) [23:01:03] PROBLEM host: deployment-cache-upload-test2.pmtpa.wmflabs is DOWN address: 10.4.1.55 CRITICAL - Host Unreachable (10.4.1.55) [23:03:21] Ryan_Lane: could you possibly use the console to fix up the /etc/fstab on deployment-cache-upload-test2.pmtpa.wmflabs ? :D [23:03:32] please -;] [23:04:05] is the next step for me to select some meaningful stuff from https://labsconsole.wikimedia.org/w/index.php?title=Special:NovaInstance&action=configure&instanceid=a92fd355-ae8d-4a7e-be33-7dfd29e02580&project=openid®ion=pmtpa [23:04:07] ? [23:04:32] Wikinaut: follow the instructions here: https://labsconsole.wikimedia.org/wiki/Help:Single_Node_MediaWiki [23:04:36] ty [23:06:28] tl,dr [23:06:31] tl;dr [23:06:53] thisi seems ok https://labsconsole.wikimedia.org/w/index.php?title=Special:NovaInstance&action=configure&instanceid=a92fd355-ae8d-4a7e-be33-7dfd29e02580&project=openid®ion=pmtpa [23:11:30] It is not clear to me what exactly the FQDN of my instance is (to be set up as hostname see point 4 in https://labsconsole.wikimedia.org/wiki/Help:Single_Node_MediaWiki#How_to_create_a_single-node_MediaWiki_server [23:12:03] s/FQDN/fully qualified hostname/ [23:12:24] wikinaut.pmtpa.wmflabs ? :-] [23:12:46] could be accessed via http://wikinaut.instance-proxy.wmflabs.org/ (granted you have something listening on port 80) [23:12:51] hashar: no [23:13:00] that's not the instance name [23:13:13] oh [23:13:19] it's openid-wiki2.pmtpa.wmflabs [23:13:37] Wikinaut: you should either leave it blank, or use openid-wiki2.instance-proxy.wmflabs.org [23:13:42] depends on how you want to access it [23:13:58] I honestly don't think it matters, though [23:17:14] ok. The "sudo puppetd -tv" can be (please) put beneath the "submit" button on https://labsconsole.wikimedia.org/w/index.php?title=Special:NovaInstance&action=configure&instanceid=a92fd355-ae8d-4a7e-be33-7dfd29e02580&project=openid®ion=pmtpa [23:17:56] or better not, because it's dangerous [23:18:04] I withdraw my ad-hoc proposal [23:25:01] Ryan_Lane: would you be able to fix the /etc/fstab or should I ask andrew tomorrow ? ;) [23:31:22] PROBLEM host: deployment-cache-upload-test2.pmtpa.wmflabs is DOWN address: 10.4.1.55 CRITICAL - Host Unreachable (10.4.1.55) [23:33:40] [bz] (8NEW - created by: 2T. Gries, priority: 4Unprioritized - 6normal) [Bug 45214] Suggestion: when installing instances, starting puppet runs etc.: ping the developer by mail about the status - https://bugzilla.wikimedia.org/show_bug.cgi?id=45214 [23:33:58] :-) [23:34:35] otice: Finished catalog run in 845.85 seconds [23:34:37] wikinaut@openid-wiki2:/srv$ ls -l [23:34:39] total 4 [23:34:40] drwxr-xr-x 16 root root 4096 Feb 20 23:30 mediawiki [23:34:41] wikinaut@openid-wiki2:/srv$ [23:34:43] :-) :-) :-) [23:38:39] SOS: /srv/mediawiki is there, but Error404 when accessing http://openid-wiki2.instance-proxy.wmflabs.org/wiki [23:38:46] what I'm doing wrong ? [23:39:02] perhaps apache restart [23:39:03] mom [23:40:38] no, that was it not... [23:41:00] Ryan_Lane: master, something is missing [23:41:12] RECOVERY Total processes is now: OK on parsoid-roundtrip4-8core.pmtpa.wmflabs 10.4.0.39 output: PROCS OK: 144 processes [23:42:33] missing symlinks ??? 10.4.1.15] File does not exist: /var/www/wiki [23:43:25] Ryan_Lane: i have read something about. I think, the puppet class should /but did not/ set up the symlinks [23:43:42] bug or feature? [23:43:56] symlinks to /srv/mediawiki [23:44:03] filing a bug.... [23:46:00] [bz] (8NEW - created by: 2T. Gries, priority: 4Unprioritized - 6normal) [Bug 45215] role::mediawiki-install::labs lacks setting up of symlinks from /var/www/wiki to /srv/mediawiki - https://bugzilla.wikimedia.org/show_bug.cgi?id=45215 [23:48:13] Here we have also a problem that in the CHat the links in brackets [Bug xxxxx] are incorrectly linked ! To Mozilla !! [23:48:42] (do I get an award for finding three bugs in two minutes ?) [23:54:12] PROBLEM Total processes is now: WARNING on parsoid-roundtrip4-8core.pmtpa.wmflabs 10.4.0.39 output: PROCS WARNING: 152 processes [23:58:35] it shouldn't need a symlink [23:58:49] it uses aliases [23:58:54] Wikinaut: re-run puppet [23:59:05] fixed [23:59:08] I was impatient [23:59:35] and: I had to change the $wgServer to this .. [23:59:38] ## The protocol and server name to use in fully-qualified URLs [23:59:40] ## $wgServer = "http://openid-wiki2.pmtpa.wmflabs"; [23:59:41] $wgServer = "http://openid-wiki2.instance-proxy.wmflabs.org"; [23:59:49] to get the css files