[00:01:28] and as for security rules? [00:01:54] security groups are firewall rules [00:02:09] I knew that [00:02:10] basics of labs is that you can only use it for wikimedia related purposes [00:06:28] So I make a rule with 80 as the port, and incoming connections will be accepted on it? [00:07:10] (I'm sorry, I'm new to OpenStack and the like.) [00:07:15] * Ryan_Lane nods [00:07:19] CIDR should be 0.0.0.0/0 [00:07:25] protocol should be tcp [00:07:30] from 80 to 80 [00:07:45] Thanks [00:07:48] yw [00:09:13] Ryan_Lane, where does nova-manage look for my admin password? [00:09:28] nova-manage? [00:09:33] umm [00:09:43] I thought nova-manage doesn't take credentials at all [00:09:57] I am trying to create a project. It's failing, and the logfile says 'INVALID_CREDENTIALS' [00:10:05] ah [00:10:10] #nova-manage project create --project demo --user andrew [00:10:29] if it is talking to ldap, then it's the LDAP configuration in nova.conf that needs to be fixed [00:13:14] ldap_password in nova.conf looks right to me. But, could be my ldap is still screwed up. [00:13:31] can you do an ldap search against it manually using the credentials? [00:15:37] PROBLEM host: simplewikt is DOWN address: simplewikt PING CRITICAL - Packet loss = 100% [00:15:37] RECOVERY dpkg-check is now: OK on diablo-n-gluster diablo-n-gluster output: All packages OK [00:20:08] hm... indeed, ldap is broken. "SASL(-4): no mechanism available" [00:20:30] use -x [00:20:36] which is simple, rather than SASL [00:21:04] Ugh! [00:21:16] Jews: ? [00:21:26] Not talking to you [00:21:31] Ah, ok, that gets me a normal 'Invalid credentials" [00:21:53] Jews: why not? did I do something to offend you? [00:21:58] No. [00:21:59] * Ryan_Lane likes trolling [00:22:08] lol :) [00:23:14] instance is timing out [00:23:35] stuck on pending... lol [00:23:44] sometimes it takes a bit [00:23:55] unless the system ran out of memory [00:24:19] k [00:24:52] hmm. I need to more move instances to virt0 [00:24:59] err. virt1 [00:25:41] well, virt2 is reporting only about 1GB of memory left [00:26:08] which instance is showing as pending/ [00:26:20] oh [00:26:24] you mean this page? [00:26:29] !instance I-00000148 [00:26:29] https://labsconsole.wikimedia.org/wiki/Help:Instances [00:26:33] yes [00:26:33] bah [00:26:43] that's because that attribute doesn't properly get updated [00:26:54] @instance [00:27:01] @search instance [00:27:01] Results (found 4): instancelist, instance-json, access, instance, [00:27:12] !resource I-00000148 [00:27:18] * Ryan_Lane rolls his eyes [00:27:26] the instance list shows running [00:28:06] Timed out for me some time ago\ [00:28:27] still timing out [00:28:38] did you reboot it? [00:28:51] i'll try [00:28:56] no. don't [00:28:58] I was asking [00:29:08] already did [00:29:14] It might help [00:29:41] yep. :O [00:29:59] well, I still can't ping it [00:30:07] did you remove rules from the default security group? [00:30:10] I just deleted it [00:30:16] I didn't, I added rules [00:30:25] why delete it? [00:31:03] what made it die all of a sudden? [00:31:19] my reboot [00:31:25] it came back up [00:31:32] it wasn't replying to ping [00:31:47] Meh, I'm confused. [00:32:00] Ryan_Lane: I'm still jabbing at ldap but, meanwhile -- it occurs to me that I probably need some mysql stuff installed as well. Do you know which classes I need for nova? [00:32:09] Ah, now I see [00:32:10] so, before, when I asked if you had rebooted, had you *already* rebooted? [00:32:19] port 22 is not opened and all that fun stuff [00:32:25] it should be [00:32:28] fixing [00:32:35] it's open.... [00:32:50] the default security group has port 22 [00:32:51] and icmp [00:33:12] andrewbogott: see nova-production1's configuration [00:33:17] I used a different ruleset [00:33:24] 'k [00:33:32] Now fixed [00:33:39] Jews: *always* keep the default group checked [00:33:44] Okay. [00:33:49] * Jews forgot [00:33:54] please read the docs :) [00:33:55] !instance [00:33:55] https://labsconsole.wikimedia.org/wiki/Help:Instances [00:34:32] "In general" -- change it to "always" :P [00:36:00] well, there's cases where you wouldn't want to [00:36:20] but unless you know why you shouldn't, you should ;) [00:36:37] back in a little bit [00:44:56] PROBLEM host: simplewikt is DOWN address: simplewikt CRITICAL - Host Unreachable (simplewikt) [00:49:56] Looks done now. [00:50:48] !log simplewiki After some confusion, added firewall rules correctly and made new instance. Hope it works now... [00:50:49] Logged the message, Master [00:52:28] I'm ready for the IP now [00:52:50] Also, https://labsconsole.wikimedia.org/wiki/Nova_Resource:I-00000145 seems stray [00:53:48] stray? [00:53:50] what do you mean? [00:54:15] https://labsconsole.wikimedia.org/wiki/Nova_Resource:I-00000145 [00:54:27] yeah. I don't understand what you mean by stray? [00:55:17] Says pending and I can't SSH in, and my small instance is working now [00:58:14] i think I'm figuring out gerrit a little bit…just figured out that it wanted me to interactively merge all of my previous master commits as part of a rebase just to merge from master into my branch [00:58:20] now i want to merge my branch back to master [00:58:38] but i'm afraid that it will try to do some kind of crazy rebase + review thing [00:58:45] if I've already had my stuff reviewed in my branch [00:58:48] and approved [00:58:49] Ryan_Lane: Before you vanish for good... any ideas about what's going on with ldap on my instance? (I've twiddled so many puppet settings at this point that I'm tempted to start over from scratch.) [00:58:52] how do I now put it back in master [00:59:24] or. where is the appropriate place to ask this question? [01:06:10] ottomata: This is not a terrible place to ask, but I don't know the answer, and it's after 5pm in California which means you may need to re-ask next week. [01:06:43] Ryan_Lane, main point is that my simplewikt instance is ready to demo [01:08:12] thanks andrewbogott [01:10:46] RECOVERY host: simplewikt is UP address: simplewikt PING OK - Packet loss = 0%, RTA = 6.57 ms [01:33:15] hello? [02:39:06] RECOVERY Free ram is now: OK on bots-sql3 bots-sql3 output: OK: 20% free memory [02:41:46] RECOVERY Free ram is now: OK on puppet-lucid puppet-lucid output: OK: 20% free memory [02:42:36] RECOVERY Current Users is now: OK on testing-ldap-build testing-ldap-build output: USERS OK - 0 users currently logged in [02:44:06] RECOVERY Disk Space is now: OK on testing-ldap-build testing-ldap-build output: DISK OK [02:44:06] RECOVERY dpkg-check is now: OK on testing-ldap-build testing-ldap-build output: All packages OK [02:45:26] RECOVERY Free ram is now: OK on testing-ldap-build testing-ldap-build output: OK: 61% free memory [02:45:36] RECOVERY Total Processes is now: OK on testing-ldap-build testing-ldap-build output: PROCS OK: 80 processes [02:46:46] RECOVERY Current Load is now: OK on testing-ldap-build testing-ldap-build output: OK - load average: 0.01, 0.09, 0.08 [02:57:39] Ryan_Lane: Are you there? [02:59:36] Hydriz: yes, but not for long [02:59:50] Can I get access back? [02:59:57] Or are you still on vacation :( [03:00:22] got back recently [03:12:06] PROBLEM Free ram is now: WARNING on bots-sql3 bots-sql3 output: Warning: 18% free memory [08:40:52] Hello, I'm admin of the French Wikisource and I would like to have a write access to http://wikisource-dev.wmflabs.org in order to try MW extensions before ask their installation in Wikisource. Zaran say me that I've to ask here. What should I do ? [14:47:32] https://gerrit.wikimedia.org/r/gitweb?p=operations/puppet.git;a=commit;h=f6359eefead213987ca623553b5571d28ac8fe8a < Ryans meeting with the legal department not go well. [15:13:45] PROBLEM Current Load is now: CRITICAL on diablo diablo output: CHECK_NRPE: Error - Could not complete SSL handshake. [15:17:35] PROBLEM dpkg-check is now: CRITICAL on diablo diablo output: DPKG CRITICAL dpkg reports broken packages [15:18:45] RECOVERY Current Load is now: OK on diablo diablo output: OK - load average: 0.31, 1.05, 0.85 [15:22:35] RECOVERY dpkg-check is now: OK on diablo diablo output: All packages OK [16:03:48] PROBLEM Current Load is now: CRITICAL on gluster-devstack gluster-devstack output: Connection refused by host [16:04:28] PROBLEM Current Users is now: CRITICAL on gluster-devstack gluster-devstack output: Connection refused by host [16:05:08] PROBLEM Disk Space is now: CRITICAL on gluster-devstack gluster-devstack output: Connection refused by host [16:05:48] PROBLEM Free ram is now: CRITICAL on gluster-devstack gluster-devstack output: Connection refused by host [16:06:58] PROBLEM Total Processes is now: CRITICAL on gluster-devstack gluster-devstack output: CHECK_NRPE: Error - Could not complete SSL handshake. [16:07:38] PROBLEM dpkg-check is now: CRITICAL on gluster-devstack gluster-devstack output: CHECK_NRPE: Error - Could not complete SSL handshake. [16:54:28] RECOVERY Current Users is now: OK on gluster-devstack gluster-devstack output: USERS OK - 2 users currently logged in [16:55:08] RECOVERY Disk Space is now: OK on gluster-devstack gluster-devstack output: DISK OK [16:55:38] I really hope that Ryan is going to arrive [16:55:48] RECOVERY Free ram is now: OK on gluster-devstack gluster-devstack output: OK: 67% free memory [16:56:58] RECOVERY Total Processes is now: OK on gluster-devstack gluster-devstack output: PROCS OK: 134 processes [16:58:48] RECOVERY Current Load is now: OK on gluster-devstack gluster-devstack output: OK - load average: 0.00, 0.10, 0.07 [17:01:00] * andrewbogott wishes that nagios didn't send messages to IRC because of my standard development cycle. [17:22:07] Ryan_Lane: Not enjoy your meeting yesterday? https://gerrit.wikimedia.org/r/gitweb?p=operations/puppet.git;a=commit;h=f6359eefead213987ca623553b5571d28ac8fe8a [17:22:26] heh [17:22:34] nah. thought my laptop was stolen [17:22:37] seems it wasn't [17:23:33] Ah, that would of sucked [17:26:52] ji [17:26:53] hi [17:26:59] :) [17:27:29] howdy [17:27:38] cool [17:28:06] gonna write notes http://etherpad.wikimedia.org/LabsIrcConf [17:28:09] from conference [17:28:14] * Ryan_Lane nods [17:28:18] k [17:29:12] It's in half an hour or an hour an a half? Silly timezones [17:32:42] huh [17:35:54] I thought it was in 30 mins? [17:37:46] yeah. it's in 30 mins [17:37:56] it's 17:37 GMT right now [17:56:23] I hope that people who do not already use wmf labs are going to join us too, at some point it should be some intro too [17:58:16] I hope they're not all bot owners as 2/3 are raped daily :P [17:58:26] really? [17:58:41] I hope some "they" are here :) [17:58:52] Well actually it might be fixed now as Beetstra was sorting her..(his?) bot. [17:58:54] or it's gonna be a meeting of us 3 [18:00:09] *gong* [18:00:14] :D [18:00:15] this they (singular) is somewhat around (having dinner) [18:00:18] I assume it did get posted to say wikitech? [18:00:23] yes [18:00:24] it was [18:00:42] Mmm dinner I wish but that would take me like 40min, might be lazy and just order pizza. [18:01:04] petan: so, this is your thing, want to start it off? [18:01:09] ok we can discuss food later :) [18:01:23] Why .. are bots more important than food? [18:01:30] right I never really did any conference, online or any other, but I guess yes [18:01:31] * Ryan_Lane is hungry and wants breakfast ;) [18:01:45] and I just ate supper... [18:02:03] I think we can start talk all points we have in pad [18:02:26] So you're handing back to Ryan_Lane for 'tech specs'? :P [18:02:27] the first part was meant to be some intro for newbies but I don't know if some are around [18:02:38] link? [18:02:44] http://etherpad.wikimedia.org/LabsIrcConf [18:03:18] technical specs. right now we have 4 compute nodes, and one controller [18:03:33] ok the compute nodes are "physical servers" right? [18:03:43] the controller is as well [18:03:52] what are technical specs of these servers? [18:04:07] I've noticed there is nothing about it on wikitech [18:04:24] the compute nodes have 48GB memory, two processors with 6 cores each [18:04:53] and 1.2 TB of storage per node [18:05:04] the storage is in a raid1 [18:05:05] err [18:05:07] raid10 [18:05:14] ok, so we can have around 24 instances on one server [18:05:17] What does labs relay on mwf production for (apart from network), just apt? [18:05:32] and the nodes share storage with each other using gluster [18:05:37] basically just apt [18:05:46] why is svn firewalled on labs? [18:05:49] ssh one [18:05:52] ssh [18:06:07] we may open that back up, or may just keep it closed until git transition [18:06:20] all of labs is blocked from production for ssh [18:06:47] it's to avoid hijacking of forwarded keys [18:06:59] Is there still a rough plan to switch to gluster for project storage and move away from assigning public IPs directly to have a domain based forwarding on nginx/varnish/some nice proxy? [18:07:15] both, ues [18:07:17] *yes [18:07:35] I have the gluster storage nodes installed, and they are running a glusterfs cluster that is peered [18:07:43] ok, let's have a bit of explanation of vm's we have, the open stack support moving instances from one node to other right? [18:07:48] I just need to write some code to make them share to the projects now [18:07:55] :) [18:08:14] work on a proxy solution hasn't started at all, but is planned [18:08:21] petan: yes [18:08:25] we can do live migrations [18:08:31] I need to do some right now, in fact [18:08:53] virt1 still doesn't have nearly as many instances as the others [18:08:54] so in case that one node needs to be restarted it's possible to move all instances to running one [18:09:01] yes, but it's painful [18:09:13] current nova support for live migrations is kind of crap [18:09:28] Can there not be logic that if we 'shutdown' a node via openstack it handles shifting all the vms off. Like XenServer does if you mark a server as broke [18:09:34] you basically need to do one instance at a time, or the migrations fail in really shitty ways [18:09:37] but in case there is outage of some node it's possible to quickly recover it? [18:09:44] so there should not be long outages, taking days [18:09:50] Damianz: I could write a script for it, yes [18:10:29] well, I'm not sure if it's possible to do cold migrations [18:10:47] I'll need to ask the openstack people about it. I don't see it in the docs and I haven't dug into the code [18:11:03] in that situation, right now, we'd need to fix the node [18:11:08] for example if one instance is rebooted on loaded node is it possible to move it automaticaly to node with low load? [18:11:29] automatically? no [18:11:33] ok [18:11:35] it's a manual process [18:11:44] so ballancing load between nodes is manual [18:11:46] this will be less problematic with the next set of hardware [18:12:10] I don't have the specs of the new hardware, but they are kind of great :) [18:12:19] 256GB of memory or so [18:12:30] In relation to if it's broken - do we plan to have some form of labs<>production mailing list/jira instance etc (or do we have it now) where by say a node is physically down rather than nagios moaning someone from mwf ops will sort it. Not sure how to explain it but rather than the TS model of admins manage have a community/admins manage is strange for certain things. [18:12:38] memory is our limiting factor right now [18:12:57] CPU utilization is very low, but we are overallocated on memory, and the hosts are swapping [18:13:15] In relation to re-balencing/migration - Once ganglia is live we probably should look at something where by if one node has some very cpu intence vms of live migrating them automagically. [18:13:43] Damianz: labs is considered "quasi-production". we'll try to fix it quickly, but it may not be to the support level of production [18:14:01] or do you mean things inside of labs? [18:14:16] everything on the instances is community managed [18:14:22] Well I think things inside say a project is down to the project manager/members which is community [18:14:29] But the community can't fix say switches exploding. [18:14:48] right. hardware and infrastructure that supports labs will be handled by ops [18:15:16] I think there are many communication methods either bugzilla or mailing list we have [18:15:33] indeed [18:15:48] I believe that labs-l is watched by ops [18:16:00] if something breaks with the hardware, we'll get alerted by production nagios [18:16:09] another thing I wanted to sort out [18:16:15] who all from ops is responsible for labs [18:16:25] basically just me [18:16:25] So say for a feature request - are we looking for bug tracker or sending a review of a puppet diff for merging into production? [18:16:39] we are hiring another position for labs [18:16:52] ok, so in case there is a problem with labs, we need to wait for you? [18:17:02] including requests for new accounts etc [18:17:05] Damianz: I'd say both, usually [18:17:21] I'm more thinking we have 2 groups of people - those who can write up manifests and they will get merged and those who notice bugs and not be able to fix them (and those that ask ops for roject issues..) [18:17:22] anyone with the ability to make svn accounts can make labs accounts [18:17:33] the dev process and labs are kind of mingled together [18:18:15] also, if there is a problem with labs, and I'm not around, someone else on ops will likely take a look at it [18:18:27] everyone works together on breakages, usually [18:18:33] ok, so when we need to have an account or project created, we can basically ping anyone from ops? or certain people [18:18:57] well, so far only a few have made projects and such, but they all have permissions [18:19:22] We could look at something like AIAV or whatever it's called where account/project requests go and someone with access picks it up. [18:19:27] we should really have some way of requesting project creations [18:19:32] maybe on bz? [18:19:44] I was also thinking of allowing anyone who is netadmin and sysadmin in a project to be able to create projects for others [18:20:05] seems a good idea [18:20:07] bz could work for it, yeah [18:20:10] that would make sense but it's a bit security issue [18:20:15] I think creating projects should b ereally easy [18:20:23] well, actually not much of a security issue [18:20:25] if u create 50 projects and 400 instance [18:20:32] ah. true [18:20:33] you get labs down [18:20:34] You could easily limit projects/instances though [18:20:42] right now projects are quota'd [18:20:49] 10 instances per project [18:20:49] petan, if they can make a vm in an existing project, what do they gain creating projectgs? [18:20:59] Platonides: override quota [18:21:01] the ability to create an unlimited number [18:21:13] throttling [18:21:39] well, it's something to think about [18:21:44] Maybe have a voting system or something where 2 people who are netadmin in projects can create a new project... or have an ccount over a certain age. [18:22:25] it could be permission based, and auto-confirm based, yeah [18:22:35] Tbh though if it's something that forces us to use the wiki then I'd rather not :P [18:22:47] I wouldn't want to do that with other things, but I see no reason to limit project creation to the web interface [18:22:52] err [18:22:54] *not to limit [18:23:15] I don't see any realistic way of letting people create projects from the cli [18:23:17] Yeah, that's more 'specialised' though - day to day stuff would be a pain to restrict ourselves to the web interface. [18:23:49] yes, we're working on moving OpenStackManager specific things into openstack so that we can allow people to use the cli [18:24:14] but, project creation isn't one we plan on changing [18:24:28] we skipped the rules part :) [18:24:36] Rules? how boring :P [18:24:39] heh [18:24:44] well, we do need to discuss that [18:24:50] so, I had a meeting with legal yesterday [18:25:03] people in general don't know how to name their instances so we sometimes end up with a crazy ones, check nagios to see what all we have there [18:25:03] we'll be writing up a terms of service, and a privacy policy soon [18:25:19] so we could make a naming conventions for instances [18:25:19] petan: Could we group nagios by project? :) [18:25:23] Damianz: yes [18:25:29] people will be required to accept the terms of service to login [18:25:30] will be working on that [18:25:51] also, it's looking like we won't require people to identify [18:25:54] ok, can you make a list of most important parts of these terms? [18:26:29] labs project will be required to display the privacy policy and terms of use, and they must show a warning to users anywhere information can be collected [18:26:47] also, privacy information should simply not be kept where possible [18:27:06] ok [18:27:23] so that people can become a check users on test site, even if not identified to foundation [18:27:35] example [18:27:48] hmm... that could be unexpected [18:27:49] we'll likely need to remove public IPs from projects that don't properly display terms of use and privacy policy, or don't display warnings, until the project complies [18:27:56] if the users login in there aren't aware of that [18:28:05] Will there be any project based terms or just labs? [18:28:18] Platonides: it'll be displayed on the login page, in that situation [18:28:36] Damianz: what do you mean? [18:29:08] Ryan_Lane: is it possible to make exception in case all members of project are identified? [18:29:17] so that there is no requirement to inform people etc [18:29:27] I think we'd prefer to not have to identify people [18:29:54] ok, but if the members of project chose to do that rather than changing code of sw to show the warnings etc [18:29:57] it's an administrative burden [18:29:59] For any projects taht say require storing private information - will they be restricted with seperate usage/terms stuff or not. For example (eventhough it shouldn't be dealing with live data) the donations stuff every year. [18:31:03] so, access to any project that may hold production private data will require a signed confidentiality agreement with the foundation [18:31:17] Makes sense [18:31:46] access to those projects will require another role, so it won't technically be possible to log in, even if someone accidentally adds another person [18:31:49] ok [18:32:07] it also may be a separate hardware cluster, etc, etc [18:32:18] we're still working out very sensitive data and labs [18:32:44] can we get a basic rules of what is allowed / disallowed and how people should name their instances :) [18:32:45] That won't affect the general stuff though? Like deleted diffs and having access to the production dbs for say bots. [18:33:05] probably db should not content private data [18:33:07] like TS [18:33:20] sure. of course, this is all a precursor to the terms of service [18:33:27] 1. no hacking [18:33:41] Ryan_Lane: *ehhem* cracking [18:34:09] of course, authorized pen tests for things inside of labs may be OK, but only if an exception has been granted. meaning, you need to ask ops first. [18:34:14] malicious hacking ;) [18:34:31] 2. No proprietary software [18:34:44] of course, it should also be possible to get an exception to that rule too [18:35:01] there's some cases where we need proprietary software for testing. like db2 or oracle [18:35:13] ok [18:35:17] 3. No copyrighted data [18:35:45] but even the gpl sw has some kind of copyright or not? [18:35:49] again, exceptions can be made, but people need to ask [18:35:55] fair use is also ok [18:36:06] right [18:36:08] You mean copyright without decent access like gpld stuff that someone holds the copyright for? [18:36:12] sorry [18:36:15] let me clarify :) [18:36:27] only open source software, and open content licenses [18:36:35] OSI approved open source licenses [18:36:59] Awwww [18:37:04] ok [18:37:04] WTFPL isn't accepted by the OSI IIRC [18:37:10] :P [18:37:12] heh [18:37:33] basically, we'd like to avoid getting sued by the RIAA or MPAA [18:37:46] People get sued? I thought you just got raided by the FBI [18:37:52] I have a feeling we didn't make friends with them with that whole SOPA thing ;) [18:37:53] heh [18:38:10] 4. No torrenting [18:38:19] Downloading or? [18:38:22] either [18:38:25] :( [18:38:28] no bittorent [18:38:37] even legal? [18:38:42] Not even opensource stuff? Some things are only avaible for torrent download [18:38:45] torrents of linux :) [18:38:49] heh [18:38:55] of course, there can be exceptions, but it'll require approaval [18:38:59] No offering of illegal content [18:39:04] I wrote this [18:39:16] well, that goes into 3. [18:39:17] feel free to fix it [18:39:24] ok I write to epad [18:39:44] but yes, no illegal content, as defined within the United States. Legal may change all of this wording, btw ;) [18:39:50] s/may/likely will/ [18:40:09] 4. No tor nodes [18:40:11] Yeah I assume everything is 'as in the US' as the servers are over there and mwf is over there mainly. [18:40:20] again, exceptions can be made to this policy [18:40:25] Damianz: yep [18:40:35] and we have no plans to have labs infrastructure outside of the US [18:40:46] this is difficult enough, legally, without involving other countries [18:41:26] 5. Only Wikimedia related work is allowed [18:41:31] Any thoughts on proxies in general? (not say socks for accessing labs but web etc, more along the lines of tor). [18:41:35] of course, that's pretty broad [18:42:12] #4 was no tor nodes ;) [18:42:20] or anything related to that [18:42:28] Open web proxies etc I assume you fall under tor [18:42:36] no VPNs from labs to other networks, etc [18:42:45] no open proxies. right [18:43:44] if you guys could think of other things I should specifically mention, let me know [18:43:53] I may have a bit of a question here .. I am storing 'external link additions' in SQL, linked to who added which external link in which diff on Wikimedia projects .. that may be sensitive data [18:43:55] there's no reason only the WMF can draft these :) [18:44:08] Beetstra: is it publically available? [18:44:13] 'Abuse of recourses' which is very wide and could be nazi but just in gneeral if someone is taking the mickey [18:44:23] meaning, is that data you are pulling from public sources? [18:44:36] In principle, yes .. I am pulling it out the live diffs [18:45:28] 6. Actions deemed abuse can result in loss of privileges. This is at the discretion of the Labs staff and community. [18:45:40] 7 [18:45:45] yay for catch-alls! :D [18:45:53] Beetstra: then it isn't private [18:45:55] :P [18:46:10] I still want a wikimedia email :P [18:46:18] yes. we're working on that [18:46:31] yay [18:46:32] It should be possible to do penetesting as there are a few (ehm) securityholes [18:46:33] I want it too :D [18:46:34] :D [18:46:34] things are harder when thinking multi-tenant :) [18:46:59] jeblad: yes, as mentioned, there can be exceptions for pen testing, but it will require approval [18:47:42] in other words, we need to know it's going on, otherwise it looks like malicious hacking [18:47:42] OK [18:48:02] general rules to instances? how they should be named, resources etc [18:48:03] ok,.. [18:48:08] so, the wikimedia related work rule needs to be clarified some [18:48:14] If you want to do that it should probably go on the mailing list for community/op approval [18:48:37] working on mediawiki, even if not work Wikimedia related projects, is still Wikimedia related, for instance [18:49:08] Yeah... [18:49:10] so, if anyone is confused about whether their work is Wikimedia related or not, they should ask :) [18:49:15] But then it could be a patch of something commercial [18:49:19] Just as a side comment, I'm not sure that anyone would detect the more advanced penetrations.. Just my 5¢ [18:49:30] jeblad: I don't disagree [18:49:31] s/of/or/ [18:49:43] which actually brings up something that I wanted to mention :) [18:49:59] everyone should be self-policing their projects [18:50:14] if you add someone to a project, you should ensure they are doing what they said they'd be doing [18:50:20] will labs be limited to mw related stuff or can it be used more like testing interesting technologies? [18:50:32] jeblad: not limited to mediawiki [18:50:46] anything that could be used for wikimedia related work [18:50:47] Interesting tech stuff might be related to mwf work for ops etc [18:51:03] For example, I have some wild ideas about stats and need a mongodb .. oki [18:51:12] yep. that's totally fine [18:51:36] so, going back to self-policing projects... [18:51:46] I do some policing, but I don't scale terribly well [18:51:55] * Damianz makes Ryan_Lane multi threaded [18:52:00] heh [18:52:09] would it then be a vm and config left to the interested ones or would it be necessary to set på complete clone? [18:52:28] jeblad: what do you mean? [18:52:51] how much must be done by an ops at wmf [18:52:57] nothing [18:52:57] In regards to policing and going back to creating projects - who polices projects? ops, community or everyone? [18:53:09] in labs, project members can do everything [18:53:23] well, people with sysadmin and netadmin in a project can do everything [18:53:26] excluding one thing [18:53:36] they can't allocate public IP addresses [18:53:46] since we are very short on them [18:53:52] Damianz: everyone should [18:53:54] * jeblad plans to shot himself in both feets and then some more feets [18:53:58] heh [18:54:26] if someone sees something odd going on, send an email to security@wikimedia.org [18:54:33] and we'll investigate further [18:54:37] It might be interesting to have per project status and maybe some news of 'whats been going on in labs' or such. [18:54:44] yeah, would be cool [18:54:57] we have SMW on labs... [18:55:19] so, we could add them there, and have it compile a monthly/weekly report [18:55:20] + SAL [18:55:29] that somehow reflect status heh [18:55:34] yeah [18:55:54] nice to see in human readable terms what people are doing, for the wider community [18:56:09] so the idea of status updates is kind of cool [18:56:17] not required, of course :) [18:56:36] so, instance naming conventions.... [18:56:47] we should probably have some conventions :D [18:56:55] I guess people somehow must hook up with eachother. Is there any plan on how to accomplish that? [18:56:57] right now it's basically a free-for-all [18:57:14] jeblad: Yell for help and someone will do stuff [18:57:21] peaople = those interested in a specific project [18:57:22] jeblad: well, projects are documented with who's a member, basically, you just check out the project page and talk to some of the members [18:57:26] !project bots [18:57:26] https://labsconsole.wikimedia.org/wiki/Nova_Resource:bots [18:57:32] ^^ see that page, for instance [18:58:06] Reminds me I should change CBNGs mysql pass... [18:58:08] I don't necessarily have an issue with the instance naming free-for-all, but there's a few cases where instances are really poorly named [18:58:28] Damianz: did it get logged in the SAL or something? [18:58:39] Then a last question and I shall shut up, how will this interact with toolserver.org? [18:58:44] Nah it's just something really crappy and we have a bots phpmyadmin somewhere. [18:58:57] Not important, times up [18:59:07] jeblad: as of now, it doesn't [18:59:14] Not like I have it running mysqldumps on cron either :( [18:59:21] an end-goal is a toolserver-like environment in labs [18:59:24] Also is there any plan in regards to backups genreally? [18:59:31] so, it's a replacement, more than anything [18:59:42] Damianz: nope. in general we won't backup most things [19:00:02] instances will never be backed up. they should be puppetized [19:00:06] Most stuff should be re-dployable from puppet but stuff like mysql I assume will fall to per project/member [19:00:14] mysql servers will have backup [19:00:25] we'll do that on the server side [19:00:37] :) [19:00:46] we plan on having hardware mysql that's managable via a service [19:01:03] Would be nice, vm disks are slow. [19:01:16] yes. IO in virtualization sucks [19:01:16] what kind of backup for sql [19:01:19] online backups? [19:01:23] or daily [19:01:25] likely LVM snapshots [19:01:28] ok [19:01:36] though they'll probably be rotated often [19:02:32] if project members want more long-lived backups, they can do dumps and store them in their gluster storage [19:02:45] of course, that storage will be quota'd [19:02:48] what if someone will te4st io intensive projects? [19:02:54] Just thinking I'd rather not loose 2months of data for a bot again... [19:02:59] jeblad: what do you mean? [19:03:07] Damianz: yeah, I'd also prefer that :) [19:03:19] I dislike data-loss with a passion [19:03:31] Mhm [19:03:35] the gluster storage can act as a storage area for backups of a project [19:03:44] but, ops isn't going to do it for you ;) [19:03:51] server - client traffic like a gadget interacting with a mongodb [19:04:22] jeblad: in general, instances that eat a lot of resources mostly only affect themselves [19:04:38] though if there are too many of those on one host, it's problematic [19:04:47] we'll move the instances around to spread the load [19:05:00] I have a question/suggesting regarding instance naming. I've been annotating my instances with usage notes on the summary wiki (e.g. https://labsconsole.wikimedia.org/wiki/Nova_Resource:I-00000131) [19:05:16] My question is -- is that page actually persistent, or will my notes get clobbered at random times? [19:05:24] clobbered, unfortunately [19:05:35] I'll add sub-page support for instance notes [19:05:37] OK, that's not a good place to store that info, then :) [19:05:39] so that it won't be clobbered [19:05:41] Cool. [19:05:51] similar to how documentation is done on project pages [19:06:41] anything else to talk about? :) [19:06:52] documentation? [19:06:56] I'm going to assume probably not right now but do we have a way of deploying out the test cluster or are we looking to puppetize it properly when we split up the repo into project branches and re-do the merging stuff? [19:07:21] we'd *very* much like to have it all puppetized fully [19:07:24] yes [19:07:40] because we plan on using it for automated infrastructure testing [19:07:44] btw there is still a lot of stuff to talk :) [19:08:01] oh yeah. project branches [19:08:14] the test branch is basically fucked beyond repair right now [19:08:30] It would be nice to be able to have say Jenkins build a cluster, deploy out mw how it in production, test stuff then murder its self. [19:08:43] Damianz: yeah. that's the goal [19:09:02] that way we can test to ensure both ops changes, and major mediawiki changes don't totally break things [19:09:19] that would be an incredibly expensive test, so would likely not get run for most changes [19:09:38] It would be good for testing major arch changes though [19:09:43] yep [19:10:12] so, with regards to the test branch... [19:10:21] the way we are currently doing things sucks [19:10:28] mhm [19:10:42] I want to move us to per-project branches, where instances are configured to use that branch [19:10:50] but also, where anyone can switch to another branch [19:11:09] ideally, instances will usually run the test branch [19:11:10] Having a centralised puppet server for labs with all branches there if needed? [19:11:22] we'd like to get rid of the centralized puppet server [19:11:34] and instead have the repo live on each instance [19:12:00] How would that work for certain things like passwords/contacts which arn't in the repo but are linked when we merge production back into test? [19:12:13] also, we'd like changes to project branches to be able to skip review [19:12:38] Damianz: we're going to have to figure that out [19:12:52] but we'll likely also just have the private repo on all instances too [19:13:04] no reason we can't. everyone has access to it anyway [19:13:23] (private labs repo, production one is only accessible to ops) [19:13:34] s/,/./ [19:14:00] Yeah, we'd just need some op input to ensure production changes are replicated in labs so puppet doesn't freak out. [19:14:19] well, puppet would be local [19:14:27] so, when you ran it, you'd see it freak out [19:14:43] if we see long-broken puppet in instances, we'll just switch the branch to test [19:15:09] test will always be assumed to work [19:15:22] and changes in test will be quickly cherry-picked to production [19:15:42] I was more thinking if you stuck another var in private which the manifests refered to it would need someone to add it to the private labs repo so puppet didn't break when the manifests where merged back - if we're trying to keep test mergable with production. [19:15:58] yep [19:16:13] that'll be part of the testing process ;) [19:17:20] petan: what's the next topic? [19:17:22] let's move to tools? [19:17:34] I mean the tool lab part [19:17:47] there are no tools like on toolserver now because of lack of db [19:18:22] but once we replicate production db to labs so that people access it, people would be able to create toolserver like tools on labs [19:18:24] Db access would be highly amazing... relaying on the TS for stuff on labs just for out of date dbs is annoying. [19:18:49] is it going to be managed like a project per tool [19:19:04] because that would mean an instance per tool [19:19:16] so a lot of resources while not needed [19:19:26] DB access would be amazing... [19:19:44] we'll have replicated database access [19:19:59] we haven't planned it out terribly well, yet [19:20:12] Ryan_Lane: Hopefully it wont be hours out of date or broken many times a month :P [19:20:13] and yes, we're planning an instance per-tool [19:20:31] I believe we'll be monitoring it like the rest of the infrastructure [19:20:41] and when master switches happen, it'll be updated with everything else [19:20:48] :)) [19:20:51] :D [19:20:56] * hoo|away is in love with Ryan_Lane again... [19:21:08] it's likely going to take a while for us to get to this [19:21:31] way too much stuff to do right now :) [19:21:32] petan: 'Shared instances' apart from the bots that have their own instances so don't crash the others from being a usage whore? [19:21:47] it's not requirement to have a shared instance [19:21:50] Damianz: that's one of our reasons for wanting separate instances [19:22:05] also, the hardware we are getting in will support thousands of instances [19:22:13] but if a bot doesn't eat a lot of cpu / memory there is no need to have it on separate instance [19:22:23] like wm-bot [19:22:39] bots can more easily be handled in a shared way [19:22:46] tools are more difficult [19:22:49] Tbf I'd like in tests to be able to trash and re-launch my bot instances as needed to test stuff then merge into the labs 'production' stuff and have it update by puppet. [19:23:01] "the others from being a usage whore?" - /me stares at ClueBot :P [19:23:02] we should have a scheduler which run a bot on isntance with lowest load [19:23:09] +1 [19:23:20] SGE ? [19:23:32] that may be excessive ;) [19:23:35] but it's possible [19:23:44] methecooldude: CB is going to get packages when I get bored so we can shift it about if needed. [19:23:49] s/packages/packaged/ [19:24:26] I'd really like all bots to be packaged and puppetized [19:24:28] In regards to bots - is the standard to run them under their own users or labs users? I'd say the former if we're going towards puppet. [19:24:30] with init scripts and such [19:24:43] their own system users [19:24:52] if they are packaged, that's easy :) [19:24:58] Yeah [19:25:04] ok [19:25:05] also, they should be installed on all bot-runners [19:25:08] Like shifting cbng from supervisord to upstart would be cool [19:25:27] then moving a bot is a matter of stopping it on one instance, and starting it on another [19:25:30] It would be kinda nice if based on ganglia/nagios we could move or restart bots [19:25:44] yeah. I could see that going badly, though ;) [19:25:49] lol [19:25:51] probably easier to move them manually [19:26:18] we should move log to bots-1 btw [19:26:30] because it's running on overloaded bots-2 and crashing [19:26:38] Have a server dead, oh these bots are on there -> start them on other instance wouldn't be too bad... we'd need a STONITH approach though [19:26:39] !log test [19:26:41] like now [19:26:43] it's gone [19:27:23] Damianz: delete the instance ;) [19:27:32] if all of the instances are the same, why not? [19:27:38] Which as another note - do we plan on supporting being able to have access keys to openstack where we can have scripts pull stats/do stuff without our user key? [19:27:44] that's about as STONITH as you can get ;) [19:27:48] :D [19:27:52] DIE! [19:28:01] yes. cli access is in the plans [19:28:04] it's hard, though [19:28:24] because openstack doesn't support puppet or dns (though thanks to andrewbogott it will in essex) [19:28:35] * Damianz gives andrewbogott a cookie [19:28:43] also, it doesn't support writing to mediawiki [19:28:46] andrewbogott is working on all of these things [19:28:58] as well as some other great things :) [19:29:04] Actually, right now I'm eating lunch. [19:29:08] Oh [19:29:13] * Damianz takes the cookie off andrewbogott until he does work [19:29:14] :D [19:30:12] petan: hm. the bot is dead again? that's odd [19:30:30] I'd love to be able to restart it from labsconsole [19:30:30] it's because it's running on heavily loaded server [19:30:33] yeah [19:30:33] It seems to be randomly diying for the past couple of weeks. [19:30:44] it should be moved to bots-1 [19:30:58] another reason to have a load balancer [19:31:05] Ryan_Lane: I'd like to be able to see all bot/tool status centrally and cull as required but that's more a project thing. [19:31:09] I'll put some effort in to update the package (since I modified the code some, since then) and move it [19:31:15] Damianz: nagios? [19:31:27] we'll have ganglia soon [19:31:29] petan: Are any bots monitored in nagios? [19:31:35] I hope it will be [19:31:37] Actual bots, I know mind arn't. [19:31:56] if we create a class to nova, people could definte own services using console [19:32:01] labs console :) [19:32:15] nagios parses stuff using smw from wiki [19:32:20] Hmmm [19:32:43] so if there was a generic service where people would just pick and define the check it would be possible [19:32:54] That would be shiny [19:33:16] also I want to make access to nagios using ldap [19:33:23] * Ryan_Lane nods [19:33:24] however I didn't find out how [19:33:36] we'll have to talk about that [19:33:40] ok [19:33:56] we really need some form of web sso [19:33:59] In regards to ldap, don't we have 2 ldap servers atm? As one that does some stuff and one that does other stuff rather than instances replicated. [19:34:14] our ldap situation right now kind of sucks [19:34:29] I need to rearchitect everything in production [19:34:34] SSO for mw would be nice, being able to use wikipedia details reliablty would be amazing for project stuff :( [19:34:38] for labs we currently have one LDAP server [19:35:26] we also only have one DNS server ;) [19:35:38] Oh awesome [19:35:41] when I bring things up in eqiad, it'll all be redundant [19:36:38] building a robust infrastructure takes time :) [19:36:44] there's a reason we're still in closed beta. heh [19:37:06] ok so let's discuss the sql server now [19:37:12] * Ryan_Lane nods [19:37:22] so, we'll likely have two clusters [19:37:35] one that is replicated from production, like toolserver [19:37:46] and another that allows users to create/manage their own databases [19:38:07] ok, is it going to be only one sql service, or would it be possible to have a server per project [19:38:11] makes sense [19:38:11] could they be joined [19:38:13] ? [19:38:17] so that people can have unique db's etc [19:38:38] Is the plan to run Mariadb over MySQL? [19:38:42] maria [19:38:44] Platonides: hm. that may be difficult [19:38:56] it was maria [19:39:00] I'm not totally sure now [19:39:20] I don't want people storing their ldap password on a filesystem [19:39:35] and if we have LDAP auth, it's likely [19:39:59] I think I'd really prefer a service that gives the user a password that is specific to the database, and can easily be changed [19:40:09] Ryan_Lane, one big feature is to be able to join a user table with a cluster one [19:40:12] Having something on labsconsole that lets you make a user wouldn't be that hard though. [19:40:15] Platonides: ah [19:40:24] Platonides: I'll keep that in mind, then [19:40:45] Damianz: yeah, something along those lines would be nice [19:41:00] of course, we need to ensure people can't delete/modify other people's databases... [19:41:03] we also need quotas [19:41:04] the toolserver does that with -user servers of the cluster, where you can create dbs [19:41:12] having this per-project would be ideal [19:41:36] Platonides: is it per-request, or do all users have the ability to make as many dbs as possible? [19:41:38] so there is no plan of having a sql server per project? [19:41:50] like people could have own databases and root access [19:41:53] It would be kinda nice if you're a member of the project you can access all the dbs but I'm not sure how that would work auth wise. [19:42:19] Ryan_Lane, for each wmf cluster you can connect with dns like s1-rr or s1-user [19:42:28] I can't imagine how we make more production like wiki installations on shared sql server [19:42:33] depending if you want to use the user server or are just interested in the data [19:42:39] because db names will conflict etc [19:42:51] I think you can create new databases if they begin with your user name [19:43:05] petan: as mentioned, having it per-project would be nice [19:43:11] ah [19:43:16] We could do something crazy like have a proxy infront of them and per project prefixes on dbs but that would be insane. [19:43:25] :D [19:43:31] well, we do plan on using tungsten [19:43:32] see https://wiki.toolserver.org/view/Database_access#User_databases [19:44:09] Forced prefixes are annoying [19:44:25] what's the problem in requiring to configure dbs with the proper name on each instance? [19:44:26] indeed [19:44:27] easiest way to avoid name conflicts, though [19:44:37] Platonides: performance [19:44:54] you can't have a dbs on instance because it's slow [19:44:54] uh? [19:44:59] I don't really like per-user databases... [19:45:01] Also at some point we'd have to shared the physcail servers if labs gets busy..er. [19:45:12] I mean, we have servers now and they are really slow [19:45:16] petan, I don't mean installing a mysqld on each instance [19:45:19] ah ok [19:45:22] in fact, I *really* don't like it. I'd much prefer per-project databases [19:45:23] nvm [19:45:33] was talinking about configuring on each instance/project to which db it should connect [19:45:48] instead of doing ugly things with a proxy prefixing dbs [19:45:54] ah. yeah [19:45:58] that's possible [19:46:20] anyway, if anyone wants to work on this before we get to it… :) [19:46:26] if needed, the app could eg. extract the prefix from the uname, or something similar [19:46:58] I am a bit lost now, if someone could insert it to epad, would be cool [19:47:27] one question; I was under the impression that labs was testing, but am I correct if it is more like "the place for all weird stuff"? [19:47:39] heh [19:47:47] define weird [19:47:50] I think with some projects we're going to not fit into that model though - they are probably better using hadoop or some nosql server if they want 200GB dbs though I guess. [19:48:15] jeblad: it's a community maintained environment that is mostly geared around testing [19:48:17] s/we're/are'nt'/ [19:48:20] Really can't type tonight gah [19:48:38] Closed beta testing env running production stuff! [19:48:47] jeblad: some aspects of it are quasi-production [19:48:55] well, not production from the WMF's POV ;) [19:49:26] ok, .. quasi-production from vm-servers.. :-] [19:49:34] If it was production we could wake Ryan_Lane up at 2am to fix stuff :D [19:49:39] :) [19:49:48] :d [19:49:49] err [19:49:50] :D [19:50:04] notice none of you have my cell phone number :) [19:50:08] we need to implement sms nagios :) [19:50:23] Phone? So last year, I'd just annoy you on g+ :P [19:50:25] I do get alerts if labs is done [19:50:31] *down [19:50:43] we'll test it at 2 am [19:50:47] no thanks :) [19:51:03] I like the production nagios alerts based on timezone [19:51:05] * jeblad writes note about writing an app for wake up call to Ryan_Lane with a big red button for every member on "labs" [19:51:18] Damianz: yeah, we added that recently [19:51:25] we have enough people for that to work now [19:51:49] Yeah... I need more ops people in the us for that to work, my entire team is uk based :( [19:51:49] jeblad: I'll retaliate with a stab button :) [19:52:22] ok, so what are the expected limits on sql server [19:52:28] I suppose certain queries will not be allowed [19:52:42] + there will be timeouts for queries [19:52:44] I unfortunately suck at mysql.... [19:52:57] MySQL is kinda dependant on what other commodity services we have though IMO. [19:53:01] so hopefully someone else will be helping with this [19:53:25] I haven't put a ton of thought into the mysql support [19:53:39] ok [19:54:08] we're hiring, though, so hopefully that person will have better knowledge of it. or maybe asher will help out. [19:54:51] right let's discuss gluster storage now [19:55:02] Tungsten looks cool, can't say I've ever used it but I've got some MySQL replication stuff coming up in projects :D [19:55:11] I don't really know how gluster works at all [19:55:19] Damianz: yeah, we want it for the multiplexing and filtering support [19:55:30] Yeah binlog filtering sucks ass. [19:55:50] gluster is a distributed block-based filesystem [19:56:12] ok so there will be a huge filesystem right? [19:56:16] Yeah [19:56:20] yes [19:56:24] we have 4 nodes [19:56:26] each project will have assigned storage there [19:56:30] with some quota [19:56:34] each node has 40TB of storage, roughly [19:56:45] gluster acts like a raid-1 across the network [19:56:50] ok [19:56:51] 8D [19:56:52] so, we have roughly 80TB of storage [19:57:12] would it be possible to change quota on fly? [19:57:24] like this project needs 80gb+ [19:57:25] we don't have any software written for any of this yet :) [19:57:39] right now I'm going to write a script like we have for home directories [19:57:47] Are the nodes in a 1<>1 or peered together (so less like a standard raid1 config)? [19:57:52] andrewbogott is writing a nova volume driver for it, though [19:58:02] peered [19:58:03] ok, the performance of this gluster storage is going to be better than performance of current fs [19:58:04] on instances [19:58:13] very likely, yes [19:58:41] how is it going to be mounted to intances [19:58:50] there will be like a shared mountpoint on all instances? [19:58:54] the way things look right now, it's: ext3 -> qcow2 -> gluster -> xfs -> lvm -> raid10 [19:58:56] like a nfs server? [19:59:17] with volume storage it'll be: gluster -> xfs -> lvm -> raid10 [19:59:26] yes [19:59:36] it'll be automounted, and that config will be in LDAP [19:59:45] so basically each project will have a big folder mounted to each instance [19:59:50] yep [19:59:59] is it possible to split it [20:00:21] in which way? [20:00:21] like to have 2 separate gluster storages for 1 project [20:00:26] ah [20:00:31] I assume the nova driver will connect out to all nodes so if we shoot one in the foot it won't take down all the instances. [20:00:35] if you for instance fill one up, the other still has space [20:00:37] ..well bots. [20:00:37] I kind of wanted the ability for projects to have x number of volumes [20:00:43] where x is quota'd [20:00:44] ok [20:00:52] and each volume would be quota'd [20:01:21] Damianz: well, the nova driver won't actually mount it on the instances [20:01:28] it'll just make it available to the instances [20:02:02] Ah [20:02:03] automount will mount it [20:02:16] Are you planning on keeping instance storage local or shifting it over? [20:02:21] local [20:02:28] :) [20:02:29] I want them to be separate clusters [20:02:37] also, it just makes a lot of sense :) [20:02:45] Yeah [20:05:52] ok, last point is documentation heh [20:05:59] I started help:contents on labs [20:06:00] We have documentation? [20:06:09] I think it would be cool to move all to Help space [20:06:16] Damianz: some :) [20:06:21] petan: agreed [20:06:31] also to insert it to Sidebar [20:06:35] yep [20:06:37] like a link to Contents [20:06:51] it would be cool to have a link on each nova special page [20:06:55] to specific manual page [20:07:42] ah. yeah, that would be useful [20:07:46] petan: I changed sidebar [20:07:49] cool [20:07:55] people can actually do this [20:08:01] @search security [20:08:01] Results (found 1): security, [20:08:04] !security [20:08:04] https://labsconsole.wikimedia.org/wiki/Security_Groups [20:08:08] :) [20:08:17] but having a link directly on a page would be faster [20:08:20] I'd kinda like per project or in the bots case, per bot docs. [20:08:33] I meant a documentation for labs now [20:08:34] Ideally anyone in a project should be able to fix other peoples stuff. [20:08:39] but projects could have own docs too [20:08:43] Unfortunately, we need to limit sysop access on labsconsole [20:08:56] so, any admin related actions need to go through ops [20:09:19] or just create a new usergourp in wiki, like with editinterface etc [20:09:37] but I don't really think there is too much to do what require a sysop [20:09:47] writing documentation can be done by regular users :) [20:09:52] https://bugzilla.wikimedia.org/show_bug.cgi?id=34500 [20:10:06] petan: ah. right [20:10:15] I forget that's possible :) [20:11:21] is there anything you guys hate about the openstackmanager interface? features you'd like added? [20:11:34] I hate it randomly decides people don't exist [20:11:41] yeah. I hate that too [20:11:50] I still haven't been able to track that bug down [20:11:58] I think we should have improved interface so it can handle like thousands of projects [20:12:02] Also being able to list instances in other projects would be nice [20:12:07] in future it will be hard to navigate through [20:12:07] petan: yep. working on that now :) [20:12:11] I can find them by searching for their pages but no list :( [20:12:19] by default no projects will be shown [20:12:30] we'll have a project filter where you can select which projects you'd like shown [20:12:42] ok [20:12:49] maybe better error messages heh [20:12:57] with instructions what is wrong [20:12:57] yeah. the messages suck :( [20:13:12] like people are coming here asking why they can't do this and that [20:13:14] maybe I can pass through the openstack error messages, when they exist [20:13:15] Also what's 'managed sudo policies' for? [20:13:26] Damianz: it's an admin specific thing [20:13:29] Error: you can't insert new security rule because you aren't a netadmin in project "YAAY" [20:13:29] we can manage sudo via LDAP [20:13:40] instead of "Fail" [20:13:44] ah [20:13:55] well, I'm actually just removing those options, when a user doesn't have the rights :) [20:14:07] :) [20:14:15] it's a lot more user-friendly [20:14:29] Apart from listing instances in other projects and the horrid table in table in table project screen I can't think of much else. [20:14:35] heh [20:14:43] I'm also changing the project list [20:14:46] to be headings [20:14:51] projects can be categorized in future [20:15:04] like a matrix of projects etc [20:15:20] people could click through to find an interesting project to work on [20:15:25] ah. hm. I wonder how I can implement that... [20:15:27] There should be some docs about how to be involved [20:15:33] jeblad: indeed [20:15:38] there should be some docs :) [20:15:40] +1 [20:15:46] :) [20:15:47] documentation is hard :) [20:15:56] Code is self documenting! [20:16:02] I disagree :D [20:16:03] meh [20:16:14] Now it seems like it is only about what you wnat to achieve and that it is a community effort, but nothing about how to involve the community [20:16:15] Damianz: you didn't see my code [20:16:24] Well.. sometimes it's quicker to browse the code than read the docs :P [20:16:46] for (x in blah Blah) etc (really I name variables blah sometimes) [20:16:50] jeblad: It's kinda like you annoy Ryan_Lane then someone else then do stuff... organisation is lacking. [20:17:04] yep [20:17:16] I know one possible help [20:17:24] liquid threads, for one [20:17:32] and project access requests on project talk pages [20:17:39] That would be nice [20:17:40] people should know that Ryan isn't only person who can help [20:17:44] !ryan [20:17:44] yes [20:17:47] !Ryan [20:17:47] man of the all answers ever [20:17:48] A general 'I have no idea what I'm doing' page would be good too [20:17:51] !Ryan del [20:17:51] Successfully removed Ryan [20:17:54] heh [20:18:31] !Ryan is man of all answers ever (but there are others :)) [20:18:31] Key was added! [20:18:39] there's a log way back where I explained in great detail how things work [20:18:43] in this channel [20:18:51] I should take that and turn it into documentation :) [20:18:59] petan: Oh also could you move the un-foldered logs into the folders... it's confusing :P [20:19:15] we should link to this channel's IRC logs from labsconsole too [20:19:24] unfoldered logs are from time we had wm-bot [20:19:27] * Damianz really should stop logging into apache1 and grep'ing petan's home dir for referce. [20:19:29] * before we had [20:19:45] Damianz: there is a project to make a search engine [20:19:50] :D [20:19:54] it's in wikimedia svn [20:20:00] if you want to help, please make it :) [20:20:04] Ooh shiny except I don't have a svn account :P [20:20:12] that's a pain to ask for one [20:20:30] well, you have a labs account, you just need to be added into an svn group [20:20:38] it usually takes few weeks and you have to pass a test that you are proper programmer heh [20:20:40] I'm going to let the devs handle that, though :) [20:20:42] Though you might have given me an excuse to read how to ask for it. [20:20:50] Meh [20:20:53] petan: with the switch to git, that all goes away [20:20:57] I hope [20:21:05] it's review before merge [20:21:15] Pretty sure all my wikipedia rights have been from annoying admins I know :P [20:21:45] so it doesn't matter if someone has no programming skills, we'll see it in the code [20:21:46] * petan thinks that rollback bit is less destructive than sudo reboot [20:21:53] Ryan_Lane: Can't the svn stuff that's turning into git just use git review? [20:22:08] yes [20:22:45] :) [20:22:58] Damianz: ur grepping my home? [20:23:01] wondering what is there [20:23:08] logs are in /mnt/public_html [20:23:09] petan: Well public_html dir [20:23:16] symlink :) [20:23:19] true [20:23:35] I was lazy to fix a bot so I fixed it using a link [20:23:42] Lol [20:23:51] We should patch Ryan_Lane's code to mount bind them :P [20:25:50] Ryan_Lane: can you make an account to one huggle project coder [20:25:56] for labs [20:26:09] yeah. where's he at? [20:26:16] I hope he come soon [20:26:28] I need to eat lunch :) [20:26:37] is meeting over? [20:26:51] I guess yes unless someone has more questions [20:27:04] I will try to summarize it all [20:27:42] I need food :D [20:28:09] I think the next meeting needs some noobs, maybe once we're not in closed beta anymore. [20:28:11] cool. good meeting guys. [20:28:20] want to have one every month or so? [20:28:27] it would be cool [20:28:35] like to have all people online :) [20:28:44] hello [20:28:48] Ryan_Lane: mmovchin is that guy [20:28:50] howdy [20:28:55] mmovchin: you'll be working on huggle? [20:29:08] !account-questions | mmovchin [20:29:08] mmovchin: I need the following info from you: 1. Your preferred wiki user name. This will also be your git username, so if you'd prefer this to be your real name, then provide your real name. 2. Your SVN account name, or your preferred shell account name, if you do not have SVN access. 3. Your preferred email address. [20:29:15] hello Ryan [20:29:20] yes, you're right [20:29:21] he is actually developer of huggle now, but would like to have access to our instance [20:29:26] ah. cool [20:29:53] Can 1 and 2 be the same? [20:29:57] yes [20:30:04] 1+2) mmovchin [20:30:09] 3) michael@movchin.de [20:30:10] Thank you [20:30:26] Damianz: can you improve the part of docs in pad [20:30:31] the project docs [20:30:38] I don't know what was your idea [20:31:13] I was thinking more per project so for example how bots works, what runs on it etc so if something is borked I can go ah it's probably this [20:31:28] that should be a subpage of project I guess [20:31:39] or maybe each project could have a page in main space [20:31:49] !initial-login | mmovchin [20:31:49] mmovchin: https://labsconsole.wikimedia.org/wiki/Access#Initial_log_in [20:32:03] ok. I'm off to get food and such [20:32:07] * Ryan_Lane waves [20:32:15] ok :) [20:32:47] thank you, ryan [20:32:50] yw [20:32:52] * Damianz waves at Ryan_Lane|food [20:33:10] petan: See #huggle [20:35:53] mmovchin: let me know if you need any help [20:35:59] with access to instance etc [20:36:12] 02/18/2012 - 20:36:11 - Creating a home directory for mmovchin at /export/home/bastion/mmovchin [20:38:12] 02/18/2012 - 20:38:12 - Creating a home directory for mmovchin at /export/home/huggle/mmovchin [20:39:03] petan: thanks [20:39:12] 02/18/2012 - 20:39:11 - Updating keys for mmovchin [20:39:49] mmovchin: what if we move to wikimedia bz [20:39:58] rather than making own one [20:40:06] wikimedia bz? [20:40:09] bz? [20:40:09] bugzilla [20:40:15] ah [20:40:23] That would be great [20:40:28] ok [20:40:31] So we don't have to set up a own one [20:40:35] yes [20:40:42] Could you realize that? [20:40:49] :) [20:40:54] Or is everyone free to request a project on bugzilla? [20:40:56] you mean if I can set it up? [20:41:00] yes [20:41:05] on wm bz [20:41:07] I can ask someone who can [20:41:12] thank you :) [20:41:26] It would be great if it would be ready until the meeting [20:41:50] So I can move all issue we have on Googlecode to Bugzilla [20:43:15] *issues [20:43:27] ok [20:44:17] thanks [20:44:24] FYI: Just send out the mail [22:12:00] can i get an public ip for my nova instance now [22:30:30] Athlon: What instance is that? [22:34:16] simplewikt [22:36:06] Athlon: I don't have access to that, so I can't help you sorry [23:00:43] Athlon: Short answer, no [23:00:47] Long answer, ask Ryan. [23:01:19] Athlon: have you tested with a socks-proxy to make sure it works? [23:02:10] yes [23:02:38] ok. gimme a little bit :) [23:02:48] in the middle of somethign else [23:28:03] what's your tool called? [23:28:50] Athlon: ^^ [23:31:12] simplewikt [23:33:08] simplewikt.tools.wmflabs.org [23:34:03] hm. can't ping it [23:35:29] heh [23:35:35] it's outside of the subnet [23:35:41] we're out of public IPs for now [23:35:44] so, you'll need to wait. [23:36:11] Ryan_Lane: You should just get an ipv6 tunnel on one of the isntances and give out ips for free :P [23:36:28] I need to enable IPv6 support in nova [23:38:14] well, that surely didn't take long [23:38:18] seems I need to work on that proxy soon [23:39:10] Assuming we only want to support http/https..maybe ftp a varnish/nginx proxy wouldn't be that much work. [23:39:32] yeah, I'd likely do an nginx proxy [23:39:41] transparent. no caching [23:40:45] We could probably do it with puppet and variables but that could get really messy if you wanted hundreds in I guess. [23:40:55] nah. I'd prefer to make it like an openstack service [23:41:30] hm. there's a load balancing service that already exists... [23:41:41] I wonder if it can just to simple proxying [23:41:54] http://wiki.openstack.org/Atlas-LB [23:46:19] that doesn't seem appropriate. guess I'll need to write something [23:47:32] Hmm that could be interesting [23:48:00] It would be nice if you could have a bunch of boxes setup with like cryosync/pacemaker and make a cloudy lb/proxy thing. [23:48:11] I hate pacemaker [23:48:25] we use LVS for load balancing [23:48:31] and BGP to failover between them [23:48:45] it's reliable and simple :) [23:49:38] BGP is a funny protocol [23:49:48] Kinda really simple yet can cause weird routing issues [23:50:27] * Ryan_Lane nods [23:50:39] Doing it in BGP in theory should have a lower risk factor as switching ips around/restarting services has a time betwean down and up... and also leaves the capacity side on one live box. [23:51:25] yep [23:52:55] What do you use LVS for then? Doing the monitoring with a hook to update BGP or a bunch of LVS instances above the squid boxes? I assume the squid then round robin on the apache instances or do they go back to routed ips based on lvs too? [23:53:05] * Damianz thinks he read about this ages ago on wiki blog tech somewhere [23:53:10] BGP directs the traffic to the LVS server [23:53:17] the LVS server directs it to the real servers [23:53:26] the real servers return traffic to the client [23:53:46] So you're using direct routing in LVS rather than NAT/Proxying it it. [23:53:54] yep [23:53:59] Cool [23:58:29] !account-questions [23:58:29] I need the following info from you: 1. Your preferred wiki user name. This will also be your git username, so if you'd prefer this to be your real name, then provide your real name. 2. Your SVN account name, or your preferred shell account name, if you do not have SVN access. 3. Your preferred email address. [23:58:44] yes, I'm lazy enough to copy/paste this into an email [23:59:17] You should just make wm-bot do the emailing :P [23:59:33] heh