14:02:07 #startmeeting tripleo 14:02:08 Meeting started Tue Sep 22 14:02:07 2015 UTC and is due to finish in 60 minutes. The chair is slagle. Information about MeetBot at http://wiki.debian.org/MeetBot. 14:02:09 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 14:02:11 The meeting name has been set to 'tripleo' 14:02:18 yo 14:02:29 hi 14:02:41 hi everyone 14:02:48 \o 14:03:26 hello 14:03:42 * dprince loves the new meeting time already 14:03:43 #topic agenda 14:03:43 * bugs 14:03:43 * reviews 14:03:43 * Projects needing releases 14:03:43 * CI 14:03:46 * Specs 14:03:48 * one off agenda items 14:03:51 * open discussion 14:04:10 you will notice i added a standing item for one off agenda items before open discussion 14:04:33 #info add one off agenda items to the wiki before each meeting if you have them 14:04:57 #link https://wiki.openstack.org/wiki/Meetings/TripleO meeting wiki page 14:05:14 #topic bugs 14:05:35 #link https://bugs.launchpad.net/tripleo/ 14:05:35 #link https://bugs.launchpad.net/diskimage-builder/ 14:05:35 #link https://bugs.launchpad.net/os-refresh-config 14:05:35 #link https://bugs.launchpad.net/os-apply-config 14:05:35 #link https://bugs.launchpad.net/os-collect-config 14:05:37 #link https://bugs.launchpad.net/os-cloud-config 14:05:40 #link https://bugs.launchpad.net/os-net-config 14:05:42 #link https://bugs.launchpad.net/tuskar 14:05:45 #link https://bugs.launchpad.net/python-tuskarclient 14:06:06 we had a lot of untriaged bugs last week, i believe i went through and hit them all 14:06:32 A general question about bugs I have is if something is added to the #tripleo trello do we also want to create LP bugs? 14:06:55 sometimes I just look at LP bugs when there is an issue 14:07:01 probably if it's a bug, then yes 14:07:48 if it's added to the trello board as a task that someone is doing, then i dont feel those need bugs as well 14:08:04 does that make sense? 14:08:13 that list of bug urls needs a refresh in the wiki ;) 14:08:24 yeah i was going to ask about removing the last two shardy 14:08:26 sure 14:08:44 ok so this was me, I tend usually to wait until I have enough info befor I open a bug, since I started using trello I havn't been creating a bug /me will do better 14:09:06 shardy: is there something missing? 14:09:13 derekh: yeah, and I haven't started using the trello yet ;) 14:09:15 dprince: i'd say, nice to have, since the trello boards may not have that great visibility outside tripleo (people would expect lp is what i'm saying i guess) 14:09:22 derekh: so I've duplicated some work... 14:09:39 slagle: the new repos we recently added aren't there 14:09:39 dprince: your not the only one, maybe it was a stupid idea 14:10:10 slagle: or are they all tracked via the tripleo LP org? 14:10:12 derekh: not stupid, I'm willing to try it. Just not part of my daily checks with regards to project visibility yet 14:10:13 shardy: i dont think we have launchpad projects for those, but we should decide if we want them 14:10:22 slagle: aha, Ok, my mistake 14:10:38 We should probably explicitly state that in the wiki 14:10:38 dprince: ok 14:10:46 Use triple for all projects other than ... 14:11:06 #action d0ugal to update wiki about using the tripleo project 14:11:13 you see how that works? :) 14:11:56 if we want to add new projects too, we can 14:11:59 slagle: damn :) 14:13:48 afaics, all the critical bugs are assigned and being worked on 14:14:02 is there any to draw attention to? 14:14:46 ok, moving on 14:14:48 #topic reviews 14:15:25 #info There's a dashboard linked from https://wiki.openstack.org/wiki/TripleO#Review_team - look for "TripleO Inbox Dashboard" 14:15:45 That should include all of the new projects now. 14:16:03 i was going to point that out :) 14:16:18 #info bnemec updated the review dashboard for all the new projects 14:16:18 I have a review up to add them to the dashboard configuration too. 14:16:32 https://review.openstack.org/#/c/223861/ 14:16:41 in case anyone wants to generate a custom dashboard. 14:16:53 #info if you're using custom queries or your own searches, be sure they include all the new repos 14:17:38 i also want to point out that we are in despearate need of people to review more, across the board 14:18:01 * dtantsur tries to review moar of tripleo stuff nowadays 14:18:08 there are a couple of new specs proposed, the common library, stable branches, container support 14:18:23 ya lots of new code in tripleo, so lots more to review 14:18:35 i don't realistically see us able to push forward on any of those without a serious step up in reviews 14:19:34 i don't think a "required minimum" number of reviews per day really works or is an incentive, but we should be open to whatever ideas people have on how to increase review bandwidth 14:20:10 slagle: I guess improving CI stability is the main one 14:20:19 i guess i'll just say (hopefully without sounding too snarkish), if you have code you want to land, you better be reviewing other peoples code 14:20:42 slagle: I'm sure I'm not the only one who tends to wait for passing CI before devoting review bandwidth, I know it's been a bad few weeks on that front 14:20:51 is there anything collectively we can be doing to help with that? 14:20:53 I think we are pretty close to stable CI 14:21:07 shardy: we shouldn't do that imo (wait for ci) 14:21:08 trown: don't jinx us man 14:21:13 there will always be stuff like the puppet change today 14:21:15 lol 14:21:18 lol 14:21:51 shardy: people miss out on getting early feedback if they're cahnge sits unreviewed for a couple of days due to failing ci 14:21:58 slagle: it's tough though, when there's 200 pending reviews in your inbox, you're drawn to those you can most likely land (so they're not in your inbox anymore! :) 14:22:18 oh yea, for sure :) 14:22:26 i try to look for stuff that is ready to land 14:22:33 but i try to balance it out too 14:23:31 clearing out the way old stuff would help too here 14:23:45 +1 to clearing out old stuff (and bugs) 14:23:54 I actually did some of that yesterday. 14:23:55 shardy: as far as collectively helping with CI, let's discuss about that when we get to that topic in just a bit :) want to make sure we draw attention to it 14:24:08 slagle: sure :) 14:24:27 Yeah, I never understood why OpenStack stopped the auto-abandon thing 14:24:34 I'll try to do more reviews. The moving in is all done now, I should be able to get more sleep and do more reviews. ;-) 14:24:41 Because people's feelings got hurt. Apparently. 14:24:48 all projects are clogged with a bazillion ancient patches which is pretty distracting at times 14:25:02 ++ 14:25:13 well, i say we abandon. we can have our own policy 14:25:16 We could resurrect the bot for TripleO. I thought some other projects had done that already. 14:25:20 i think neutron has an 'autocull' feature where old .. yeah what slagle said 14:25:30 I am +1 to an auto-abandon ... it is not like it is any harder to un-abandon than it will be to rebase the ancient patch anyways 14:25:30 for ironic we decided to do it manually 14:25:37 bnemec: Yeah, I remember arguing about it on the ML, personally I didn't find clicking "restore change" occasionally all that offensive :) 14:25:49 shardy: Nor did I. 14:26:03 bnemec: would you like to lead this effort? :) 14:26:13 And I continue to say that wasting core reviewer time abandoning old patches is stupid. 14:26:16 slagle: Sure 14:26:22 I don't know what the old rules are, but auto-abandoning if the poster doesn't respond is a good idea IMO. 14:26:24 lets bring it back, I forget how many people were against it when we last brough it up 14:26:40 #action bnemec to look into auto-abandon policy for TripleO. 14:26:55 Auto abandon anything that has negative feedback and hasn't been dealt with in over 2 weeks 14:27:02 I'm not arguing it on the ML again. 14:27:13 yea, use the list at your discrtion :) 14:27:19 discretion 14:27:24 bnemec: ++ 14:27:31 I believe it was generally agreed that projects could do it if they wanted to, but it wasn't going to be re-enabled globally. 14:27:46 I think a project specific policy is fine, maybe communicate it via a tripleo tagged ML message 14:27:59 shardy: Agreed. 14:28:00 +1 14:28:15 cool, any other review business? 14:28:28 #topic releases 14:28:42 i did a few releases last week, but didnt get to all of them 14:28:55 i'm happy to go through them again this week 14:29:07 #action slagle to do releases 14:29:20 as always, please ping folks if you need a release sooner and don't feel like waiting 14:29:33 #topic CI 14:29:55 We've been hurting hard since the move to instack based ci (some of it not unexpected I think its a completly new job) 14:30:10 Over the last week, we out several outright breakages for various reasons and unreliability due to usually bad mirrors. 14:30:28 We are now pinned to specific mirrors for F21, Centos and EPEL, in jenkins nodes, delorean containers and on undercloud 14:30:49 This I think has helped a lot but people havn't noticed due to the breakages 14:31:21 dprince: i'm curious about your thoughts about this latest puppet-nova breakage 14:31:29 was that revert worthy? 14:31:32 A fix for the most recent job breakage merged a few hours ago in t-h-t but we had to wait for a new package to hit delorean/current 14:31:40 slagle: I think Depends-On is the best solution for these 14:31:49 it seems changes like that in the puppet modules is disastrous for people using master 14:32:11 slagle: we could also easily pin to an older puppet-nova in our CI while we get proper review time to think about it 14:32:33 derekh: good update. i think the pins will help as well 14:32:50 derekh: are there things you can think of where people can contribute to help out CI? 14:33:14 slagle: I've been adding taks to the trello boards as they pop into my head If anybody has time to help out see here https://trello.com/b/dFPVlpFZ/tripleo-short-term-tasks 14:33:24 *tasks 14:33:43 #link https://trello.com/b/dFPVlpFZ/tripleo-short-term-tasks ways to help out ci 14:33:45 Maybe I should just change that trello board to ci-tasks or something, then others can have other boards if they want 14:33:55 I've been working on these but outages have been taking priority when they happen so its been slow progress 14:34:18 i like the use of the [toci] tag myself 14:34:28 slagle: ack 14:35:48 FYI: the jobs overcloud-f21-nonha and overcloud-f21puppet-nonha are identical 14:35:51 i also wanted to point i'm trying to come up with a script that will work for ci and local environments, https://review.openstack.org/225096 14:36:00 so we can have some consistency there 14:36:05 slagle: ++ 14:36:17 we also have a symlink for the ci delorean repo ow 14:36:18 now 14:36:36 slagle: yup, switching over to that script is a must 14:36:47 #link http://trunk.rdoproject.org/centos7/current-tripleo/ delorean repo that toci is using 14:37:10 #info FYI: the jobs overcloud-f21-nonha and overcloud-f21puppet-nonha are identical 14:37:15 we still need a mechanism to automate that link 14:37:26 sounds like we got a volunteer! 14:37:38 I can take that 14:37:43 :) 14:37:49 trown: https://trello.com/c/RXaUOfPf/14-toci-bump-delorean-repo-url 14:37:53 I had an idea about delorean feature, or maybe it is a new (small) project called delorean-mirror 14:37:55 #action trown to look into automation around updating the delorean repo symlink 14:38:16 essentially, it would mirror the latest current Delorean link either locally, or to another cloud for use by our CI. 14:38:51 this would avoid the downtime associated w/ trunk.rdoproject.org 14:39:12 thoughts? worth putting effort into this 14:39:37 I think we should be putting effort into not having downtime w/ trunk.rdoproject.org 14:39:38 +1 to mirror all the things. That's basically infra's policy on it too. 14:39:44 but mirroring is also good 14:39:46 we could probably script it such that if we bump our DELOREAN_REPO_URL in tripleo-ci we automitically refresh the repo 14:40:05 i'm not 100% convinced it's needed. the last downtime was painful, but other than that, how much downtime has there been? 14:40:34 trown: better uptime for trunk.rdoproject.org is good but the maintenance there is beyond the scope of upstream tripleo 14:40:46 dprince: +.5, I'm not sure the recent outages are a fair representation to what we can expect 14:40:48 i guess it depends on the effort 14:40:53 slagle: this isn't new, mirrors are always a good idea IMO 14:41:00 dprince: ya, but upstream tripleo is relying on it, so it is a bit of a strange relationship 14:41:02 derekh: ^^ 14:41:17 derekh: lets not focus on trunk.rdoproject.org then 14:41:24 derekh: we just need mirrors for all the things I think 14:41:27 dprince: ok 14:41:33 i agree mirrors are a good idea 14:41:35 otherwise our downtime is just going to be always bad 14:41:44 CI downtime 14:41:50 ya I am +1 to mirrors in general 14:41:57 just depends on the effort to setup i guess 14:42:04 Yup, I think we could just have a "mirror" server to replace out old bandersnatch mirror 14:42:19 the idea w/ delorean-mirror would be that it is special, and requires a bit of extra work to mirror properly (perhaps in a lightweight fashion that is also useful to local devs) 14:42:22 then aim to get as much as possible on it, no need to single anything out 14:43:13 #agreed mirror as much as possible 14:43:24 delorean-mirror -> [fedora@bandersnatch ~]$ sudo wget --no-parent -r http://trunk.rdoproject.org/centos7/df/03/df0377d64e1ef0b53c4e78a8ff6a50159de5131a_733f1417/ 14:43:37 right, nothing is single out. I for example already have local mirrors for Centos and Fedora I use 14:44:18 any other ci business? 14:44:18 * bnemec should do that 14:44:31 There was a big discussion about not "forking" delorean... i.e. running your own locally. And that is what I do. The delorean-mirror thing was an idea that might be more paletable for some as it provides the same redundency... but is just a mirror 14:44:43 I would like to get inspection back running in CI https://review.openstack.org/#/c/210436/ 14:44:50 and dependent patches 14:45:04 we removed it to get past the outage at the end of last week 14:45:25 trown: I'd like introspection on for just a single job if possible 14:45:25 trown: +1 14:45:59 dprince: it takes almost no time now, and I put in line to change the timeout to 10m 14:46:00 trown: the extra CI time isn't strictly required for all jobs right? Perhaps combining it with another job to minimize the # of jobs too 14:46:14 trown: if it takes more than 5 minutes it is worth it IMO 14:46:39 it will only take more than 5 minutes on failure 14:46:46 dprince: are you ok with turning it back on in the job we have now? 14:46:50 and add the other job later? 14:47:00 slagle: yep, just thinking ahea 14:47:02 but I am ok just putting it on the nonha job as derekh suggested 14:47:05 slagle: I don't think it should be required 14:47:20 ok 14:47:20 in reality, even with it turned on we are not actually using it 14:47:36 as in we do not use the data it collects 14:47:40 it's 2 minutes for dsvm, might be slightly more for IPA 14:47:54 trown: the test could make some basic assertions about what's in the swift data tho? 14:48:15 #info turn introspection back on in ci, setup a separate job in the future 14:48:19 it is actually more than we think though because if we skip it we could skip building the extra ramdisk too right? 14:48:28 nope 14:48:40 dprince, we're switching to IPA, which will be one ramdisk 14:49:07 ya we could skip building it today...but it would not help once deploy also uses IPA 14:49:07 dtantsur: ah, good point. But until then it would save a bit... 14:49:27 #info switch to IPA will mean consolidated deploy and introspecton ramdisk 14:49:33 I hope that complete switch to IPA will happen in 1-2 weeks, not more 14:49:37 dtantsur: all the more reason to switch to IP sooner then :) 14:49:40 What is IPA? 14:49:42 old deploy ramdisk is deprecated 14:49:49 * d0ugal is here to ask the silly questions 14:49:51 d0ugal, ironic-python-agent, our new ramdisk 14:49:56 not beer unfortunately :) 14:50:01 dtantsur: aha, shame. 14:50:01 d0ugal: its a beer that does discovery 14:50:05 lol 14:50:08 It's not bash. 14:50:12 That's all I care about. :-) 14:50:15 lol 14:50:28 A step closer to beer then 14:51:06 upstream deprecated the bash ramdisk in liberty, so that is all that really matters :) 14:51:23 ok, are we all good to move on to specs? (9 mins left) 14:51:31 #topic specs 14:51:48 So I'm interested to get more feedback on the release branch spec 14:51:59 particularly if folks think we could implement it from Liberty 14:52:08 I had a question on the release branches that IIRC is not covered 14:52:13 what do we do with requirements? 14:52:18 i think we could, but we gotta get the ci setup 14:52:25 e.g it'd be great to do an announcement around the liberty release pointing to "stable tripleo" install docs 14:52:51 #link https://review.openstack.org/#/c/221811/ stable release branch policy 14:53:12 shardy: i'll review your latest ps today 14:53:17 dtantsur: the release branches have their own requirements, so I guess we'd need to get syncs from openstack/requirements stable branch 14:53:32 shardy, then we'll have hard time backporting features 14:53:40 (and least might have) 14:53:56 dtantsur: That is a good point, and is probably a constraint we'll have to consider wrt features that are feasable 14:54:20 That's how all of the other projects are doing it now though, so we won't have any unique problems around that. 14:54:20 dtantsur: impossible requirements is probably a valid No-Backport justification 14:54:30 this would be good to add to the spec 14:54:37 bnemec, other porjects don't backport features... 14:54:39 slagle: yup, will do 14:55:01 dtantsur: There's probably a reason for that. ;-) 14:55:20 I know, but we plan on backporting features, right? 14:55:26 and we're the only project that will do it 14:55:37 Why would we backport features? 14:55:38 #topic specs / open discussion 14:55:58 I think trying to backport features is really ambitious...probably too much so for liberty 14:56:12 er, can I also ask for additional eyes on the overcloud deployment library spec? 14:56:15 Especially if it puts us in an impossible situation wrt requirements. 14:56:24 bnemec, that's what is written on the spec :) please vote there 14:56:31 Ah 14:56:36 * bnemec hasn't read that spec yet :-( 14:56:52 bnemec: the idea is that if, say, nova adds a new feature, we don't prohibit e.g backporting the t-h-t patches which add support for it 14:56:59 yea the reasoning is laid out in the spec 14:57:04 tzumainn: link would be good thing to have https://review.openstack.org/#/c/219754 14:57:06 obviously it has to be supported in stable/liberty puppet-tripleo tho 14:57:13 whoops, thank you marios! 14:57:32 we just don't block things which would otherwise work, just because they're adding functionality vs bugfixes 14:57:43 the other option of course is just do a normal stable branc 14:57:46 branch 14:57:52 That sounds like a gigantic headache and diverges from the rest of OpenStack's stable policy. 14:57:53 #info reviews please on Mainn's library/api spec @ https://review.openstack.org/#/c/219754 14:58:11 which might be easier, but risks making folks wait longer when we're playing catch-up with new features added to projects 14:58:30 it also calls for much tougher reviews of feature backports, and we are not reviewing new stuff enough 14:58:52 bnemec, trown: feedback on the spec would be great, let's discuss it there :) 14:58:52 yea, i have that concern as well. 14:58:55 I'll look at the review today. We should discuss this there anyway. 14:59:00 +1 :-) 14:59:04 shardy: ya I will look at the spec today 14:59:12 I'm quite willing to just do the normal stable thing if that's the consensus 14:59:30 but I really really want *a* way to deploy stable openstack w/upstream tripleo 14:59:42 ++ for that 14:59:48 ya that seems required really 14:59:51 derekh: (waiting for appropriate moment but almost out of time...) hey how do we join that board? (or the group more generally https://trello.com/tripleo/members ) 15:00:23 thanks for the participation today folks! 15:00:27 #endmeeting