16:00:15 #startmeeting fuel 16:00:15 Meeting started Thu Sep 10 16:00:15 2015 UTC and is due to finish in 60 minutes. The chair is xarses. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:16 #chair xarses 16:00:16 Todays Agenda: 16:00:16 #link https://etherpad.openstack.org/p/fuel-weekly-meeting-agenda 16:00:16 Who's here? 16:00:17 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:19 The meeting name has been set to 'fuel' 16:00:20 Current chairs: xarses 16:00:25 o/ 16:00:27 hi 16:00:29 o/ 16:00:31 hi 16:00:33 hi 16:00:52 hi 16:00:54 hi 16:00:54 hi 16:00:55 рш 16:00:57 hi 16:00:58 hi 16:01:08 hi 16:01:12 o/ 16:01:21 does etherpad work for you? 16:01:26 no 16:01:27 its very slow 16:01:28 nope 16:01:35 issue with srt infrastructure 16:01:37 I have the agenda up 16:01:47 it was bad in the U.S. too 16:01:52 can't blame srt for this, can't open it in US, either 16:02:13 Hi all. 16:02:19 xarses: lets start? 16:02:23 #topic community builds from master (someone from Infra team?) 16:02:38 angdraug: ^? 16:02:39 bookwar: ^ 16:02:41 hi 16:02:44 =) 16:02:57 hi 16:02:58 we have builds enabled, available here https://ci.fuel-infra.org/view/ISO/ 16:03:23 are they using Liberty packages? 16:03:28 they are built from current up-to day mos8.0 repo of ubuntu packages 16:03:40 which means that most of the packages are still from kilo 16:04:03 but as packaging team does their job they add packages from liberty to this repo 16:04:08 bvt2 actually runs against these builds from master there? 16:04:19 so we expect that this built will start to fail soon 16:04:22 yes 16:04:22 when do you see that we will have the packages in place? 16:04:56 that a question for packaging team, i don't have enough data 16:05:14 bookwar: thanks bookwar. It's great to see that we enabled master builds very soon after we've created stable/7.0 16:05:45 question about liberty packages 16:05:57 will we keep kilo builds running? 16:06:19 yes until liberty builds stabilize 16:06:36 angdraug: great, thanks 16:06:37 Fuel CI will stay with kilo until then, too 16:06:45 yes, we keep 8.0-kilo builds in place, and we will use them for fuel-library ci 16:07:04 you can see that currently fuel-library ci uses 8.0-kilo-1 iso 16:07:23 status is here https://ci.fuel-infra.org/ in the Status section 16:08:16 moving on? 16:08:22 thank you guys 16:08:24 yes please. great job bookwar! 16:08:30 #topic elections of nailgun, fuel-library component leads; PTL (mihgen) 16:08:57 I just wanted to reiterate this. There was email to openstack-dev about code review process changes 16:09:11 no one objects of having elections for component leads 16:09:33 and we have two largest pieces in Fuel, which are nailgun & fuel-library 16:09:55 so I'd like to hear your opinions guys whether we can move forward with it 16:10:10 it was also suggested to run elections at the same time as for PTL 16:10:17 how we're going to elect component leads? from the cores? 16:10:19 using standard openstack procedures 16:10:46 So nominate your self and share your platform? 16:10:50 so, we have the same nomination deadlines and voting deadlines as in upstream OSt ? 16:10:58 announcement of the PTL elections on openstack-dev: 16:11:02 #link http://lists.openstack.org/pipermail/openstack-dev/2015-September/074037.html 16:11:16 aglarendil: yes 16:11:19 what if no one among cores wants to be a component lead? xD 16:11:20 component leads - from cores, I suppose. Frankly, I'd still need to read how openstack does it currently in other projects 16:11:42 ikalnitsky: that's gonna be bad, that's what we want to avoid ) 16:12:00 I think just following the PTL process should work for component leads, too 16:12:08 no need to reinvent the bicycle 16:12:10 two component leads will basically get freedom from feature-work assignments and can influence arhitecture of the component 16:12:37 angdraug: that's my assumption as well. So I think we'd need to learn more, and propose it in openstack-dev 16:12:53 (I'd need to learn more at least) 16:13:08 and fast, nomination period opens tomorrow 16:13:26 nomination period where? 16:13:29 voting period is September 17-24 16:13:34 #link https://wiki.openstack.org/wiki/PTL_Elections_September_2015 16:13:36 #link https://wiki.openstack.org/wiki/PTL_Elections_September_2015 16:14:10 so we have until 17th to persuade as many Fuel cores as possible to nominate themselves :) 16:14:50 oh. then we'd need to try to get in the same schedule 16:14:51 so one core will not work on features? 16:15:02 salmon_: yes. 16:15:12 folks, one question, it seems that OSt has its own voting platform 16:15:15 salmon_: component lead should not work on features 16:15:16 salmon_: yes except that will be doing specs/design review for all features for that component 16:15:25 did we arrange all the details that we will have the same platform used? 16:15:39 or will we conduct votese calculation on ourselves? 16:15:45 nailgun is to big for it 16:16:09 salmon_: too big for what? 16:16:27 for having one Lead, says salmon_ 16:16:30 mihgen: to have on lead 16:16:36 *one 16:16:39 are we going to re-elect component leads on a periodic basis? 16:16:51 alex_didenko: according to OSt workflow - yes 16:16:54 each six months 16:17:03 aglarendil: I don't know frankly, not sure if angdraug is aware. As I said, we'd need to analyze how it works now in openstack 16:17:29 mihgen: angdraug we would need to rush with this 16:17:30 at the very least we can reuse the voting system, anyone can use that 16:17:38 aglarendil: true 16:18:04 angdraug: can I give you an action to follow up on that? 16:18:04 mihgen what is your expectations from component lead. bullet points... 16:18:13 salmon_: well, may be. But with modularization we've started a work on, I'd expect more and more modules to become independent expertise areas, where we may grow other leads 16:18:31 I hope so :) 16:18:43 xarses: yes 16:18:46 #action angdraug will follow up on usage of upstream voting system 16:19:03 maximov: you can find it in my email about code review... and we'd need another one with more data 16:19:22 in short - architecture, quality, making it easy for others to contribute the code 16:19:43 mike, lets create a new mail thread back in the ML and request candidates 16:19:52 catch majority of bugs on review/gating, not a month after code merge 16:19:58 mihgen ok, but what about community work. will component lead drive communication w/ community? 16:20:11 maximov: yes. 16:20:12 I think we missed this part 16:20:26 it's going to be way more important for PTL 16:20:43 but PTL would rely on component leads for specific areas 16:21:05 #action mihgen will create a new ML covering duties, duration, and nomination pool for component leads and request candidates 16:21:07 for instance, fuel-library component lead will have to work closely with puppet-openstack, ansible-openstack if needed, etc. 16:21:25 and what about nailgun lead? 16:21:29 we don't have community 16:21:41 thats not true 16:21:57 ikalnitsky: we will be building one + there are other areas of Fuel, where integration is needed 16:22:13 also they would need to work with the current compute project so that the API's are aligned. I forgot their name 16:22:14 plugin developers are part of nailgun community 16:22:30 xarses: tuskar 16:22:31 all these integration questions with whatever other components / pieces, plugable architecture falls into this category 16:22:32 ? 16:22:41 angdraug: thanks, yes 16:23:49 there should be a balance, common sense in prioritization of work 16:24:20 mihgen I only have one concern, component lead will be doing a lot of review, communication, architecture etc.. but what about writing code, what would be a trigger for component lead to contribute to the code? 16:24:46 maximov: mostly, they won't be, they won't really have time 16:24:57 angdraug: are we waiting for your message or that was question mark for something else? 16:25:12 angdraug: that was question to xarses, he replied already 16:25:17 maximov: that's ok if component lead will write way less of code 16:25:29 lets go back to the ML and discuss further, we need to move along through the schedule for today 16:25:30 xarses well, don't you think that in this case component lead will loose his expertise.. 16:25:42 but it's important to write code which defines framework for others to work within 16:25:51 folks, sorry to interrupt, can we discuss details of elections and roles in the ML? we spent almost 25 minutes on that 16:25:57 maximov: by reviewing, it shouldn't be much of an issue 16:26:06 let's move on 16:26:16 #topic backlog of small enhancements (gaps in functionality) to work on: http://bit.ly/1Nk9rmI (mihgen) 16:26:27 #link http://bit.ly/1Nk9rmI 16:26:33 that's what we have in our launchpad 16:26:40 bugs with "feature" tag 16:26:55 they in fact form our backlog of small enhancements 16:27:08 so my proposal is to continue using launchpad for small feature gaps 16:27:21 +1 16:27:22 and let enhancements team go through this backlog 16:27:26 +1 16:27:28 in kanban-like model 16:27:28 one small note - we have many different tags, "fueature-*" 16:27:39 alex_didenko: nope those are different 16:27:43 so it's not very convenient to filter all of them 16:27:54 feature-* is about bugs introduced by a feature 16:27:56 those are for taxonomy of bugs 16:28:17 alex_didenko: those should be for bugs that relate only to that feature (usually because it caused them) 16:28:19 "feature" is for bugs which require a requests for enhancement 16:28:36 mihgen: there should be more of them. we need to go thoroughly through bugs and retriage all of them 16:28:41 I deffinitely saw new feature requests in bugs with feature-* tags 16:28:42 some of them miss feature tag 16:28:47 and some are duplicates 16:29:07 then we'd need to do triaging, and mark accordingly 16:29:30 alex_didenko: we'd need to kick them back to 'feature' then 16:29:37 and inform out engineers about tags 16:29:50 but I hope there are no objections on using launchpad. We already have many feature requests in there, and introducing something else like trello would be an overkill really 16:29:50 we need to gather info about tags in some single place, like wiki 16:30:01 alex_didenko: +1 16:30:01 ashtokolov is working on the process, how we will organize this work 16:30:29 I think there is wiki page for tags, at least for nailgun 16:30:42 salmon_: thats what I thought 16:31:07 who will take action for preparing this mail since it still causes confusion? 16:31:08 maximov mihgen I'm working on it 16:31:26 ashtokolov: since you are working on it, can you please then let everyone know in openstack-dev about "feature" tag, etc. ? 16:31:56 mihgen ok I got it 16:31:57 salmon_: I think it should be somewhere on this page https://wiki.openstack.org/wiki/Fuel/How_to_contribute#Test_and_report_bugs and for whole fuel 16:32:10 xarses: action item for ashtokolov .. ? 16:32:14 alex_didenko: +1 16:32:16 alex_didenko: it's on our internal wiki :/ 16:32:30 should be on wiki.openstack.org/Fuel 16:32:34 yeah, we should remove such infro from internal resources and update public wiki 16:32:39 #action ashtokolov will send out ML about usage of 'feature' tag; will start conversation about usage of other tags 16:32:51 great. moving on? 16:33:03 sorry guys for abusing half of the meeting on my topics 16:33:23 #topic https://bugs.launchpad.net/mos/7.0.x/+bug/1491576 (holser) 16:33:26 Launchpad bug 7 in Launchpad itself "Need help for novice translators" [High,Fix released] - Assigned to Данило Шеган (danilo) 16:34:18 folks, I need to get off the desktop - I will be online on the smartphone. Please do not forget to mention me 16:34:18 sgolovatiuk ^^ 16:34:23 Looks like the issue is resolved already. Is there anything else to discuss sgolovatiuk ? 16:34:39 we still use holser for sgolovatiuk ) 16:34:54 I wouls say it's workarounded 16:34:57 :) 16:35:11 randomized logrotate time? 16:35:14 mihgen: it would work fine if he has his irc client set with the nickname 16:35:25 is it backported to stable/7.0 ? 16:35:27 the problem exists - sometimes mod_wsgo seg fault on USR1 signal 16:35:35 let's use copytruncate and do not reload apache at all 16:35:46 randomized logrotate helped 16:36:14 so we will never restart all controllers at the same time 16:36:37 however, alex_didenko proposed to remove postinstall and replace it with copytruncate 16:36:49 in this case apache won't be restarted at all 16:37:12 this would be certainly better, but I'd leave it for master only. let's stop for stable/7.0.. 16:37:21 aglarendil is still investigating scale up/down issue with apache and keystone 16:37:27 k 16:37:30 we'd need to create a bug for it though so not to forget about it 16:37:38 fo 7.0 we won't be adding anything 16:37:54 we stopped playing with fantom bugs 16:38:03 phantom 16:38:14 random is enough for now 16:38:17 thnx xarses 16:38:21 sgolovatiuk: so do we have bug for 8.0 about copytruncate? 16:38:32 we have bug for 8.0 16:38:40 ok great then 16:38:45 thanks sgolovatiuk 16:38:52 #topic https://bugs.launchpad.net/fuel/7.0.x/+bug/1493483 (smakar) 16:38:55 Launchpad bug 7 in Launchpad itself "Need help for novice translators" [High,Fix released] - Assigned to Данило Шеган (danilo) 16:39:46 what is Данило Шеган ?) 16:39:58 the url parsing is bad 16:39:58 * sgolovatiuk shrugs 16:40:01 but in bot 16:40:14 it's getting fuel/7.0.x/... and thinking it's bug #7 16:40:15 bug 7 in Launchpad itself "Need help for novice translators" [High,Fix released] https://launchpad.net/bugs/7 - Assigned to Данило Шеган (danilo) 16:40:16 I just poked infra guys about it 16:40:34 :D 16:40:39 bug #1493483 16:40:41 bug 1493483 in Fuel for OpenStack "Pacemaker service is unable to be started in time when zabbix plugins enabled" [Critical,Incomplete] https://launchpad.net/bugs/1493483 - Assigned to Ksenia Demina (kdemina) 16:41:17 why it's assigned on Ksenia? 16:41:27 aglarendil: dilyin do you guys know context here? ^^ 16:41:38 we asked her to retest this case 16:41:45 or, miss that 16:41:45 without zabbiz plugin 16:42:12 so it seems like more incomplete.. ? 16:42:19 well, yes 16:42:47 bug log clearly shows error in deployment 16:42:55 which is unlikely to be related to zabbix.. 16:43:28 but we have similar tests in swarm 16:43:31 which works 16:43:42 from smakar there: "It has just been reproduced again" 16:43:49 the only difference is zabbix 16:43:57 "the root cause that pacemaker could not start on node-13 node-14" 16:44:09 it might be a problem which is not always reproduced 16:44:30 This is an old update 16:44:31 mihgen: that lab was broken also 16:44:54 I proposed to retry on clean env with recent ISO 16:45:11 yes. at least with 288 16:45:19 ok. We need a lab I think guys 16:45:29 so if she reproduces a bug, we need someone to take a look... 16:45:47 someone experienced in our HA.. 16:45:57 k 16:46:00 logs don't seem to be enough.. 16:46:23 ok.. thanks for the update on this bug. moving on? 16:46:29 yep 16:46:32 #topic https://bugs.launchpad.net/fuel/+bug/1493390 (ikalnitsky) 16:46:35 Launchpad bug 1493390 in Fuel for OpenStack 8.0.x "Deployment failed on connectivity check to http://mirror.fuel-infra.org/mos-repos/ubuntu/7.0/ after stop deployment action " [High,Confirmed] - Assigned to Fuel build team (fuel-build) 16:46:57 actually, guys, the issue is pretty simple. sometimes we face a connectivity issue in our system tests before deployment 16:47:01 unfortunately, we don't even log a response status code, and after env revert (obviously) everything's working fine. 16:47:04 i remember we had a similar issue last release. 16:47:07 so i've prepared a patch that adds 3 retries and logs response status code 16:47:10 #link https://review.openstack.org/#/c/222172/ 16:47:14 but i'm not sure about backporting to stable/7.0 16:47:17 i think it's medium if it's not often 16:47:20 is there any QA around? could someone comment how often we face it? 16:47:23 that's it :) 16:47:29 ikalnitsky: I'd say it's medium 16:47:55 so mark it as won't fix for 7.0 ? 16:48:18 ikalnitsky: we faced it about once a week, maybe a little more often 16:48:23 more over, if it is reproduced only at certain time (like stop during certain phase), then I can't really say it's high 16:48:38 It very much could be the issue with default route that we fixed 16:48:45 to me, frankly, stop of deployment in general is not a blocking from release feature 16:48:52 mihgen: we do have this check only in before deployment checks.. when you add new nodes 16:49:06 mihgeh: if our repositories are failed for customers - it isn't looks like production ready 16:49:13 but if it fails, you can retry it manually again 16:49:35 Folks Can we try to reproduce it with the latest ISO? 16:49:50 if there is an easy workaround - it's not Critical then for sure 16:50:03 and now we don't accept anything but critical bugfixes to stable/7.0 16:50:36 so I don't know why we need to even think about backporting, if it's not a critical bug .. 16:50:44 mihgen: agree 16:50:55 why cant we propose it to stable/7.0 16:51:02 then if we cut a new iso it's there 16:51:10 otherwise don't cut a new iso just for this 16:51:14 xarses: because we are in HCF 16:51:20 I think it's important to have retries here 16:51:32 xarses: I'd be rather strict 16:51:32 one shot is, well quite lame 16:51:39 and follow agreement 16:51:53 we have enough backports,which introduce regressions sometimes 16:51:57 It could be fixed by l3 clear route patch 16:52:15 xarses: it can also be targeted to updates 16:52:21 xarses: +1 to mihgen, even a trivial change introduces a chance of regressions, we have to be very conservative 16:52:24 and delivered in updates channel after release 16:52:30 +1 here 16:52:40 let's move it to updates, if someone think it's important 16:52:45 but i'd remove from 7.0 at all 16:53:13 ok great. 16:53:17 moving on? 16:53:29 #topic https://bugs.launchpad.net/fuel/+bug/1494314 (idv1985) 16:53:31 Launchpad bug 1494314 in Fuel for OpenStack "Add of detached keystone node to existing cluster failed with Call cib_apply_diff failed (-205): Update was older than existing configuration" [High,Confirmed] - Assigned to Dmitry Ilyin (idv1985) 16:55:10 this bug is because we have not refactored haproxy_ocf to use pacemaker wrappers as all other services do. Because of that cluster-haproxy task is failing when it's being run on non-controller nodes as detached keystone plugin does 16:55:58 we can either remove hardcoded primary-controller role from the task or finish the task refactoring 16:56:16 dilyin: I suppose we just fix haproxy tasj 16:56:34 I'd vote for proper fix in master, and don't backport anything to stable/7.0 now 16:56:39 actually i'm going to do both 16:56:42 It is a simple similar change that can be automatically fixed by our ci 16:56:42 since it's not critical.. 16:56:43 looks intrusive to me, do we really want to make commits like that in 7.0 before HCF? 16:56:57 It is actually critical mihgen 16:57:10 It breaks detached services feature 16:57:13 "before HCF" ? 16:57:23 err, I meant after HCF of course 16:57:31 aglarendil: for all services or just keystone? 16:57:47 And the change is similar to the same for other services 16:58:11 i'll post the quick fix to make it work and we'll be able to inish refactoring on the next release 16:58:43 I'd consider delivering of a fix in updates channel 16:59:01 2 min 16:59:05 1 16:59:06 this feature is fairly complex and requires a lot of additional work anyway 16:59:26 time 16:59:41 #endmeeting