21:00:11 #startmeeting containers 21:00:12 Meeting started Tue Oct 23 21:00:11 2018 UTC and is due to finish in 60 minutes. The chair is strigazi. Information about MeetBot at http://wiki.debian.org/MeetBot. 21:00:14 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 21:00:16 The meeting name has been set to 'containers' 21:00:19 #topic Roll Call 21:00:29 o/ 21:00:32 o/ 21:01:17 o/ 21:02:47 o/ 21:03:14 brtknr: you here? 21:03:17 o/ 21:03:53 Thanks for joining the meeting ttsioutsflwang imdigitaljim eandersson 21:03:59 Agenda: 21:04:05 #link https://wiki.openstack.org/wiki/Meetings/Containers#Agenda_for_2018-10-23_2100_UTC 21:04:16 it has some items 21:04:27 #topic Stories/Tasks 21:04:51 1. node groups https://review.openstack.org/#/c/607363/ 21:05:13 I think we are pretty close to the final state of the spec 21:05:21 please take a look 21:05:44 strigazi: tomorrow I will push again 21:06:02 adapting ricardo's comments 21:06:41 oh I thought you pushed today, ok so take a looks guys take a look tmr as well :) 21:07:22 :) tmr better 21:07:24 @all do you want to discuss anything about nodegroups today? 21:07:46 questions about nodegroups? 21:08:23 ok, next then 21:08:26 o/ sorry for lateness, but yes 21:09:06 schaney: hello. you have smth about nodegroups? 21:09:08 any mechanism to interact with NGs individually? 21:09:29 opposed to the top level cluster or stack 21:09:57 the api will be like: 21:10:01 schaney: do you mean updating a specific nodegroup? 21:10:20 yes for scaling or the like 21:10:33 cluster//nodegroup/ 21:10:48 so PATCH cluster//nodegroup/ 21:10:48 https://review.openstack.org/#/c/607363/2/specs/stein/magnum-nodegroups.rst@117 21:11:37 sorry i'm late! 21:11:49 colin-: welcome 21:12:44 oh gotcha, ill have to dig into the work a bit, under the hood magnum just targets the name of the node group represented by the heat parameter though? 21:13:05 ttsiouts: the node groups is basically the same thing like node pool in GKE, right? 21:13:32 schaney: that's the plan 21:13:48 flwang: exactly 21:14:05 ttsiouts: cool 21:14:37 i see, did there happen to be any work with the "random node scale down" when magnum shrinks the cluster? 21:14:40 ttsiouts: i will review the spec first 21:14:56 from the API route it would seem so? 21:14:59 i think we probably better read the spec first and put comments in the code 21:15:07 instead of discussing design details here 21:15:12 good call 21:16:05 schaney: we want to add a CLI for removing specific nodes from the cluster 21:16:18 but this will come further down the way 21:16:21 schaney: this won't be covered by this spec, but we should track it somewhere else 21:16:39 gotcha, thanks for the info 21:16:48 flwang: thanks! 21:17:05 flwang: tmr it will be more complete 21:17:17 were looking at modifying our driver to perhaps consume senlin clusters for each group 21:17:39 masters, minions-group-1, ... minions-group-n 21:17:49 imdigitaljim: is it a hard depedency? 21:17:51 would all have a senlin profile/cluster in heat 21:17:53 i mean for senlin 21:18:01 imdigitaljim: that could be done, that is why we have drivers 21:18:14 flwang: should be optional 21:18:34 like alternative 21:18:41 it would probably be the driver would take on a senlin dependency (not magnum as a whole) 21:18:44 just like octavia or not 21:19:08 when the cluster drivers where proposed senlin and ansble were the reasoning behind it 21:19:20 we're more focused on autoscaling/autohealing rather than cli manual scaling 21:20:02 the senlin PTL is here and is actively talking to the heat PTL on managing the senlin resources in heat so we might be able to inhouse create a better opportunity for senlin + heat + magnum 21:20:47 imdigitaljim: here, like in this meeting? 21:20:51 no 21:21:00 sorry i just mean he works at blizzard 21:22:19 This plan is compatible with nodegroups and nodegroups make it actually easier 21:22:55 we think so 21:23:04 I'm not aware of it in detail, but it sounds doable 21:23:17 Senlin would work well with the NG layout, one thing to note is Senlin's dedicated API 21:23:26 yeah we arent either but we'll be working out over the next couple weeks 21:23:50 and see if its within reason on feasibility 21:23:50 The Senlin PTL will be in Berlin btw 21:23:53 ^ 21:24:06 cool 21:24:16 Its honestly too early to be talking about, there's a lot of heat/senlin ground work to do first 21:24:32 But hey, it's a thing we're thinking about. 21:24:43 fair enough 21:25:38 shall we move on? 21:26:01 yes 21:26:30 I pushed two patches we were talking about, one is: 21:26:41 Add heat_container_agent_tag label https://review.openstack.org/#/c/612727 21:27:20 we discussed with flwang and eandersson already, others have a look too 21:27:52 the tag of the heat-agent was hardcoded this makes it a label. 21:28:21 the 2nd one needs some discussion, it is 21:28:30 deploy tiller by default https://review.openstack.org/#/c/612336/ 21:29:30 Shall we have it in by default or optional? 21:30:39 strigazi: any potential issue if we enable it by default? 21:30:41 and then next steps are, with tls or without? a tiller per ns or with the cluster-role? 21:31:16 flwang: the user might want a diffrent tiller config 21:31:25 strigazi: that's the problem i think 21:31:34 flwang: other than that, tiller will be there silent 21:32:16 we have seen similar 'issue' with other enabling, like the keystone auth integration 21:32:17 flwang: so you are in-favor of optional 21:32:38 a new enabling feature may introduce a bunch of config 21:32:51 but now in mangum, we don't have a good way to maintain those config 21:33:08 labels are too flexible and lose 21:33:14 i prefer optional 21:33:38 based on the feedback we got so far, most of customer just want a vanilla k8s cluster 21:34:05 if they want something, they can DIY 21:34:26 what vanilla means? api, sch, cm, kubelet, proxy, dns, cni 21:34:38 and i agree it's because we(catalyst cloud) are public cloud and our customers' requirements are vary 21:34:57 vanilla means a pure cluster, not too much plugins/addons 21:35:11 flwang: We've been getting some similar feedback from power users too 21:35:17 for private cloud, the thing maybe different 21:35:40 but I think that's expected from power users that are used to DYI 21:35:48 most of the customers of k8s know how to play with it 21:36:13 what they want is just a stable k8s cluster with good HA and integration with the underhood cloud provider 21:36:48 cbrumm__: what do you mean 'power users'? 21:37:00 so optional 21:37:21 people who've used k8s before that was not a managed service 21:37:34 cbrumm__: ok, i see, thx 21:38:24 flwang: any argument against having this optional? 21:39:07 strigazi: TBH, I would suggest we start to define a better addon architecture 21:39:10 ^ 21:39:29 like refactor the labels 21:39:37 thats one of the goals our driver plans to be solving 21:39:53 imdigitaljim: show me the code ;) 21:40:21 so we don't add anything until we refactor? 21:40:25 i have heard about the v2 driver million times, i want to see the code :D 21:40:33 strigazi: i'm not saying that 21:40:38 :D when i get some free cycles and feel like its in a great uploading spot 21:40:42 i just say we should be more careful 21:42:49 the current model is not unreasonable. we need to define careful and not make the service a framework 21:43:47 I think for v1 which can be refactored only replaced/deprecated the model of addons is there 21:44:00 s/can/can't/ 21:44:27 labels for on/off and tags 21:44:57 imo we should look into a config file 21:45:13 kubernetes followed the same pattern when they realized there were too many flags 21:45:38 config file to create the labels/fields or a config file to pass to the cluster 21:45:42 ? 21:46:12 Merged openstack/magnum stable/queens: Provide a region to the K8S Fedora Atomic config https://review.openstack.org/612199 21:46:19 openstack coe cluster template create --config myvalues.yaml 21:46:48 these values are like labels? 21:46:53 could be everything 21:46:57 could be just labels 21:47:02 code too? 21:47:13 like to code? 21:47:19 ? 21:47:20 or link to code? 21:47:53 instead of like --labels octavia_enabled=true, etc etc 21:47:56 it could be 21:48:03 [Loadbalancer} 21:48:09 [LoadBalancer] 21:48:12 got it 21:48:15 octavia_enabled=true 21:48:27 but you could also do 21:48:29 [Network] 21:48:55 floating_ ... fixed_network= .. fixed_subnet= 21:48:58 yaml or the ini format 21:49:02 either or 21:49:02 any 21:49:04 json 21:49:09 yep 21:49:09 doesnt matter however we want to do it 21:49:18 agree 21:49:21 imo im a fan of that model 21:49:43 and with that case, we can publish sample config files 21:49:48 flwang: what would cover your concern about the loose label design? 21:49:52 and user can decide how to combine the config 21:50:11 strigazi: yep, that's basically the arch in my mind 21:51:05 #action strigazi to draft a spec and story for creating cluster with a config file. 21:51:13 strigazi: can we discuss this one https://github.com/kubernetes/cloud-provider-openstack/issues/280 next? 21:51:30 I'll try to bring smth in the next meeting for this 21:51:39 strigazi: thanks 21:52:24 before going into the bug of CPO 21:53:02 @all take a look to the rest of the list of review in the agenda. they are ready to go in 21:53:07 https://wiki.openstack.org/wiki/Meetings/Containers#Agenda_for_2018-10-23_2100_UTC 21:53:18 will do 21:54:00 flwang: what might help a little is using the config drive before the metadata service 21:54:05 colin-: thx 21:55:03 strigazi: does that need change in magnum? 21:55:14 imdigitaljim: did you ever see this issue https://github.com/kubernetes/cloud-provider-openstack/issues/280 ? 21:55:16 imdigitaljim: eandersson colin- what is your experience with CPO 21:55:31 worker nodes are missing ips 21:55:35 i have not experienced this issue 21:55:36 flwang: in the config of the CPO 21:55:48 imdigitaljim: probably because you're using v1.12? 21:55:57 we have an internal downstream with a few patches on CPO 21:56:06 were waiting on upstream commit permission for kubernetes org 21:56:12 (blizzard admin stuff) 21:56:20 we're on 1.12.1 correct 21:56:24 patches regarding this bug? 21:56:30 no 21:56:51 UDP support, LBaaS naming, and a LBaaS edge case 21:56:59 https://github.com/kubernetes/kubernetes/pull/65226#issuecomment-431933545 21:57:04 to jim, outloud just now i said "it's been better than starting from scratch"? :) 21:57:19 found it useful and it has definitely saved us some time, but as he said we've also found some gaps we want to address 21:57:34 it seems like a very common, high-chance problem 21:58:10 colin-: you talk about k/cpo? 21:58:15 yes 21:58:34 strigazi: when you say 'config of CPO', does that mean we at least have to use cm+CPO mode? 21:59:27 flwang: https://github.com/openstack/magnum/blob/master/magnum/drivers/common/templates/kubernetes/fragments/write-kube-os-config.sh#L12 22:00:08 it is adding [Metadata]search-order=configDriver,metadataService 22:01:11 strigazi: got it. but based on https://github.com/kubernetes/cloud-provider-openstack/issues/280#issuecomment-427416908 22:01:57 does that mean we only need add this one [Metadata]search-order=configDriver,metadataService ? 22:02:25 flwang: the code is... even if you disable set [Metadata]search-order=configDrive it still calls the metadataservice 22:03:11 I tried with configdrive only and it was still doing calls to the APIs. 22:03:30 i'm confused 22:03:56 does that need any change in nova? 22:04:04 i mean nova config 22:04:23 I'll end the meeting to be ~1 hour and we continue 22:04:33 cool, thanks 22:04:40 Merged openstack/magnum master: Minor fixes to re-align with Ironic https://review.openstack.org/612748 22:04:53 @all thanks for joining the meeting 22:05:02 see you next week 22:05:22 #endmeeting