14:00:59 #startmeeting neutron_qos 14:01:00 Meeting started Wed Jun 1 14:00:59 2016 UTC and is due to finish in 60 minutes. The chair is njohnston. Information about MeetBot at http://wiki.debian.org/MeetBot. 14:01:01 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 14:01:04 The meeting name has been set to 'neutron_qos' 14:01:08 All: I am having some trouble with my network; can someone else drive the meeting? I have the agenda laid out at https://etherpad.openstack.org/p/neutron_qos_meeting_chair 14:01:37 I am signed on from my iPhone which makes for poor typing and worse copy and paste 14:01:46 sure njohnston :) 14:01:50 #chair ajo_ 14:01:51 Current chairs: ajo_ njohnston 14:02:25 not sure if the attendance is very high today :) 14:02:33 Hey 14:02:34 slaweq was not able to join 14:02:38 hi davidsha :) 14:02:39 nice 14:02:44 hello 14:03:04 hi ralonsoh :) 14:03:10 ping ihrachys :) 14:04:00 I will revert the order of bugs / RFEs if that works for you njohnston 14:04:13 By all means ajo_ 14:04:17 thanks :) 14:04:25 revert = invert 14:04:36 we have: 14:04:38 #topic Approved RFEs 14:04:38 #link http://tinyurl.com/qos-rfe-approved 14:05:05 a new approved RFE (min bandwidth egress on the hypervisor) :) 14:05:16 ralonsoh++ 14:05:21 working on it 14:05:23 and ihrachys++ as approver 14:05:30 I'll upload the last spec revision in one hour 14:05:53 regarding that, are you aware of the discussion around the ingress/egress direction field, and extended rule validation stuff? 14:06:02 yes 14:06:07 that's my question 14:06:14 there is nothing yet defined 14:06:18 ok, good, those are things that we can handle/switch at the last moment 14:06:27 before merging it, the change shall not be big 14:06:45 i'll upload only the changes for ovs 14:06:50 then LB and then sr-iov 14:06:55 For those not aware, we're talking about : 14:06:59 hi 14:07:01 I have a question about PoC for minimum bandwith. 14:07:04 #link https://bugs.launchpad.net/neutron/+bug/1586056 extended rule validation 14:07:04 Launchpad bug 1586056 in neutron "[RFE] Improved validation mechanism for QoS rules with port types" [Wishlist,In progress] - Assigned to Slawek Kaplonski (slaweq) 14:07:08 I think that linux-htb must be set to patch-tun or int-br- in https://review.openstack.org/#/c/318531/2/neutron/plugins/ml2/drivers/openvswitch/agent/extension_drivers/qos_driver.py 14:07:11 hi hichihara go ahead :) 14:07:37 hichihara i made a comment 14:07:37 ajo_: what do you think? 14:07:58 ralonsoh: I saw it but I cannot agree 14:07:58 hichihara, patch-tun does not support htb/queues unless we convert it to a veth 14:08:02 ok 14:08:03 which is less performant. 14:08:10 but I haven't looked at the code details yet 14:08:21 hichihara, that's the way I did it on my experiments 14:08:27 something else that was suggested in the ovn/ovs list 14:08:45 is that we could create the queues in the physical interfaces which we have tunnel traffic going through 14:08:50 But each taps linux-htb really work fine? 14:08:54 and then (somehow, I don't know how yet) attach those queues to ovs 14:09:13 let me look at the current code & discussion, 1 sec 14:09:25 but may be we can sort that out in the review. Probably we have enough time today. 14:09:54 ah 14:09:56 I see 14:10:06 ralonsoh, hichira is right, that doesn't work, I believe 14:10:08 ralonsoh: I'd like to run the code so please upload a code working 14:10:17 for linux htb to work... 14:10:26 Yeah! 14:10:29 you need to have some hierarchy on the queues over the same device 14:10:30 for example 14:10:46 imagine we did it on the patch port that goes to br-tun 14:10:55 (but that it was a veth, otherwise it does not work). 14:10:57 yes 14:11:14 yes, we need use veth 14:11:23 ok, I'll review this, but I don't understand your request 14:11:45 you need first, to create the top queue, where you specify the "max" of the link (this is the amount the kernel uses to make sure it substract bandwidth from other ports... beause it's reaching the "top queue" max) 14:11:50 and 14:11:59 under that queue, you add up every separate queue, for every port 14:12:00 so 14:12:02 effectively 14:12:03 you have 14:12:12 ok 14:12:16 patch-port-to-br-tun : "queue" max: 10Gbps (for example) 14:12:19 ajo_: ++ 14:12:20 under that queue (hierarchically) 14:12:38 but how can I specify the max QoS for tunnel or physical nics? 14:12:40 |_ port 1 queue , min: 1Gbps 14:12:52 |_ port 2 queue, min: 3Gbps 14:12:53 etc... 14:12:55 ok 14:13:07 otherwise the linux kernel does not know how to do it 14:13:11 let me send you a link, 1 sec 14:13:30 ajo_: Agree! 14:14:01 ralonsoh, hichihara : https://github.com/mangelajo/ovs-experiments/blob/master/qos/qos_complex_traffic_shapping.sh#L277 14:14:29 nice 14:14:40 ajo: perfect, thanks for the info 14:14:44 i'll submit a new code 14:14:46 and this link( http://horms.net/projects/openvswitch/2010-10/openvswitch.en.pdf ) helped to understand for me 14:14:50 btw 14:14:59 it seems 14:15:03 (per ovn list conversation) 14:15:10 that we don't necessarily need to create veth ports, 14:15:12 let me explain 14:15:30 for tunnels, we have the physical interface that will handle the traffic (eth0, eth1, ... etc) 14:15:44 so, we create those queues there (apparently, it's possible) 14:15:54 and then that's it 14:16:20 I wonder if we will have to do any magic trick to allow a port which is not directly attached to OVS bridge, to be used to create queues 14:16:29 if that's not possible, we will have to use veths, until that's possible 14:16:45 ajo_: ok 14:16:46 But how we know the physical interface? 14:17:10 hichihara, by looking at the tunnel endpoint IP, the default routes, and the system interfaces 14:17:33 I'm OK. We first can try your suggestion. And second plan is veth. 14:17:34 ajo_: the tunnel nic is easy 14:17:36 in the case of provider networks, it's easier, because the interface is directly attached to the br-* bridge 14:17:43 but not the physical bridge nic 14:17:50 let me look for the ovs mail list message 14:19:42 [ovs-dev] [RFC] QOS using queues in OVN 14:19:43 [ovs-dev] [RFC] QOS using queues in OVN 14:19:44 '':D 14:19:48 my copy paste is crazy 14:20:01 :) 14:20:02 https://www.mail-archive.com/dev@openvswitch.org/msg62077.html 14:20:26 ok, so let's move on the topic 14:20:47 and let's keep working on the technical details, ralonsoh thanks for tackling this blueprint! :) 14:20:59 ajo_: Thanks. I will check your link :) 14:21:15 talking again about 14:21:27 #link https://bugs.launchpad.net/neutron/+bug/1586056 extended rule validation 14:21:27 Launchpad bug 1586056 in neutron "[RFE] Improved validation mechanism for QoS rules with port types" [Wishlist,In progress] - Assigned to Slawek Kaplonski (slaweq) 14:22:02 slawek and I created a POC https://review.openstack.org/#/c/319694/ 14:22:15 it's still a very early version 14:22:28 but that would allow us to introduce new fields on the rules 14:22:43 and every mechanism driver / binding type to declare which rules & fields are supported. 14:23:01 for example: 14:23:02 https://review.openstack.org/#/c/319694/8/neutron/plugins/ml2/drivers/openvswitch/mech_driver/mech_openvswitch.py 14:23:09 supported_qos_rule_types = { 14:23:09 qos_consts.RULE_TYPE_BANDWIDTH_LIMIT: { 14:23:09 "max_kbps": [], 14:23:09 "max_burst_kbps": [] 14:23:09 }, 14:23:10 qos_consts.RULE_TYPE_DSCP_MARK: { 14:23:11 "dscp_mark": [] 14:23:15 } 14:23:17 } 14:23:23 if we added direction to bandwidth limit, we could have 14:23:39 "direction": ["egress"], 14:23:44 for sr-iov 14:23:45 and 14:23:52 "direction" : [] 14:23:58 for ovs 14:24:16 and we would add two things 14:24:20 1) validation at API call levels 14:24:40 to... rule-update, rule-create, port-update (policy), network-update (policy) 14:25:11 so, if we detect an incompatible combination of port + policy-rule(including arguments), we fail the API call with a reasonable explanation for the user/admin. 14:25:23 2) reporting 14:25:27 by introducing a new call 14:25:50 which would report the binding types (or vnic_type) to rule&argument combinations. 14:25:59 --- 14:26:15 This would enable us to incrementally enhance our rules as necessary 14:26:45 while we support the heterogeneous nature of the technologies we support 14:26:52 (support support... 'X) 14:26:59 --- 14:27:02 the other option would be 14:27:13 we rename the bandwidth-limit rule to egress-bandwidth-limit 14:27:19 and introduce a new ingress-bandwidth-limit 14:27:37 and, same thing goes to all other directional rules (like the min bw ralonsoh is working on) 14:27:40 ------ 14:27:52 but we would hit the same problem the day we introduce support for traffic classifiers... 14:28:12 if we attach a classifier to rules, ... some technologies will support that, some will not 14:28:21 ---- 14:28:29 ralonsoh, njohnston , hichihara , ihrachys thoughts? 14:29:00 I need to review the code and ground any idea... 14:29:18 ^^ what he said :-) 14:29:22 lol ':D 14:29:33 heavy duty stuff, sorry guys :D 14:29:39 But I think the first option (with directions) is better 14:29:46 if you have time to review the RFE and the ideas... 14:29:52 I mean: have a parameter with the direction 14:30:02 yes, that's the basis, if we believe that extending rules that way is more healthy... then that's the way 14:30:04 I believe that 14:30:23 all the above stuff boils down to... 14:30:41 imagine that you have a network with 4 ports, 2 OVS ports, and 2 SR-IOV ports 14:30:56 and that network has a policy "A", with a bw limit rule on egress, 10Mbps 14:31:04 admin comes via api and tries 14:31:18 neutron qos-policy-rule-update "A" --direction ingress 14:31:25 it should fail with a meaningful message 14:31:46 "port X binded by sr-iov, does not support direction: ingress for this rule.." 14:32:06 I should have started by that ':D 14:32:18 hehehe agree! 14:32:37 I agree the arrangement. QoS is complexed for me. We need to make it more uniform. 14:33:22 thanks hichihara , yes, it's indeed a complex topic, because not all technologies can implement the same features. 14:33:46 and if we stick to the minimum subset, we just ended with DSCP & max bw egress... :) 14:34:17 ok, let's move on 14:34:23 #topic non approved RFE's 14:34:25 https://bugs.launchpad.net/neutron/+bugs?field.tag=qos+rfe+&field.tags_combinator=ALL 14:34:30 let's see if we have something new around 14:35:08 it looks like not 14:35:15 ok, 14:35:50 Can the traffic classification RFE be moved out of incomplete status? 14:35:51 let me link this tangential RFE: https://bugs.launchpad.net/neutron/+bug/1563967 14:35:51 Launchpad bug 1563967 in neutron "[RFE] Extend l2-extensions-api for flow management " [Wishlist,Triaged] - Assigned to Miguel Angel Ajo (mangelajo) 14:36:02 njohnston, let me see, sec 14:38:37 so they moved it to incomplete because we were prioritizing other things: https://bugs.launchpad.net/neutron/+bug/1563967 14:38:37 Launchpad bug 1563967 in neutron "[RFE] Extend l2-extensions-api for flow management " [Wishlist,Triaged] - Assigned to Miguel Angel Ajo (mangelajo) 14:38:37 sorry 14:38:42 http://eavesdrop.openstack.org/meetings/neutron_drivers/2016/neutron_drivers.2016-03-24-22.01.log.html#l-104 14:39:08 the RFE is bound to having classifiers, and I guess we need to solve that first 14:39:20 davidsha, how do you see it ? 14:40:17 ajo_: I agree, priority queuing requires traffic to be classified and funneled into 4 queues, so the Common classifier is a requirment 14:40:57 Ok got it 14:41:07 davidsha, may be we could move it back to Wishlist, but I wonder how do we do if it's dependent into classifiers 14:41:20 I wouldn't make sense to get it approved without the dependency first 14:41:31 davidsha, any advance on the classifier front ? I've lost track of that thread 14:43:01 ajo_: the rfe was filed, thats all thats been done so far 14:43:54 davidsha, ok, let's keep tracking it 14:44:06 ajo_: there is an old spec for neutron_classifier that we may refurbish for this. 14:44:13 :) 14:44:50 davidsha, about the L2 flow extensibility: https://bugs.launchpad.net/neutron/+bug/1563967 any thoughts which could be important? reviews on the spec are welcomed from anywone: 14:44:50 Launchpad bug 1563967 in neutron "[RFE] Extend l2-extensions-api for flow management " [Wishlist,Triaged] - Assigned to Miguel Angel Ajo (mangelajo) 14:45:08 ajo_: I have the PoC done, I'll put it up now. 14:45:11 https://review.openstack.org/#/c/320439/ 14:45:18 but I know it's a heavy duty topic too 14:45:22 davidsha++ 14:45:41 I think our highest priority things now shall be the extended validation and this, 14:46:03 first is a blocker for other rfes to merge 14:46:24 and this one, not 100% a blocker, but it will be important for QoS to interoperate with other extensions 14:46:26 gracefully 14:47:16 probably we should move to bugs 14:47:23 kk, there aren't any tests worth mentioning yet, but I've set up dscp to use it and make sure it can work 14:47:39 niiice 14:47:53 looking forward to see it 14:48:11 #topic bugs 14:48:14 https://bugs.launchpad.net/openstack/+bugs?field.searchtext=&orderby=-importance&field.status:list=NEW&field.status:list=OPINION&field.status:list=EXPIRED&field.status:list=CONFIRMED&field.status:list=TRIAGED&field.status:list=INPROGRESS&field.status:list=FIXCOMMITTED&field.status:list=INCOMPLETE_WITH_RESPONSE&field.status:list=INCOMPLETE_WITHOUT_RESPONSE&assignee_option=any&field.assignee=&field.bug_reporter=&field.bug_commenter=&fie 14:48:14 ld.subscriber=&field.structural_subscriber=&field.tag=qos+-rfe&field.tags_combinator=ALL&field.has_cve.used=&field.omit_dupes.used=&field.omit_dupes=on&field.affects_me.used=&field.has_patch.used=&field.has_branches.used=&field.has_branches=on&field.has_no_branches.used=&field.has_no_branches=on&field.has_blueprints.used=&field.has_blueprints=on&field.has_no_blueprints.used=&field.has_no_blueprints=on&search=Search 14:48:18 woops sorry 14:48:25 Whoa 14:48:28 #link http://bit.ly/1WhXlzm 14:48:40 thank god njohnston had a shorter version of it :) 14:48:58 I'm working on : https://bugs.launchpad.net/networking-ovn/+bug/1584204 14:48:58 Launchpad bug 1584204 in neutron "VersionsCallbackNotFound exception when using QoS" [Undecided,In progress] - Assigned to John Kasperski (jckasper) 14:49:10 it shall be almost there 14:49:31 njohnston, how's it going with: https://bugs.launchpad.net/neutron/+bug/1486607 14:49:31 Launchpad bug 1486607 in neutron "tenants seem like they were able to detach admin enforced QoS policies from ports or networks" [Medium,In progress] - Assigned to Nate Johnston (nate-johnston) 14:49:45 probably you're waiting for review.... :'( 14:49:56 I think it just needs +2s 14:50:28 https://review.openstack.org/#/c/217092/ 14:50:30 the patch, 14:50:43 yes I remember the complications because of the no exchangeability of policy.json to check it 14:50:51 I will prioritize that again, sorry for the delay 14:50:57 Thanks! 14:51:27 thank you! :D 14:52:19 the pecan bug https://bugs.launchpad.net/neutron/+bug/1568400 could be fixed by: https://review.openstack.org/#/c/314886/ 14:52:19 Launchpad bug 1568400 in neutron "Pecan does not route to QoS extension" [Medium,In progress] - Assigned to Brandon Logan (brandon-logan) 14:52:22 we probably need to check 14:53:51 the patch for https://bugs.launchpad.net/neutron/+bug/1515564 needs to be rebuilt on top of the great refactor by mfranc213 :D 14:53:51 Launchpad bug 1515564 in neutron "Internal server error when running qos-bandwidth-limit-rule-update as a tenant" [Low,In progress] - Assigned to Liyingjun (liyingjun) 14:54:16 mfranc213, did a good job with the qos_plugin.py refactor 14:54:40 ahh 14:54:42 there's a new version 14:54:48 in fact it seems like it's already been rebased 14:55:04 so, adding to my queue too 14:56:31 ajo_: Before end meeting, can I ask you about Strict minimum bandwidth? 14:56:41 https://bugs.launchpad.net/networking-qos/+bug/1585373 14:56:41 Launchpad bug 1585373 in networking-qos "qos-policy update without specify --shared causing it change to default False" [Undecided,Confirmed] - Assigned to ugvddm (271025598-9) 14:56:48 this was on the wrong project ^fixing 14:57:00 hichihara, ah, that one depends on the generic resource pool implementation at the nova side 14:57:08 I suspect we could implement it on next cycle 14:57:16 However, Do we keep to pray these good progress only? 14:57:21 based on the mailing list discussions (linked form the RFE) 14:57:23 Don't we need to propose nova spec about new resource(NIC_BW?) for generic resource pool like SRIOV? 14:58:07 hichihara, not sure if those will need to be declared in code, or via API, hichihara do you have time to check that with Jaypipes ? 14:58:23 it's good to keep an eye on that, 14:58:28 ajo_: OK. 14:58:37 also, about non-strict, we may need to look at SR-IOV 14:59:01 I think it's rather easy to do, but we may need to ask the intel or melanox guys about the link settings for min bw, etc 14:59:05 ralonsoh, ^ 14:59:08 ajo_: I will consider what we need. 14:59:11 and may be loop somebody else to help with that 14:59:13 I know 14:59:16 thanks :) 14:59:24 I suspect we should endmeeting 14:59:30 bye! 14:59:35 so, again thanks a lot everybody 14:59:35 Thanks! 14:59:38 bye! 14:59:43 o/ :) 14:59:51 #endmeeting