15:00:24 #startmeeting monasca 15:00:25 Meeting started Wed Dec 5 15:00:24 2018 UTC and is due to finish in 60 minutes. The chair is witek. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:26 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:28 The meeting name has been set to 'monasca' 15:00:36 Hello everyone! 15:01:25 Hey all 15:01:32 hello 15:01:39 hi dougsz and joadavis 15:02:18 I've created last minute agenda 15:02:26 https://etherpad.openstack.org/p/monasca-team-meeting-agenda 15:02:50 #topic Events Listener 15:02:56 https://review.openstack.org/583803 15:03:03 Thanks again for the reviews 15:03:20 I haven't groked all the nova cells concepts yet 15:03:39 we have some constructive discussion in review 15:04:05 I am pretty sketchy myself - I've mostly read the docs so far 15:04:58 It's something we are adding support for in Kolla, and it would be nice to make sure it works well with the Events listener 15:05:12 yes, agree 15:06:33 For the first implementation, if we can get all the messages sent to the kafka monevents topic that may work 15:07:17 I don't know if we need to separate events based on which cell they come from. Assuming there is some data in the dimensions that would identify the source cell for the event 15:08:14 the issue dougsz is bringing up is, that there might be no connectivity between Cell message queue and Monasca cluster 15:08:46 I had thought we could just use existing code for connecting to rabbitmq, but will need to do a quick check if it can handle multiple rabbitmq connections. But if there is no network connection allowed, that is a problem. 15:09:56 ideas for options if there is no connectivity? 15:11:26 I was thinking this might be a use case where posting events to the Monasca Events API would work well 15:13:00 yeah, that might be the way. Would we need a monasca service in each cell to listen to the cell's RabbitMQ for events then post them to the Events API? 15:13:13 I'd suggest to seek some more expert knowledge on Nova Cells 15:13:24 That was the way I imagined it could work. 15:13:25 monasca service or modified monasca agent 15:13:26 how they are supposed to be deployed 15:13:38 yes 15:14:02 We might be able to get some info from Cern 15:14:07 I will ask around 15:14:12 thanks 15:14:15 good idea, thanks 15:15:02 let's continue the discussion in review 15:15:36 the last unclarified question seems to be, how the listener is supposed to be delpoyed 15:16:18 as part of API, or as separate new Monasca service 15:16:52 no, actually not the only one 15:17:08 should it send to Kafka, or to Events API? 15:18:22 #topic Python 3 support 15:18:34 sending to Kafka seems more efficient. Though we can use the API if needed for cells 15:19:06 I took a quick look at the py3 cassandra failure you mentioned in the agenda 15:19:27 That is a strange tempest error, and not obviously related to cassandra 15:19:52 but I pinged sgrasley and jgu to see if they know of any issues with cassandra clients and py3 15:20:43 adriancz has mentioned issues with string handling 15:21:18 he has set up CI jobs for running tempest tests with Python 3 15:21:22 https://review.openstack.org/619987 15:22:47 the jobs are marked as non-voting for now, but we should fix them as soon as possible 15:23:02 thanks joadavis for forwarding this 15:24:04 agreed. (though I wish all zuul tests were more predictable - we still get many random failures) 15:24:19 (but I'll stop complaining now) 15:24:50 :) 15:25:24 I have observed long queues of Zuul jobs recently 15:26:00 at some point it has taken several hours before the jobs were even started 15:26:14 I tried initiating 3 rechecks over in ceilometer repos yesterday for different random failures 15:26:57 :( 15:27:12 did it improve anything? 15:27:40 at least one passed (it was the mailing list change). I need to read my email and see about the rest 15:28:33 I think something changed in Zuul. in the past when you did a recheck it would post back a message indicating it started the gates, but now I don't see one until it finishes 15:29:23 but you can check the status on http://zuul.openstack.org/status 15:29:31 yes 15:29:49 let's move on 15:30:06 #topic project health (feedback for TC) 15:30:32 gmann has contacted me and asked for feedback regarding Monasca project health 15:30:52 TC tries to track potential issues in projects 15:31:01 https://wiki.openstack.org/wiki/OpenStack_health_tracker 15:31:48 he has asked among others if there is anything TC could do to support our project? 15:33:18 please let me know, if you have ideas 15:33:48 #topic Trove + Monasca + Vitrage + Mistral 15:33:50 For us, I think we have a relatively healthy community, though would always like more contributors (as all the projects say) 15:34:25 joadavis: agree, we have very few active contributors 15:34:26 I think the TC could help in coordinating and supporting other services that consume monasca data, such as watcher. Or vitrage 15:34:42 and the SIGs help with that too 15:34:52 and that transitions to our next topic. :) 15:36:29 thanks joadavis, I thinks we've said something similar to dhellmann during the Denver PTG 15:37:08 dumb question: is Trove still around? I hadn't heard about it in a year or so 15:37:32 OK, let's move to the Trove topic 15:37:48 I think the project was sort of abandoned 15:37:57 and then Samsung jumped in 15:39:01 they have given a presentation in Berlin 15:39:11 I recall Bartosz Żurkowski asking about using Monasca to monitor it before the Berlin summit 15:39:14 https://www.openstack.org/summit/berlin-2018/summit-schedule/events/22273/towards-production-grade-database-as-a-service-in-openstack 15:39:20 ( a Trove core) 15:39:31 Not sure where he got to 15:39:48 Ah, you just answered that :) 15:40:00 I have missed that presentation 15:40:32 but they presented POC for integration between Trove, Monasca, Vitrage and Mistral 15:40:42 cool 15:41:10 a draft for Monasca datasource in Vitrage has been pushed just today 15:41:18 https://review.openstack.org/622899 15:41:52 Ifat has reported this in the self-healing meeting today 15:42:20 and they are willing to help to get this merged 15:43:16 Do they need any work on the Monasca side for this? other than configuring a webhook to publish to Vitrage? 15:43:51 I think not 15:44:58 we have discussed an option for sending Monasca notifications via message queue 15:45:09 which is kind of default in Vitrage 15:45:33 but I think it's not necessary at the end 15:46:24 that's all from my side 15:46:46 I didn't come to merging notification changes 15:47:12 but I pushed the Alembic devstack change, to have it all complete 15:47:22 turned out to be pretty easy 15:47:31 https://review.openstack.org/622361 15:48:03 thanks witek 15:48:19 I filed two new stories yesterday. One is a simple cleanup of some ceilometer monitoring in the agent. 15:48:43 The other is https://storyboard.openstack.org/#!/story/2004539 and might take a bit more work 15:49:15 I don't know if removing an alarm configuration for http_alive was supported well. 15:49:31 I may follow up with sc on this to see if he has any ideas. 15:50:08 alarm or collector configuration? 15:50:33 sorry, for the agent collector configuration 15:51:23 will you or someone from SUSE work on this? 15:51:36 the use case is that a compute host has been removed, so we want to do cleanup. but at this point, we just know the hostname and not all the target_hostnames used in setting up the configuration 15:51:45 I'm working on it currently 15:51:53 OK, thanks 15:52:33 I'm sharing in case anyone has insight in to the agent. 15:52:49 otherwise, that is enough from me 15:52:51 sorry, haven't hit this use case yet 15:53:39 we put a lot of work in to getting a cloud up, not as much in to removing parts of it. :) 15:53:47 :) 15:55:21 OK, I'll be wrapping up if there is nothing else 15:55:36 thank you for joining 15:55:50 and see you next week 15:56:05 for interested 15:56:05 goodbye 15:56:19 second round of self-healing meeting starts in one hour 15:56:37 in #openstack-self-healing 15:56:47 bye bye 15:56:48 bye all 15:56:53 #endmeeting