Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

initial prometheus2 #95

Merged
merged 3 commits into from
Nov 17, 2017
Merged

initial prometheus2 #95

merged 3 commits into from
Nov 17, 2017

Conversation

auhlig
Copy link
Member

@auhlig auhlig commented Nov 10, 2017

Similar to prometheus collector: 1h retention, no persistence. Also limited to discover endpoint, pods in namespaces ["k-staging","k-eu-nl"]. Reminder to self : Do not hardcode ns.
Should I add some metrics to the k*s controllers?
Also do we want node-, openstack exporter or kube-state-metrics? @BugRoger @databus23

If you want to take a look: Prometheus2 ( collector + frontend) is running in admin cluster.

@BugRoger
Copy link
Contributor

BugRoger commented Nov 13, 2017

Let's skip the collector. The main reason for the double-Prometheus setup was that we had to aggregate the CPU metrics with the 80 CPUs blades. Should not be a problem in this case anymore.

Similar to prometheus collector: 1h retention, no persistence

For the frontend we need persistence. Similar retention should be fine.

Also limited to discover endpoint, pods in namespaces ["k-staging","k-eu-nl"]

Just collect everything.

Should I add some metrics to the k*s controllers?

yes!

Also do we want node-, openstack exporter or kube-state-metrics?

Node and Kube-State-Metrics certainly won't hurt.

@BugRoger
Copy link
Contributor

Would it be possible to ingress the Prometheus UIs to kubernikus.<region>.cloud.sap/prometheus?

@auhlig auhlig force-pushed the prometheus2 branch 2 times, most recently from b0586f3 to f398435 Compare November 13, 2017 22:08
@auhlig auhlig changed the title **WIP** initial prometheus2 initial prometheus2 Nov 13, 2017
@auhlig
Copy link
Member Author

auhlig commented Nov 13, 2017

All done. Not sure if 9b81c6c works, but I'll give it a try.
I'll create a separate PR for the metrics and the exporter.

@BugRoger
Copy link
Contributor

The config is pretty specific for the bare-metal control planes and most of the jobs don't apply here. But np, we can iterate and use it as a start.

I'm unsure about putting the chart into vendor. Wouldn't it be cleaner to just have it as a sub-chart? Having it in vendor suggests it's a generic chart that is reused multiple times.

@auhlig
Copy link
Member Author

auhlig commented Nov 14, 2017

I just copied the config. Will remove and cleanup afterwards. Thought that might be easier.
Since it will eventually differ from the generic one it makes sense to mv it out of the vendor folder.

@BugRoger
Copy link
Contributor

👍

@auhlig
Copy link
Member Author

auhlig commented Nov 15, 2017

rebased and added kube-state-metrics (vendor) after all

@auhlig
Copy link
Member Author

auhlig commented Nov 15, 2017

This is good to go from my side. Please merge @BugRoger. See: https://kubernikus.staging.cloud.sap/prometheus

@auhlig
Copy link
Member Author

auhlig commented Nov 15, 2017

Do we want dedicated channels in slack, like #kubernikus-{critical|warning|info}? Or everything to the #kubernikus ch?

@BugRoger
Copy link
Contributor

Very cool. Eventually we need to wall it off with authentication, but cool for now.

I'd prefer to have a similar channel setup. Makes it easier to tune the notification settings.

@auhlig
Copy link
Member Author

auhlig commented Nov 17, 2017

am routing and channels setup done. invited you all. if no one objects, I'm going to merge this tomorrow.
regarding authentication: simple nginx reverse proxy or maia style with keystone integration?

@auhlig auhlig merged commit 2f4e1c5 into master Nov 17, 2017
@auhlig auhlig deleted the prometheus2 branch November 17, 2017 14:21
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants