-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[2.4.2] [Bug] Monitoring is not deployed correctly on new cluster creation. #26440
Comments
same here |
Having same issues |
The bug is reproduced in v2.4.2 single install when adding a cluster with cluster monitoring enabled by editing the cluster as YAML file Rancher logs
Workaround: More Info: |
The issue is also reproduced when creating a cluster using cluster template with monitoring enabled. [While creating a cluster template, enable monitoring by editing the yaml and setting
|
Seems an issue on metricserver that impacts Helm: helm/helm#6361 Then we need to solve the issue on metricserver api returning false (FailedDiscoveryCheck)
|
The helm bug is in 2.16 till 2.16.3 and there is a fix for the upstream helm bug in helm 2.16.5 |
The bug fix is validated in v2.4-head 63a490f and master-head 00ff159 Steps:
Results:
Update: |
@jiaqiluo how much time does it take for metrics to show up if monitoring is enabled after the cluster is active? Is the time similar for a 2.3 setup if you enable monitoring during cluster create? |
@mrajashree it is usually about 3 to 5 mins wait for the apps to be active, then another 3 to 5 mins for the metics showing up. |
Validated the fix on master-head(00ff159) and v2.4-head(cf5ab1d)
In both master-head and v2.4-head it took around 4 minutes for the monitoring app to be active and around 5 minutes for the metrics to show up after the app was active. |
Here is the difference I noticed: When enabling the cluster monitoring in the provisioning configuration, the UI will keep showing the While if the cluster monitoring is enabled after the cluster is active, the |
@jiaqiluo If this is the same behavior you see for 2.3 setup then it's fine |
Why is this reopened? monitoring does work. If the delay is not same for a 2.3 setup, we can open a new issue, but now the monitoring app no longer fails to deploy |
this bug is confirmed to be fixed. different behavior is observed in v2.3.6, so a new issue is made to track that: #26692 |
What kind of request is this (question/bug/enhancement/feature request):
Bug
Steps to reproduce (least amount of steps as possible):
Create a new cluster using the terraform provider (might not be relevant) with monitoring enabled.
Result:
The monitoring apps do not deploy correctly. Looking in the clusters System Apps both show this error:
Other details that may be helpful:
Here are the rancher logs. It might be they are created before the CRDs are installed by the agent? If I kick them off again by forcing an upgrade they seem to install fine so seems like a timing issue. Seems like the monitoring-operator does automatically get kicked off again and deploys correctly, not sure if cluster-monitoring will.
Environment information
rancher/rancher
/rancher/server
image tag or shown bottom left in the UI):Cluster information
kubectl version
): 1.17The text was updated successfully, but these errors were encountered: