-
Notifications
You must be signed in to change notification settings - Fork 243
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] crc start error doesn't exit with non-zero exit code #4284
Comments
@bobbygryzynger this we display as error but actually it is mostly a warning (we might change the messaging) sometime it happens that due to slow IO or cert regenerate it takes more than expected time for operator reconciliation but that's doesn't mean the cluster is completely useless. |
Thanks @praveenkumar, understood. For this particular error, the cluster was unstable after this. A suggestion: make anything that's error level exit with a non-zero. This particular issue could just be a warning, but my experience with it suggests it should still be an error. Maybe it could remain an error with a suggestion logged on how to increase the timeout (if that's possible). |
@bobbygryzynger Because for you the operator never become stable? |
@praveenkumar, that's right. When I saw this, even after waiting a bit, the operators were still unstable. |
In that case, if you able to access the kube api then try to use https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-operator-issues.html to see why an operator is not stable. |
@praveenkumar the operators being unstable is really a separate issue from what I'm requesting here. All I'd really like to see here is that when errors occur, a non-zero code is produced so that my scripts can pick up on that. |
@bobbygryzynger Right and as I said what you observe as error should be a warning and it is issue with our messaging apart from that we are already returning non-zero exit code when error happen. |
It seems to me that the cluster being unstable is worthy of an error, but I won't belabor the point. |
@bobbygryzynger I am reopening this issue since we didn't make change in the messaging yet.
At some point I do agree but if we error out and not execute next steps which actually update the kubeconfig with user context then there is no easy way to access the cluster api to debug which clusteroperator is behaving differently and that's the reason I consider this as warning and let user use the API for debugging or may be the operator which is misbehaving is not even required by user then this can be ignored. |
I think the request is that when there is this "cluster not ready" message, after |
General information
crc setup
before starting it (Yes/No)? YesCRC version
CRC status
$ crc status --log-level debug DEBU CRC version: 2.22.1+e8068b4 DEBU OpenShift version: 4.13.3 DEBU Podman version: 4.4.4 DEBU Running 'crc status' CRC VM: Running OpenShift: Starting (v4.13.3) RAM Usage: 9.412GB of 16.77GB Disk Usage: 20.8GB of 79.93GB (Inside the CRC VM) Cache Usage: 202.3GB Cache Directory: /home/bgryzyng/.crc/cache
CRC config
Host Operating System
$ cat /etc/os-release NAME="Fedora Linux" ...
Steps to reproduce
crc start
Expected
If an error occurs, exit code should be 1 or greater
Actual
Exit code is zero
Logs
The text was updated successfully, but these errors were encountered: