GUI acceptance tests using environment deployed from packages.

Build: #2108 failed

Job: Onezone data discovery failed

Stages & jobs

  1. Acceptance Test

Job result summary

Completed
Duration
8 minutes
Revision
616ef003a0a36dc00e59f722ca7550385601e480
Fixed in
#2109 (Child of ODSRV-OZP-1738)
No failed test found. A possible compilation error occurred.

Error summary

The build generated some errors. See the full build log for more details.

  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  42300      0 --:--:-- --:--:-- --:--:-- 42300
Error: Kubernetes cluster unreachable: Get "http://localhost:8080/version?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
The connection to the server localhost:8080 was refused - did you specify the right host or port?
The connection to the server localhost:8080 was refused - did you specify the right host or port?
The connection to the server localhost:8080 was refused - did you specify the right host or port?
The connection to the server localhost:8080 was refused - did you specify the right host or port?
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: Cannot kill container: 56fa77a10043: Container 56fa77a100430c284acd3af277198cf494e85cf823e7b112097bd49052405b26 is not running
Error response from daemon: Cannot kill container: 70b3c7053a89: Container 70b3c7053a891422f3be2eac2bf211d5239951889d5b1c5bc8d9d0b1fc73f339 is not running
Error response from daemon: Cannot kill container: 3b91aa950914: Container 3b91aa95091411e2e347cb2eb4f58567a12f8c7648c81c2d5630ad24655359fd is not running
Error response from daemon: Cannot kill container: edcdd16a01b5: Container edcdd16a01b5e11a90b063ece50052b3f40b91fed88c3224820b18168b33b428 is not running
Error response from daemon: Cannot kill container: 3c614a804f5c: Container 3c614a804f5c485f643e9e2c046249d4c725e876420cac4736892658aac586fe is not running
                timed out waiting for the condition

        This error is likely caused by:
                - The kubelet is not running
                - The kubelet is unhealthy due to a misconfiguration of the node in some way (required cgroups disabled)

        If you are on a systemd-powered system, you can try to troubleshoot the error with the following commands:
                - 'systemctl status kubelet'
                - 'journalctl -xeu kubelet'

        Additionally, a control plane component may have crashed or exited when started by the container runtime.
        To troubleshoot, list all containers using your preferred container runtimes CLI.

        Here is one example how you may list all Kubernetes containers running in docker:
                - 'docker ps -a | grep kube | grep -v pause'
                Once you have found the failing container, you can inspect its logs with:
                - 'docker logs CONTAINERID'


stderr:
        [WARNING IsDockerSystemdCheck]: detected "cgroupfs" as the Docker cgroup driver. The recommended driver is "systemd". Please follow the guide at https://kubernetes.io/docs/setup/cri/
        [WARNING FileExisting-socat]: socat not found in system path
        [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
error execution phase wait-control-plane: couldn't initialize a Kubernetes cluster
To see the stack trace of this error execute with --v=5 or higher

*
╭──────────────────────────────────────────────────────────────────╮
│                                                                  │
│    * If the above advice does not help, please let us know:      │
│      https://github.com/kubernetes/minikube/issues/new/choose    │
│                                                                  │
│    * Please attach the following file to the GitHub issue:       │
│    * - /root/.minikube/logs/lastStart.txt                        │
│                                                                  │
╰──────────────────────────────────────────────────────────────────╯
X Exiting due to K8S_KUBELET_NOT_RUNNING: wait: /bin/bash -c "sudo env PATH=/var/lib/minikube/binaries/v1.21.2:$PATH kubeadm init --config /var/tmp/minikube/kubeadm.yaml  --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,Mem": exit status 1
stdout:
[init] Using Kubernetes version: v1.21.2
[preflight] Running pre-flight checks
[preflight] Pulling images required for setting up a Kubernetes cluster
[preflight] This might take a minute or two, depending on the speed of your internet connection
[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
[certs] Using certificateDir folder "/var/lib/minikube/certs"
[certs] Using existing ca certificate authority
[certs] Using existing apiserver certificate and key on disk
[certs] Using existing apiserver-kubelet-client certificate and key on disk
[certs] Using existing front-proxy-ca certificate authority
[certs] Using existing front-proxy-client certificate and key on disk
[certs] Using existing etcd/ca certificate authority
[certs] Using existing etcd/server certificate and key on disk
[certs] Using existing etcd/peer certificate and key on disk
[certs] Using existing etcd/healthcheck-client certificate and key on disk
[certs] Using existing apiserver-etcd-client certificate and key on disk
[certs] Using the existing "sa" key
[kubeconfig] Using kubeconfig folder "/etc/kubernetes"
[kubeconfig] Writing "admin.conf" kubeconfig file
[kubeconfig] Writing "kubelet.conf" kubeconfig file
[kubeconfig] Writing "controller-manager.conf" kubeconfig file
[kubeconfig] Writing "scheduler.conf" kubeconfig file
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Starting the kubelet
[control-plane] Using manifest folder "/etc/kubernetes/manifests"
[control-plane] Creating static Pod manifest for "kube-apiserver"
[control-plane] Creating static Pod manifest for "kube-controller-manager"
[control-plane] Creating static Pod manifest for "kube-scheduler"
[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
[kubelet-check] Initial timeout of 40s passed.

        Unfortunately, an error has occurred:
                timed out waiting for the condition

        This error is likely caused by:
                - The kubelet is not running
                - The kubelet is unhealthy due to a misconfiguration of the node in some way (required cgroups disabled)

        If you are on a systemd-powered system, you can try to troubleshoot the error with the following commands:
                - 'systemctl status kubelet'
                - 'journalctl -xeu kubelet'

        Additionally, a control plane component may have crashed or exited when started by the container runtime.
        To troubleshoot, list all containers using your preferred container runtimes CLI.

        Here is one example how you may list all Kubernetes containers running in docker:
                - 'docker ps -a | grep kube | grep -v pause'
                Once you have found the failing container, you can inspect its logs with:
                - 'docker logs CONTAINERID'


stderr:
        [WARNING IsDockerSystemdCheck]: detected "cgroupfs" as the Docker cgroup driver. The recommended driver is "systemd". Please follow the guide at https://kubernetes.io/docs/setup/cri/
        [WARNING FileExisting-socat]: socat not found in system path
        [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
error execution phase wait-control-plane: couldn't initialize a Kubernetes cluster
To see the stack trace of this error execute with --v=5 or higher

* Suggestion: Check output of 'journalctl -xeu kubelet', try passing --extra-config=kubelet.cgroup-driver=systemd to minikube start
* Related issue: https://github.com/kubernetes/minikube/issues/4172
cp: cannot stat 'onedata/test-reports/*.xml': No such file or directory
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
mv: cannot stat 'onedata/tests/gui/logs/report.*': No such file or directory
mv: cannot stat 'onedata/tests/gui/logs/gui_report/images.yaml': No such file or directory
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  41544      0 --:--:-- --:--:-- --:--:-- 41544
Error: Kubernetes cluster unreachable: Get "http://localhost:8080/version?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
The connection to the server localhost:8080 was refused - did you specify the right host or port?
The connection to the server localhost:8080 was refused - did you specify the right host or port?
The connection to the server localhost:8080 was refused - did you specify the right host or port?
The connection to the server localhost:8080 was refused - did you specify the right host or port?
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: readlink /mnt/storage/docker/overlay2/l: invalid argument
Error response from daemon: Cannot kill container: 9ae068bcba3b: Container 9ae068bcba3b566b97040a39960539a2268c1d56d53f45b37dbaea9e18cb271d is not running
Error response from daemon: Cannot kill container: 5115f0ca7db9: Container 5115f0ca7db933febdcf4f4060a73bfca9593fe038f34b94cbe28508a461234a is not running
Error response from daemon: Cannot kill container: 6426b5ba8f7b: Container 6426b5ba8f7b787dd88a086339697d24ee486b56e38b8a988bba10d46db33804 is not running
Error response from daemon: Cannot kill container: 78ddae5b5c69: Container 78ddae5b5c69e8193e6d487608f91c976a47918c409ca03a20188d6e40a3671b is not running
Error response from daemon: Cannot kill container: 4c5e9cc2b586: Container 4c5e9cc2b586e9ec69e8ccfc5fd60f3bf02ca4eb271599c8e6358ceda6b0c498 is not running