Acceptance tests using different clients concurrently. Environment deployed from packages.
Build: #3228 failed
Job: Archive privileges was successful
Job result summary
- Completed
- Duration
- 20 minutes
- Agent
- bamboo-agent-os-11
- Revision
-
d704831f382422ac250de6a2a1b215bef0202e9c
- Total tests
- 10
- Successful since
- #3058 ()
Tests
- 10 tests in total
- 9 minutes taken in total.
Error summary
The build generated some errors. See the full build log for more details.
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4822 100 4822 0 0 54795 0 --:--:-- --:--:-- --:--:-- 54795
Unable to find image 'alpine:latest' locally
latest: Pulling from library/alpine
c6a83fedfae6: Pulling fs layer
c6a83fedfae6: Download complete
c6a83fedfae6: Pull complete
Digest: sha256:0a4eaa0eecf5f8c050e5bba433f58c052be7587ee8af3e8b3910ef9ab5fbe9f5
Status: Downloaded newer image for alpine:latest
Error: Kubernetes cluster unreachable: Get "http://localhost:8080/version?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.467355 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.468012 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.469450 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.470903 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.472457 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
E0905 10:45:21.533838 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.534269 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.535729 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.537530 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.538949 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
E0905 10:45:21.602319 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.602562 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.603917 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.605265 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.606710 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
E0905 10:45:21.665685 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.666428 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.667846 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.669300 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.670787 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
E0905 10:45:21.735327 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.735653 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.737206 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.738892 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.740569 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
Error response from daemon: Cannot kill container: 39d27b86ff35: Container 39d27b86ff3578ae519ebda8091c2255b83e28b151f65b579e5122b3dae68cf0 is not running
Unable to find image 'ubuntu:14.10' locally
14.10: Pulling from library/ubuntu
[DEPRECATION NOTICE] Docker Image Format v1, and Docker Image manifest version 2, schema 1 support will be removed in an upcoming release. Suggest the author of docker.io/library/ubuntu:14.10 to upgrade the image to the OCI Format, or Docker Image manifest v2, schema 2. More information at https://docs.docker.com/go/deprecated-image-specs/
b0efe5c05b4c: Pulling fs layer
0a1f1b169319: Pulling fs layer
1ceb0a3c7c48: Pulling fs layer
a3ed95caeb02: Pulling fs layer
a3ed95caeb02: Waiting
1ceb0a3c7c48: Verifying Checksum
1ceb0a3c7c48: Download complete
0a1f1b169319: Verifying Checksum
0a1f1b169319: Download complete
a3ed95caeb02: Verifying Checksum
a3ed95caeb02: Download complete
b0efe5c05b4c: Verifying Checksum
b0efe5c05b4c: Download complete
b0efe5c05b4c: Pull complete
0a1f1b169319: Pull complete
1ceb0a3c7c48: Pull complete
a3ed95caeb02: Pull complete
Digest: sha256:6341c688b4b0b82ec735389b3c97df8cf2831b8cb8bd1856779130a86574ac5c
Status: Downloaded newer image for ubuntu:14.10
W0905 10:47:44.609363 12042 out.go:239] ! kubectl and minikube configuration will be stored in /root
! kubectl and minikube configuration will be stored in /root
W0905 10:47:44.609372 12042 out.go:239] ! To use kubectl or minikube commands as your own user, you may need to relocate them. For example, to overwrite your own settings, run:
! To use kubectl or minikube commands as your own user, you may need to relocate them. For example, to overwrite your own settings, run:
W0905 10:47:44.609378 12042 out.go:239] *
*
W0905 10:47:44.609399 12042 out.go:239] - sudo mv /root/.kube /root/.minikube $HOME
- sudo mv /root/.kube /root/.minikube $HOME
W0905 10:47:44.609406 12042 out.go:239] - sudo chown -R $USER $HOME/.kube $HOME/.minikube
- sudo chown -R $USER $HOME/.kube $HOME/.minikube
W0905 10:47:44.609412 12042 out.go:239] *
*
W0905 10:47:44.609418 12042 out.go:239] * This can also be done automatically by setting the env var CHANGE_MINIKUBE_NONE_USER=true
* This can also be done automatically by setting the env var CHANGE_MINIKUBE_NONE_USER=true
I0905 10:47:44.610260 12042 addons.go:69] Setting storage-provisioner=true in profile "minikube"
I0905 10:47:44.610293 12042 addons.go:231] Setting addon storage-provisioner=true in "minikube"
I0905 10:47:44.610374 12042 host.go:66] Checking if "minikube" exists ...
I0905 10:47:44.610962 12042 kubeconfig.go:92] found "minikube" server: "https://10.87.23.65:8443"
I0905 10:47:44.610986 12042 api_server.go:166] Checking apiserver status ...
I0905 10:47:44.611016 12042 exec_runner.go:51] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0905 10:47:44.611065 12042 addons.go:69] Setting default-storageclass=true in profile "minikube"
I0905 10:47:44.611078 12042 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "minikube"
I0905 10:47:44.611487 12042 kubeconfig.go:92] found "minikube" server: "https://10.87.23.65:8443"
I0905 10:47:44.611508 12042 api_server.go:166] Checking apiserver status ...
I0905 10:47:44.611546 12042 exec_runner.go:51] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0905 10:47:44.658774 12042 exec_runner.go:51] Run: sudo egrep ^[0-9]+:freezer: /proc/13529/cgroup
I0905 10:47:44.668228 12042 exec_runner.go:51] Run: sudo egrep ^[0-9]+:freezer: /proc/13529/cgroup
I0905 10:47:44.674828 12042 api_server.go:182] apiserver freezer: "12:freezer:/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd"
I0905 10:47:44.674886 12042 exec_runner.go:51] Run: sudo cat /sys/fs/cgroup/freezer/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd/freezer.state
I0905 10:47:44.677184 12042 api_server.go:182] apiserver freezer: "12:freezer:/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd"
I0905 10:47:44.677234 12042 exec_runner.go:51] Run: sudo cat /sys/fs/cgroup/freezer/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd/freezer.state
I0905 10:47:44.684705 12042 api_server.go:204] freezer state: "THAWED"
I0905 10:47:44.684733 12042 api_server.go:253] Checking apiserver healthz at https://10.87.23.65:8443/healthz ...
I0905 10:47:44.689436 12042 api_server.go:204] freezer state: "THAWED"
I0905 10:47:44.689467 12042 api_server.go:253] Checking apiserver healthz at https://10.87.23.65:8443/healthz ...
I0905 10:47:44.696793 12042 api_server.go:279] https://10.87.23.65:8443/healthz returned 200:
ok
I0905 10:47:44.727140 12042 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0905 10:47:44.718761 12042 api_server.go:279] https://10.87.23.65:8443/healthz returned 200:
ok
I0905 10:47:44.721478 12042 kapi.go:248] "coredns" deployment in "kube-system" namespace and "minikube" context rescaled to 1 replicas
I0905 10:47:44.740746 12042 start.go:223] Will wait 6m0s for node &{Name: IP:10.87.23.65 Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0905 10:47:44.763840 12042 out.go:177] * Verifying Kubernetes components...
I0905 10:47:44.740889 12042 addons.go:423] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0905 10:47:44.744487 12042 addons.go:231] Setting addon default-storageclass=true in "minikube"
I0905 10:47:44.750991 12042 exec_runner.go:51] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 127.0.0.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0905 10:47:44.796055 12042 exec_runner.go:144] found /etc/kubernetes/addons/storage-provisioner.yaml, removing ...
I0905 10:47:44.796078 12042 exec_runner.go:203] rm: /etc/kubernetes/addons/storage-provisioner.yaml
I0905 10:47:44.796137 12042 host.go:66] Checking if "minikube" exists ...
I0905 10:47:44.796145 12042 exec_runner.go:151] cp: memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0905 10:47:44.796323 12042 exec_runner.go:51] Run: sudo cp -a /tmp/minikube236238774 /etc/kubernetes/addons/storage-provisioner.yaml
I0905 10:47:44.796527 12042 exec_runner.go:51] Run: sudo systemctl is-active --quiet service kubelet
I0905 10:47:44.796819 12042 kubeconfig.go:92] found "minikube" server: "https://10.87.23.65:8443"
I0905 10:47:44.796840 12042 api_server.go:166] Checking apiserver status ...
I0905 10:47:44.796867 12042 exec_runner.go:51] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0905 10:47:44.809992 12042 exec_runner.go:51] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0905 10:47:44.816748 12042 api_server.go:52] waiting for apiserver process to appear ...
I0905 10:47:44.816819 12042 exec_runner.go:51] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0905 10:47:44.830808 12042 exec_runner.go:51] Run: sudo egrep ^[0-9]+:freezer: /proc/13529/cgroup
I0905 10:47:44.851557 12042 api_server.go:72] duration metric: took 110.749241ms to wait for apiserver process to appear ...
I0905 10:47:44.851581 12042 api_server.go:88] waiting for apiserver healthz status ...
I0905 10:47:44.851602 12042 api_server.go:253] Checking apiserver healthz at https://10.87.23.65:8443/healthz ...
I0905 10:47:44.853596 12042 api_server.go:182] apiserver freezer: "12:freezer:/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd"
I0905 10:47:44.853647 12042 exec_runner.go:51] Run: sudo cat /sys/fs/cgroup/freezer/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd/freezer.state
I0905 10:47:44.859248 12042 api_server.go:279] https://10.87.23.65:8443/healthz returned 200:
ok
I0905 10:47:44.860491 12042 api_server.go:141] control plane version: v1.28.1
I0905 10:47:44.860524 12042 api_server.go:131] duration metric: took 8.922271ms to wait for apiserver health ...
I0905 10:47:44.860535 12042 system_pods.go:43] waiting for kube-system pods to appear ...
I0905 10:47:44.867464 12042 system_pods.go:59] 4 kube-system pods found
I0905 10:47:44.867498 12042 system_pods.go:61] "etcd-bamboo-agent-os-11" [77a62927-cdc6-432e-bc82-5b03bc54789c] Pending
I0905 10:47:44.867507 12042 system_pods.go:61] "kube-apiserver-bamboo-agent-os-11" [60e878a3-660e-4e93-bacf-8e8c427d99de] Pending
I0905 10:47:44.867523 12042 system_pods.go:61] "kube-controller-manager-bamboo-agent-os-11" [6cfe5949-1d72-4f25-861e-3dbc56884997] Pending
I0905 10:47:44.867531 12042 system_pods.go:61] "kube-scheduler-bamboo-agent-os-11" [125f01c2-1452-44d8-855d-8db6bc9e446b] Pending
I0905 10:47:44.867537 12042 system_pods.go:74] duration metric: took 6.995783ms to wait for pod list to return data ...
I0905 10:47:44.867547 12042 kubeadm.go:581] duration metric: took 126.746683ms to wait for : map[apiserver:true system_pods:true] ...
I0905 10:47:44.867561 12042 node_conditions.go:102] verifying NodePressure condition ...
I0905 10:47:44.869710 12042 api_server.go:204] freezer state: "THAWED"
I0905 10:47:44.869738 12042 api_server.go:253] Checking apiserver healthz at https://10.87.23.65:8443/healthz ...
I0905 10:47:44.870937 12042 node_conditions.go:122] node storage ephemeral capacity is 25215872Ki
I0905 10:47:44.870973 12042 node_conditions.go:123] node cpu capacity is 4
I0905 10:47:44.870986 12042 node_conditions.go:105] duration metric: took 3.419832ms to run NodePressure ...
I0905 10:47:44.871000 12042 start.go:228] waiting for startup goroutines ...
I0905 10:47:44.877309 12042 api_server.go:279] https://10.87.23.65:8443/healthz returned 200:
ok
I0905 10:47:44.877380 12042 addons.go:423] installing /etc/kubernetes/addons/storageclass.yaml
I0905 10:47:44.877410 12042 exec_runner.go:151] cp: memory --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0905 10:47:44.877563 12042 exec_runner.go:51] Run: sudo cp -a /tmp/minikube3722303512 /etc/kubernetes/addons/storageclass.yaml
I0905 10:47:44.886175 12042 exec_runner.go:51] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0905 10:47:45.879552 12042 exec_runner.go:84] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 127.0.0.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.083534584s)
I0905 10:47:45.879582 12042 start.go:901] {"host.minikube.internal": 127.0.0.1} host record injected into CoreDNS's ConfigMap
I0905 10:47:45.888717 12042 exec_runner.go:84] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (1.002502528s)
I0905 10:47:46.067776 12042 exec_runner.go:84] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (1.257746993s)
I0905 10:47:46.088186 12042 out.go:177] * Enabled addons: default-storageclass, storage-provisioner
I0905 10:47:46.094818 12042 addons.go:502] enable addons completed in 1.501075578s: enabled=[default-storageclass storage-provisioner]
I0905 10:47:46.094878 12042 start.go:233] waiting for cluster config update ...
I0905 10:47:46.094894 12042 start.go:242] writing updated cluster config ...
I0905 10:47:46.095259 12042 exec_runner.go:51] Run: rm -f paused
I0905 10:47:46.159829 12042 start.go:600] kubectl: 1.28.1, cluster: 1.28.1 (minor skew: 0)
I0905 10:47:46.163896 12042 out.go:177] * Done! kubectl is now configured to use "minikube" cluster and "default" namespace by default
Submodule 'automation-examples' (ssh://git@git.onedata.org:7999/vfs/automation-examples.git) registered for path 'automation-examples'
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/automation-examples'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/onezone_swagger/bamboos'...
Unable to find image 'docker.onedata.org/one_env:v39' locally
v39: Pulling from one_env
41af1b5f0f51: Pulling fs layer
da812d3cf979: Pulling fs layer
6ea2f9c9eaa1: Pulling fs layer
ba301be5517a: Pulling fs layer
c41cf17cec29: Pulling fs layer
718cbc35a0e0: Pulling fs layer
7640532158c9: Pulling fs layer
6eba1fc2d1b6: Pulling fs layer
caf52bb4b365: Pulling fs layer
3dfbdb225edd: Pulling fs layer
b62f693fa900: Pulling fs layer
79ae03a86415: Pulling fs layer
ba301be5517a: Waiting
6eba1fc2d1b6: Waiting
caf52bb4b365: Waiting
3dfbdb225edd: Waiting
b62f693fa900: Waiting
79ae03a86415: Waiting
c41cf17cec29: Waiting
7640532158c9: Waiting
718cbc35a0e0: Waiting
6ea2f9c9eaa1: Verifying Checksum
6ea2f9c9eaa1: Download complete
ba301be5517a: Verifying Checksum
ba301be5517a: Download complete
c41cf17cec29: Verifying Checksum
c41cf17cec29: Download complete
41af1b5f0f51: Download complete
7640532158c9: Verifying Checksum
7640532158c9: Download complete
6eba1fc2d1b6: Verifying Checksum
6eba1fc2d1b6: Download complete
caf52bb4b365: Verifying Checksum
caf52bb4b365: Download complete
3dfbdb225edd: Verifying Checksum
b62f693fa900: Verifying Checksum
b62f693fa900: Download complete
79ae03a86415: Verifying Checksum
79ae03a86415: Download complete
41af1b5f0f51: Pull complete
718cbc35a0e0: Verifying Checksum
718cbc35a0e0: Download complete
da812d3cf979: Verifying Checksum
da812d3cf979: Download complete
da812d3cf979: Pull complete
6ea2f9c9eaa1: Pull complete
ba301be5517a: Pull complete
c41cf17cec29: Pull complete
718cbc35a0e0: Pull complete
7640532158c9: Pull complete
6eba1fc2d1b6: Pull complete
caf52bb4b365: Pull complete
3dfbdb225edd: Pull complete
b62f693fa900: Pull complete
79ae03a86415: Pull complete
Digest: sha256:e39dbddda3d96a874a4a2aebf2f188bfeb8ffdeb25e3fd27678ab5197f387940
Status: Downloaded newer image for docker.onedata.org/one_env:v39
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Unable to find image 'docker.onedata.org/swagger-aggregator:1.5.0' locally
1.5.0: Pulling from swagger-aggregator
3e51df1a11ab: Pulling fs layer
afe1a672db39: Pulling fs layer
80bad90eec10: Pulling fs layer
ca328de5836f: Pulling fs layer
4f4fb700ef54: Pulling fs layer
f9032f2640a0: Pulling fs layer
982a4aef97ed: Pulling fs layer
aca789ddad1c: Pulling fs layer
a3e17b3016f6: Pulling fs layer
b67d253c6e22: Pulling fs layer
302c9eabb2bb: Pulling fs layer
bb6e4824f633: Pulling fs layer
ca328de5836f: Waiting
4f4fb700ef54: Waiting
f9032f2640a0: Waiting
982a4aef97ed: Waiting
aca789ddad1c: Waiting
a3e17b3016f6: Waiting
b67d253c6e22: Waiting
302c9eabb2bb: Waiting
bb6e4824f633: Waiting
80bad90eec10: Verifying Checksum
80bad90eec10: Download complete
afe1a672db39: Verifying Checksum
afe1a672db39: Download complete
ca328de5836f: Verifying Checksum
ca328de5836f: Download complete
4f4fb700ef54: Verifying Checksum
4f4fb700ef54: Download complete
f9032f2640a0: Verifying Checksum
f9032f2640a0: Download complete
aca789ddad1c: Verifying Checksum
aca789ddad1c: Download complete
a3e17b3016f6: Verifying Checksum
a3e17b3016f6: Download complete
3e51df1a11ab: Verifying Checksum
3e51df1a11ab: Download complete
b67d253c6e22: Verifying Checksum
b67d253c6e22: Download complete
bb6e4824f633: Verifying Checksum
bb6e4824f633: Download complete
302c9eabb2bb: Verifying Checksum
302c9eabb2bb: Download complete
982a4aef97ed: Verifying Checksum
982a4aef97ed: Download complete
3e51df1a11ab: Pull complete
afe1a672db39: Pull complete
80bad90eec10: Pull complete
ca328de5836f: Pull complete
4f4fb700ef54: Pull complete
f9032f2640a0: Pull complete
982a4aef97ed: Pull complete
aca789ddad1c: Pull complete
a3e17b3016f6: Pull complete
b67d253c6e22: Pull complete
302c9eabb2bb: Pull complete
bb6e4824f633: Pull complete
Digest: sha256:e2e8e762a03a0acdd49e63c4168157cb4e0e79f31f4e815561e9f4c65dbf8ac8
Status: Downloaded newer image for docker.onedata.org/swagger-aggregator:1.5.0
Unable to find image 'swaggerapi/swagger-codegen-cli:2.4.20' locally
2.4.20: Pulling from swaggerapi/swagger-codegen-cli
e7c96db7181b: Pulling fs layer
f910a506b6cb: Pulling fs layer
b6abafe80f63: Pulling fs layer
0d9127f69a1f: Pulling fs layer
0d9127f69a1f: Waiting
f910a506b6cb: Verifying Checksum
f910a506b6cb: Download complete
e7c96db7181b: Download complete
0d9127f69a1f: Download complete
e7c96db7181b: Pull complete
f910a506b6cb: Pull complete
b6abafe80f63: Verifying Checksum
b6abafe80f63: Download complete
b6abafe80f63: Pull complete
0d9127f69a1f: Pull complete
Digest: sha256:e961c734f4a232ea050293e9b16aed4cc131ffecf4a7d8671f15f1d79bca8796
Status: Downloaded newer image for swaggerapi/swagger-codegen-cli:2.4.20
/bin/sh: 2: [[: not found
635970228e5b: Download complete
v12: Pulling from onedata/acceptance_mixed
d5fd17ec1767: Pulling fs layer
635970228e5b: Pulling fs layer
e6b8b518c4b7: Pulling fs layer
262a85701f26: Pulling fs layer
56fb40ac7433: Pulling fs layer
72938d7adcac: Pulling fs layer
cdf99cc302c9: Pulling fs layer
99e960dff935: Pulling fs layer
543cdc7d97b9: Pulling fs layer
c5eae658fb83: Pulling fs layer
e8c6e757d9d8: Pulling fs layer
c313dba7f09c: Pulling fs layer
262a85701f26: Waiting
cdf99cc302c9: Waiting
99e960dff935: Waiting
543cdc7d97b9: Waiting
c5eae658fb83: Waiting
e8c6e757d9d8: Waiting
c313dba7f09c: Waiting
56fb40ac7433: Waiting
72938d7adcac: Waiting
e6b8b518c4b7: Verifying Checksum
e6b8b518c4b7: Download complete
d5fd17ec1767: Verifying Checksum
d5fd17ec1767: Download complete
56fb40ac7433: Verifying Checksum
56fb40ac7433: Download complete
262a85701f26: Verifying Checksum
262a85701f26: Download complete
72938d7adcac: Verifying Checksum
72938d7adcac: Download complete
d5fd17ec1767: Pull complete
99e960dff935: Download complete
543cdc7d97b9: Verifying Checksum
543cdc7d97b9: Download complete
cdf99cc302c9: Download complete
e8c6e757d9d8: Verifying Checksum
e8c6e757d9d8: Download complete
c313dba7f09c: Verifying Checksum
c313dba7f09c: Download complete
c5eae658fb83: Verifying Checksum
c5eae658fb83: Download complete
635970228e5b: Verifying Checksum
635970228e5b: Download complete
635970228e5b: Pull complete
e6b8b518c4b7: Pull complete
262a85701f26: Pull complete
56fb40ac7433: Pull complete
72938d7adcac: Pull complete
cdf99cc302c9: Pull complete
99e960dff935: Pull complete
543cdc7d97b9: Pull complete
c5eae658fb83: Pull complete
e8c6e757d9d8: Pull complete
c313dba7f09c: Pull complete
Digest: sha256:96db83e9518bd8b75168f158bad519b54bb4b61336e7ba414e24be2483b152bb
Status: Downloaded newer image for onedata/acceptance_mixed:v12
/usr/local/lib/python3.8/dist-packages/pytest_selenium/drivers/crossbrowsertesting.py:72: SyntaxWarning: "is not" with a literal. Did you mean "!="?
if report.when == 'setup' or info.get('test_score') is not 'fail':
/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/tests/gui/steps/oneprovider/qos.py:357: SyntaxWarning: assertion is always true, perhaps remove parentheses?
assert ('qos-status-impossible' in
/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/tests/gui/steps/oneprovider/qos.py:360: SyntaxWarning: assertion is always true, perhaps remove parentheses?
assert ('qos-status-fulfilled' in
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
id: ‘user1’: no such user
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
id: ‘user2’: no such user
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(828) [sender=3.1.3]
command terminated with exit code 3
cp: cannot stat 'onedata-acceptance/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4822 100 4822 0 0 20175 0 --:--:-- --:--:-- --:--:-- 20175
Error response from daemon: Cannot kill container: be02da3d39b1: Container be02da3d39b1c337afd45c665804e9f1a5f5660a0f3074dd16af569b206786d7 is not running
Error response from daemon: Cannot kill container: b4ad1150a7cc: Container b4ad1150a7cc73ecc7753c664cac4e36d347c0bb971bc362d0e767fe4d8d7dc7 is not running
Error response from daemon: Cannot kill container: 62b1e783c557: Container 62b1e783c557c5b939a584411c0f04194dbdc82ba856d5a3792a760a886019c9 is not running
Error response from daemon: Cannot kill container: 734b5b709295: Container 734b5b709295eb84df89ebece5f448c4ff06c3b8c8e0741a4dbfa2082955cd94 is not running
Error response from daemon: Cannot kill container: 58946ebe3882: Container 58946ebe3882144c4d29c366be93d04787d13c589948a328f8bdd21fd6ac6eac is not running
Error response from daemon: Cannot kill container: 8fde8e230758: Container 8fde8e23075878eef70025b0ffeb6837cd5caef6ccd4371b3880f238ce87e398 is not running
Error response from daemon: Cannot kill container: 0d791010eab1: Container 0d791010eab1e4a7143b3e130f6e7918be51ee0c239c4dd68b0dd1ae8713a1e8 is not running
Error response from daemon: Cannot kill container: 41415982881d: Container 41415982881da6c8252d44d89905408d1e48cb3e589c90eba60ac82143bd21b5 is not running
Error response from daemon: Cannot kill container: 73a58464ef97: Container 73a58464ef971224fd278b0fab83aa68d64da1224486141aa7f389e683bd3e58 is not running
Error response from daemon: Cannot kill container: 6d047936ad0f: Container 6d047936ad0f9c5ac5c22fdbc6f182c3f21ee5c813e60b14487f356dd6ca816b is not running
Error response from daemon: Cannot kill container: 731b33f5b1b3: Container 731b33f5b1b3029450cee6dd1546ddf155e21e0ddcbbb3ee88065ba16ab31731 is not running
Error response from daemon: Cannot kill container: bedd723ea081: Container bedd723ea081dbd4ff5eeec0674d158b6917116e182aaac53dcc72ebc60253ee is not running
Error response from daemon: Cannot kill container: addae66781c1: Container addae66781c114250673632fc6fc9468a75d98a6244ffe698a5a81d71d63e277 is not running
Error response from daemon: Cannot kill container: 5d104dce513f: Container 5d104dce513feb5d3bdaf3bae9f0e29b5081a8f46cb5de8b1b54fab200542d08 is not running
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4822 100 4822 0 0 54795 0 --:--:-- --:--:-- --:--:-- 54795
Unable to find image 'alpine:latest' locally
latest: Pulling from library/alpine
c6a83fedfae6: Pulling fs layer
c6a83fedfae6: Download complete
c6a83fedfae6: Pull complete
Digest: sha256:0a4eaa0eecf5f8c050e5bba433f58c052be7587ee8af3e8b3910ef9ab5fbe9f5
Status: Downloaded newer image for alpine:latest
Error: Kubernetes cluster unreachable: Get "http://localhost:8080/version?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.467355 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.468012 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.469450 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.470903 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.472457 7761 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
E0905 10:45:21.533838 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.534269 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.535729 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.537530 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.538949 7771 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
E0905 10:45:21.602319 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.602562 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.603917 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.605265 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.606710 7777 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
E0905 10:45:21.665685 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.666428 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.667846 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.669300 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.670787 7783 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
E0905 10:45:21.735327 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.735653 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.737206 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.738892 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
E0905 10:45:21.740569 7790 memcache.go:265] couldn't get current server API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp 127.0.0.1:8080: connect: connection refused
The connection to the server localhost:8080 was refused - did you specify the right host or port?
Error response from daemon: Cannot kill container: 39d27b86ff35: Container 39d27b86ff3578ae519ebda8091c2255b83e28b151f65b579e5122b3dae68cf0 is not running
Unable to find image 'ubuntu:14.10' locally
14.10: Pulling from library/ubuntu
[DEPRECATION NOTICE] Docker Image Format v1, and Docker Image manifest version 2, schema 1 support will be removed in an upcoming release. Suggest the author of docker.io/library/ubuntu:14.10 to upgrade the image to the OCI Format, or Docker Image manifest v2, schema 2. More information at https://docs.docker.com/go/deprecated-image-specs/
b0efe5c05b4c: Pulling fs layer
0a1f1b169319: Pulling fs layer
1ceb0a3c7c48: Pulling fs layer
a3ed95caeb02: Pulling fs layer
a3ed95caeb02: Waiting
1ceb0a3c7c48: Verifying Checksum
1ceb0a3c7c48: Download complete
0a1f1b169319: Verifying Checksum
0a1f1b169319: Download complete
a3ed95caeb02: Verifying Checksum
a3ed95caeb02: Download complete
b0efe5c05b4c: Verifying Checksum
b0efe5c05b4c: Download complete
b0efe5c05b4c: Pull complete
0a1f1b169319: Pull complete
1ceb0a3c7c48: Pull complete
a3ed95caeb02: Pull complete
Digest: sha256:6341c688b4b0b82ec735389b3c97df8cf2831b8cb8bd1856779130a86574ac5c
Status: Downloaded newer image for ubuntu:14.10
W0905 10:47:44.609363 12042 out.go:239] ! kubectl and minikube configuration will be stored in /root
! kubectl and minikube configuration will be stored in /root
W0905 10:47:44.609372 12042 out.go:239] ! To use kubectl or minikube commands as your own user, you may need to relocate them. For example, to overwrite your own settings, run:
! To use kubectl or minikube commands as your own user, you may need to relocate them. For example, to overwrite your own settings, run:
W0905 10:47:44.609378 12042 out.go:239] *
*
W0905 10:47:44.609399 12042 out.go:239] - sudo mv /root/.kube /root/.minikube $HOME
- sudo mv /root/.kube /root/.minikube $HOME
W0905 10:47:44.609406 12042 out.go:239] - sudo chown -R $USER $HOME/.kube $HOME/.minikube
- sudo chown -R $USER $HOME/.kube $HOME/.minikube
W0905 10:47:44.609412 12042 out.go:239] *
*
W0905 10:47:44.609418 12042 out.go:239] * This can also be done automatically by setting the env var CHANGE_MINIKUBE_NONE_USER=true
* This can also be done automatically by setting the env var CHANGE_MINIKUBE_NONE_USER=true
I0905 10:47:44.610260 12042 addons.go:69] Setting storage-provisioner=true in profile "minikube"
I0905 10:47:44.610293 12042 addons.go:231] Setting addon storage-provisioner=true in "minikube"
I0905 10:47:44.610374 12042 host.go:66] Checking if "minikube" exists ...
I0905 10:47:44.610962 12042 kubeconfig.go:92] found "minikube" server: "https://10.87.23.65:8443"
I0905 10:47:44.610986 12042 api_server.go:166] Checking apiserver status ...
I0905 10:47:44.611016 12042 exec_runner.go:51] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0905 10:47:44.611065 12042 addons.go:69] Setting default-storageclass=true in profile "minikube"
I0905 10:47:44.611078 12042 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "minikube"
I0905 10:47:44.611487 12042 kubeconfig.go:92] found "minikube" server: "https://10.87.23.65:8443"
I0905 10:47:44.611508 12042 api_server.go:166] Checking apiserver status ...
I0905 10:47:44.611546 12042 exec_runner.go:51] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0905 10:47:44.658774 12042 exec_runner.go:51] Run: sudo egrep ^[0-9]+:freezer: /proc/13529/cgroup
I0905 10:47:44.668228 12042 exec_runner.go:51] Run: sudo egrep ^[0-9]+:freezer: /proc/13529/cgroup
I0905 10:47:44.674828 12042 api_server.go:182] apiserver freezer: "12:freezer:/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd"
I0905 10:47:44.674886 12042 exec_runner.go:51] Run: sudo cat /sys/fs/cgroup/freezer/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd/freezer.state
I0905 10:47:44.677184 12042 api_server.go:182] apiserver freezer: "12:freezer:/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd"
I0905 10:47:44.677234 12042 exec_runner.go:51] Run: sudo cat /sys/fs/cgroup/freezer/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd/freezer.state
I0905 10:47:44.684705 12042 api_server.go:204] freezer state: "THAWED"
I0905 10:47:44.684733 12042 api_server.go:253] Checking apiserver healthz at https://10.87.23.65:8443/healthz ...
I0905 10:47:44.689436 12042 api_server.go:204] freezer state: "THAWED"
I0905 10:47:44.689467 12042 api_server.go:253] Checking apiserver healthz at https://10.87.23.65:8443/healthz ...
I0905 10:47:44.696793 12042 api_server.go:279] https://10.87.23.65:8443/healthz returned 200:
ok
I0905 10:47:44.727140 12042 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0905 10:47:44.718761 12042 api_server.go:279] https://10.87.23.65:8443/healthz returned 200:
ok
I0905 10:47:44.721478 12042 kapi.go:248] "coredns" deployment in "kube-system" namespace and "minikube" context rescaled to 1 replicas
I0905 10:47:44.740746 12042 start.go:223] Will wait 6m0s for node &{Name: IP:10.87.23.65 Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0905 10:47:44.763840 12042 out.go:177] * Verifying Kubernetes components...
I0905 10:47:44.740889 12042 addons.go:423] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0905 10:47:44.744487 12042 addons.go:231] Setting addon default-storageclass=true in "minikube"
I0905 10:47:44.750991 12042 exec_runner.go:51] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 127.0.0.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0905 10:47:44.796055 12042 exec_runner.go:144] found /etc/kubernetes/addons/storage-provisioner.yaml, removing ...
I0905 10:47:44.796078 12042 exec_runner.go:203] rm: /etc/kubernetes/addons/storage-provisioner.yaml
I0905 10:47:44.796137 12042 host.go:66] Checking if "minikube" exists ...
I0905 10:47:44.796145 12042 exec_runner.go:151] cp: memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0905 10:47:44.796323 12042 exec_runner.go:51] Run: sudo cp -a /tmp/minikube236238774 /etc/kubernetes/addons/storage-provisioner.yaml
I0905 10:47:44.796527 12042 exec_runner.go:51] Run: sudo systemctl is-active --quiet service kubelet
I0905 10:47:44.796819 12042 kubeconfig.go:92] found "minikube" server: "https://10.87.23.65:8443"
I0905 10:47:44.796840 12042 api_server.go:166] Checking apiserver status ...
I0905 10:47:44.796867 12042 exec_runner.go:51] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0905 10:47:44.809992 12042 exec_runner.go:51] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0905 10:47:44.816748 12042 api_server.go:52] waiting for apiserver process to appear ...
I0905 10:47:44.816819 12042 exec_runner.go:51] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0905 10:47:44.830808 12042 exec_runner.go:51] Run: sudo egrep ^[0-9]+:freezer: /proc/13529/cgroup
I0905 10:47:44.851557 12042 api_server.go:72] duration metric: took 110.749241ms to wait for apiserver process to appear ...
I0905 10:47:44.851581 12042 api_server.go:88] waiting for apiserver healthz status ...
I0905 10:47:44.851602 12042 api_server.go:253] Checking apiserver healthz at https://10.87.23.65:8443/healthz ...
I0905 10:47:44.853596 12042 api_server.go:182] apiserver freezer: "12:freezer:/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd"
I0905 10:47:44.853647 12042 exec_runner.go:51] Run: sudo cat /sys/fs/cgroup/freezer/kubepods/burstable/pod95b55205f706ebe8ff1e4ae81be83996/501acc8cb96ed092770c2f6f5dda1dcb2e6d44fe33d2348b77a96fa5736c94bd/freezer.state
I0905 10:47:44.859248 12042 api_server.go:279] https://10.87.23.65:8443/healthz returned 200:
ok
I0905 10:47:44.860491 12042 api_server.go:141] control plane version: v1.28.1
I0905 10:47:44.860524 12042 api_server.go:131] duration metric: took 8.922271ms to wait for apiserver health ...
I0905 10:47:44.860535 12042 system_pods.go:43] waiting for kube-system pods to appear ...
I0905 10:47:44.867464 12042 system_pods.go:59] 4 kube-system pods found
I0905 10:47:44.867498 12042 system_pods.go:61] "etcd-bamboo-agent-os-11" [77a62927-cdc6-432e-bc82-5b03bc54789c] Pending
I0905 10:47:44.867507 12042 system_pods.go:61] "kube-apiserver-bamboo-agent-os-11" [60e878a3-660e-4e93-bacf-8e8c427d99de] Pending
I0905 10:47:44.867523 12042 system_pods.go:61] "kube-controller-manager-bamboo-agent-os-11" [6cfe5949-1d72-4f25-861e-3dbc56884997] Pending
I0905 10:47:44.867531 12042 system_pods.go:61] "kube-scheduler-bamboo-agent-os-11" [125f01c2-1452-44d8-855d-8db6bc9e446b] Pending
I0905 10:47:44.867537 12042 system_pods.go:74] duration metric: took 6.995783ms to wait for pod list to return data ...
I0905 10:47:44.867547 12042 kubeadm.go:581] duration metric: took 126.746683ms to wait for : map[apiserver:true system_pods:true] ...
I0905 10:47:44.867561 12042 node_conditions.go:102] verifying NodePressure condition ...
I0905 10:47:44.869710 12042 api_server.go:204] freezer state: "THAWED"
I0905 10:47:44.869738 12042 api_server.go:253] Checking apiserver healthz at https://10.87.23.65:8443/healthz ...
I0905 10:47:44.870937 12042 node_conditions.go:122] node storage ephemeral capacity is 25215872Ki
I0905 10:47:44.870973 12042 node_conditions.go:123] node cpu capacity is 4
I0905 10:47:44.870986 12042 node_conditions.go:105] duration metric: took 3.419832ms to run NodePressure ...
I0905 10:47:44.871000 12042 start.go:228] waiting for startup goroutines ...
I0905 10:47:44.877309 12042 api_server.go:279] https://10.87.23.65:8443/healthz returned 200:
ok
I0905 10:47:44.877380 12042 addons.go:423] installing /etc/kubernetes/addons/storageclass.yaml
I0905 10:47:44.877410 12042 exec_runner.go:151] cp: memory --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0905 10:47:44.877563 12042 exec_runner.go:51] Run: sudo cp -a /tmp/minikube3722303512 /etc/kubernetes/addons/storageclass.yaml
I0905 10:47:44.886175 12042 exec_runner.go:51] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0905 10:47:45.879552 12042 exec_runner.go:84] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 127.0.0.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.083534584s)
I0905 10:47:45.879582 12042 start.go:901] {"host.minikube.internal": 127.0.0.1} host record injected into CoreDNS's ConfigMap
I0905 10:47:45.888717 12042 exec_runner.go:84] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (1.002502528s)
I0905 10:47:46.067776 12042 exec_runner.go:84] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (1.257746993s)
I0905 10:47:46.088186 12042 out.go:177] * Enabled addons: default-storageclass, storage-provisioner
I0905 10:47:46.094818 12042 addons.go:502] enable addons completed in 1.501075578s: enabled=[default-storageclass storage-provisioner]
I0905 10:47:46.094878 12042 start.go:233] waiting for cluster config update ...
I0905 10:47:46.094894 12042 start.go:242] writing updated cluster config ...
I0905 10:47:46.095259 12042 exec_runner.go:51] Run: rm -f paused
I0905 10:47:46.159829 12042 start.go:600] kubectl: 1.28.1, cluster: 1.28.1 (minor skew: 0)
I0905 10:47:46.163896 12042 out.go:177] * Done! kubectl is now configured to use "minikube" cluster and "default" namespace by default
Submodule 'automation-examples' (ssh://git@git.onedata.org:7999/vfs/automation-examples.git) registered for path 'automation-examples'
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/automation-examples'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/onezone_swagger/bamboos'...
Unable to find image 'docker.onedata.org/one_env:v39' locally
v39: Pulling from one_env
41af1b5f0f51: Pulling fs layer
da812d3cf979: Pulling fs layer
6ea2f9c9eaa1: Pulling fs layer
ba301be5517a: Pulling fs layer
c41cf17cec29: Pulling fs layer
718cbc35a0e0: Pulling fs layer
7640532158c9: Pulling fs layer
6eba1fc2d1b6: Pulling fs layer
caf52bb4b365: Pulling fs layer
3dfbdb225edd: Pulling fs layer
b62f693fa900: Pulling fs layer
79ae03a86415: Pulling fs layer
ba301be5517a: Waiting
6eba1fc2d1b6: Waiting
caf52bb4b365: Waiting
3dfbdb225edd: Waiting
b62f693fa900: Waiting
79ae03a86415: Waiting
c41cf17cec29: Waiting
7640532158c9: Waiting
718cbc35a0e0: Waiting
6ea2f9c9eaa1: Verifying Checksum
6ea2f9c9eaa1: Download complete
ba301be5517a: Verifying Checksum
ba301be5517a: Download complete
c41cf17cec29: Verifying Checksum
c41cf17cec29: Download complete
41af1b5f0f51: Download complete
7640532158c9: Verifying Checksum
7640532158c9: Download complete
6eba1fc2d1b6: Verifying Checksum
6eba1fc2d1b6: Download complete
caf52bb4b365: Verifying Checksum
caf52bb4b365: Download complete
3dfbdb225edd: Verifying Checksum
b62f693fa900: Verifying Checksum
b62f693fa900: Download complete
79ae03a86415: Verifying Checksum
79ae03a86415: Download complete
41af1b5f0f51: Pull complete
718cbc35a0e0: Verifying Checksum
718cbc35a0e0: Download complete
da812d3cf979: Verifying Checksum
da812d3cf979: Download complete
da812d3cf979: Pull complete
6ea2f9c9eaa1: Pull complete
ba301be5517a: Pull complete
c41cf17cec29: Pull complete
718cbc35a0e0: Pull complete
7640532158c9: Pull complete
6eba1fc2d1b6: Pull complete
caf52bb4b365: Pull complete
3dfbdb225edd: Pull complete
b62f693fa900: Pull complete
79ae03a86415: Pull complete
Digest: sha256:e39dbddda3d96a874a4a2aebf2f188bfeb8ffdeb25e3fd27678ab5197f387940
Status: Downloaded newer image for docker.onedata.org/one_env:v39
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Unable to find image 'docker.onedata.org/swagger-aggregator:1.5.0' locally
1.5.0: Pulling from swagger-aggregator
3e51df1a11ab: Pulling fs layer
afe1a672db39: Pulling fs layer
80bad90eec10: Pulling fs layer
ca328de5836f: Pulling fs layer
4f4fb700ef54: Pulling fs layer
f9032f2640a0: Pulling fs layer
982a4aef97ed: Pulling fs layer
aca789ddad1c: Pulling fs layer
a3e17b3016f6: Pulling fs layer
b67d253c6e22: Pulling fs layer
302c9eabb2bb: Pulling fs layer
bb6e4824f633: Pulling fs layer
ca328de5836f: Waiting
4f4fb700ef54: Waiting
f9032f2640a0: Waiting
982a4aef97ed: Waiting
aca789ddad1c: Waiting
a3e17b3016f6: Waiting
b67d253c6e22: Waiting
302c9eabb2bb: Waiting
bb6e4824f633: Waiting
80bad90eec10: Verifying Checksum
80bad90eec10: Download complete
afe1a672db39: Verifying Checksum
afe1a672db39: Download complete
ca328de5836f: Verifying Checksum
ca328de5836f: Download complete
4f4fb700ef54: Verifying Checksum
4f4fb700ef54: Download complete
f9032f2640a0: Verifying Checksum
f9032f2640a0: Download complete
aca789ddad1c: Verifying Checksum
aca789ddad1c: Download complete
a3e17b3016f6: Verifying Checksum
a3e17b3016f6: Download complete
3e51df1a11ab: Verifying Checksum
3e51df1a11ab: Download complete
b67d253c6e22: Verifying Checksum
b67d253c6e22: Download complete
bb6e4824f633: Verifying Checksum
bb6e4824f633: Download complete
302c9eabb2bb: Verifying Checksum
302c9eabb2bb: Download complete
982a4aef97ed: Verifying Checksum
982a4aef97ed: Download complete
3e51df1a11ab: Pull complete
afe1a672db39: Pull complete
80bad90eec10: Pull complete
ca328de5836f: Pull complete
4f4fb700ef54: Pull complete
f9032f2640a0: Pull complete
982a4aef97ed: Pull complete
aca789ddad1c: Pull complete
a3e17b3016f6: Pull complete
b67d253c6e22: Pull complete
302c9eabb2bb: Pull complete
bb6e4824f633: Pull complete
Digest: sha256:e2e8e762a03a0acdd49e63c4168157cb4e0e79f31f4e815561e9f4c65dbf8ac8
Status: Downloaded newer image for docker.onedata.org/swagger-aggregator:1.5.0
Unable to find image 'swaggerapi/swagger-codegen-cli:2.4.20' locally
2.4.20: Pulling from swaggerapi/swagger-codegen-cli
e7c96db7181b: Pulling fs layer
f910a506b6cb: Pulling fs layer
b6abafe80f63: Pulling fs layer
0d9127f69a1f: Pulling fs layer
0d9127f69a1f: Waiting
f910a506b6cb: Verifying Checksum
f910a506b6cb: Download complete
e7c96db7181b: Download complete
0d9127f69a1f: Download complete
e7c96db7181b: Pull complete
f910a506b6cb: Pull complete
b6abafe80f63: Verifying Checksum
b6abafe80f63: Download complete
b6abafe80f63: Pull complete
0d9127f69a1f: Pull complete
Digest: sha256:e961c734f4a232ea050293e9b16aed4cc131ffecf4a7d8671f15f1d79bca8796
Status: Downloaded newer image for swaggerapi/swagger-codegen-cli:2.4.20
/bin/sh: 2: [[: not found
635970228e5b: Download complete
v12: Pulling from onedata/acceptance_mixed
d5fd17ec1767: Pulling fs layer
635970228e5b: Pulling fs layer
e6b8b518c4b7: Pulling fs layer
262a85701f26: Pulling fs layer
56fb40ac7433: Pulling fs layer
72938d7adcac: Pulling fs layer
cdf99cc302c9: Pulling fs layer
99e960dff935: Pulling fs layer
543cdc7d97b9: Pulling fs layer
c5eae658fb83: Pulling fs layer
e8c6e757d9d8: Pulling fs layer
c313dba7f09c: Pulling fs layer
262a85701f26: Waiting
cdf99cc302c9: Waiting
99e960dff935: Waiting
543cdc7d97b9: Waiting
c5eae658fb83: Waiting
e8c6e757d9d8: Waiting
c313dba7f09c: Waiting
56fb40ac7433: Waiting
72938d7adcac: Waiting
e6b8b518c4b7: Verifying Checksum
e6b8b518c4b7: Download complete
d5fd17ec1767: Verifying Checksum
d5fd17ec1767: Download complete
56fb40ac7433: Verifying Checksum
56fb40ac7433: Download complete
262a85701f26: Verifying Checksum
262a85701f26: Download complete
72938d7adcac: Verifying Checksum
72938d7adcac: Download complete
d5fd17ec1767: Pull complete
99e960dff935: Download complete
543cdc7d97b9: Verifying Checksum
543cdc7d97b9: Download complete
cdf99cc302c9: Download complete
e8c6e757d9d8: Verifying Checksum
e8c6e757d9d8: Download complete
c313dba7f09c: Verifying Checksum
c313dba7f09c: Download complete
c5eae658fb83: Verifying Checksum
c5eae658fb83: Download complete
635970228e5b: Verifying Checksum
635970228e5b: Download complete
635970228e5b: Pull complete
e6b8b518c4b7: Pull complete
262a85701f26: Pull complete
56fb40ac7433: Pull complete
72938d7adcac: Pull complete
cdf99cc302c9: Pull complete
99e960dff935: Pull complete
543cdc7d97b9: Pull complete
c5eae658fb83: Pull complete
e8c6e757d9d8: Pull complete
c313dba7f09c: Pull complete
Digest: sha256:96db83e9518bd8b75168f158bad519b54bb4b61336e7ba414e24be2483b152bb
Status: Downloaded newer image for onedata/acceptance_mixed:v12
/usr/local/lib/python3.8/dist-packages/pytest_selenium/drivers/crossbrowsertesting.py:72: SyntaxWarning: "is not" with a literal. Did you mean "!="?
if report.when == 'setup' or info.get('test_score') is not 'fail':
/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/tests/gui/steps/oneprovider/qos.py:357: SyntaxWarning: assertion is always true, perhaps remove parentheses?
assert ('qos-status-impossible' in
/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAP/onedata-acceptance/tests/gui/steps/oneprovider/qos.py:360: SyntaxWarning: assertion is always true, perhaps remove parentheses?
assert ('qos-status-fulfilled' in
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
id: ‘user1’: no such user
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
id: ‘user2’: no such user
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(828) [sender=3.1.3]
command terminated with exit code 3
cp: cannot stat 'onedata-acceptance/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4822 100 4822 0 0 20175 0 --:--:-- --:--:-- --:--:-- 20175
Error response from daemon: Cannot kill container: be02da3d39b1: Container be02da3d39b1c337afd45c665804e9f1a5f5660a0f3074dd16af569b206786d7 is not running
Error response from daemon: Cannot kill container: b4ad1150a7cc: Container b4ad1150a7cc73ecc7753c664cac4e36d347c0bb971bc362d0e767fe4d8d7dc7 is not running
Error response from daemon: Cannot kill container: 62b1e783c557: Container 62b1e783c557c5b939a584411c0f04194dbdc82ba856d5a3792a760a886019c9 is not running
Error response from daemon: Cannot kill container: 734b5b709295: Container 734b5b709295eb84df89ebece5f448c4ff06c3b8c8e0741a4dbfa2082955cd94 is not running
Error response from daemon: Cannot kill container: 58946ebe3882: Container 58946ebe3882144c4d29c366be93d04787d13c589948a328f8bdd21fd6ac6eac is not running
Error response from daemon: Cannot kill container: 8fde8e230758: Container 8fde8e23075878eef70025b0ffeb6837cd5caef6ccd4371b3880f238ce87e398 is not running
Error response from daemon: Cannot kill container: 0d791010eab1: Container 0d791010eab1e4a7143b3e130f6e7918be51ee0c239c4dd68b0dd1ae8713a1e8 is not running
Error response from daemon: Cannot kill container: 41415982881d: Container 41415982881da6c8252d44d89905408d1e48cb3e589c90eba60ac82143bd21b5 is not running
Error response from daemon: Cannot kill container: 73a58464ef97: Container 73a58464ef971224fd278b0fab83aa68d64da1224486141aa7f389e683bd3e58 is not running
Error response from daemon: Cannot kill container: 6d047936ad0f: Container 6d047936ad0f9c5ac5c22fdbc6f182c3f21ee5c813e60b14487f356dd6ca816b is not running
Error response from daemon: Cannot kill container: 731b33f5b1b3: Container 731b33f5b1b3029450cee6dd1546ddf155e21e0ddcbbb3ee88065ba16ab31731 is not running
Error response from daemon: Cannot kill container: bedd723ea081: Container bedd723ea081dbd4ff5eeec0674d158b6917116e182aaac53dcc72ebc60253ee is not running
Error response from daemon: Cannot kill container: addae66781c1: Container addae66781c114250673632fc6fc9468a75d98a6244ffe698a5a81d71d63e277 is not running
Error response from daemon: Cannot kill container: 5d104dce513f: Container 5d104dce513feb5d3bdaf3bae9f0e29b5081a8f46cb5de8b1b54fab200542d08 is not running