Acceptance tests using different clients concurrently. Environment deployed from packages.
Build: #2465 failed
Job: Directories multi failed
Job result summary
- Completed
- Duration
- 6 minutes
- Revision
-
32a8637a9a01b830c75d008c42ff27e7892513dd
- Total tests
- 12
- Fixed in
- #2466 (Child of ODSRV-OPRPM-2186)
Tests
- 12 tests in total
- 12 tests failed
- 12 failures are new
- 2 minutes taken in total.
Status | Test | Duration | |
---|---|---|---|
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-REST-oneclient2]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-REST-oneclient2]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-REST-web GUI]
|
1 min | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-oneclient1-REST]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-oneclient1-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-oneclient1-web GUI]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-oneclient1-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-web GUI-REST]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-web GUI-oneclient2]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_removes_it_using_client2[1oz_1op_2oc-web GUI-oneclient2]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-REST-oneclient2]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-REST-oneclient2]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-REST-web GUI]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-oneclient1-REST]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-oneclient1-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-oneclient1-web GUI]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-oneclient1-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-web GUI-REST]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_directories_multi
test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-web GUI-oneclient2]
|
18 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user1_creates_directory_using_client1_and_user2_renames_it_using_client2[1oz_1op_2oc-web GUI-oneclient2]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
Error summary
The build generated some errors. See the full build log for more details.
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 47479 0 --:--:-- --:--:-- --:--:-- 47479
Error response from daemon: Cannot kill container: 36131dfda771: Container 36131dfda7715726ac373fd5ec10e0c3d35b448c200c5d1cb625f2d109760c61 is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
/bin/sh: 2: [[: not found
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 43896 0 --:--:-- --:--:-- --:--:-- 44314
Error from server (NotFound): pods "dev-oneclient-krakow-7cc544c745-wjbxb" not found
Error from server (NotFound): pods "dev-oneprovider-krakow-0" not found
Error: No such object: 265dbf0afde6
Error: No such object: 265dbf0afde6
Error: No such object: 51a2a32b97cd
Error: No such object: 51a2a32b97cd
Error response from daemon: Cannot kill container: bfafc6291ca0: Container bfafc6291ca010db294c3faaf83e7c052898a6420dff562651eec0e0bd71e494 is not running
Error response from daemon: Cannot kill container: 9e08ca370520: No such container: 9e08ca370520
Error: No such container: 9e08ca370520
Error response from daemon: Cannot kill container: 2fd1788a2cd2: No such container: 2fd1788a2cd2
Error: No such container: 2fd1788a2cd2
Error response from daemon: Cannot kill container: ae8be3b39025: No such container: ae8be3b39025
Error: No such container: ae8be3b39025
Error response from daemon: Cannot kill container: 9ba1fa42d3cf: No such container: 9ba1fa42d3cf
Error: No such container: 9ba1fa42d3cf
Error response from daemon: Cannot kill container: ac49e331bc45: No such container: ac49e331bc45
Error: No such container: ac49e331bc45
Error response from daemon: Cannot kill container: d8b4150c7673: No such container: d8b4150c7673
Error: No such container: d8b4150c7673
Error response from daemon: Cannot kill container: 7e441479078d: No such container: 7e441479078d
Error: No such container: 7e441479078d
Error response from daemon: Cannot kill container: 8daa824790a9: No such container: 8daa824790a9
Error: No such container: 8daa824790a9
Error response from daemon: Cannot kill container: 517ac7a498cf: No such container: 517ac7a498cf
Error: No such container: 517ac7a498cf
Error response from daemon: Cannot kill container: b440f242047e: No such container: b440f242047e
Error: No such container: b440f242047e
Error response from daemon: Cannot kill container: cd6f62f89353: No such container: cd6f62f89353
Error: No such container: cd6f62f89353
Error response from daemon: Cannot kill container: aaf676c50f72: No such container: aaf676c50f72
Error: No such container: aaf676c50f72
Error response from daemon: Cannot kill container: 254032d008c8: No such container: 254032d008c8
Error: No such container: 254032d008c8
Error response from daemon: Cannot kill container: 265dbf0afde6: No such container: 265dbf0afde6
Error: No such container: 265dbf0afde6
Error response from daemon: Cannot kill container: 34e46eab7281: No such container: 34e46eab7281
Error: No such container: 34e46eab7281
Error response from daemon: Cannot kill container: 51a2a32b97cd: No such container: 51a2a32b97cd
Error: No such container: 51a2a32b97cd
Error response from daemon: Cannot kill container: fc4cfae5d8b9: Container fc4cfae5d8b9e1d3253069928e0593d4fb4a1f490a3198a00999fa2167ca87d8 is not running
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 47479 0 --:--:-- --:--:-- --:--:-- 47479
Error response from daemon: Cannot kill container: 36131dfda771: Container 36131dfda7715726ac373fd5ec10e0c3d35b448c200c5d1cb625f2d109760c61 is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CDMT/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
/bin/sh: 2: [[: not found
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 43896 0 --:--:-- --:--:-- --:--:-- 44314
Error from server (NotFound): pods "dev-oneclient-krakow-7cc544c745-wjbxb" not found
Error from server (NotFound): pods "dev-oneprovider-krakow-0" not found
Error: No such object: 265dbf0afde6
Error: No such object: 265dbf0afde6
Error: No such object: 51a2a32b97cd
Error: No such object: 51a2a32b97cd
Error response from daemon: Cannot kill container: bfafc6291ca0: Container bfafc6291ca010db294c3faaf83e7c052898a6420dff562651eec0e0bd71e494 is not running
Error response from daemon: Cannot kill container: 9e08ca370520: No such container: 9e08ca370520
Error: No such container: 9e08ca370520
Error response from daemon: Cannot kill container: 2fd1788a2cd2: No such container: 2fd1788a2cd2
Error: No such container: 2fd1788a2cd2
Error response from daemon: Cannot kill container: ae8be3b39025: No such container: ae8be3b39025
Error: No such container: ae8be3b39025
Error response from daemon: Cannot kill container: 9ba1fa42d3cf: No such container: 9ba1fa42d3cf
Error: No such container: 9ba1fa42d3cf
Error response from daemon: Cannot kill container: ac49e331bc45: No such container: ac49e331bc45
Error: No such container: ac49e331bc45
Error response from daemon: Cannot kill container: d8b4150c7673: No such container: d8b4150c7673
Error: No such container: d8b4150c7673
Error response from daemon: Cannot kill container: 7e441479078d: No such container: 7e441479078d
Error: No such container: 7e441479078d
Error response from daemon: Cannot kill container: 8daa824790a9: No such container: 8daa824790a9
Error: No such container: 8daa824790a9
Error response from daemon: Cannot kill container: 517ac7a498cf: No such container: 517ac7a498cf
Error: No such container: 517ac7a498cf
Error response from daemon: Cannot kill container: b440f242047e: No such container: b440f242047e
Error: No such container: b440f242047e
Error response from daemon: Cannot kill container: cd6f62f89353: No such container: cd6f62f89353
Error: No such container: cd6f62f89353
Error response from daemon: Cannot kill container: aaf676c50f72: No such container: aaf676c50f72
Error: No such container: aaf676c50f72
Error response from daemon: Cannot kill container: 254032d008c8: No such container: 254032d008c8
Error: No such container: 254032d008c8
Error response from daemon: Cannot kill container: 265dbf0afde6: No such container: 265dbf0afde6
Error: No such container: 265dbf0afde6
Error response from daemon: Cannot kill container: 34e46eab7281: No such container: 34e46eab7281
Error: No such container: 34e46eab7281
Error response from daemon: Cannot kill container: 51a2a32b97cd: No such container: 51a2a32b97cd
Error: No such container: 51a2a32b97cd
Error response from daemon: Cannot kill container: fc4cfae5d8b9: Container fc4cfae5d8b9e1d3253069928e0593d4fb4a1f490a3198a00999fa2167ca87d8 is not running