GUI acceptance tests using environment deployed from packages.
Build: #2261 failed
Job: Oneprovider ACL subfiles failed
Job result summary
- Completed
- Duration
- 5 minutes
- Revision
-
32a8637a9a01b830c75d008c42ff27e7892513dd
- Total tests
- 10
- Fixed in
- #2262 (Child of ODSRV-OPRPM-2186)
Tests
- 10 tests in total
- 10 tests failed
- 10 failures are new
- 2 minutes taken in total.
Status | Test | Duration | |
---|---|---|---|
Collapse |
test_oneprovider_acl_subfile
test_remove_subfile[1oz_1op_deployed-group-group1-fails-all except [data:delete child]]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_remove_subfile[1oz_1op_deployed-group-group1-fails-all except [data:delete child]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_oneprovider_acl_subfile
test_remove_subfile[1oz_1op_deployed-group-group1-succeeds-[data:delete child, data:traverse directory, data:list files]]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_remove_subfile[1oz_1op_deployed-group-group1-succeeds-[data:delete child, data:traverse directory, data:list files]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_oneprovider_acl_subfile
test_remove_subfile[1oz_1op_deployed-user-user1-fails-all except [data:delete child]]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_remove_subfile[1oz_1op_deployed-user-user1-fails-all except [data:delete child]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_oneprovider_acl_subfile
test_remove_subfile[1oz_1op_deployed-user-user1-succeeds-[data:delete child, data:traverse directory, data:list files]]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_remove_subfile[1oz_1op_deployed-user-user1-succeeds-[data:delete child, data:traverse directory, data:list files]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_oneprovider_acl_subfile
test_rename_subfile[1oz_1op_deployed-group-group1-fails-all except [data:add files]]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_rename_subfile[1oz_1op_deployed-group-group1-fails-all except [data:add files]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_oneprovider_acl_subfile
test_rename_subfile[1oz_1op_deployed-group-group1-fails-all except [data:delete child]]
|
17 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_rename_subfile[1oz_1op_deployed-group-group1-fails-all except [data:delete child]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_oneprovider_acl_subfile
test_rename_subfile[1oz_1op_deployed-group-group1-succeeds-[data:list files, data:delete child, data:traverse directory, data:add files]]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_rename_subfile[1oz_1op_deployed-group-group1-succeeds-[data:list files, data:delete child, data:traverse directory, data:add files]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_oneprovider_acl_subfile
test_rename_subfile[1oz_1op_deployed-user-user1-fails-all except [data:add files]]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_rename_subfile[1oz_1op_deployed-user-user1-fails-all except [data:add files]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_oneprovider_acl_subfile
test_rename_subfile[1oz_1op_deployed-user-user1-fails-all except [data:delete child]]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_rename_subfile[1oz_1op_deployed-user-user1-fails-all except [data:delete child]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_oneprovider_acl_subfile
test_rename_subfile[1oz_1op_deployed-user-user1-succeeds-[data:list files, data:delete child, data:traverse directory, data:add files]]
|
1 min | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_rename_subfile[1oz_1op_deployed-user-user1-succeeds-[data:list files, data:delete child, data:traverse directory, data:add files]]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
Error summary
The build generated some errors. See the full build log for more details.
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 43896 0 --:--:-- --:--:-- --:--:-- 43896
Error response from daemon: Cannot kill container: 85441dc036ac: Container 85441dc036ac8d64a2f203effd802e4e3612b6cd4a9af9f331b53407e82bccd5 is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
W0816 19:58:15.786304 729 warnings.go:70] policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 48468 0 --:--:-- --:--:-- --:--:-- 48978
Error from server (NotFound): pods "dev-oneprovider-krakow-0" not found
Error response from daemon: Cannot kill container: c53cb42a6d97: Container c53cb42a6d9785b2bb22d79341ac75ed86a50324fbdc66aa4bfca92f79295383 is not running
Error response from daemon: Cannot kill container: 7b2359ecd664: Container 7b2359ecd6648e3430c20dd6bc497d3ec474f9c5e154a2659227c6eb1f787d8f is not running
Error response from daemon: Cannot kill container: 472add51ce17: Container 472add51ce17253e4d63b240d391741ea5164d1ce5bb2e0b1a00289ccf9ad1ea is not running
Error response from daemon: Cannot kill container: 1b4941fa16c4: Container 1b4941fa16c4f527ea541dc1d62b5e2801d5318d1d17eaaaf430d0ed1c09e1bb is not running
Error response from daemon: Cannot kill container: f64844986eab: Container f64844986eab9295747ac4481f87dff9a652dd6c96d69c286f1f6225fe974bd1 is not running
Error response from daemon: Cannot kill container: d21bde3e6fb6: Container d21bde3e6fb662554e7f0fa02d9aa2f5b8c7a121cfb1cbe408a4888520311da8 is not running
Error response from daemon: Cannot kill container: daed5cd07a46: Container daed5cd07a4635261c6c91cedb67317ea05050476590844f3ac8b71bae4f096c is not running
Error response from daemon: Cannot kill container: 9e5b651c35dc: Container 9e5b651c35dcc815cfe80bf9a2abd779b182fa5ede96a89cfc7b4499429c3bc0 is not running
Error response from daemon: Cannot kill container: b672b786c2bf: Container b672b786c2bf1758646c3d2968a80ce6055d3a7eeae4d4d89973835ecec79487 is not running
Error response from daemon: Cannot kill container: 102a0acfe265: Container 102a0acfe2654f5531926c01fa8893aaa681a8f7e9f41e84c75b71db4b81e2a8 is not running
Error response from daemon: Cannot kill container: bf2a4a43d03b: Container bf2a4a43d03b42b21ba104522ffd33e7f1042e3ad9781b454e28f125657694f7 is not running
Error response from daemon: Cannot kill container: 4798d5f66d9b: Container 4798d5f66d9ba23c733e54e504682cef3586e60a9ffbffc783a54bf4b42b8993 is not running
Error response from daemon: Cannot kill container: a7aa380d55b5: Container a7aa380d55b5f190803a66b1a5874587a13faf62f8259c0c652c8aa48cf8ca53 is not running
Error response from daemon: Cannot kill container: a7a1adb8e76e: Container a7a1adb8e76ef22c6ef95486d93fabdd240272e7c44bcce3dacf527d68e58195 is not running
Error response from daemon: Cannot kill container: 3a5d5f6c0422: Container 3a5d5f6c042293087c9ab87c9c3d3e0336811dbb513c9be143e5d8fdba50c60a is not running
Error response from daemon: Cannot kill container: 8fda04276ef5: Container 8fda04276ef58038a59873fdb1a2e7e6462e0a4cf8b496c9ba5cd35313f5941f is not running
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 43896 0 --:--:-- --:--:-- --:--:-- 43896
Error response from daemon: Cannot kill container: 85441dc036ac: Container 85441dc036ac8d64a2f203effd802e4e3612b6cd4a9af9f331b53407e82bccd5 is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COASF/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
W0816 19:58:15.786304 729 warnings.go:70] policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 48468 0 --:--:-- --:--:-- --:--:-- 48978
Error from server (NotFound): pods "dev-oneprovider-krakow-0" not found
Error response from daemon: Cannot kill container: c53cb42a6d97: Container c53cb42a6d9785b2bb22d79341ac75ed86a50324fbdc66aa4bfca92f79295383 is not running
Error response from daemon: Cannot kill container: 7b2359ecd664: Container 7b2359ecd6648e3430c20dd6bc497d3ec474f9c5e154a2659227c6eb1f787d8f is not running
Error response from daemon: Cannot kill container: 472add51ce17: Container 472add51ce17253e4d63b240d391741ea5164d1ce5bb2e0b1a00289ccf9ad1ea is not running
Error response from daemon: Cannot kill container: 1b4941fa16c4: Container 1b4941fa16c4f527ea541dc1d62b5e2801d5318d1d17eaaaf430d0ed1c09e1bb is not running
Error response from daemon: Cannot kill container: f64844986eab: Container f64844986eab9295747ac4481f87dff9a652dd6c96d69c286f1f6225fe974bd1 is not running
Error response from daemon: Cannot kill container: d21bde3e6fb6: Container d21bde3e6fb662554e7f0fa02d9aa2f5b8c7a121cfb1cbe408a4888520311da8 is not running
Error response from daemon: Cannot kill container: daed5cd07a46: Container daed5cd07a4635261c6c91cedb67317ea05050476590844f3ac8b71bae4f096c is not running
Error response from daemon: Cannot kill container: 9e5b651c35dc: Container 9e5b651c35dcc815cfe80bf9a2abd779b182fa5ede96a89cfc7b4499429c3bc0 is not running
Error response from daemon: Cannot kill container: b672b786c2bf: Container b672b786c2bf1758646c3d2968a80ce6055d3a7eeae4d4d89973835ecec79487 is not running
Error response from daemon: Cannot kill container: 102a0acfe265: Container 102a0acfe2654f5531926c01fa8893aaa681a8f7e9f41e84c75b71db4b81e2a8 is not running
Error response from daemon: Cannot kill container: bf2a4a43d03b: Container bf2a4a43d03b42b21ba104522ffd33e7f1042e3ad9781b454e28f125657694f7 is not running
Error response from daemon: Cannot kill container: 4798d5f66d9b: Container 4798d5f66d9ba23c733e54e504682cef3586e60a9ffbffc783a54bf4b42b8993 is not running
Error response from daemon: Cannot kill container: a7aa380d55b5: Container a7aa380d55b5f190803a66b1a5874587a13faf62f8259c0c652c8aa48cf8ca53 is not running
Error response from daemon: Cannot kill container: a7a1adb8e76e: Container a7a1adb8e76ef22c6ef95486d93fabdd240272e7c44bcce3dacf527d68e58195 is not running
Error response from daemon: Cannot kill container: 3a5d5f6c0422: Container 3a5d5f6c042293087c9ab87c9c3d3e0336811dbb513c9be143e5d8fdba50c60a is not running
Error response from daemon: Cannot kill container: 8fda04276ef5: Container 8fda04276ef58038a59873fdb1a2e7e6462e0a4cf8b496c9ba5cd35313f5941f is not running