GUI acceptance tests using environment deployed from packages.
Build: #2261 failed
Job: Onezone providers basic failed
Job result summary
- Completed
- Duration
- 4 minutes
- Revision
-
32a8637a9a01b830c75d008c42ff27e7892513dd
- Total tests
- 8
- Fixed in
- #2262 (Child of ODSRV-OPRPM-2186)
Tests
- 8 tests in total
- 7 tests failed
- 7 failures are new
- 1 test was quarantined / skipped
- 1 minute taken in total.
Status | Test | Duration | |
---|---|---|---|
Collapse |
test_onezone_providers_basic
test_user_can_go_to_oneprovider_by_clicking_on_visit_provider_in_providers_popup[1oz_1op_deployed]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_can_go_to_oneprovider_by_clicking_on_visit_provider_in_providers_popup[1oz_1op_deployed]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_onezone_providers_basic
test_user_opens_provider_popup_by_clicking_on_supporting_provider_in_data_page[1oz_1op_deployed]
|
1 min | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_opens_provider_popup_by_clicking_on_supporting_provider_in_data_page[1oz_1op_deployed]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_onezone_providers_basic
test_user_sees_all_your_providers_are_offline_message_when_no_provider_is_online[1oz_1op_deployed]
|
13 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_sees_all_your_providers_are_offline_message_when_no_provider_is_online[1oz_1op_deployed]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_onezone_providers_basic
test_user_sees_provider_on_the_space_providers_map_after_supporting[1oz_1op_deployed]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_sees_provider_on_the_space_providers_map_after_supporting[1oz_1op_deployed]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_onezone_providers_basic
test_user_sees_that_after_unsupporting_space_number_displayed_in_space_counter_for_given_provider_decreases[1oz_1op_deployed]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_sees_that_after_unsupporting_space_number_displayed_in_space_counter_for_given_provider_decreases[1oz_1op_deployed]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_onezone_providers_basic
test_user_sees_that_if_space_is_displayed_in_provider_in_data_that_provider_is_also_displayed_in_provider_in_spaces[1oz_1op_deployed]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_sees_that_if_space_is_displayed_in_provider_in_data_that_provider_is_also_displayed_in_provider_in_spaces[1oz_1op_deployed]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_onezone_providers_basic
test_user_sees_that_provider_popup_can_be_closed_with_clicking_on_map[1oz_1op_deployed]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_sees_that_provider_popup_can_be_closed_with_clicking_on_map[1oz_1op_deployed]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
Error summary
The build generated some errors. See the full build log for more details.
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 45617 0 --:--:-- --:--:-- --:--:-- 45617
Error response from daemon: Cannot kill container: 271e23838f25: Container 271e23838f257b6ec05dfafd54be43c1ac9fada0f4f6dc1220b918d27a2dd7b2 is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
W0816 20:12:53.202276 736 warnings.go:70] policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 45174 0 --:--:-- --:--:-- --:--:-- 45174
Error response from daemon: Cannot kill container: 4172cfd13c6a: Container 4172cfd13c6a5a7e9d49e89012ad81dff1ed58ec86f8f64a1a60a5b03c167a1e is not running
Error response from daemon: Cannot kill container: 8c2c1a9685e4: Container 8c2c1a9685e4d8b08790acdaf34d970684a119ed4a6cb6da34e4ed588ac8c793 is not running
Error response from daemon: Cannot kill container: 6593aa5eb720: Container 6593aa5eb720692483af4a3b037d69cbcebb44c22603a71c5afc2f02d4f25cb5 is not running
Error response from daemon: Cannot kill container: b36f0e9e6e28: Container b36f0e9e6e2882ea943337189174cfc795228a62f33116f0ed94b082cf7cfd88 is not running
Error response from daemon: Cannot kill container: 4ac83bb75795: Container 4ac83bb75795664e2085c36e41ffaf6a323ff56386886efae00cc43bcbc2fed5 is not running
Error response from daemon: Cannot kill container: 21ec94789c21: Container 21ec94789c2102e1f27715446e402779f68ae909404574b2c85c490ff3f71fb7 is not running
Error response from daemon: Cannot kill container: 7683c0279b33: Container 7683c0279b33662dc20ca1360971333f42ae60c623d3f88e10f02f352124cff2 is not running
Error response from daemon: Cannot kill container: 9d7b84cae698: Container 9d7b84cae6987551cb2d62750d13fc304767b9e5bf153b369cf900be123a720d is not running
Error response from daemon: Cannot kill container: 6b092b89a2bb: Container 6b092b89a2bbde7d83c6396ca92cea84db9ade871a297adb68e7dba58e3400d0 is not running
Error response from daemon: Cannot kill container: 3c62474ab6d0: Container 3c62474ab6d0c04ffc82dd911af974479eee30305e72cb8dbcff28833d96d2ca is not running
Error response from daemon: Cannot kill container: 03e9d14cd678: Container 03e9d14cd678e0108fb12b555e8cb9ac16fb10f2f320b42cb73bad4fa97636c8 is not running
Error response from daemon: Cannot kill container: aa66553ee638: Container aa66553ee638af3a5edea70878903e5ac18c194903d3f40ceeb2850936fe39b4 is not running
Error response from daemon: Cannot kill container: a74a2d991f60: Container a74a2d991f60295f3f80e858f6817f4bf81b28b24cda36c8fb730c338e6ccbc7 is not running
Error response from daemon: Cannot kill container: 7388ae8bc69c: Container 7388ae8bc69cd6c3a82529bcd766a6cbd36ac6c0ffd88a0c93db960b62ff61e8 is not running
Error response from daemon: Cannot kill container: 836d3404bbb9: Container 836d3404bbb938ff852e414986b08581f3f09203f788327c7d051585a65af5e9 is not running
Error response from daemon: Cannot kill container: 923db8a9a5bb: Container 923db8a9a5bb78e509158c947b667012451be7df6ff9311ebfd8071f4688ada5 is not running
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 45617 0 --:--:-- --:--:-- --:--:-- 45617
Error response from daemon: Cannot kill container: 271e23838f25: Container 271e23838f257b6ec05dfafd54be43c1ac9fada0f4f6dc1220b918d27a2dd7b2 is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COPBT/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
W0816 20:12:53.202276 736 warnings.go:70] policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 45174 0 --:--:-- --:--:-- --:--:-- 45174
Error response from daemon: Cannot kill container: 4172cfd13c6a: Container 4172cfd13c6a5a7e9d49e89012ad81dff1ed58ec86f8f64a1a60a5b03c167a1e is not running
Error response from daemon: Cannot kill container: 8c2c1a9685e4: Container 8c2c1a9685e4d8b08790acdaf34d970684a119ed4a6cb6da34e4ed588ac8c793 is not running
Error response from daemon: Cannot kill container: 6593aa5eb720: Container 6593aa5eb720692483af4a3b037d69cbcebb44c22603a71c5afc2f02d4f25cb5 is not running
Error response from daemon: Cannot kill container: b36f0e9e6e28: Container b36f0e9e6e2882ea943337189174cfc795228a62f33116f0ed94b082cf7cfd88 is not running
Error response from daemon: Cannot kill container: 4ac83bb75795: Container 4ac83bb75795664e2085c36e41ffaf6a323ff56386886efae00cc43bcbc2fed5 is not running
Error response from daemon: Cannot kill container: 21ec94789c21: Container 21ec94789c2102e1f27715446e402779f68ae909404574b2c85c490ff3f71fb7 is not running
Error response from daemon: Cannot kill container: 7683c0279b33: Container 7683c0279b33662dc20ca1360971333f42ae60c623d3f88e10f02f352124cff2 is not running
Error response from daemon: Cannot kill container: 9d7b84cae698: Container 9d7b84cae6987551cb2d62750d13fc304767b9e5bf153b369cf900be123a720d is not running
Error response from daemon: Cannot kill container: 6b092b89a2bb: Container 6b092b89a2bbde7d83c6396ca92cea84db9ade871a297adb68e7dba58e3400d0 is not running
Error response from daemon: Cannot kill container: 3c62474ab6d0: Container 3c62474ab6d0c04ffc82dd911af974479eee30305e72cb8dbcff28833d96d2ca is not running
Error response from daemon: Cannot kill container: 03e9d14cd678: Container 03e9d14cd678e0108fb12b555e8cb9ac16fb10f2f320b42cb73bad4fa97636c8 is not running
Error response from daemon: Cannot kill container: aa66553ee638: Container aa66553ee638af3a5edea70878903e5ac18c194903d3f40ceeb2850936fe39b4 is not running
Error response from daemon: Cannot kill container: a74a2d991f60: Container a74a2d991f60295f3f80e858f6817f4bf81b28b24cda36c8fb730c338e6ccbc7 is not running
Error response from daemon: Cannot kill container: 7388ae8bc69c: Container 7388ae8bc69cd6c3a82529bcd766a6cbd36ac6c0ffd88a0c93db960b62ff61e8 is not running
Error response from daemon: Cannot kill container: 836d3404bbb9: Container 836d3404bbb938ff852e414986b08581f3f09203f788327c7d051585a65af5e9 is not running
Error response from daemon: Cannot kill container: 923db8a9a5bb: Container 923db8a9a5bb78e509158c947b667012451be7df6ff9311ebfd8071f4688ada5 is not running