Acceptance tests using different clients concurrently. Environment deployed from packages.
Build: #2465 failed
Job: Archives failed
Job result summary
- Completed
- Duration
- 8 minutes
- Revision
-
32a8637a9a01b830c75d008c42ff27e7892513dd
- Total tests
- 21
- Fixed in
- #2466 (Child of ODSRV-OPRPM-2186)
Tests
- 21 tests in total
- 21 tests failed
- 21 failures are new
- 2 minutes taken in total.
Status | Test | Duration | |
---|---|---|---|
Collapse |
test_archives
test_user_of_client_checking_does_not_see_archive_removed_previously_via_client_removing[1oz_1op_1oc-REST-web GUI]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_does_not_see_archive_removed_previously_via_client_removing[1oz_1op_1oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_does_not_see_archive_removed_previously_via_client_removing[1oz_1op_1oc-web GUI-REST]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_does_not_see_archive_removed_previously_via_client_removing[1oz_1op_1oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_archive_created_previously_via_client_creating[1oz_1op_1oc-REST-web GUI]
|
1 min | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_archive_created_previously_via_client_creating[1oz_1op_1oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_archive_created_previously_via_client_creating[1oz_1op_1oc-web GUI-REST]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_archive_created_previously_via_client_creating[1oz_1op_1oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_bagit_archive_created_previously_via_client_creating[1oz_1op_1oc-REST-web GUI]
|
1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_bagit_archive_created_previously_via_client_creating[1oz_1op_1oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_bagit_archive_created_previously_via_client_creating[1oz_1op_1oc-web GUI-REST]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_bagit_archive_created_previously_via_client_creating[1oz_1op_1oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_dip_archive_created_previously_via_client_creating[1oz_1op_1oc-REST-web GUI]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_dip_archive_created_previously_via_client_creating[1oz_1op_1oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_dip_archive_created_previously_via_client_creating[1oz_1op_1oc-web GUI-REST]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_dip_archive_created_previously_via_client_creating[1oz_1op_1oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_new_deleted_callback_url_after_changing_it_using_rest[1oz_1op_1oc-REST]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_new_deleted_callback_url_after_changing_it_using_rest[1oz_1op_1oc-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_new_deleted_callback_url_after_changing_it_using_rest[1oz_1op_1oc-web GUI]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_new_deleted_callback_url_after_changing_it_using_rest[1oz_1op_1oc-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_new_preserved_callback_url_after_changing_it_using_rest[1oz_1op_1oc-REST]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_new_preserved_callback_url_after_changing_it_using_rest[1oz_1op_1oc-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_new_preserved_callback_url_after_changing_it_using_rest[1oz_1op_1oc-web GUI]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_new_preserved_callback_url_after_changing_it_using_rest[1oz_1op_1oc-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_that_archive_has_base_archive_after_client_creating_created_incremental_archive[1oz_1op_1oc-REST-web GUI]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_that_archive_has_base_archive_after_client_creating_created_incremental_archive[1oz_1op_1oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_that_archive_has_base_archive_after_client_creating_created_incremental_archive[1oz_1op_1oc-web GUI-REST]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_that_archive_has_base_archive_after_client_creating_created_incremental_archive[1oz_1op_1oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_that_archive_has_been_recalled_after_client_recalling_recalled_archive[1oz_1op_1oc-REST-web GUI]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_that_archive_has_been_recalled_after_client_recalling_recalled_archive[1oz_1op_1oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_that_archive_has_been_recalled_after_client_recalling_recalled_archive[1oz_1op_1oc-web GUI-REST]
|
16 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_that_archive_has_been_recalled_after_client_recalling_recalled_archive[1oz_1op_1oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_that_dataset_has_more_archives_than_its_parent_after_user_of_client_creating_created_nested_archive_on_child_dataset[1oz_1op_1oc-REST-web GUI]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_that_dataset_has_more_archives_than_its_parent_after_user_of_client_creating_created_nested_archive_on_child_dataset[1oz_1op_1oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_client_checking_sees_that_dataset_has_more_archives_than_its_parent_after_user_of_client_creating_created_nested_archive_on_child_dataset[1oz_1op_1oc-web GUI-REST]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_client_checking_sees_that_dataset_has_more_archives_than_its_parent_after_user_of_client_creating_created_nested_archive_on_child_dataset[1oz_1op_1oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_rest_cancels_archive_recall_after_user_of_web_gui_recalled_archive[1oz_1op_1oc]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_rest_cancels_archive_recall_after_user_of_web_gui_recalled_archive[1oz_1op_1oc]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_user_of_rest_sees_progress_of_archive_recall_after_user_of_web_gui_recalled_archive[1oz_1op_1oc]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_of_rest_sees_progress_of_archive_recall_after_user_of_web_gui_recalled_archive[1oz_1op_1oc]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_archives
test_using_web_gui_user1_sees_that_archive_description_has_been_changed_after_user1_changed_it_using_rest[1oz_1op_1oc]
|
< 1 sec | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_web_gui_user1_sees_that_archive_description_has_been_changed_after_user1_changed_it_using_rest[1oz_1op_1oc]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
Error summary
The build generated some errors. See the full build log for more details.
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 40460 0 --:--:-- --:--:-- --:--:-- 40112
Error response from daemon: Cannot kill container: 480aa3e6b8f4: Container 480aa3e6b8f4925507d1b8afa4d704a38ecf1dadc58585ac45e350d1a7ed98ec is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Unable to find image 'docker.onedata.org/swagger-aggregator:1.5.0' locally
1.5.0: Pulling from swagger-aggregator
3e51df1a11ab: Pulling fs layer
afe1a672db39: Pulling fs layer
80bad90eec10: Pulling fs layer
ca328de5836f: Pulling fs layer
4f4fb700ef54: Pulling fs layer
f9032f2640a0: Pulling fs layer
982a4aef97ed: Pulling fs layer
aca789ddad1c: Pulling fs layer
f9032f2640a0: Waiting
982a4aef97ed: Waiting
ca328de5836f: Waiting
4f4fb700ef54: Waiting
aca789ddad1c: Waiting
a3e17b3016f6: Pulling fs layer
b67d253c6e22: Pulling fs layer
302c9eabb2bb: Pulling fs layer
bb6e4824f633: Pulling fs layer
b67d253c6e22: Waiting
a3e17b3016f6: Waiting
302c9eabb2bb: Waiting
bb6e4824f633: Waiting
afe1a672db39: Verifying Checksum
afe1a672db39: Download complete
80bad90eec10: Verifying Checksum
80bad90eec10: Download complete
4f4fb700ef54: Verifying Checksum
4f4fb700ef54: Download complete
ca328de5836f: Verifying Checksum
ca328de5836f: Download complete
3e51df1a11ab: Verifying Checksum
3e51df1a11ab: Download complete
f9032f2640a0: Verifying Checksum
f9032f2640a0: Download complete
aca789ddad1c: Verifying Checksum
aca789ddad1c: Download complete
b67d253c6e22: Download complete
a3e17b3016f6: Download complete
bb6e4824f633: Verifying Checksum
bb6e4824f633: Download complete
302c9eabb2bb: Verifying Checksum
302c9eabb2bb: Download complete
982a4aef97ed: Verifying Checksum
982a4aef97ed: Download complete
3e51df1a11ab: Pull complete
afe1a672db39: Pull complete
80bad90eec10: Pull complete
ca328de5836f: Pull complete
4f4fb700ef54: Pull complete
f9032f2640a0: Pull complete
982a4aef97ed: Pull complete
aca789ddad1c: Pull complete
a3e17b3016f6: Pull complete
b67d253c6e22: Pull complete
302c9eabb2bb: Pull complete
bb6e4824f633: Pull complete
Digest: sha256:e2e8e762a03a0acdd49e63c4168157cb4e0e79f31f4e815561e9f4c65dbf8ac8
Status: Downloaded newer image for docker.onedata.org/swagger-aggregator:1.5.0
Unable to find image 'swaggerapi/swagger-codegen-cli:2.4.20' locally
2.4.20: Pulling from swaggerapi/swagger-codegen-cli
e7c96db7181b: Pulling fs layer
f910a506b6cb: Pulling fs layer
b6abafe80f63: Pulling fs layer
0d9127f69a1f: Pulling fs layer
f910a506b6cb: Download complete
e7c96db7181b: Verifying Checksum
e7c96db7181b: Download complete
e7c96db7181b: Pull complete
f910a506b6cb: Pull complete
0d9127f69a1f: Verifying Checksum
0d9127f69a1f: Download complete
b6abafe80f63: Verifying Checksum
b6abafe80f63: Download complete
b6abafe80f63: Pull complete
0d9127f69a1f: Pull complete
Digest: sha256:e961c734f4a232ea050293e9b16aed4cc131ffecf4a7d8671f15f1d79bca8796
Status: Downloaded newer image for swaggerapi/swagger-codegen-cli:2.4.20
/bin/sh: 2: [[: not found
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 46069 0 --:--:-- --:--:-- --:--:-- 46530
Error response from daemon: Cannot kill container: e6dc07c66c8e: Container e6dc07c66c8e0f9f77b4dad5279bcb296da325fbab3599619d6108b97bc77d58 is not running
Error response from daemon: Cannot kill container: d038bb4f9e6b: Container d038bb4f9e6ba271195e943ab4a35c7e2fd908cd3c3e3d4214bdff89d54fc486 is not running
Error response from daemon: Cannot kill container: 1dec137e8673: Container 1dec137e8673eedc592a2a6646c6523d120e11557e4fc82af44dc5184a519363 is not running
Error response from daemon: Cannot kill container: a4b64b7b8be9: Container a4b64b7b8be933d34e94c56080f2da1074caab486db3feee180f6693e9d56f0a is not running
Error response from daemon: Cannot kill container: 0388e2f0b3e4: Container 0388e2f0b3e4ae3481dfa04357af7abf561f41a5aa09ee5fe06f7483a31bb9d5 is not running
Error response from daemon: Cannot kill container: 7e707d467393: Container 7e707d467393ef1909649119055a83bcb2adc4a36d8be94397766956f8aee339 is not running
Error response from daemon: Cannot kill container: 552c5d26890b: Container 552c5d26890beb1d05631bff142e2fb1acd755f4ee44556dce5bbea1282155bb is not running
Error response from daemon: Cannot kill container: a90c91f516d8: Container a90c91f516d8ecf59c8b39994741e54bff30bc6a9dd2be12e1636b500c41b6fb is not running
Error response from daemon: Cannot kill container: f0ef581304dd: Container f0ef581304dd2c733fb8ac95d9deb32e435c8185a3fe5a12cb250ffc650e620d is not running
Error response from daemon: Cannot kill container: f2d3954d7636: Container f2d3954d7636072f6375cc22159998ae72e5917598baf5d6d8e94ba9bf28d3af is not running
Error response from daemon: Cannot kill container: 646446054c27: Container 646446054c2744ee6bdc9a063fc4c53e6e6c60102d10070a35f1f9e883919a57 is not running
Error response from daemon: Cannot kill container: 95542cf0af11: Container 95542cf0af11abcb7b004b5e94ab2d5002d7fe2ae4054b28e9f106dc6380f26d is not running
Error response from daemon: Cannot kill container: 8bbc4f6b4331: Container 8bbc4f6b4331db5fd4481b5f88e7066360290d33a0bea14f9fdbebed368608f0 is not running
Error response from daemon: Cannot kill container: c479fd6f73cf: Container c479fd6f73cfdbb760ab9db35d986d5fa91eb695560376e7c0c16f1f08519fa2 is not running
Error response from daemon: Cannot kill container: df6554719931: Container df6554719931a4cc1d4182cbd01538467e76c6b8918b819317a6817dbca8b2e5 is not running
Error response from daemon: Cannot kill container: 64cfc64a1125: Container 64cfc64a112531ba605a58de5d8e6ac408857ebc6e2d9f442f97ab498f326512 is not running
Error response from daemon: Cannot kill container: ef83e41ced51: Container ef83e41ced51031b1c13a58160cd4069603208652e2a10a98e278a24ea6c964a is not running
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 40460 0 --:--:-- --:--:-- --:--:-- 40112
Error response from daemon: Cannot kill container: 480aa3e6b8f4: Container 480aa3e6b8f4925507d1b8afa4d704a38ecf1dadc58585ac45e350d1a7ed98ec is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CAT3/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Unable to find image 'docker.onedata.org/swagger-aggregator:1.5.0' locally
1.5.0: Pulling from swagger-aggregator
3e51df1a11ab: Pulling fs layer
afe1a672db39: Pulling fs layer
80bad90eec10: Pulling fs layer
ca328de5836f: Pulling fs layer
4f4fb700ef54: Pulling fs layer
f9032f2640a0: Pulling fs layer
982a4aef97ed: Pulling fs layer
aca789ddad1c: Pulling fs layer
f9032f2640a0: Waiting
982a4aef97ed: Waiting
ca328de5836f: Waiting
4f4fb700ef54: Waiting
aca789ddad1c: Waiting
a3e17b3016f6: Pulling fs layer
b67d253c6e22: Pulling fs layer
302c9eabb2bb: Pulling fs layer
bb6e4824f633: Pulling fs layer
b67d253c6e22: Waiting
a3e17b3016f6: Waiting
302c9eabb2bb: Waiting
bb6e4824f633: Waiting
afe1a672db39: Verifying Checksum
afe1a672db39: Download complete
80bad90eec10: Verifying Checksum
80bad90eec10: Download complete
4f4fb700ef54: Verifying Checksum
4f4fb700ef54: Download complete
ca328de5836f: Verifying Checksum
ca328de5836f: Download complete
3e51df1a11ab: Verifying Checksum
3e51df1a11ab: Download complete
f9032f2640a0: Verifying Checksum
f9032f2640a0: Download complete
aca789ddad1c: Verifying Checksum
aca789ddad1c: Download complete
b67d253c6e22: Download complete
a3e17b3016f6: Download complete
bb6e4824f633: Verifying Checksum
bb6e4824f633: Download complete
302c9eabb2bb: Verifying Checksum
302c9eabb2bb: Download complete
982a4aef97ed: Verifying Checksum
982a4aef97ed: Download complete
3e51df1a11ab: Pull complete
afe1a672db39: Pull complete
80bad90eec10: Pull complete
ca328de5836f: Pull complete
4f4fb700ef54: Pull complete
f9032f2640a0: Pull complete
982a4aef97ed: Pull complete
aca789ddad1c: Pull complete
a3e17b3016f6: Pull complete
b67d253c6e22: Pull complete
302c9eabb2bb: Pull complete
bb6e4824f633: Pull complete
Digest: sha256:e2e8e762a03a0acdd49e63c4168157cb4e0e79f31f4e815561e9f4c65dbf8ac8
Status: Downloaded newer image for docker.onedata.org/swagger-aggregator:1.5.0
Unable to find image 'swaggerapi/swagger-codegen-cli:2.4.20' locally
2.4.20: Pulling from swaggerapi/swagger-codegen-cli
e7c96db7181b: Pulling fs layer
f910a506b6cb: Pulling fs layer
b6abafe80f63: Pulling fs layer
0d9127f69a1f: Pulling fs layer
f910a506b6cb: Download complete
e7c96db7181b: Verifying Checksum
e7c96db7181b: Download complete
e7c96db7181b: Pull complete
f910a506b6cb: Pull complete
0d9127f69a1f: Verifying Checksum
0d9127f69a1f: Download complete
b6abafe80f63: Verifying Checksum
b6abafe80f63: Download complete
b6abafe80f63: Pull complete
0d9127f69a1f: Pull complete
Digest: sha256:e961c734f4a232ea050293e9b16aed4cc131ffecf4a7d8671f15f1d79bca8796
Status: Downloaded newer image for swaggerapi/swagger-codegen-cli:2.4.20
/bin/sh: 2: [[: not found
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 46069 0 --:--:-- --:--:-- --:--:-- 46530
Error response from daemon: Cannot kill container: e6dc07c66c8e: Container e6dc07c66c8e0f9f77b4dad5279bcb296da325fbab3599619d6108b97bc77d58 is not running
Error response from daemon: Cannot kill container: d038bb4f9e6b: Container d038bb4f9e6ba271195e943ab4a35c7e2fd908cd3c3e3d4214bdff89d54fc486 is not running
Error response from daemon: Cannot kill container: 1dec137e8673: Container 1dec137e8673eedc592a2a6646c6523d120e11557e4fc82af44dc5184a519363 is not running
Error response from daemon: Cannot kill container: a4b64b7b8be9: Container a4b64b7b8be933d34e94c56080f2da1074caab486db3feee180f6693e9d56f0a is not running
Error response from daemon: Cannot kill container: 0388e2f0b3e4: Container 0388e2f0b3e4ae3481dfa04357af7abf561f41a5aa09ee5fe06f7483a31bb9d5 is not running
Error response from daemon: Cannot kill container: 7e707d467393: Container 7e707d467393ef1909649119055a83bcb2adc4a36d8be94397766956f8aee339 is not running
Error response from daemon: Cannot kill container: 552c5d26890b: Container 552c5d26890beb1d05631bff142e2fb1acd755f4ee44556dce5bbea1282155bb is not running
Error response from daemon: Cannot kill container: a90c91f516d8: Container a90c91f516d8ecf59c8b39994741e54bff30bc6a9dd2be12e1636b500c41b6fb is not running
Error response from daemon: Cannot kill container: f0ef581304dd: Container f0ef581304dd2c733fb8ac95d9deb32e435c8185a3fe5a12cb250ffc650e620d is not running
Error response from daemon: Cannot kill container: f2d3954d7636: Container f2d3954d7636072f6375cc22159998ae72e5917598baf5d6d8e94ba9bf28d3af is not running
Error response from daemon: Cannot kill container: 646446054c27: Container 646446054c2744ee6bdc9a063fc4c53e6e6c60102d10070a35f1f9e883919a57 is not running
Error response from daemon: Cannot kill container: 95542cf0af11: Container 95542cf0af11abcb7b004b5e94ab2d5002d7fe2ae4054b28e9f106dc6380f26d is not running
Error response from daemon: Cannot kill container: 8bbc4f6b4331: Container 8bbc4f6b4331db5fd4481b5f88e7066360290d33a0bea14f9fdbebed368608f0 is not running
Error response from daemon: Cannot kill container: c479fd6f73cf: Container c479fd6f73cfdbb760ab9db35d986d5fa91eb695560376e7c0c16f1f08519fa2 is not running
Error response from daemon: Cannot kill container: df6554719931: Container df6554719931a4cc1d4182cbd01538467e76c6b8918b819317a6817dbca8b2e5 is not running
Error response from daemon: Cannot kill container: 64cfc64a1125: Container 64cfc64a112531ba605a58de5d8e6ac408857ebc6e2d9f442f97ab498f326512 is not running
Error response from daemon: Cannot kill container: ef83e41ced51: Container ef83e41ced51031b1c13a58160cd4069603208652e2a10a98e278a24ea6c964a is not running