Acceptance tests using different clients concurrently. Environment deployed from packages.
Build: #2465 failed
Job: LUMA failed
Job result summary
- Completed
- Duration
- 13 minutes
- Revision
-
32a8637a9a01b830c75d008c42ff27e7892513dd
- Total tests
- 26
- Fixed in
- #2466 (Child of ODSRV-OPRPM-2186)
Tests
- 26 tests in total
- 26 tests failed
- 26 failures are new
- 4 minutes taken in total.
Status | Test | Duration | |
---|---|---|---|
Collapse |
test_access_tokens
test_user_can_create_a_file_in_a_space_using_client1_authenticating_themselves_with_an_access_token_created_by_client2[1oz_1op_1oc-REST-web GUI]
|
3 mins | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_can_create_a_file_in_a_space_using_client1_authenticating_themselves_with_an_access_token_created_by_client2[1oz_1op_1oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_can_create_a_file_in_a_space_using_client1_authenticating_themselves_with_an_access_token_created_by_client2[1oz_1op_1oc-oneclient1-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_can_create_a_file_in_a_space_using_client1_authenticating_themselves_with_an_access_token_created_by_client2[1oz_1op_1oc-oneclient1-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_can_create_a_file_in_a_space_using_client1_authenticating_themselves_with_an_access_token_created_by_client2[1oz_1op_1oc-oneclient1-web GUI]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_can_create_a_file_in_a_space_using_client1_authenticating_themselves_with_an_access_token_created_by_client2[1oz_1op_1oc-oneclient1-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_cannot_use_an_access_token_created_with_client1_and_revoked_with_client2_when_using_client3[1oz_1op_1oc-REST-web GUI-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_cannot_use_an_access_token_created_with_client1_and_revoked_with_client2_when_using_client3[1oz_1op_1oc-REST-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_cannot_use_an_access_token_created_with_client1_and_revoked_with_client2_when_using_client3[1oz_1op_1oc-REST-web GUI-oneclient1]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_cannot_use_an_access_token_created_with_client1_and_revoked_with_client2_when_using_client3[1oz_1op_1oc-REST-web GUI-oneclient1]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_cannot_use_an_access_token_created_with_client1_and_revoked_with_client2_when_using_client3[1oz_1op_1oc-web GUI-REST-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_cannot_use_an_access_token_created_with_client1_and_revoked_with_client2_when_using_client3[1oz_1op_1oc-web GUI-REST-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_cannot_use_an_access_token_created_with_client1_and_revoked_with_client2_when_using_client3[1oz_1op_1oc-web GUI-REST-oneclient1]
|
30 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_cannot_use_an_access_token_created_with_client1_and_revoked_with_client2_when_using_client3[1oz_1op_1oc-web GUI-REST-oneclient1]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_creates_a_file_with_oneclient_but_not_with_rest_using_token_with_caveat_set_for_oneclient_interface[1oz_1op_1oc]
|
3 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_creates_a_file_with_oneclient_but_not_with_rest_using_token_with_caveat_set_for_oneclient_interface[1oz_1op_1oc]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_creates_a_file_with_rest_but_not_with_oneclient_using_token_with_caveat_set_for_rest_interface[1oz_1op_1oc]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_creates_a_file_with_rest_but_not_with_oneclient_using_token_with_caveat_set_for_rest_interface[1oz_1op_1oc]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_sees_all_token_caveats_in_token_configuration_using_client1_after_setting_them_in_new_access_token_with_client2[1oz_1op_1oc-REST-web GUI]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_sees_all_token_caveats_in_token_configuration_using_client1_after_setting_them_in_new_access_token_with_client2[1oz_1op_1oc-REST-web GUI]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_user_sees_all_token_caveats_in_token_configuration_using_client1_after_setting_them_in_new_access_token_with_client2[1oz_1op_1oc-web GUI-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_user_sees_all_token_caveats_in_token_configuration_using_client1_after_setting_them_in_new_access_token_with_client2[1oz_1op_1oc-web GUI-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_can_create_file_after_getting_token_with_caveat_set_for_oneprovider_service_created_by_web_gui[1oz_1op_1oc-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_can_create_file_after_getting_token_with_caveat_set_for_oneprovider_service_created_by_web_gui[1oz_1op_1oc-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_can_create_file_after_getting_token_with_caveat_set_for_oneprovider_service_created_by_web_gui[1oz_1op_1oc-oneclient1]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_can_create_file_after_getting_token_with_caveat_set_for_oneprovider_service_created_by_web_gui[1oz_1op_1oc-oneclient1]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_can_create_file_after_getting_token_with_expiration_time_set_in_caveat_created_by_web_gui[1oz_1op_1oc-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_can_create_file_after_getting_token_with_expiration_time_set_in_caveat_created_by_web_gui[1oz_1op_1oc-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_can_create_file_after_getting_token_with_expiration_time_set_in_caveat_created_by_web_gui[1oz_1op_1oc-oneclient1]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_can_create_file_after_getting_token_with_expiration_time_set_in_caveat_created_by_web_gui[1oz_1op_1oc-oneclient1]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_can_see_file_after_getting_token_with_caveat_set_for_read_only_created_by_web_gui[1oz_1op_1oc-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_can_see_file_after_getting_token_with_caveat_set_for_read_only_created_by_web_gui[1oz_1op_1oc-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_can_see_file_after_getting_token_with_caveat_set_for_read_only_created_by_web_gui[1oz_1op_1oc-oneclient1]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_can_see_file_after_getting_token_with_caveat_set_for_read_only_created_by_web_gui[1oz_1op_1oc-oneclient1]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_cannot_create_file_after_getting_token_with_caveat_set_only_for_onezone_service_created_by_web_gui[1oz_1op_1oc-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_cannot_create_file_after_getting_token_with_caveat_set_only_for_onezone_service_created_by_web_gui[1oz_1op_1oc-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_cannot_create_file_after_getting_token_with_caveat_set_only_for_onezone_service_created_by_web_gui[1oz_1op_1oc-oneclient1]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_cannot_create_file_after_getting_token_with_caveat_set_only_for_onezone_service_created_by_web_gui[1oz_1op_1oc-oneclient1]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_cannot_create_file_after_getting_token_with_expiration_time_set_in_caveat_created_by_web_gui[1oz_1op_1oc-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_cannot_create_file_after_getting_token_with_expiration_time_set_in_caveat_created_by_web_gui[1oz_1op_1oc-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_cannot_create_file_after_getting_token_with_expiration_time_set_in_caveat_created_by_web_gui[1oz_1op_1oc-oneclient1]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_cannot_create_file_after_getting_token_with_expiration_time_set_in_caveat_created_by_web_gui[1oz_1op_1oc-oneclient1]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_fails_to_create_file_after_getting_token_with_caveat_set_only_for_read_only_created_by_web_gui[1oz_1op_1oc-REST]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_fails_to_create_file_after_getting_token_with_caveat_set_only_for_read_only_created_by_web_gui[1oz_1op_1oc-REST]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_client1_user_fails_to_create_file_after_getting_token_with_caveat_set_only_for_read_only_created_by_web_gui[1oz_1op_1oc-oneclient1]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_client1_user_fails_to_create_file_after_getting_token_with_caveat_set_only_for_read_only_created_by_web_gui[1oz_1op_1oc-oneclient1]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_oneclient1_user_can_see_file_using_token_with_caveat_set_for_path_created_by_web_gui_after_owner_renames_file[1oz_1op_1oc]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_oneclient1_user_can_see_file_using_token_with_caveat_set_for_path_created_by_web_gui_after_owner_renames_file[1oz_1op_1oc]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_rest_user_can_create_a_group_in_a_space_after_getting_token_with_caveat_set_for_onezone_service_created_by_web_gui[1oz_1op_1oc]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_rest_user_can_create_a_group_in_a_space_after_getting_token_with_caveat_set_for_onezone_service_created_by_web_gui[1oz_1op_1oc]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
|||
Collapse |
test_access_tokens
test_using_rest_user_cannot_create_a_group_in_a_space_after_getting_token_with_caveat_set_only_for_oneprovider_service_created_by_web_gui[1oz_1op_1oc]
|
2 secs | |
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}} request = <FixtureRequest for <Function 'test_using_rest_user_cannot_create_a_group_in_a_space_after_getting_token_with_caveat_set_only_for_oneprovider_service_created_by_web_gui[1oz_1op_1oc]'>> @pytest.mark.usefixtures(*function_args) def scenario_wrapper(request): > _execute_scenario(feature, scenario, request, encoding) /usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: (47 more lines...) |
Error summary
The build generated some errors. See the full build log for more details.
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 45617 0 --:--:-- --:--:-- --:--:-- 45617
Error response from daemon: Cannot kill container: 3151d97bcc4e: Container 3151d97bcc4e34aa63ab16939df3d7a785ee309c34e0a1e2b9813282f93c8465 is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Unable to find image 'docker.onedata.org/swagger-aggregator:1.5.0' locally
1.5.0: Pulling from swagger-aggregator
3e51df1a11ab: Pulling fs layer
afe1a672db39: Pulling fs layer
80bad90eec10: Pulling fs layer
ca328de5836f: Pulling fs layer
4f4fb700ef54: Pulling fs layer
f9032f2640a0: Pulling fs layer
982a4aef97ed: Pulling fs layer
aca789ddad1c: Pulling fs layer
a3e17b3016f6: Pulling fs layer
b67d253c6e22: Pulling fs layer
302c9eabb2bb: Pulling fs layer
bb6e4824f633: Pulling fs layer
4f4fb700ef54: Waiting
f9032f2640a0: Waiting
982a4aef97ed: Waiting
aca789ddad1c: Waiting
a3e17b3016f6: Waiting
b67d253c6e22: Waiting
302c9eabb2bb: Waiting
bb6e4824f633: Waiting
ca328de5836f: Waiting
afe1a672db39: Verifying Checksum
afe1a672db39: Download complete
80bad90eec10: Verifying Checksum
80bad90eec10: Download complete
ca328de5836f: Verifying Checksum
ca328de5836f: Download complete
4f4fb700ef54: Verifying Checksum
4f4fb700ef54: Download complete
3e51df1a11ab: Verifying Checksum
3e51df1a11ab: Download complete
aca789ddad1c: Verifying Checksum
aca789ddad1c: Download complete
a3e17b3016f6: Verifying Checksum
a3e17b3016f6: Download complete
b67d253c6e22: Verifying Checksum
b67d253c6e22: Download complete
f9032f2640a0: Download complete
302c9eabb2bb: Verifying Checksum
302c9eabb2bb: Download complete
bb6e4824f633: Verifying Checksum
bb6e4824f633: Download complete
982a4aef97ed: Verifying Checksum
982a4aef97ed: Download complete
3e51df1a11ab: Pull complete
afe1a672db39: Pull complete
80bad90eec10: Pull complete
ca328de5836f: Pull complete
4f4fb700ef54: Pull complete
f9032f2640a0: Pull complete
982a4aef97ed: Pull complete
aca789ddad1c: Pull complete
a3e17b3016f6: Pull complete
b67d253c6e22: Pull complete
302c9eabb2bb: Pull complete
bb6e4824f633: Pull complete
Digest: sha256:e2e8e762a03a0acdd49e63c4168157cb4e0e79f31f4e815561e9f4c65dbf8ac8
Status: Downloaded newer image for docker.onedata.org/swagger-aggregator:1.5.0
Unable to find image 'swaggerapi/swagger-codegen-cli:2.4.20' locally
2.4.20: Pulling from swaggerapi/swagger-codegen-cli
e7c96db7181b: Pulling fs layer
f910a506b6cb: Pulling fs layer
b6abafe80f63: Pulling fs layer
0d9127f69a1f: Pulling fs layer
0d9127f69a1f: Waiting
f910a506b6cb: Verifying Checksum
f910a506b6cb: Download complete
e7c96db7181b: Verifying Checksum
e7c96db7181b: Download complete
e7c96db7181b: Pull complete
0d9127f69a1f: Verifying Checksum
0d9127f69a1f: Download complete
f910a506b6cb: Pull complete
b6abafe80f63: Verifying Checksum
b6abafe80f63: Download complete
b6abafe80f63: Pull complete
0d9127f69a1f: Pull complete
Digest: sha256:e961c734f4a232ea050293e9b16aed4cc131ffecf4a7d8671f15f1d79bca8796
Status: Downloaded newer image for swaggerapi/swagger-codegen-cli:2.4.20
/bin/sh: 2: [[: not found
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 41918 0 --:--:-- --:--:-- --:--:-- 41918
Error response from daemon: Cannot kill container: 02eaba3fdcca: Container 02eaba3fdcca6361b6572e79c77bc7cc92117cb43d63bb7b5e3bb14e0bb3910a is not running
Error response from daemon: Cannot kill container: 88de1beb74e5: Container 88de1beb74e59e068a0dbde14c63bb2c25bcc84eac6c80a73a37e7fa6ed960ed is not running
Error response from daemon: Cannot kill container: db5e1014422f: Container db5e1014422f593d088e0a908753e9ab870c48e5a3733d8a3d49d35ab4b92b2a is not running
Error response from daemon: Cannot kill container: 9b36c6aa2a85: Container 9b36c6aa2a85320a5b4c8f7f71d21ae4ed066dc45778dab4451b04051c9fafc7 is not running
Error response from daemon: Cannot kill container: d7c78246bec1: Container d7c78246bec111a9f38ed0259fed944176712a36185f26cd8eaa5e5329a52739 is not running
Error response from daemon: Cannot kill container: b11736dc5cbb: Container b11736dc5cbb8bac02ae10141a8d291ffcd65a53a6fa16517e9d85f8d84774cf is not running
Error response from daemon: Cannot kill container: e0a5c09f1f3b: Container e0a5c09f1f3bc892318a5157b0064fc7917c715b780cd1235e777a695d4a6df6 is not running
Error response from daemon: Cannot kill container: 7ad84add25f7: Container 7ad84add25f707100b558a1661a6141e8142e44b54b6bf5f4fc7f13cc27cbd0e is not running
Error response from daemon: Cannot kill container: b26c3a4839b5: Container b26c3a4839b5a0851003dabbd0ec64dfe063da843b10683b324082d337e1c863 is not running
Error response from daemon: Cannot kill container: 781858df708e: Container 781858df708e9b940d4de52d8cfb1da17139fc56b43df917170ba4030f70e664 is not running
Error response from daemon: Cannot kill container: a3bc2c037fea: Container a3bc2c037feaadf5a4b848786321c8838b6b08b7ca224d0af667e65ff0dc69ef is not running
Error response from daemon: Cannot kill container: 6990e1d607ce: Container 6990e1d607ce65fc64bdac2941ce61dfc3bac7a6330b27720c83a2fa549f0e01 is not running
Error response from daemon: Cannot kill container: c3ffaaf71ae0: Container c3ffaaf71ae01209398fbd439350daa2dbb42f81aeba107fa1cb614b8aa8de09 is not running
Error response from daemon: Cannot kill container: 67eab605dd9d: Container 67eab605dd9d990bb0ba9642bcc12d998297c6b2bd313bd436d6b854412a6a46 is not running
Error response from daemon: Cannot kill container: af26137ecbea: Container af26137ecbeaee93aa60eeffba09f4606d8d4e7baf39d63232f16c2229ed8e8c is not running
Error response from daemon: Cannot kill container: a62ffe83b5b8: Container a62ffe83b5b8e0abcbf100b23c328deca53779f442ea1b228ece0031da477c52 is not running
Error response from daemon: Cannot kill container: e6557c76419a: Container e6557c76419ab3fba327af6118ccca11234ce11244933ee9aa59c064645defe6 is not running
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 45617 0 --:--:-- --:--:-- --:--:-- 45617
Error response from daemon: Cannot kill container: 3151d97bcc4e: Container 3151d97bcc4e34aa63ab16939df3d7a785ee309c34e0a1e2b9813282f93c8465 is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-MAOPT-CATC/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Unable to find image 'docker.onedata.org/swagger-aggregator:1.5.0' locally
1.5.0: Pulling from swagger-aggregator
3e51df1a11ab: Pulling fs layer
afe1a672db39: Pulling fs layer
80bad90eec10: Pulling fs layer
ca328de5836f: Pulling fs layer
4f4fb700ef54: Pulling fs layer
f9032f2640a0: Pulling fs layer
982a4aef97ed: Pulling fs layer
aca789ddad1c: Pulling fs layer
a3e17b3016f6: Pulling fs layer
b67d253c6e22: Pulling fs layer
302c9eabb2bb: Pulling fs layer
bb6e4824f633: Pulling fs layer
4f4fb700ef54: Waiting
f9032f2640a0: Waiting
982a4aef97ed: Waiting
aca789ddad1c: Waiting
a3e17b3016f6: Waiting
b67d253c6e22: Waiting
302c9eabb2bb: Waiting
bb6e4824f633: Waiting
ca328de5836f: Waiting
afe1a672db39: Verifying Checksum
afe1a672db39: Download complete
80bad90eec10: Verifying Checksum
80bad90eec10: Download complete
ca328de5836f: Verifying Checksum
ca328de5836f: Download complete
4f4fb700ef54: Verifying Checksum
4f4fb700ef54: Download complete
3e51df1a11ab: Verifying Checksum
3e51df1a11ab: Download complete
aca789ddad1c: Verifying Checksum
aca789ddad1c: Download complete
a3e17b3016f6: Verifying Checksum
a3e17b3016f6: Download complete
b67d253c6e22: Verifying Checksum
b67d253c6e22: Download complete
f9032f2640a0: Download complete
302c9eabb2bb: Verifying Checksum
302c9eabb2bb: Download complete
bb6e4824f633: Verifying Checksum
bb6e4824f633: Download complete
982a4aef97ed: Verifying Checksum
982a4aef97ed: Download complete
3e51df1a11ab: Pull complete
afe1a672db39: Pull complete
80bad90eec10: Pull complete
ca328de5836f: Pull complete
4f4fb700ef54: Pull complete
f9032f2640a0: Pull complete
982a4aef97ed: Pull complete
aca789ddad1c: Pull complete
a3e17b3016f6: Pull complete
b67d253c6e22: Pull complete
302c9eabb2bb: Pull complete
bb6e4824f633: Pull complete
Digest: sha256:e2e8e762a03a0acdd49e63c4168157cb4e0e79f31f4e815561e9f4c65dbf8ac8
Status: Downloaded newer image for docker.onedata.org/swagger-aggregator:1.5.0
Unable to find image 'swaggerapi/swagger-codegen-cli:2.4.20' locally
2.4.20: Pulling from swaggerapi/swagger-codegen-cli
e7c96db7181b: Pulling fs layer
f910a506b6cb: Pulling fs layer
b6abafe80f63: Pulling fs layer
0d9127f69a1f: Pulling fs layer
0d9127f69a1f: Waiting
f910a506b6cb: Verifying Checksum
f910a506b6cb: Download complete
e7c96db7181b: Verifying Checksum
e7c96db7181b: Download complete
e7c96db7181b: Pull complete
0d9127f69a1f: Verifying Checksum
0d9127f69a1f: Download complete
f910a506b6cb: Pull complete
b6abafe80f63: Verifying Checksum
b6abafe80f63: Download complete
b6abafe80f63: Pull complete
0d9127f69a1f: Pull complete
Digest: sha256:e961c734f4a232ea050293e9b16aed4cc131ffecf4a7d8671f15f1d79bca8796
Status: Downloaded newer image for swaggerapi/swagger-codegen-cli:2.4.20
/bin/sh: 2: [[: not found
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
id: ‘user1’: no such user
command terminated with exit code 1
id: ‘user2’: no such user
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
command terminated with exit code 1
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Defaulted container "oneclient" out of: oneclient, wait-for-onezone (init), wait-for-token-dispenser (init)
rsync: change_dir#3 "/tmp/logs" failed: No such file or directory (2)
rsync error: errors selecting input/output files, dirs (code 3) at main.c(819) [sender=3.1.2]
command terminated with exit code 3
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 4653 100 4653 0 0 41918 0 --:--:-- --:--:-- --:--:-- 41918
Error response from daemon: Cannot kill container: 02eaba3fdcca: Container 02eaba3fdcca6361b6572e79c77bc7cc92117cb43d63bb7b5e3bb14e0bb3910a is not running
Error response from daemon: Cannot kill container: 88de1beb74e5: Container 88de1beb74e59e068a0dbde14c63bb2c25bcc84eac6c80a73a37e7fa6ed960ed is not running
Error response from daemon: Cannot kill container: db5e1014422f: Container db5e1014422f593d088e0a908753e9ab870c48e5a3733d8a3d49d35ab4b92b2a is not running
Error response from daemon: Cannot kill container: 9b36c6aa2a85: Container 9b36c6aa2a85320a5b4c8f7f71d21ae4ed066dc45778dab4451b04051c9fafc7 is not running
Error response from daemon: Cannot kill container: d7c78246bec1: Container d7c78246bec111a9f38ed0259fed944176712a36185f26cd8eaa5e5329a52739 is not running
Error response from daemon: Cannot kill container: b11736dc5cbb: Container b11736dc5cbb8bac02ae10141a8d291ffcd65a53a6fa16517e9d85f8d84774cf is not running
Error response from daemon: Cannot kill container: e0a5c09f1f3b: Container e0a5c09f1f3bc892318a5157b0064fc7917c715b780cd1235e777a695d4a6df6 is not running
Error response from daemon: Cannot kill container: 7ad84add25f7: Container 7ad84add25f707100b558a1661a6141e8142e44b54b6bf5f4fc7f13cc27cbd0e is not running
Error response from daemon: Cannot kill container: b26c3a4839b5: Container b26c3a4839b5a0851003dabbd0ec64dfe063da843b10683b324082d337e1c863 is not running
Error response from daemon: Cannot kill container: 781858df708e: Container 781858df708e9b940d4de52d8cfb1da17139fc56b43df917170ba4030f70e664 is not running
Error response from daemon: Cannot kill container: a3bc2c037fea: Container a3bc2c037feaadf5a4b848786321c8838b6b08b7ca224d0af667e65ff0dc69ef is not running
Error response from daemon: Cannot kill container: 6990e1d607ce: Container 6990e1d607ce65fc64bdac2941ce61dfc3bac7a6330b27720c83a2fa549f0e01 is not running
Error response from daemon: Cannot kill container: c3ffaaf71ae0: Container c3ffaaf71ae01209398fbd439350daa2dbb42f81aeba107fa1cb614b8aa8de09 is not running
Error response from daemon: Cannot kill container: 67eab605dd9d: Container 67eab605dd9d990bb0ba9642bcc12d998297c6b2bd313bd436d6b854412a6a46 is not running
Error response from daemon: Cannot kill container: af26137ecbea: Container af26137ecbeaee93aa60eeffba09f4606d8d4e7baf39d63232f16c2229ed8e8c is not running
Error response from daemon: Cannot kill container: a62ffe83b5b8: Container a62ffe83b5b8e0abcbf100b23c328deca53779f442ea1b228ece0031da477c52 is not running
Error response from daemon: Cannot kill container: e6557c76419a: Container e6557c76419ab3fba327af6118ccca11234ce11244933ee9aa59c064645defe6 is not running