GUI acceptance tests using environment deployed from packages.

Build: #2261 failed

Job: Multiprovider QoS storage failed

Stages & jobs

  1. Acceptance Test

Job result summary

Completed
Duration
6 minutes
Revision
32a8637a9a01b830c75d008c42ff27e7892513dd
Total tests
5
Fixed in
#2262 (Child of ODSRV-OPRPM-2186)

Tests

  • 5 tests in total
  • 5 tests failed
  • 5 failures are new
  • 4 minutes taken in total.
New test failures 5
Status Test Duration
Collapse Failed test_multiprovider_qos_storage test_every_possible_storage_matches_when_any_storage_condition_is_chosen[1oz_2op_deployed]
2 secs
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_every_possible_storage_matches_when_any_storage_condition_is_chosen[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_multiprovider_qos_storage test_file_is_replicated_after_migration_from_one_storage_with_qos_requirement[1oz_2op_deployed]
3 mins
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_file_is_replicated_after_migration_from_one_storage_with_qos_requirement[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_multiprovider_qos_storage test_user_can_select_one_of_storages_of_supporting_providers_in_qos_graphical_editor_and_it_causes_to_match_this_storage[1oz_2op_deployed]
3 secs
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_can_select_one_of_storages_of_supporting_providers_in_qos_graphical_editor_and_it_causes_to_match_this_storage[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_multiprovider_qos_storage test_user_sees_matching_storages_count_changing_while_editing_nested_expression_in_qos_visual_editor_and_submits_the_expression_successfully[1oz_2op_deployed]
32 secs
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_sees_matching_storages_count_changing_while_editing_nested_expression_in_qos_visual_editor_and_submits_the_expression_successfully[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_multiprovider_qos_storage test_user_successfully_adds_anystorage__storageid_qos_requirement[1oz_2op_deployed]
3 secs
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_successfully_adds_anystorage__storageid_qos_requirement[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)

Error summary

The build generated some errors. See the full build log for more details.

  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  43083      0 --:--:-- --:--:-- --:--:-- 43485
Error response from daemon: Cannot kill container: cfe7404ab493: Container cfe7404ab493211a5d0b8db0e2d8714b516b0753dc65773638d2b4a1e17f107c is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CMQS/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CMQS/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CMQS/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CMQS/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CMQS/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CMQS/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CMQS/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CMQS/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CMQS/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  40460      0 --:--:-- --:--:-- --:--:-- 40460
Error response from daemon: Cannot kill container: bf8640803d7a: Container bf8640803d7a9c70d3c4c1d663d4c3114df8a382a176730eb5dee4270d40826b is not running
Error response from daemon: Cannot kill container: a9303fefc94b: Container a9303fefc94b79ca6f09432f7e2323e3ce61d530520b2c1d60b4989325d4d95a is not running
Error response from daemon: Cannot kill container: 6dc56c285f20: Container 6dc56c285f20a1908962fdd36a4c8399c9df4bea8ec60271d911e3677ac93aba is not running
Error response from daemon: Cannot kill container: 9977e0303947: Container 9977e03039470b355d478e03d58f83048fdf0941d9652f38370adb9c267f497c is not running
Error response from daemon: Cannot kill container: ad24a316ca86: Container ad24a316ca8645713ef7fe9bc82b6b1302cfb27694c20fbcf5c9a720a84ff52a is not running
Error response from daemon: Cannot kill container: 2b2d2fe97250: Container 2b2d2fe972505d1302c56ef9f6aa07e9679242df6178e80517d17c828b92d753 is not running
Error response from daemon: Cannot kill container: 825a3be999ae: Container 825a3be999aeb5976d1f1d231ee1dd4c9a8950675cc20291a8bf11303aaeaecd is not running
Error response from daemon: Cannot kill container: 1a9f16a4448a: Container 1a9f16a4448a9c99b3a99d28fc6feb63a4bf0a4461da0e68937d55cb48ad5604 is not running
Error response from daemon: Cannot kill container: 5d004601717e: Container 5d004601717e19e5a1304da6b8ba798728e4a4d7866e6dfe12e3d13a48dcf064 is not running
Error response from daemon: Cannot kill container: 2daded278ccf: Container 2daded278ccfec43a86f7f4950f5af91de0b4d034b79b0192070e51e32d97bdd is not running
Error response from daemon: Cannot kill container: a5bd861472cd: Container a5bd861472cd532dc989c08675695ff45c3bbea4d311d25e0de5b354cdea9e19 is not running
Error response from daemon: Cannot kill container: 6271e0f6da12: Container 6271e0f6da12067b300bb1a43b936d2db4423766063a65c3b4e6e1c0fd3abb92 is not running
Error response from daemon: Cannot kill container: 4768a9f4eead: Container 4768a9f4eeadf3bdf47eb44fa43087951f3a3f42090b2f85b4e6c68bcf7742ab is not running
Error response from daemon: Cannot kill container: 49f8aecc60a6: Container 49f8aecc60a64859a9059b490dd813949d2985f7c4d5ba66873723a190941283 is not running
Error response from daemon: Cannot kill container: 484c4f1f1544: Container 484c4f1f15440b91384d11619a89c3b163ff8eea0796f687a36723e76ea4fc11 is not running
Error response from daemon: Cannot kill container: 431edc037e61: Container 431edc037e61f7a8dd72277ee21d407f52723c6004251cbaf03964a696635f06 is not running
Error response from daemon: Cannot kill container: 7babb2ed7cfd: Container 7babb2ed7cfd9265ac7b6946b9978ccbbaf5674fd028686c3fd1a92e053e5d52 is not running
Error response from daemon: Cannot kill container: fc956c9d1adb: Container fc956c9d1adbf8a09908b26db4f3019fd3241dbeb74215b87bd4e5afe8fa8ca8 is not running
Error response from daemon: Cannot kill container: 10f0b0c50d39: Container 10f0b0c50d392a588edf70ad168a6a41ec8bf7e71a99998a8c3dc30be0a65596 is not running