GUI acceptance tests using environment deployed from packages.

Build: #2261 failed

Job: Onepanel basic failed

Stages & jobs

  1. Acceptance Test

Job result summary

Completed
Duration
34 minutes
Revision
32a8637a9a01b830c75d008c42ff27e7892513dd
Total tests
17
Fixed in
#2262 (Child of ODSRV-OPRPM-2186)

Tests

  • 17 tests in total
  • 9 tests failed
  • 9 failures are new
  • 32 minutes taken in total.
New test failures 9
Status Test Duration
Collapse Failed test_onepanel_basic test_support_space[1oz_1op_deployed-browser1]
4 mins
AssertionError: no info notify with ".*[Aa]dded.*support.*space.*" msg found
request = <FixtureRequest for <Function 'test_support_space[1oz_1op_deployed-browser1]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(32 more lines...)
Collapse Failed test_onepanel_basic test_support_space[1oz_1op_deployed-browser2]
3 mins
AssertionError: no info notify with ".*[Aa]dded.*support.*space.*" msg found
request = <FixtureRequest for <Function 'test_support_space[1oz_1op_deployed-browser2]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(32 more lines...)
Collapse Failed test_onepanel_basic test_user_changes_provider_name_and_domain[1oz_1op_deployed]
2 secs
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_changes_provider_name_and_domain[1oz_1op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_onepanel_basic test_user_deregisters_provider_registers_it_again_and_sees_that_provider_is_working[1oz_1op_deployed]
14 secs
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_deregisters_provider_registers_it_again_and_sees_that_provider_is_working[1oz_1op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_onepanel_basic test_user_fails_to_update_import_in_storage_that_is_not_importenabled[1oz_1op_deployed]
2 mins
RuntimeError: space5 space not found
oz_page = <class 'tests.gui.utils.onezone.OZLoggedIn'>
driver = <selenium.webdriver.chrome.webdriver.WebDriver (session="df6cd3c48fc59e92fed45ac8379cf041")>
name = 'space5'

    def _choose_space_from_menu_list(oz_page, driver, name):
        option = 'data'
        try:
(74 more lines...)
Collapse Failed test_onepanel_basic test_user_sees_that_synchronization_autoupdate_still_works_after_changing_mount_point_for_storage[1oz_1op_deployed]
3 mins
RuntimeError: space3 space not found
oz_page = <class 'tests.gui.utils.onezone.OZLoggedIn'>
driver = <selenium.webdriver.chrome.webdriver.WebDriver (session="0d68a580d3dd895f995956971ed6e762")>
name = 'space3'

    def _choose_space_from_menu_list(oz_page, driver, name):
        option = 'data'
        try:
(74 more lines...)
Collapse Failed test_onepanel_basic test_user_successfully_deletes_space_instead_of_revoking_it[1oz_1op_deployed]
2 mins
AssertionError: no info notify with ".*[Aa]dded.*support.*space.*" msg found
request = <FixtureRequest for <Function 'test_user_successfully_deletes_space_instead_of_revoking_it[1oz_1op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(40 more lines...)
Collapse Failed test_onepanel_basic test_user_uploads_files_on_freshly_supported_space_on_newly_created_storage[1oz_1op_deployed-browser_emergency-new_storage2]
3 mins
RuntimeError: no "new_storage2" found in StorageContentPage in content in Oneprovider Panel 21.02.0-alpha26
request = <FixtureRequest for <Function 'test_user_uploads_files_on_freshly_supported_space_on_newly_created_storage[1oz_1op_deployed-browser_emergency-new_storage2]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(39 more lines...)
Collapse Failed test_onepanel_basic test_user_uploads_files_on_freshly_supported_space_on_newly_created_storage[1oz_1op_deployed-browser_unified-new_storage1]
3 mins
RuntimeError: no "new_storage1" found in StorageContentPage in content in dev-oneprovider-krakow Oneprovider Panel 21.02.0-alpha26
request = <FixtureRequest for <Function 'test_user_uploads_files_on_freshly_supported_space_on_newly_created_storage[1oz_1op_deployed-browser_unified-new_storage1]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(40 more lines...)

Error summary

The build generated some errors. See the full build log for more details.

  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  48468      0 --:--:-- --:--:-- --:--:-- 48468
Error response from daemon: Cannot kill container: 139558058e08: Container 139558058e08e2a9fea228d91ea0ad1ec6852a8a4fa3defe5369a0ee92b7b30a is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COBT/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COBT/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COBT/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COBT/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COBT/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COBT/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COBT/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COBT/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COBT/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
W0816 20:25:28.497837     716 warnings.go:70] policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  47969      0 --:--:-- --:--:-- --:--:-- 47969
Error from server (NotFound): pods "dev-oneprovider-krakow-0" not found
Error response from daemon: Cannot kill container: 00a6e7e85e2b: Container 00a6e7e85e2b9dc7fc8033dc623454e944b9f4f21d407baa87e7feb40111f338 is not running
Error response from daemon: Cannot kill container: a35051eb6abd: Container a35051eb6abdaba9c2832da7847932cfd448c3caa328f5fe9a28b013279f7ed5 is not running
Error response from daemon: Cannot kill container: d80e81b2cbd5: Container d80e81b2cbd5b3d63d1c785dd8c057eb08914063805b7d0206f04d161396f20d is not running
Error response from daemon: Cannot kill container: 8947628e7f78: Container 8947628e7f78359f0312400b83dd3f5b93dcf3787cde8fbae53cdf77653d92e5 is not running
Error response from daemon: Cannot kill container: 07021d55a046: Container 07021d55a046f872f67a4047ae20038dcf7ebbbc1516368aa8282c58a43f0516 is not running
Error response from daemon: Cannot kill container: 6cd26b296541: Container 6cd26b29654134e37f12347d5f7acf9970d386e4005432acae2bb1bd034bd42d is not running
Error response from daemon: Cannot kill container: d57305a6c1ac: Container d57305a6c1accb872951101edc93a1893ba0d4e2481d58dfdb59a29327afe359 is not running
Error response from daemon: Cannot kill container: eb9b9ab512d9: Container eb9b9ab512d900baca0b8cab399e69fe9494ccf42ab3e2fd2922c239f91b09bb is not running
Error response from daemon: Cannot kill container: 6c297440e58c: Container 6c297440e58c80dead549e8f00dbd746295c2baf6ca415965738664ad9654f8e is not running
Error response from daemon: Cannot kill container: 49c6af755c65: Container 49c6af755c655d11aaaea5327f4502a0de1cc5fa6b70779d11322b224082a3bd is not running
Error response from daemon: Cannot kill container: 438f6887971f: Container 438f6887971f85843c3ba74485efc0b9b88c7cacebeb1050767fb7b9feac9ebc is not running
Error response from daemon: Cannot kill container: b567eb91a14c: Container b567eb91a14c03b9955432c56f05de20bd3b6dcc9aa487e946bc28f57e32e119 is not running
Error response from daemon: Cannot kill container: 5c094930da0a: Container 5c094930da0acfa6d9e91ca8d82af09fef2629cd5ba8bed5a55e32dabc89a673 is not running
Error response from daemon: Cannot kill container: 68d0c3dc1613: Container 68d0c3dc16132cafadec8fb4d433a9ce3a17747f1299903f254ced565365d4c9 is not running
Error response from daemon: Cannot kill container: 1cad765701ec: Container 1cad765701ec82d22709662cdf348959ea44a719a17a34246869a02497c9d293 is not running
Error response from daemon: Cannot kill container: 04680798ee0f: Container 04680798ee0f1aa80b4d605bcc7ea628153b7495164a14f1604656a4b68a5abc is not running