GUI acceptance tests using environment deployed from packages.

Build: #2261 failed

Job: Oneprovider ACL files failed

Stages & jobs

  1. Acceptance Test

Job result summary

Completed
Duration
4 minutes
Revision
32a8637a9a01b830c75d008c42ff27e7892513dd
Total tests
8
Fixed in
#2262 (Child of ODSRV-OPRPM-2186)

Tests

  • 8 tests in total
  • 8 tests failed
  • 8 failures are new
  • 1 minute taken in total.
New test failures 8
Status Test Duration
Collapse Failed test_oneprovider_acl_files test_read_files_acl[1oz_1op_deployed-group-group1-fails-all except [acl:read acl]]
13 secs
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_read_files_acl[1oz_1op_deployed-group-group1-fails-all except [acl:read acl]]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_acl_files test_read_files_acl[1oz_1op_deployed-group-group1-succeeds-[acl:read acl]]
1 sec
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_read_files_acl[1oz_1op_deployed-group-group1-succeeds-[acl:read acl]]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_acl_files test_read_files_acl[1oz_1op_deployed-user-user1-fails-all except [acl:read acl]]
1 sec
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_read_files_acl[1oz_1op_deployed-user-user1-fails-all except [acl:read acl]]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_acl_files test_read_files_acl[1oz_1op_deployed-user-user1-succeeds-[acl:read acl]]
1 sec
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_read_files_acl[1oz_1op_deployed-user-user1-succeeds-[acl:read acl]]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_acl_files test_rename_file[1oz_1op_deployed-group-group1-fails-all except [general:delete]]
1 sec
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_rename_file[1oz_1op_deployed-group-group1-fails-all except [general:delete]]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_acl_files test_rename_file[1oz_1op_deployed-group-group1-succeeds-[general:delete]]
1 sec
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_rename_file[1oz_1op_deployed-group-group1-succeeds-[general:delete]]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_acl_files test_rename_file[1oz_1op_deployed-user-user1-fails-all except [general:delete]]
1 sec
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_rename_file[1oz_1op_deployed-user-user1-fails-all except [general:delete]]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_acl_files test_rename_file[1oz_1op_deployed-user-user1-succeeds-[general:delete]]
1 min
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_rename_file[1oz_1op_deployed-user-user1-succeeds-[general:delete]]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)

Error summary

The build generated some errors. See the full build log for more details.

  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  29826      0 --:--:-- --:--:-- --:--:-- 29826
Error response from daemon: Cannot kill container: fb30617e392b: Container fb30617e392bd7bf19286b260b0cc6bd118f83ce39a55e0d466809258ea7c24c is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COAF/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COAF/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COAF/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COAF/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COAF/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COAF/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COAF/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COAF/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-COAF/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
W0816 20:02:03.543857     720 warnings.go:70] policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  36928      0 --:--:-- --:--:-- --:--:-- 36928
Error from server (NotFound): pods "dev-oneprovider-krakow-0" not found
Error response from daemon: Cannot kill container: 4828b3f8bdf4: Container 4828b3f8bdf4f1927f1b8dd5786c2d080a3a0a7a295d404cd8646eb32665cae6 is not running
Error response from daemon: Cannot kill container: 75340e006732: Container 75340e0067325512174fca1a96359817609bb70a3c33e0face7bdca5ba09d068 is not running
Error response from daemon: Cannot kill container: aa1fd4ddff41: Container aa1fd4ddff41e92b75335e3a200526b56fd1b9df802bcf970bc0573e7f2c83aa is not running
Error response from daemon: Cannot kill container: 7e92b066da6b: Container 7e92b066da6b206de7b283b9b07f1e0a63776a1614c5244556df165415b0c14f is not running
Error response from daemon: Cannot kill container: 7a21e2412add: Container 7a21e2412addee69a34e7be7a524a5c7719a01f28b228198f1b3547e325294ed is not running
Error response from daemon: Cannot kill container: f66633bfabee: Container f66633bfabeef420da1e6649d627e75b46f254fc3293083117be2622c1f96621 is not running
Error response from daemon: Cannot kill container: bb0c3a4b70cd: Container bb0c3a4b70cd73d7aabc050388d358a9e3da182e22a239e302a691380f269f08 is not running
Error response from daemon: Cannot kill container: adcf4e261dad: Container adcf4e261dad3decca5d409de270f2cccc29de52d01eae2ec02e5f50825f2b06 is not running
Error response from daemon: Cannot kill container: ab695f9e0b7c: Container ab695f9e0b7cc9611df0aae9c9d5fe7d21e91df3856dccf5ca66286f1821ca1b is not running
Error response from daemon: Cannot kill container: 799ddf321b85: Container 799ddf321b8560eb1fe842d2233424e8c748257bea685f298524f9ea733dcd41 is not running
Error response from daemon: Cannot kill container: 2b6a64a634cc: Container 2b6a64a634cc14a278eead82356c9493aee4370fcce11ad1c0af16cc8e7c5f3f is not running
Error response from daemon: Cannot kill container: e454d55f4199: Container e454d55f4199f34a3fd347596327687305df7429c0b7bc95467345c218673e55 is not running
Error response from daemon: Cannot kill container: 1e942052188a: Container 1e942052188a183d24667784f7dd9a992e6ad0d1403a133a9dfb95322cb3af2e is not running
Error response from daemon: Cannot kill container: ddf73c000e55: Container ddf73c000e5572dea30e0b30597b7f3aaea1d20f45fde58e3fb3cffa1a40782e is not running
Error response from daemon: Cannot kill container: 3e84c01e562d: Container 3e84c01e562d2156c49f032d34ba557326552aab8c9881ad559314fe6d570d15 is not running
Error response from daemon: Cannot kill container: da627030ed47: Container da627030ed47234de7744667e95b3528abb9d718d204df3a113b07f2350dbfb4 is not running