GUI acceptance tests using environment deployed from packages.

Build: #2261 failed

Job: Transfers directories failed

Stages & jobs

  1. Acceptance Test

Job result summary

Completed
Duration
3 minutes
Revision
32a8637a9a01b830c75d008c42ff27e7892513dd
Total tests
5
Fixed in
#2262 (Child of ODSRV-OPRPM-2186)

Tests

  • 5 tests in total
  • 5 tests failed
  • 5 failures are new
  • 2 minutes taken in total.
New test failures 5
Status Test Duration
Collapse Failed test_oneprovider_transfers_directories test_user_migrates_directory_to_remote_provider[1oz_2op_deployed]
15 secs
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_migrates_directory_to_remote_provider[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_transfers_directories test_user_replicates_directory_to_remote_provider[1oz_2op_deployed]
1 min
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_replicates_directory_to_remote_provider[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_transfers_directories test_user_replicates_directory_with_file_on_current_provider_to_the_same_provider[1oz_2op_deployed]
1 sec
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_replicates_directory_with_file_on_current_provider_to_the_same_provider[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_transfers_directories test_user_tries_to_migrate_directory_to_too_small_space_on_remote_provider[1oz_2op_deployed]
1 sec
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_tries_to_migrate_directory_to_too_small_space_on_remote_provider[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)
Collapse Failed test_oneprovider_transfers_directories test_user_tries_to_replicate_directory_to_too_small_space_on_remote_provider[1oz_2op_deployed]
1 sec
tests.utils.http_exceptions.HTTPBadRequest: [400] Bad Request: {"error":{"id":"errorOnNodes","details":{"hostnames":["dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local"],"error":{"id":"badValueToken","details":{"tokenError":{"id":"badToken","description":"Provided token could not be understood by the server."},"key":"token"},"description":"Bad value: provided \"token\" is not a valid token (see details)."}},"description":"Error on nodes dev-oneprovider-krakow-0.dev-oneprovider-krakow.default.svc.cluster.local: Bad value: provided \"token\" is not a valid token (see details)."}}
request = <FixtureRequest for <Function 'test_user_tries_to_replicate_directory_to_too_small_space_on_remote_provider[1oz_2op_deployed]'>>

    @pytest.mark.usefixtures(*function_args)
    def scenario_wrapper(request):
>       _execute_scenario(feature, scenario, request, encoding)

/usr/local/lib/python3.6/dist-packages/pytest_bdd/scenario.py:227: 
(47 more lines...)

Error summary

The build generated some errors. See the full build log for more details.

  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  19632      0 --:--:-- --:--:-- --:--:-- 19632
Error response from daemon: Cannot kill container: 4edc40880aa7: Container 4edc40880aa70b1c7f6ffef504e740909cf8f4a16ae5dd52283759c889a585b1 is not running
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'bamboos'
Submodule 'cdmi_swagger' (ssh://git@git.onedata.org:7999/vfs/cdmi-swagger.git) registered for path 'cdmi_swagger'
Submodule 'one_env' (ssh://git@git.onedata.org:7999/vfs/one-env.git) registered for path 'one_env'
Submodule 'onepanel_swagger' (ssh://git@git.onedata.org:7999/vfs/onepanel-swagger.git) registered for path 'onepanel_swagger'
Submodule 'oneprovider_swagger' (ssh://git@git.onedata.org:7999/vfs/oneprovider-swagger.git) registered for path 'oneprovider_swagger'
Submodule 'onezone_swagger' (ssh://git@git.onedata.org:7999/vfs/onezone-swagger.git) registered for path 'onezone_swagger'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CTD/onedata/bamboos'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CTD/onedata/cdmi_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CTD/onedata/one_env'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CTD/onedata/onepanel_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CTD/onedata/oneprovider_swagger'...
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CTD/onedata/onezone_swagger'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onepanel_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CTD/onedata/onepanel_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'oneprovider_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CTD/onedata/oneprovider_swagger/bamboos'...
Submodule 'bamboos' (ssh://git@git.onedata.org:7999/vfs/bamboos.git) registered for path 'onezone_swagger/bamboos'
Cloning into '/mnt/storage/bamboo-agent-home/xml-data/build-dir/ODSRV-GAPT-CTD/onedata/onezone_swagger/bamboos'...
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
Error: could not find tiller
Error: could not find tiller
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
error: failed to create clusterrolebinding: clusterrolebindings.rbac.authorization.k8s.io "serviceaccounts-cluster-admin" already exists
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(178) [Receiver=3.1.2]
command terminated with exit code 126
cp: cannot stat 'onedata/one_env/sources_info.yaml': No such file or directory
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100  4653  100  4653    0     0  46069      0 --:--:-- --:--:-- --:--:-- 45617
Error from server (NotFound): pods "dev-oneprovider-paris-0" not found
Error response from daemon: Cannot kill container: 2d129225ff1d: Container 2d129225ff1dd2d68b7761fd6000ba07135b913937020e8bacfd2f3b271f46ff is not running
Error response from daemon: Cannot kill container: f5d0f5696e3f: Container f5d0f5696e3f5b90d6cd63afdb5f2a129c8083ed0104fb78800302ae7001cb76 is not running
Error response from daemon: Cannot kill container: b9eafffc0a53: Container b9eafffc0a5396d1b41e64d7351ce47f0c340d4cf1b944ac12f61c71c504f43a is not running
Error response from daemon: Cannot kill container: df71d3a2dec1: Container df71d3a2dec19aa7b90cf0fc12c8eb7f5fee818d4f6a3f69a6a35810812988e5 is not running
Error response from daemon: Cannot kill container: 633cd1c2c125: Container 633cd1c2c125037cc2c526feb5393d192ef47f8c92fa548fbb78e07e562ec892 is not running
Error response from daemon: Cannot kill container: dfcfdf948a81: Container dfcfdf948a81faf2556dcf5b1876ed15f743d9013e9ee07f43940acb54e32f8a is not running
Error response from daemon: Cannot kill container: c515c2b68229: Container c515c2b6822935da5855c1229efff3092f9e3685856142e8b944f13ffd725ad6 is not running
Error response from daemon: Cannot kill container: df22142e18ed: Container df22142e18ed8a8b6e8261fd8fdedec7d57ed8f0b6318d520b8d0fc73a3693af is not running
Error response from daemon: Cannot kill container: 13919577cd47: Container 13919577cd47a462c42cc7f85caa7e6788bbac7d229b2de7e8f2a314213fd547 is not running
Error response from daemon: Cannot kill container: 88e7bdae0c59: Container 88e7bdae0c595b8140ad21a4a9ddc24a56dadb3ee0a74ac87354d7f58e883407 is not running
Error response from daemon: Cannot kill container: ff31e3c1cbc3: Container ff31e3c1cbc3014abbb44e11b0ee478a35b4100515db38ba074c04ce7715dd63 is not running
Error response from daemon: Cannot kill container: 0afe66945b12: Container 0afe66945b1264d3de5a20e33f2f1feee77466e69f4a7ad4b7344032014ea859 is not running
Error response from daemon: Cannot kill container: ea0733903a6e: Container ea0733903a6e8749724ef32b7bea6da54d4a126101de31179006a93aad7e0f9e is not running
Error response from daemon: Cannot kill container: d36955c5aff0: Container d36955c5aff055c6f58de34f3c6bcf6437e70fe6af8a8b038c73ac318dffe5f1 is not running
Error response from daemon: Cannot kill container: f8b9bd708663: Container f8b9bd708663bbf7507cb75e2f3ebfe9317449278462efa95530fc9f7e377f3b is not running
Error response from daemon: Cannot kill container: 08a35bf51bbc: Container 08a35bf51bbc21bd8d569db5feed771ed45b6427e1f1cd937d81743b9093d1d0 is not running
Error response from daemon: Cannot kill container: b5735fe12e48: Container b5735fe12e48e1d4f7723452b0d20a9367321ee88d961525ddbcb4458b9f9ad3 is not running
Error response from daemon: Cannot kill container: b9adc47e3fae: Container b9adc47e3fae55d8dcc5a5ed3197eea5d4df8d29eb0eb86e497b503aa49fa135 is not running
Error response from daemon: Cannot kill container: 9e933a8c487c: Container 9e933a8c487ce43af9ef6223d9cd955dd9965783ffaa139c1d1335fb882ea70a is not running