Tests scenarios for client

Scenarios for HAL- on HAL’s platform

The same procedure is used for all tests:

Software Author:

  1. prepare content

  2. fill out form

  3. submit

HAL moderator:

  1. review content submitted

  2. check metadata fields on HAL

  3. validate submission

SWH side:

  1. check content in SWH:

  • directory was created

  • revision was created

  • release was created when releaseNotes and softwareVersion was included (new feature!)

  • origin corresponds to HAL url

  1. check metadata fields on SWH (in revision)

  2. check directory

  3. check swh-id on HAL

  4. check browsability when entering SWH artifact from HAL

  5. check vault artifact recreation

  6. access deposit’s origin from SWH

scenario

test case

data

result

exceptions or specific checks

submit code

content: .tar.gz

.zip

success

submit code

content: .zip

.tar.gz

success

submit code

content: no content

empty

fail

blocked on HAL

submit code

content: double compression (.zip in .zip)

.zip x 2

fail

status failed on SWH

submit code

all metadata-single entry

metadata

success

check that all metadata is transmitted

submit code

multiple entries

metadata

success

languages / authors / descriptions

new version

new content- same metadata

content

success

check new swh-id in SWH and HAL

new version

same content- new metadata

metadata

?

dead angle- doesn’t arrives to SWH

new version

new content-new metadata

C & M

success

check artifacts history in revisions

submit code

deposit on another hal platform

C & M

success

Past known bugs:

  • v2 problem, where swh-id from first version is kept in the second version instead of the new swh-id.

  • when deposit workers are down- error 500 is returned on HAL without real explanation (because there is no error on SWH- deposit status stays deposited).