i.e. by sending zipped cucumber.json or junit files to the appropriate endpoint.
1. All automated runs creates new cycle with the name "Automated build" always. It's hard to differentiate builds when working with multiple projects in parallel. I should be able control the created cycle name.
2. Details of environment, version etc are not updated from the results file
3. Details of failure are not captured in Zephyr scale even though it's on the the result file. The test simply fails without giving details of why it failed. It need to be able to update as many meaningful information as possible to the failed tests in Zephyr Scale
I am aware that Zephyr apis [https://support.smartbear.com/zephyr-scale-server/api-docs/v1/]allows us to build a bespoke solution to fix the above problems. However, i am cautious of building additional code as wrappers and would like to use out of the box solution as much as possible for maintenance and cost purpose.
What is the recommended automated test reporting practice from Zephyr Scale team?