How attach files after an automated execution
A few weeks ago we purchased Zephyr Scale for Jira Cloud. I know how it works on Server version and now I implemented the cloud version for execution test, there I don't have any issues, however after an execution I can't upload files like the server version, where the url is like this "/testresult/{testResultId}/attachments" to post a file from a previous result. I'm trying to figure out how I can upload files, but checking the API information I could use "/automations/executions/custom" and I'm getting the following response: { "errorCode": 400, "message": "Couldn't find any mapped test cases", "status": "Bad Request" } How I can resolve this? Are there any alternatives for cloud version to accomplish this? In this place for Server version I was able to upload files after an execution with"/testresult/{testResultId}/attachments", for example:2.6KViews0likes3Commentsserver version: uploading test results to Zephyr Scale via the automation API
Hello! I'm using Zepyhr Scale Server and I would like to upload to Zephyr the results of my automation testsuite made with pytest. I've tried this POST request: post(url="https://{my-jira-host}/rest/atm/1.0/automation/execution/{projectKey}", auth=({my_username}, {my_password}), files={"file":open("test_results.zip","rb")}) but it doesn't work because the response is "errorMessages":["Invalid Custom Format JSON file"]}'. I'm uploading a zip file containing one xml file generated with pytest --junitxml=output/junitxml_report.xml as it's explained herehttps://support.smartbear.com/zephyr-scale-cloud/docs/test-automation/pytest-integration.html I've tried to make the same request with an API client (Postman) and the error is"InvalidZIPfile", even if I fail the authentication with a wrong username or even if I upload the xml file only. Maybe someone does the same thing and could help me? I'm a newbie 🙂 thanks!Solved6.6KViews0likes10CommentsZephyr Scale - Cucumber and Junit test results metadata
I am on Zephyr Scale Server. Details of automated tests results that gets sent to Zephyr Scale does not give a lot of meaningful data. Iam following Smartbear's integration guide to send results up to Zephyr Scale [https://support.smartbear.com/zephyr-scale-server/docs/test-automation/integrations/bamboo.html] i.e. by sending zippedcucumber.json or junit files to the appropriate endpoint. Problems : 1. All automated runs creates new cycle with the name "Automated build" always. It's hard to differentiate builds when working with multiple projects in parallel. I should be able control the created cycle name. 2. Details of environment, version etc are not updated from the results file 3. Details of failure are not captured in Zephyr scale even though it's on the the result file. The test simply fails without giving details of why it failed. It need to be able to update as many meaningful information as possible to the failed tests in Zephyr Scale I am aware that Zephyr apis [https://support.smartbear.com/zephyr-scale-server/api-docs/v1/]allows us to build a bespoke solution to fix the above problems. However, i am cautious of building additional code as wrappers and would like to use out of the box solution as much as possible for maintenance and cost purpose. What is the recommended automated test reporting practice from Zephyr Scale team?971Views0likes0CommentsAutomation Updates Wrong Test Version through API
In Zephyr Scale Cloud, if a test case has more than 1 version, and the test is being run through automation with the results being posted back to the test case through the API, the the results are posted to the latest version only. In some instance this is incorrect as version 2 might not yet be in use, and current executions still happen on version 1, but there is no way of indicating which version to use through the API. It appears that the automation does in fact execute the correct test externally, but just posts those results to the incorrect test version online, so the recorded evidence (and hence any reporting) is very clearly misaligned and will not make sense to anyone trying to audit results. Has anyone experienced that and have any ideas/workarounds?Solved1.7KViews0likes3Comments