<rss version="2.0" xmlns:a10="http://www.w3.org/2005/Atom"><channel><title>Inflectra Customer Forums: Test Cases which involve multiple software projects (Thread)</title><description> We have what we believe o be a very common situation:   a server which provides various REST APIs  various client apps (iOS, Android, Web, Desktop) which communicate with the server  the server and the four client apps are technically 5 different software projects with their own releases and builds  we use test automation  most of our Test Cases involve at least two of the above software projects (server + one of the apps), e.g. the login tests, where the apps are expected to check the credentials against the server  some Test Cases involve three or more of the above software projects, e.g. the iOS app sending a broadcast which should be received by the three other apps  we want to test all kinds of compatibility scenarios like can iOS app v18.3 login against a server v19.1?, can Android app 18.3 receive a broadcast sent from iOS app 19.2?      How would you do the above in SpiraTest? Were quite excited about the Test Configuration feature, where we could Test Case parameters like Client1, Client2, Server and create custom lists with all the currently supported app and server versions, but then we miss out on the traceability via Release and Build.  Alternatively, we could account for the fact that each of our 5 software projects above has their own releases and builds by creating 5 corresponding SpiraTest projects. However, how would we deal with Test Cases which involve multiple software projects (see above)? We could use the cross-project feature introduced in SpiraTest 5.1 by expressing our expectations for such multi-project scenarios as SpiraTest Requirements, copy the same Test Case into all SpiraTest Projects representing the involved software projects, and link all those Test Cases to the one Requirement. However, how would execute all those Test Cases? Using Automation, were scheduling Test Sets and have RemoteLaunch execute them. How would that work if were running a test which involves multiple software projects, represented by multiple Test Case clones across corresponding SpiraTest projects?  Any suggestions very welcome! </description><language>en-US</language><copyright>(C) Copyright 2006-2026 Inflectra Corporation.</copyright><managingEditor>support@inflectra.com</managingEditor><category domain="http://www.dmoz.org">/Computers/Software/Project_Management/</category><category domain="http://www.dmoz.org">/Computers/Software/Quality_Assurance/</category><generator>KronoDesk</generator><a10:contributor><a10:email>support@inflectra.com</a10:email></a10:contributor><a10:id>http://www.inflectra.com/kronodesk/forums/threads</a10:id><ttl>120</ttl><link>/Support/Forum/spiratest/best-practices/1980.aspx</link><item><guid isPermaLink="false">threadId=1980</guid><author>Shlomo Konwisser (shlomo.konwisser@mobileheartbeat.com)</author><category domain="http://www.inflectra.com/kronodesk/thread/tag">backwards compatibility</category><category domain="http://www.inflectra.com/kronodesk/thread/tag"> cross project</category><title>Test Cases which involve multiple software projects</title><description> We have what we believe o be a very common situation:   a server which provides various REST APIs  various client apps (iOS, Android, Web, Desktop) which communicate with the server  the server and the four client apps are technically 5 different software projects with their own releases and builds  we use test automation  most of our Test Cases involve at least two of the above software projects (server + one of the apps), e.g. the login tests, where the apps are expected to check the credentials against the server  some Test Cases involve three or more of the above software projects, e.g. the iOS app sending a broadcast which should be received by the three other apps  we want to test all kinds of compatibility scenarios like can iOS app v18.3 login against a server v19.1?, can Android app 18.3 receive a broadcast sent from iOS app 19.2?      How would you do the above in SpiraTest? Were quite excited about the Test Configuration feature, where we could Test Case parameters like Client1, Client2, Server and create custom lists with all the currently supported app and server versions, but then we miss out on the traceability via Release and Build.  Alternatively, we could account for the fact that each of our 5 software projects above has their own releases and builds by creating 5 corresponding SpiraTest projects. However, how would we deal with Test Cases which involve multiple software projects (see above)? We could use the cross-project feature introduced in SpiraTest 5.1 by expressing our expectations for such multi-project scenarios as SpiraTest Requirements, copy the same Test Case into all SpiraTest Projects representing the involved software projects, and link all those Test Cases to the one Requirement. However, how would execute all those Test Cases? Using Automation, were scheduling Test Sets and have RemoteLaunch execute them. How would that work if were running a test which involves multiple software projects, represented by multiple Test Case clones across corresponding SpiraTest projects?  Any suggestions very welcome! </description><pubDate>Fri, 01 Feb 2019 07:04:49 -0500</pubDate><a10:updated>2020-07-29T11:09:50-04:00</a10:updated><link>/Support/Forum/spiratest/best-practices/1980.aspx</link></item><item><guid isPermaLink="false">messageId=3519</guid><author>Jim R (donotreply5@kronodesk.net)</author><title> Hi Shlomo,  I think you have thought of the main options I would normally suggest:   Test Configura</title><description> Hi Shlomo,  I think you have thought of the main options I would normally suggest:   Test Configurations  Multiple Projects and Project Associations   You could just add the test case to the test set multiple times and set the parameter values for each instance. That might make the reporting back more clear (vs. test configurations).  In SpiraTeam 6.0 we will be linking test configurations to the Test Run so you can see which configuration entry was used.  Regards  Jim </description><pubDate>Wed, 06 Feb 2019 14:44:14 -0500</pubDate><a10:updated>2019-02-06T14:44:14-05:00</a10:updated><link>/Support/Forum/spiratest/best-practices/1980.aspx#reply3519</link></item></channel></rss>