-
Notifications
You must be signed in to change notification settings - Fork 2
Setting up the SpecFlowPlus Runner Server
The SpecFlow+ Runner server collects execution statistics for your tests at a central location and uses this data to improve the efficiency of your test execution. For example, you can determine which tests are executed first based on the previous execution results, e.g. execute failing tests first.
#Prerequisites
Before you can set up the server, you need to install the SpecFlow+ Runner NuGet packages that contain the server components. Information on installing the NuGet packages can be found here.
#Installing the Server To install the SpecFlow+ Runner server:
- Create a new SQL database instance used to store the execution statistics.
- Locate the "server" directory in your solution's
\packages\SpecRun.Runner.x.y.z\tools
directory (created when you install the NuGet package). Copy the contents of the “server” directory to your server. - Enter your database connection string in the
<Connection Strings>
element of the SpecRun.Server.exe.config file. - Initialise the database using
SpecRun.exe initdatabase
from the command line. - Start the service using
SpecRun.exe start
from the command line. Note: You may need to assign the service to run under a user with appropriate permissions. - The connection to the server takes place via port 6365; ensure that your firewall on the server allows connections via this port.
- Enter the server’s URL in the
.srprofile
file in your Visual Studio project (<Server ServerURL =”http://MyServer:6365” publishResults=”true”>
). - Rebuild your solution and run your tests. You can verify that you have set up everything correctly by checking if records have been added to the database on the server.
You can use these statistics to execute failing and new tests first. To do so, set the testSchedulingMode
attribute in the <Execution>
element of your .srprofile
file to Adaptive
. SpecRun then uses the statistics in the pass/fail history to determine which tests to run first. Previously failing tests and new tests are executed before successful and stable tests.