Running tests is commonly divided into a manual (without using any tool or automated script) and an automated approach.
If you want to run some manual tests (either as scripted or exploratory test procedure) then you just have to:
- Download a related branch using
composer require "mediawiki/semantic-media-wiki:dev-foo
(wherefoo
refers to the branch name) or in case you want to test the current master, use@dev
ordev-master
as version together withminimum-stability: dev
flag so that the branch/master can be fetched without any stability limitations. - Run
composer dump-autoload
to ensure that all registered classes are correctly initialized before starting any test procedure.
For the automated approach, Semantic MediaWiki relies on PHPUnit as scripted testing methodology. Scripted tests are used to verify that an expected behaviour occurs for codified requirements on whether a result can be accepted or has to be rejected for the given conditions.
- Unit test refers to a script that verifies results for a unit, module, or class against an expected outcome in an isolated environment
- Integration test (functional test) normally combines multiple components into a single process to verify results in a production like environment (DB access, sample data etc.)
- System test (and its individual modules) is treated as "black-box" to observe behaviour as a whole rather than its units
- Verify that PHUnit is installed and in case it is not use
composer require phpunit/phpunit:~4.6
to add the package - Verify that your MediaWiki installation comes with its test files and folders (e.g.
/mediawiki/tests
) in order for Semantic MediaWiki to have access to registered MW-core classes. If thetests
folder is missing, you may download it from a matched release source. - Run
composer phpunit
from the SMW base directory (e.g./extensions/SemanticMediaWiki
) using a standard command line tool which should output something similar to:
composer phpunit
MediaWiki: 1.25.2 (MediaWiki vendor autoloader)
Semantic MediaWiki: 2.3 alpha (SMWSQLStore3, mysql)
PHPUnit 4.3.5 by Sebastian Bergmann.
...
Information about PHPUnit in connection with MediaWiki can be found at smw.org and mediawiki.org.
Writing meaningful tests isn't easy nor is it complicated but it requires some diligence on how to setup a test and its environment. One simple rule is to avoid to use of hidden expectations or inheritance as remedy for the "less code is good code" aesthetics. Allow the code to be readable and if possible follow the arrange, act, assert pattern and yet again "Avoid doing magic".
For a short introduction on "How to write a test for Semantic MediaWiki", have a look at the video.
The use of MediaWikiTestCase
is discouraged as its binds tests and the test environment to MediaWiki. Generally it is best to use PHPUnit_Framework_TestCase
and in case where a MW database connection is required MwDBaseUnitTestCase
should be used instead.
QueryPrinterTestCase
base class for all query and result printersSpecialPageTestCase
derives fromSemanticMediaWikiTestCase
Integration tests are vital to confirm expected behaviour of a component from an integrative perspective that occurs through the interplay with its surroundings. SMW\Tests\Integration\
contains most of the tests that target the validation of reciprocity with MediaWiki together with listed services such as:
SPARQLStore
(fuseki
,virtuoso
,blazegraph
, orsesame
)- Other extensions that require SMW (
SM
,SESP
,SBL
etc.)
For details about the test environment see integration testing.
Integration tests can be written in a pseudo json
script in combination with a specialized TestCaseRunner
that handles the necessary object setup and tear down process for each test execution.
The script like test definition was introduced to lower the barrier of understanding of what is being tested by using a wikitext notation (internally PHPUnit is used by the ByJsonTestCaseProvider
to run/provide the actually test).
A new test file (with different test cases) is automatically loaded and run by a TestCaseRunner
as soon as it is placed in a location specified by the runner.
Each TestCaseRunner
contains a different interpretation of the json
script to keep the format straightforward but still allows for individual test assertions. Currently the following TestCaseProcessor
are provided:
RdfTestCaseProcessor
for rdf output assertionQueryTestCaseProcessor
to verify formats, queries, and conceptsParserTestCaseProcessor
to check for parser and store specific data
The section properties
and subjects
contain object entities that are planned to be used during the test which are specified by a name and a content (generally the page content in wikitext).
"properties": [ { "name": "Has description", "contents": "[[Has type::Text]]" } ], "subjects": [ { "name": "Page that contains text", "contents": "[[Has description::Foo]]" }, { "name": "Another page that contains text", "namespace": "NS_HELP", "contents": "[[Has description::Bar]]" } ]
The test result assertion is done in a very simplified way but expressive enough for users to understand the test objective and its expected results. For example, verifying that a result printer does output a certain string, one has to the define an expected output in terms of:
"expected-output": {
"to-contain": [
"<table class=\"sortable wikitable smwtable\">"
]
}
It can happen that an output is mixed with message dependent content (which when changing the site/content language will make the test script fail) and therefore it is recommended to fix the settings the test is intended for to pass with something like:
"settings": { "wgContLang": "en", "wgLang": "en", "smwgNamespacesWithSemanticLinks": { "NS_MAIN": true, "SMW_NS_PROPERTY": true } }
For other assertion options it is best to look at existing json
test files the ByJsonScript/Fixtures folder.
For details, please have a look at the benchmark guide document.
Running qunit tests in connection with MediaWiki requires to execute Special:JavaScriptTest. QUnit tests are currently not executed on Travis (see #136).
- Writing testable code
- Code coverage in a nutshell
- Test Doubles (mocks, stubs etc.) and how to write them