include(CatchMiscFunctions) if (CATCH_BUILD_SURROGATES) message(STATUS "Configuring targets for surrogate TUs") # If the folder does not exist before we ask for output redirect to # a file, it won't work. file(MAKE_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR}/surrogates) # Creates target to generate the surrogate TU for provided header. # Returns the path to the generated file. function(createSurrogateFileTarget sourceHeader pathToFile) set(pathPrefix ${PROJECT_SOURCE_DIR}/src) file(RELATIVE_PATH includePath ${pathPrefix} ${sourceHeader}) get_filename_component(basicFileName "${sourceHeader}" NAME_WE) set(surrogateFilePath ${CMAKE_CURRENT_BINARY_DIR}/surrogates/surrogate_${basicFileName}.cpp) add_custom_command( OUTPUT ${surrogateFilePath} COMMAND cmake -E echo "\#include <${includePath}>" > "${surrogateFilePath}" VERBATIM ) set(${pathToFile} ${surrogateFilePath} PARENT_SCOPE) endfunction() # Extracts all non-helper (e.g. catch_all.hpp) headers from the # Catch2 target, and returns them through the argument. function(ExtractCatch2Headers OutArg) get_target_property(targetSources Catch2 SOURCES) foreach(Source ${targetSources}) string(REGEX MATCH "^.*\\.hpp$" isHeader ${Source}) string(REGEX MATCH "_all.hpp" isAllHeader ${Source}) if(isHeader AND NOT isAllHeader) list(APPEND AllHeaders ${Source}) endif() endforeach() set(${OutArg} ${AllHeaders} PARENT_SCOPE) endfunction() ExtractCatch2Headers(mainHeaders) if (NOT mainHeaders) message(FATAL_ERROR "No headers in the main target were detected. Something is broken.") endif() foreach(header ${mainHeaders}) createSurrogateFileTarget(${header} pathToGeneratedFile) list(APPEND surrogateFiles ${pathToGeneratedFile}) endforeach() add_executable(Catch2SurrogateTarget ${surrogateFiles} ) target_link_libraries(Catch2SurrogateTarget PRIVATE Catch2WithMain) endif(CATCH_BUILD_SURROGATES) #### # Temporary workaround for VS toolset changes in 2017 # We need to disable property, but CMake doesn't support it # until 3.13 (not yet released) #### if (MSVC) configure_file(${CATCH_DIR}/tools/misc/SelfTest.vcxproj.user ${CMAKE_BINARY_DIR}/tests COPYONLY) endif(MSVC) #Temporary workaround # define the sources of the self test # Please keep these ordered alphabetically set(TEST_SOURCES ${SELF_TEST_DIR}/TestRegistrations.cpp ${SELF_TEST_DIR}/IntrospectiveTests/Clara.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/CmdLine.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/CmdLineHelpers.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/ColourImpl.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/Details.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/FloatingPoint.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/GeneratorsImpl.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/InternalBenchmark.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/PartTracker.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/RandomNumberGeneration.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/Reporters.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/Tag.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/TestCaseInfoHasher.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/TestSpecParser.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/TextFlow.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/Sharding.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/Stream.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/String.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/StringManip.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/Xml.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/ToString.tests.cpp ${SELF_TEST_DIR}/IntrospectiveTests/UniquePtr.tests.cpp ${SELF_TEST_DIR}/TimingTests/Sleep.tests.cpp ${SELF_TEST_DIR}/UsageTests/Approx.tests.cpp ${SELF_TEST_DIR}/UsageTests/BDD.tests.cpp ${SELF_TEST_DIR}/UsageTests/Benchmark.tests.cpp ${SELF_TEST_DIR}/UsageTests/Class.tests.cpp ${SELF_TEST_DIR}/UsageTests/Compilation.tests.cpp ${SELF_TEST_DIR}/UsageTests/Condition.tests.cpp ${SELF_TEST_DIR}/UsageTests/Decomposition.tests.cpp ${SELF_TEST_DIR}/UsageTests/EnumToString.tests.cpp ${SELF_TEST_DIR}/UsageTests/Exception.tests.cpp ${SELF_TEST_DIR}/UsageTests/Generators.tests.cpp ${SELF_TEST_DIR}/UsageTests/Message.tests.cpp ${SELF_TEST_DIR}/UsageTests/Misc.tests.cpp ${SELF_TEST_DIR}/UsageTests/ToStringByte.tests.cpp ${SELF_TEST_DIR}/UsageTests/ToStringChrono.tests.cpp ${SELF_TEST_DIR}/UsageTests/ToStringGeneral.tests.cpp ${SELF_TEST_DIR}/UsageTests/ToStringOptional.tests.cpp ${SELF_TEST_DIR}/UsageTests/ToStringPair.tests.cpp ${SELF_TEST_DIR}/UsageTests/ToStringTuple.tests.cpp ${SELF_TEST_DIR}/UsageTests/ToStringVariant.tests.cpp ${SELF_TEST_DIR}/UsageTests/ToStringVector.tests.cpp ${SELF_TEST_DIR}/UsageTests/ToStringWhich.tests.cpp ${SELF_TEST_DIR}/UsageTests/Tricky.tests.cpp ${SELF_TEST_DIR}/UsageTests/VariadicMacros.tests.cpp ${SELF_TEST_DIR}/UsageTests/MatchersRanges.tests.cpp ${SELF_TEST_DIR}/UsageTests/Matchers.tests.cpp ) # Specify the headers, too, so CLion recognises them as project files set(HEADERS ${TOP_LEVEL_HEADERS} ${EXTERNAL_HEADERS} ${INTERNAL_HEADERS} ${REPORTER_HEADERS} ${BENCHMARK_HEADERS} ${BENCHMARK_SOURCES} ) # Provide some groupings for IDEs #SOURCE_GROUP("benchmark" FILES ${BENCHMARK_HEADERS} ${BENCHMARK_SOURCES}) #SOURCE_GROUP("Tests" FILES ${TEST_SOURCES}) include(CTest) add_executable(SelfTest ${TEST_SOURCES}) target_link_libraries(SelfTest PRIVATE Catch2WithMain) if (BUILD_SHARED_LIBS AND WIN32) add_custom_command(TARGET SelfTest PRE_LINK COMMAND ${CMAKE_COMMAND} -E copy_if_different $ $ $ ) endif() if (CATCH_ENABLE_COVERAGE) set(ENABLE_COVERAGE ON CACHE BOOL "Enable coverage build." FORCE) find_package(codecov) add_coverage(SelfTest) list(APPEND LCOV_REMOVE_PATTERNS "'/usr/*'") coverage_evaluate() endif() # configure unit tests via CTest add_test(NAME RunTests COMMAND $ --order rand --rng-seed time) set_tests_properties(RunTests PROPERTIES FAIL_REGULAR_EXPRESSION "Filters:" COST 60 ) # Because CTest does not allow us to check both return code _and_ expected # output in one test, we run these commands twice. First time we check # the output, the second time we check the exit code. add_test(NAME List::Tests::Output COMMAND $ --list-tests --verbosity high) set_tests_properties(List::Tests::Output PROPERTIES PASS_REGULAR_EXPRESSION "[0-9]+ test cases" FAIL_REGULAR_EXPRESSION "Hidden Test" ) # This should be equivalent to the old --list-test-names-only and be usable # with --input-file. add_test(NAME List::Tests::Quiet COMMAND $ --list-tests --verbosity quiet) # Sadly we cannot ask for start-of-line and end-of-line in a ctest regex, # so we fail if we see space/tab at the start... set_tests_properties(List::Tests::Quiet PROPERTIES PASS_REGULAR_EXPRESSION "\"#1905 -- test spec parser properly clears internal state between compound tests\"[\r\n]" FAIL_REGULAR_EXPRESSION "[ \t]\"#1905 -- test spec parser properly clears internal state between compound tests\"" ) add_test(NAME List::Tests::ExitCode COMMAND $ --list-tests --verbosity high) add_test(NAME List::Tests::XmlOutput COMMAND $ --list-tests --verbosity high -r xml) set_tests_properties(List::Tests::XmlOutput PROPERTIES PASS_REGULAR_EXPRESSION "[0-9]+" FAIL_REGULAR_EXPRESSION "[0-9]+ test cases" ) add_test(NAME List::Tags::Output COMMAND $ --list-tags) set_tests_properties(List::Tags::Output PROPERTIES PASS_REGULAR_EXPRESSION "[0-9]+ tags" FAIL_REGULAR_EXPRESSION "\\[\\.\\]") add_test(NAME List::Tags::ExitCode COMMAND $ --list-tags) add_test(NAME List::Tags::XmlOutput COMMAND $ --list-tags -r xml) set_tests_properties(List::Tags::XmlOutput PROPERTIES PASS_REGULAR_EXPRESSION "18" FAIL_REGULAR_EXPRESSION "[0-9]+ tags" ) add_test(NAME List::Reporters::Output COMMAND $ --list-reporters) set_tests_properties(List::Reporters::Output PROPERTIES PASS_REGULAR_EXPRESSION "Available reporters:") add_test(NAME List::Reporters::ExitCode COMMAND $ --list-reporters) add_test(NAME List::Reporters::XmlOutput COMMAND $ --list-reporters -r xml) set_tests_properties(List::Reporters::XmlOutput PROPERTIES PASS_REGULAR_EXPRESSION "compact" FAIL_REGULAR_EXPRESSION "Available reporters:" ) add_test(NAME List::Listeners::Output COMMAND $ --list-listeners ) set_tests_properties(List::Listeners::Output PROPERTIES PASS_REGULAR_EXPRESSION "Registered listeners:" ) add_test(NAME List::Listeners::ExitCode COMMAND $ --list-listeners ) add_test(NAME List::Listeners::XmlOutput COMMAND $ --list-listeners --reporter xml ) set_tests_properties(List::Listeners::XmlOutput PROPERTIES PASS_REGULAR_EXPRESSION "" FAIL_REGULAR_EXPRESSION "Registered listeners:" ) add_test(NAME NoAssertions COMMAND $ -w NoAssertions "An empty test with no assertions") set_tests_properties(NoAssertions PROPERTIES PASS_REGULAR_EXPRESSION "No assertions in test case") # We cannot combine a regular expression on output with return code check # in one test, so we register two instead of making a checking script because # the runtime overhead is small enough. add_test(NAME TestSpecs::CombiningMatchingAndNonMatchingIsOk-1 COMMAND $ Tracker, "___nonexistent_test___") add_test(NAME TestSpecs::CombiningMatchingAndNonMatchingIsOk-2 COMMAND $ Tracker, "___nonexistent_test___") set_tests_properties(TestSpecs::CombiningMatchingAndNonMatchingIsOk-2 PROPERTIES PASS_REGULAR_EXPRESSION "No test cases matched '___nonexistent_test___'" FAIL_REGULAR_EXPRESSION "No tests ran" ) add_test(NAME TestSpecs::NoMatchedTestsFail COMMAND $ "___nonexistent_test___" ) set_tests_properties(TestSpecs::NoMatchedTestsFail PROPERTIES WILL_FAIL ON ) add_test(NAME TestSpecs::OverrideFailureWithNoMatchedTests COMMAND $ "___nonexistent_test___" --allow-running-no-tests ) add_test(NAME Warnings::UnmatchedTestSpecIsAccepted COMMAND $ Tracker --warn UnmatchedTestSpec ) set_tests_properties(Warnings::UnmatchedTestSpecIsAccepted PROPERTIES FAIL_REGULAR_EXPRESSION "Unrecognised warning option: " ) add_test(NAME Warnings::MultipleWarningsCanBeSpecified COMMAND $ Tracker --warn NoAssertions --warn UnmatchedTestSpec ) add_test(NAME TestSpecs::WarnUnmatchedTestSpecFailsWithUnmatchedTestSpec COMMAND $ Tracker, "___nonexistent_test___" --warn UnmatchedTestSpec ) set_tests_properties(TestSpecs::WarnUnmatchedTestSpecFailsWithUnmatchedTestSpec PROPERTIES WILL_FAIL ON ) add_test(NAME UnmatchedOutputFilter COMMAND $ [this-tag-does-not-exist]) set_tests_properties(UnmatchedOutputFilter PROPERTIES PASS_REGULAR_EXPRESSION "No test cases matched '\\[this-tag-does-not-exist\\]'" ) add_test(NAME FilteredSection-1 COMMAND $ \#1394 -c RunSection) set_tests_properties(FilteredSection-1 PROPERTIES FAIL_REGULAR_EXPRESSION "No tests ran") add_test(NAME FilteredSection-2 COMMAND $ \#1394\ nested -c NestedRunSection -c s1) set_tests_properties(FilteredSection-2 PROPERTIES FAIL_REGULAR_EXPRESSION "No tests ran") add_test( NAME FilteredSection::GeneratorsDontCauseInfiniteLoop-1 COMMAND $ "#2025: original repro" -c "fov_0" ) set_tests_properties(FilteredSection::GeneratorsDontCauseInfiniteLoop-1 PROPERTIES PASS_REGULAR_EXPRESSION "inside with fov: 0" # This should happen FAIL_REGULAR_EXPRESSION "inside with fov: 1" # This would mean there was no filtering ) # GENERATE between filtered sections (both are selected) add_test( NAME FilteredSection::GeneratorsDontCauseInfiniteLoop-2 COMMAND $ "#2025: same-level sections" -c "A" -c "B" --colour-mode none ) set_tests_properties(FilteredSection::GeneratorsDontCauseInfiniteLoop-2 PROPERTIES PASS_REGULAR_EXPRESSION "All tests passed \\(4 assertions in 1 test case\\)" ) # AppVeyor has a Python 2.7 in path, but doesn't have .py files as autorunnable add_test(NAME ApprovalTests COMMAND ${PYTHON_EXECUTABLE} ${CATCH_DIR}/tools/scripts/approvalTests.py $) set_tests_properties(ApprovalTests PROPERTIES FAIL_REGULAR_EXPRESSION "Results differed" COST 120 # We know that this is either the most, or second most, # expensive test in the test suite, so we give it high estimate for CI runs LABELS "uses-python" ) add_test(NAME RegressionCheck-1670 COMMAND $ "#1670 regression check" -c A -r compact) set_tests_properties(RegressionCheck-1670 PROPERTIES PASS_REGULAR_EXPRESSION "Passed 1 test case with 2 assertions.") add_test(NAME VersionCheck COMMAND $ -h) set_tests_properties(VersionCheck PROPERTIES PASS_REGULAR_EXPRESSION "Catch2 v${PROJECT_VERSION}") add_test(NAME LibIdentityTest COMMAND $ --libidentify) set_tests_properties(LibIdentityTest PROPERTIES PASS_REGULAR_EXPRESSION "description: A Catch2 test executable") add_test(NAME FilenameAsTagsTest COMMAND $ -\# --list-tags) set_tests_properties(FilenameAsTagsTest PROPERTIES PASS_REGULAR_EXPRESSION "\\[#Approx.tests\\]") # Check that the filename tags can also be matched against (#2064) add_test(NAME FilenameAsTagsMatching COMMAND $ -\# --list-tags [\#Approx.tests]) set_tests_properties(FilenameAsTagsMatching PROPERTIES PASS_REGULAR_EXPRESSION "\\[#Approx.tests\\]" # Avoids false positives by looking for start of line (newline) before the 0 FAIL_REGULAR_EXPRESSION "[\r\n]0 tag" ) add_test(NAME EscapeSpecialCharactersInTestNames COMMAND $ "Test with special\\, characters \"in name") set_tests_properties(EscapeSpecialCharactersInTestNames PROPERTIES PASS_REGULAR_EXPRESSION "1 assertion in 1 test case") add_test(NAME NegativeSpecNoHiddenTests COMMAND $ --list-tests ~[approval]) set_tests_properties(NegativeSpecNoHiddenTests PROPERTIES FAIL_REGULAR_EXPRESSION "\\[\\.\\]") add_test(NAME TestsInFile::SimpleSpecs COMMAND $ "-f ${SELF_TEST_DIR}/Misc/plain-old-tests.input") set_tests_properties(TestsInFile::SimpleSpecs PROPERTIES PASS_REGULAR_EXPRESSION "6 assertions in 2 test cases") add_test(NAME TestsInFile::EscapeSpecialCharacters COMMAND $ "-f ${SELF_TEST_DIR}/Misc/special-characters-in-file.input") set_tests_properties(TestsInFile::EscapeSpecialCharacters PROPERTIES PASS_REGULAR_EXPRESSION "1 assertion in 1 test case") add_test(NAME TestsInFile::InvalidTestNames-1 COMMAND $ "-f ${SELF_TEST_DIR}/Misc/invalid-test-names.input") set_tests_properties(TestsInFile::InvalidTestNames-1 PROPERTIES PASS_REGULAR_EXPRESSION "Invalid Filter: \"Test with special, characters in \\\\\" name\"" FAIL_REGULAR_EXPRESSION "No tests ran" ) add_test(NAME TagAlias COMMAND $ [@tricky] --list-tests) set_tests_properties(TagAlias PROPERTIES PASS_REGULAR_EXPRESSION "[0-9]+ matching test cases" FAIL_REGULAR_EXPRESSION "0 matching test cases" ) add_test(NAME RandomTestOrdering COMMAND ${PYTHON_EXECUTABLE} ${CATCH_DIR}/tests/TestScripts/testRandomOrder.py $) set_tests_properties(RandomTestOrdering PROPERTIES LABELS "uses-python" ) add_test(NAME CheckConvenienceHeaders COMMAND ${PYTHON_EXECUTABLE} ${CATCH_DIR}/tools/scripts/checkConvenienceHeaders.py ) set_tests_properties(CheckConvenienceHeaders PROPERTIES LABELS "uses-python" ) add_test(NAME "Benchmarking::SkipBenchmarkMacros" COMMAND $ "Skip benchmark macros" --reporter console --skip-benchmarks --colour-mode none ) set_tests_properties("Benchmarking::SkipBenchmarkMacros" PROPERTIES PASS_REGULAR_EXPRESSION "All tests passed \\(2 assertions in 1 test case\\)" FAIL_REGULAR_EXPRESSION "benchmark name" ) add_test(NAME "Benchmarking::FailureReporting::OptimizedOut" COMMAND $ "Failing benchmarks" -c "empty" -r xml # This test only makes sense with the optimizer being enabled when # the tests are being compiled. CONFIGURATIONS Release ) set_tests_properties("Benchmarking::FailureReporting::OptimizedOut" PROPERTIES PASS_REGULAR_EXPRESSION "could not measure benchmark\, maybe it was optimized away" FAIL_REGULAR_EXPRESSION "successes=\"1\"" ) add_test(NAME "Benchmarking::FailureReporting::ThrowingBenchmark" COMMAND $ "Failing benchmarks" -c "throw" -r xml ) set_tests_properties("Benchmarking::FailureReporting::ThrowingBenchmark" PROPERTIES PASS_REGULAR_EXPRESSION " "Failing benchmarks" -c "assert" -r xml ) set_tests_properties("Benchmarking::FailureReporting::FailedAssertion" PROPERTIES PASS_REGULAR_EXPRESSION " "Failing benchmarks" -c "fail" -r xml ) set_tests_properties("Benchmarking::FailureReporting::FailMacro" PROPERTIES PASS_REGULAR_EXPRESSION "This benchmark only fails\, nothing else" FAIL_REGULAR_EXPRESSION "successes=\"1\"" ) add_test(NAME "Benchmarking::FailureReporting::ShouldFailIsRespected" COMMAND $ "Failing benchmark respects should-fail" ) set_tests_properties("Benchmarking::FailureReporting::ShouldFailIsRespected" PROPERTIES PASS_REGULAR_EXPRESSION "1 failed as expected" ) add_test(NAME "ErrorHandling::InvalidTestSpecExitsEarly" COMMAND $ "[aa,a]" ) set_tests_properties("ErrorHandling::InvalidTestSpecExitsEarly" PROPERTIES PASS_REGULAR_EXPRESSION "Invalid Filter: \\[aa\,a\\]" FAIL_REGULAR_EXPRESSION "No tests ran" ) if (MSVC) set(_NullFile "NUL") else() set(_NullFile "/dev/null") endif() # This test checks that there is nothing written out from the process, # but if CMake is running the tests under Valgrind or similar tool, then # that will write its own output to stdout and the test would fail. if (NOT MEMORYCHECK_COMMAND) add_test(NAME "MultiReporter::CapturingReportersDontPropagateStdOut" COMMAND $ "Sends stuff to stdout and stderr" --reporter xml::out=${_NullFile} --reporter junit::out=${_NullFile} ) set_tests_properties("MultiReporter::CapturingReportersDontPropagateStdOut" PROPERTIES FAIL_REGULAR_EXPRESSION ".+" ) endif() add_test(NAME "MultiReporter::NonCapturingReportersPropagateStdout" COMMAND $ "Sends stuff to stdout and stderr" --reporter xml::out=${_NullFile} --reporter console::out=${_NullFile} ) set_tests_properties("MultiReporter::NonCapturingReportersPropagateStdout" PROPERTIES PASS_REGULAR_EXPRESSION "A string sent to stderr via clog" ) add_test(NAME "Outputs::DashAsOutLocationSendsOutputToStdout" COMMAND $ "Factorials are computed" --out=- --colour-mode none ) set_tests_properties("Outputs::DashAsOutLocationSendsOutputToStdout" PROPERTIES PASS_REGULAR_EXPRESSION "All tests passed \\(5 assertions in 1 test case\\)" ) add_test(NAME "Reporters::DashAsLocationInReporterSpecSendsOutputToStdout" COMMAND $ "Factorials are computed" --reporter console::out=- --colour-mode none ) set_tests_properties("Reporters::DashAsLocationInReporterSpecSendsOutputToStdout" PROPERTIES PASS_REGULAR_EXPRESSION "All tests passed \\(5 assertions in 1 test case\\)" ) add_test(NAME "Reporters::ReporterSpecificColourOverridesDefaultColour" COMMAND $ "Factorials are computed" --reporter console::colour-mode=ansi --colour-mode none ) set_tests_properties("Reporters::ReporterSpecificColourOverridesDefaultColour" PROPERTIES PASS_REGULAR_EXPRESSION "\\[1\;32mAll tests passed" ) add_test(NAME "Reporters::UnrecognizedOptionInSpecCausesError" COMMAND $ "Factorials are computed" --reporter console::bad-option=ansi ) set_tests_properties("Reporters::UnrecognizedOptionInSpecCausesError" PROPERTIES WILL_FAIL ON ) add_test(NAME "Colours::ColourModeCanBeExplicitlySetToAnsi" COMMAND $ "Factorials are computed" --reporter console --colour-mode ansi ) set_tests_properties("Colours::ColourModeCanBeExplicitlySetToAnsi" PROPERTIES PASS_REGULAR_EXPRESSION "\\[1\;32mAll tests passed" ) add_test(NAME "Reporters::JUnit::NamespacesAreNormalized" COMMAND $ --reporter junit "A TEST_CASE_METHOD testing junit classname normalization" ) set_tests_properties("Reporters::JUnit::NamespacesAreNormalized" PROPERTIES PASS_REGULAR_EXPRESSION "testcase classname=\"SelfTest(\.exe)?\\.A\\.B\\.TestClass\"" ) if (CATCH_ENABLE_CONFIGURE_TESTS) foreach(testName "DefaultReporter" "Disable" "DisableStringification" "ExperimentalRedirect") add_test(NAME "CMakeConfig::${testName}" COMMAND "${PYTHON_EXECUTABLE}" "${CMAKE_CURRENT_LIST_DIR}/TestScripts/testConfigure${testName}.py" "${CATCH_DIR}" "${CMAKE_CURRENT_BINARY_DIR}" ) set_tests_properties("CMakeConfig::${testName}" PROPERTIES COST 240 LABELS "uses-python" ) endforeach() endif() foreach (reporterName # "Automake" - the simple .trs format does not support any kind of comments/metadata "compact" "console" "JUnit" "SonarQube" "TAP" # "TeamCity" - does not seem to support test suite-level metadata/comments "XML") add_test(NAME "Reporters:RngSeed:${reporterName}" COMMAND $ "Factorials are computed" --reporter ${reporterName} --rng-seed 18181818 ) set_tests_properties("Reporters:RngSeed:${reporterName}" PROPERTIES PASS_REGULAR_EXPRESSION "18181818" ) endforeach() list(APPEND CATCH_WARNING_TARGETS SelfTest) set(CATCH_WARNING_TARGETS ${CATCH_WARNING_TARGETS} PARENT_SCOPE)