include(RunCTest) set(RunCMake_TEST_TIMEOUT 60) unset(ENV{CTEST_PARALLEL_LEVEL}) unset(ENV{CTEST_OUTPUT_ON_FAILURE}) set(CASE_CTEST_TEST_ARGS "") set(CASE_CTEST_TEST_LOAD "") function(run_ctest_test CASE_NAME) set(CASE_CTEST_TEST_ARGS "${ARGN}") run_ctest(${CASE_NAME}) endfunction() run_ctest_test(TestQuiet QUIET) # Tests for the 'Test Load' feature of ctest # # Spoof a load average value to make these tests more reliable. set(ENV{__CTEST_FAKE_LOAD_AVERAGE_FOR_TESTING} 5) set(RunCTest_VERBOSE_FLAG -VV) # Verify that new tests are started when the load average falls below # our threshold. run_ctest_test(TestLoadPass TEST_LOAD 6) # Verify that new tests are not started when the load average exceeds # our threshold and that they then run once the load average drops. run_ctest_test(TestLoadWait TEST_LOAD 2) # Verify that when an invalid "TEST_LOAD" value is given, a warning # message is displayed and the value is ignored. run_ctest_test(TestLoadInvalid TEST_LOAD "ERR1") # Verify that new tests are started when the load average falls below # our threshold. set(CASE_CTEST_TEST_LOAD 7) run_ctest_test(CTestTestLoadPass) # Verify that new tests are not started when the load average exceeds # our threshold and that they then run once the load average drops. set(CASE_CTEST_TEST_LOAD 4) run_ctest_test(CTestTestLoadWait) # Verify that when an invalid "CTEST_TEST_LOAD" value is given, # a warning message is displayed and the value is ignored. set(CASE_CTEST_TEST_LOAD "ERR2") run_ctest_test(CTestTestLoadInvalid) # Verify that the "TEST_LOAD" value has higher precedence than # the "CTEST_TEST_LOAD" value set(CASE_CTEST_TEST_LOAD "ERR3") run_ctest_test(TestLoadOrder TEST_LOAD "ERR4") unset(ENV{__CTEST_FAKE_LOAD_AVERAGE_FOR_TESTING}) unset(CASE_CTEST_TEST_LOAD) unset(RunCTest_VERBOSE_FLAG) function(run_TestChangeId) set(CASE_TEST_PREFIX_CODE [[ set(CTEST_CHANGE_ID "<>1") ]]) run_ctest(TestChangeId) endfunction() run_TestChangeId() function(run_TestOutputSize) set(CASE_CTEST_TEST_ARGS EXCLUDE RunCMakeVersion) set(CASE_TEST_PREFIX_CODE [[ set(CTEST_CUSTOM_MAXIMUM_PASSED_TEST_OUTPUT_SIZE 10) set(CTEST_CUSTOM_MAXIMUM_FAILED_TEST_OUTPUT_SIZE 12) ]]) set(CASE_CMAKELISTS_SUFFIX_CODE [[ add_test(NAME PassingTest COMMAND ${CMAKE_COMMAND} -E echo PassingTestOutput) add_test(NAME FailingTest COMMAND ${CMAKE_COMMAND} -E no_such_command) ]]) run_ctest(TestOutputSize) endfunction() run_TestOutputSize() run_ctest_test(TestRepeatBad1 REPEAT UNKNOWN:3) run_ctest_test(TestRepeatBad2 REPEAT UNTIL_FAIL:-1) function(run_TestRepeat case return_value ) set(CASE_CTEST_TEST_ARGS RETURN_VALUE result EXCLUDE RunCMakeVersion ${ARGN}) string(CONCAT suffix_code [[ add_test(NAME testRepeat COMMAND ${CMAKE_COMMAND} -D COUNT_FILE=${CMAKE_CURRENT_BINARY_DIR}/count.cmake -P "]] "${RunCMake_SOURCE_DIR}/TestRepeat${case}" [[.cmake") set_property(TEST testRepeat PROPERTY TIMEOUT 5) ]]) string(APPEND CASE_CMAKELISTS_SUFFIX_CODE "${suffix_code}") run_ctest(TestRepeat${case}) #write to end of the test file logic to Verify we get the expected #return code string(REPLACE "RETURN_VALUE:" "" return_value "${return_value}" ) file(APPEND "${RunCMake_BINARY_DIR}/TestRepeat${case}/test.cmake" " set(expected_result ${return_value}) message(STATUS \${result}) if(NOT result EQUAL expected_result) message(FATAL_ERROR \"expected a return value of: \${expected_result}, instead got: \${result}\") endif() " ) endfunction() run_TestRepeat(UntilFail RETURN_VALUE:1 REPEAT UNTIL_FAIL:3) run_TestRepeat(UntilPass RETURN_VALUE:0 REPEAT UNTIL_PASS:3) run_TestRepeat(AfterTimeout RETURN_VALUE:0 REPEAT AFTER_TIMEOUT:3) # test repeat and not run tests interact correctly set(CASE_CMAKELISTS_SUFFIX_CODE [[ add_test(NAME testNotRun COMMAND ${CMAKE_COMMAND}/doesnt_exist) set_property(TEST testNotRun PROPERTY TIMEOUT 5) ]]) run_TestRepeat(NotRun RETURN_VALUE:1 REPEAT UNTIL_PASS:3) unset(CASE_CMAKELISTS_SUFFIX_CODE) # test --stop-on-failure function(run_stop_on_failure) set(CASE_CTEST_TEST_ARGS EXCLUDE RunCMakeVersion) set(CASE_CMAKELISTS_SUFFIX_CODE [[ add_test(NAME StoppingTest COMMAND ${CMAKE_COMMAND} -E false) add_test(NAME NotRunTest COMMAND ${CMAKE_COMMAND} -E true) ]]) run_ctest_test(stop-on-failure STOP_ON_FAILURE) endfunction() run_stop_on_failure() # Make sure environment gets logged function(run_environment) set(ENV{BAD_ENVIRONMENT_VARIABLE} "Bad environment variable") set(CASE_CMAKELISTS_SUFFIX_CODE [[ set_property(TEST RunCMakeVersion PROPERTY ENVIRONMENT "ENV1=env1;ENV2=env2") ]]) run_ctest(TestEnvironment) endfunction() run_environment() # test for OUTPUT_JUNIT run_ctest_test(OutputJUnit OUTPUT_JUNIT junit.xml REPEAT UNTIL_FAIL:2) # Verify that extra measurements get reported. function(run_measurements) set(CASE_CMAKELISTS_SUFFIX_CODE [[ add_test( NAME double_measurement COMMAND ${CMAKE_COMMAND} -E echo 1.4847) add_test( NAME img_measurement COMMAND ${CMAKE_COMMAND} -E echo ]] ${IMAGE_DIR}/cmake-logo-16.png [[) add_test( NAME file_measurement COMMAND ${CMAKE_COMMAND} -E echo ]] ${IMAGE_DIR}/cmake-logo-16.png [[) ]]) run_ctest(TestMeasurements) endfunction() run_measurements() # Verify that test output can override the Completion Status. function(run_completion_status) set(CASE_CMAKELISTS_SUFFIX_CODE [[ add_test( NAME custom_details COMMAND ${CMAKE_COMMAND} -E echo test output\nCustomDetails\nmore output) ]]) run_ctest(TestCompletionStatus) endfunction() run_completion_status() # Verify that running ctest_test() multiple times with different label arguments # doesn't break. function(run_changing_labels) set(CASE_CMAKELISTS_SUFFIX_CODE [[ add_test(NAME a COMMAND ${CMAKE_COMMAND} -E true) set_property(TEST a PROPERTY LABELS a) add_test(NAME b COMMAND ${CMAKE_COMMAND} -E true) set_property(TEST b PROPERTY LABELS b) ]]) run_ctest(TestChangingLabels) endfunction() run_changing_labels()