| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238 | 
							- include(RunCTest)
 
- set(RunCMake_TEST_TIMEOUT 60)
 
- unset(ENV{CTEST_PARALLEL_LEVEL})
 
- unset(ENV{CTEST_OUTPUT_ON_FAILURE})
 
- set(CASE_CTEST_TEST_ARGS "")
 
- set(CASE_CTEST_TEST_LOAD "")
 
- function(run_ctest_test CASE_NAME)
 
-   set(CASE_CTEST_TEST_ARGS "${ARGN}")
 
-   run_ctest(${CASE_NAME})
 
- endfunction()
 
- run_ctest_test(TestQuiet QUIET)
 
- # Tests for the 'Test Load' feature of ctest
 
- #
 
- # Spoof a load average value to make these tests more reliable.
 
- set(ENV{__CTEST_FAKE_LOAD_AVERAGE_FOR_TESTING} 5)
 
- set(RunCTest_VERBOSE_FLAG -VV)
 
- # Verify that new tests are started when the load average falls below
 
- # our threshold.
 
- run_ctest_test(TestLoadPass TEST_LOAD 6)
 
- # Verify that new tests are not started when the load average exceeds
 
- # our threshold and that they then run once the load average drops.
 
- run_ctest_test(TestLoadWait TEST_LOAD 2)
 
- # Verify that when an invalid "TEST_LOAD" value is given, a warning
 
- # message is displayed and the value is ignored.
 
- run_ctest_test(TestLoadInvalid TEST_LOAD "ERR1")
 
- # Verify that new tests are started when the load average falls below
 
- # our threshold.
 
- set(CASE_CTEST_TEST_LOAD 7)
 
- run_ctest_test(CTestTestLoadPass)
 
- # Verify that new tests are not started when the load average exceeds
 
- # our threshold and that they then run once the load average drops.
 
- set(CASE_CTEST_TEST_LOAD 4)
 
- run_ctest_test(CTestTestLoadWait)
 
- # Verify that when an invalid "CTEST_TEST_LOAD" value is given,
 
- # a warning message is displayed and the value is ignored.
 
- set(CASE_CTEST_TEST_LOAD "ERR2")
 
- run_ctest_test(CTestTestLoadInvalid)
 
- # Verify that the "TEST_LOAD" value has higher precedence than
 
- # the "CTEST_TEST_LOAD" value
 
- set(CASE_CTEST_TEST_LOAD "ERR3")
 
- run_ctest_test(TestLoadOrder TEST_LOAD "ERR4")
 
- unset(ENV{__CTEST_FAKE_LOAD_AVERAGE_FOR_TESTING})
 
- unset(CASE_CTEST_TEST_LOAD)
 
- unset(RunCTest_VERBOSE_FLAG)
 
- function(run_TestChangeId)
 
-   set(CASE_TEST_PREFIX_CODE [[
 
-     set(CTEST_CHANGE_ID "<>1")
 
-   ]])
 
-   run_ctest(TestChangeId)
 
- endfunction()
 
- run_TestChangeId()
 
- function(run_TestOutputSize)
 
-   set(CASE_CTEST_TEST_ARGS EXCLUDE RunCMakeVersion)
 
-   set(CASE_TEST_PREFIX_CODE [[
 
- set(CTEST_CUSTOM_MAXIMUM_PASSED_TEST_OUTPUT_SIZE 10)
 
- set(CTEST_CUSTOM_MAXIMUM_FAILED_TEST_OUTPUT_SIZE 12)
 
-   ]])
 
-   set(CASE_CMAKELISTS_SUFFIX_CODE [[
 
- add_test(NAME PassingTest COMMAND ${CMAKE_COMMAND} -E echo PassingTestOutput)
 
- add_test(NAME FailingTest COMMAND ${CMAKE_COMMAND} -E no_such_command)
 
-   ]])
 
-   run_ctest(TestOutputSize)
 
- endfunction()
 
- run_TestOutputSize()
 
- # Test --test-output-truncation
 
- function(run_TestOutputTruncation mode expected)
 
-   set(CASE_CTEST_TEST_ARGS EXCLUDE RunCMakeVersion)
 
-   set(TRUNCATED_OUTPUT ${expected})  # used in TestOutputTruncation-check.cmake
 
-   set(CASE_TEST_PREFIX_CODE [[
 
- set( CTEST_CUSTOM_TEST_OUTPUT_TRUNCATION${mode})
 
-   ]])
 
-   set(CASE_CMAKELISTS_SUFFIX_CODE [[
 
- add_test(NAME Truncation_${mode} COMMAND ${CMAKE_COMMAND} -E echo 123456789)
 
-   ]])
 
-   run_ctest(TestOutputTruncation)
 
- endfunction()
 
- run_TestOutputTruncation("head" "...6789")
 
- run_TestOutputTruncation("middle" "12....*...89")
 
- run_TestOutputTruncation("tail" "12345...")
 
- run_ctest_test(TestRepeatBad1 REPEAT UNKNOWN:3)
 
- run_ctest_test(TestRepeatBad2 REPEAT UNTIL_FAIL:-1)
 
- function(run_TestRepeat case return_value )
 
-   set(CASE_CTEST_TEST_ARGS RETURN_VALUE result EXCLUDE RunCMakeVersion ${ARGN})
 
-   string(CONCAT suffix_code [[
 
- add_test(NAME testRepeat
 
-   COMMAND ${CMAKE_COMMAND} -D COUNT_FILE=${CMAKE_CURRENT_BINARY_DIR}/count.cmake
 
-                            -P "]] "${RunCMake_SOURCE_DIR}/TestRepeat${case}" [[.cmake")
 
- set_property(TEST testRepeat PROPERTY TIMEOUT 5)
 
-   ]])
 
-   string(APPEND CASE_CMAKELISTS_SUFFIX_CODE "${suffix_code}")
 
-   run_ctest(TestRepeat${case})
 
-   #write to end of the test file logic to Verify we get the expected
 
-   #return code
 
-   string(REPLACE "RETURN_VALUE:" "" return_value "${return_value}" )
 
-   file(APPEND "${RunCMake_BINARY_DIR}/TestRepeat${case}/test.cmake"
 
- "
 
-   set(expected_result ${return_value})
 
-   message(STATUS \${result})
 
-   if(NOT result EQUAL expected_result)
 
-     message(FATAL_ERROR \"expected a return value of: \${expected_result},
 
-                          instead got: \${result}\")
 
-   endif()
 
- "
 
-   )
 
- endfunction()
 
- run_TestRepeat(UntilFail RETURN_VALUE:1 REPEAT UNTIL_FAIL:3)
 
- run_TestRepeat(UntilPass RETURN_VALUE:0 REPEAT UNTIL_PASS:3)
 
- run_TestRepeat(AfterTimeout RETURN_VALUE:0 REPEAT AFTER_TIMEOUT:3)
 
- # test repeat and not run tests interact correctly
 
- set(CASE_CMAKELISTS_SUFFIX_CODE [[
 
- add_test(NAME testNotRun
 
-   COMMAND ${CMAKE_COMMAND}/doesnt_exist)
 
-   set_property(TEST testNotRun PROPERTY TIMEOUT 5)
 
-   ]])
 
- run_TestRepeat(NotRun RETURN_VALUE:1 REPEAT UNTIL_PASS:3)
 
- unset(CASE_CMAKELISTS_SUFFIX_CODE)
 
- # test --stop-on-failure
 
- function(run_stop_on_failure)
 
-   set(CASE_CTEST_TEST_ARGS EXCLUDE RunCMakeVersion)
 
-   set(CASE_CMAKELISTS_SUFFIX_CODE [[
 
- add_test(NAME StoppingTest COMMAND ${CMAKE_COMMAND} -E false)
 
- add_test(NAME NotRunTest COMMAND ${CMAKE_COMMAND} -E true)
 
-   ]])
 
-   run_ctest_test(stop-on-failure STOP_ON_FAILURE)
 
- endfunction()
 
- run_stop_on_failure()
 
- # Make sure environment gets logged
 
- function(run_environment)
 
-   set(ENV{BAD_ENVIRONMENT_VARIABLE} "Bad environment variable")
 
-   set(CASE_CMAKELISTS_SUFFIX_CODE [[
 
- set_property(TEST RunCMakeVersion PROPERTY ENVIRONMENT "ENV1=env1;ENV2=env2")
 
-   ]])
 
-   run_ctest(TestEnvironment)
 
- endfunction()
 
- run_environment()
 
- # test for OUTPUT_JUNIT
 
- run_ctest_test(OutputJUnit OUTPUT_JUNIT junit.xml REPEAT UNTIL_FAIL:2)
 
- # Verify that extra measurements get reported.
 
- function(run_measurements)
 
-   set(CASE_CMAKELISTS_SUFFIX_CODE [[
 
- add_test(
 
-   NAME double_measurement
 
-   COMMAND ${CMAKE_COMMAND} -E
 
-   echo <DartMeasurement type="numeric/double" name="my_custom_value">1.4847</DartMeasurement>)
 
- add_test(
 
-   NAME double_measurement2
 
-   COMMAND ${CMAKE_COMMAND} -E
 
-   echo <CTestMeasurement type="numeric/double" name="another_custom_value">1.8474</CTestMeasurement>)
 
- add_test(
 
-   NAME img_measurement
 
-   COMMAND ${CMAKE_COMMAND} -E
 
-   echo <DartMeasurementFile name="TestImage" type="image/png">]] ${IMAGE_DIR}/cmake-logo-16.png [[</DartMeasurementFile>)
 
- add_test(
 
-   NAME img_measurement2
 
-   COMMAND ${CMAKE_COMMAND} -E
 
-   echo <CTestMeasurementFile name="TestImage2" type="image/png">]] ${IMAGE_DIR}/cmake-logo-16.png [[</CTestMeasurementFile>)
 
- add_test(
 
-   NAME file_measurement
 
-   COMMAND ${CMAKE_COMMAND} -E
 
-   echo <DartMeasurementFile name="my_test_input_data" type="file">]] ${IMAGE_DIR}/cmake-logo-16.png [[</DartMeasurementFile>)
 
- add_test(
 
-   NAME file_measurement2
 
-   COMMAND ${CMAKE_COMMAND} -E
 
-   echo <CTestMeasurementFile name="another_test_input_data" type="file">]] ${IMAGE_DIR}/cmake-logo-16.png [[</CTestMeasurementFile>)
 
-   ]])
 
-   run_ctest(TestMeasurements)
 
- endfunction()
 
- run_measurements()
 
- # Verify that test output can override the Completion Status.
 
- function(run_completion_status)
 
-   set(CASE_CMAKELISTS_SUFFIX_CODE [[
 
- add_test(
 
-   NAME custom_details
 
-   COMMAND ${CMAKE_COMMAND} -E
 
-   echo test output\n<CTestDetails>CustomDetails</CTestDetails>\nmore output)
 
-   ]])
 
-   run_ctest(TestCompletionStatus)
 
- endfunction()
 
- run_completion_status()
 
- # Verify that running ctest_test() multiple times with different label arguments
 
- # doesn't break.
 
- function(run_changing_labels)
 
-   set(CASE_CMAKELISTS_SUFFIX_CODE [[
 
- add_test(NAME a COMMAND ${CMAKE_COMMAND} -E true)
 
- set_property(TEST a PROPERTY LABELS a)
 
- add_test(NAME b COMMAND ${CMAKE_COMMAND} -E true)
 
- set_property(TEST b PROPERTY LABELS b)
 
-   ]])
 
-   run_ctest(TestChangingLabels)
 
- endfunction()
 
- run_changing_labels()
 
- # Verify that test output can add additional labels
 
- function(run_extra_labels)
 
-   set(CASE_CMAKELISTS_SUFFIX_CODE [[
 
- add_test(
 
-   NAME custom_labels
 
-   COMMAND ${CMAKE_COMMAND} -E
 
-   echo before\n<CTestLabel>label2</CTestLabel>\n<CTestLabel>label1</CTestLabel>\n<CTestLabel>label3</CTestLabel>\n<CTestLabel>label2</CTestLabel>\nafter)
 
- set_tests_properties(custom_labels PROPERTIES LABELS "label1")
 
-   ]])
 
-   run_ctest(TestExtraLabels)
 
- endfunction()
 
- run_extra_labels()
 
 
  |