| 1 | /* Copyright (c) 2006, 2010, Oracle and/or its affiliates. |
| 2 | Copyright (c) 2011, Monty Program Ab |
| 3 | |
| 4 | This program is free software; you can redistribute it and/or modify |
| 5 | it under the terms of the GNU General Public License as published by |
| 6 | the Free Software Foundation; version 2 of the License. |
| 7 | |
| 8 | This program is distributed in the hope that it will be useful, |
| 9 | but WITHOUT ANY WARRANTY; without even the implied warranty of |
| 10 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| 11 | GNU General Public License for more details. |
| 12 | |
| 13 | You should have received a copy of the GNU General Public License |
| 14 | along with this program; if not, write to the Free Software |
| 15 | Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA |
| 16 | |
| 17 | Library for providing TAP support for testing C and C++ was written |
| 18 | by Mats Kindahl <mats@mysql.com>. |
| 19 | */ |
| 20 | |
| 21 | #include "tap.h" |
| 22 | |
| 23 | #include "my_global.h" |
| 24 | |
| 25 | #include <stdlib.h> |
| 26 | #include <stdarg.h> |
| 27 | #include <stdio.h> |
| 28 | #include <string.h> |
| 29 | #include <signal.h> |
| 30 | |
| 31 | static ulong start_timer(void); |
| 32 | static void end_timer(ulong start_time,char *buff); |
| 33 | static void nice_time(double sec,char *buff,my_bool part_second); |
| 34 | |
| 35 | /* |
| 36 | Visual Studio 2003 does not know vsnprintf but knows _vsnprintf. |
| 37 | We don't put this #define elsewhere because we prefer my_vsnprintf |
| 38 | everywhere instead, except when linking with libmysys is not |
| 39 | desirable - the case here. |
| 40 | */ |
| 41 | #if defined(_MSC_VER) && ( _MSC_VER == 1310 ) |
| 42 | #define vsnprintf _vsnprintf |
| 43 | #endif |
| 44 | |
| 45 | /** |
| 46 | @defgroup MyTAP_Internal MyTAP Internals |
| 47 | |
| 48 | Internal functions and data structures for the MyTAP implementation. |
| 49 | */ |
| 50 | |
| 51 | /** |
| 52 | Test data structure. |
| 53 | |
| 54 | Data structure containing all information about the test suite. |
| 55 | |
| 56 | @ingroup MyTAP_Internal |
| 57 | */ |
| 58 | static TEST_DATA g_test = { NO_PLAN, 0, 0, "" }; |
| 59 | |
| 60 | /** |
| 61 | Output stream for test report message. |
| 62 | |
| 63 | The macro is just a temporary solution. |
| 64 | |
| 65 | @ingroup MyTAP_Internal |
| 66 | */ |
| 67 | #define tapout stdout |
| 68 | |
| 69 | /** |
| 70 | Emit the beginning of a test line, that is: "(not) ok", test number, |
| 71 | and description. |
| 72 | |
| 73 | To emit the directive, use the emit_dir() function |
| 74 | |
| 75 | @ingroup MyTAP_Internal |
| 76 | |
| 77 | @see emit_dir |
| 78 | |
| 79 | @param pass 'true' if test passed, 'false' otherwise |
| 80 | @param fmt Description of test in printf() format. |
| 81 | @param ap Vararg list for the description string above. |
| 82 | */ |
| 83 | static void |
| 84 | vemit_tap(int pass, char const *fmt, va_list ap) |
| 85 | { |
| 86 | fprintf(tapout, "%sok %d%s" , |
| 87 | pass ? "" : "not " , |
| 88 | ++g_test.last, |
| 89 | (fmt && *fmt) ? " - " : "" ); |
| 90 | if (fmt && *fmt) |
| 91 | vfprintf(tapout, fmt, ap); |
| 92 | fflush(tapout); |
| 93 | } |
| 94 | |
| 95 | |
| 96 | /** |
| 97 | Emit a TAP directive. |
| 98 | |
| 99 | TAP directives are comments after that have the form: |
| 100 | |
| 101 | @code |
| 102 | ok 1 # skip reason for skipping |
| 103 | not ok 2 # todo some text explaining what remains |
| 104 | @endcode |
| 105 | |
| 106 | @ingroup MyTAP_Internal |
| 107 | |
| 108 | @param dir Directive as a string |
| 109 | @param why Explanation string |
| 110 | */ |
| 111 | static void |
| 112 | emit_dir(const char *dir, const char *why) |
| 113 | { |
| 114 | fprintf(tapout, " # %s %s" , dir, why); |
| 115 | fflush(tapout); |
| 116 | } |
| 117 | |
| 118 | |
| 119 | /** |
| 120 | Emit a newline to the TAP output stream. |
| 121 | |
| 122 | @ingroup MyTAP_Internal |
| 123 | */ |
| 124 | static void |
| 125 | emit_endl() |
| 126 | { |
| 127 | fprintf(tapout, "\n" ); |
| 128 | fflush(tapout); |
| 129 | } |
| 130 | |
| 131 | static void |
| 132 | handle_core_signal(int signo) |
| 133 | { |
| 134 | BAIL_OUT("Signal %d thrown\n" , signo); |
| 135 | } |
| 136 | |
| 137 | void |
| 138 | BAIL_OUT(char const *fmt, ...) |
| 139 | { |
| 140 | va_list ap; |
| 141 | va_start(ap, fmt); |
| 142 | fprintf(tapout, "Bail out! " ); |
| 143 | vfprintf(tapout, fmt, ap); |
| 144 | diag("%d tests planned, %d failed, %d was last executed" , |
| 145 | g_test.plan, g_test.failed, g_test.last); |
| 146 | emit_endl(); |
| 147 | va_end(ap); |
| 148 | exit(255); |
| 149 | } |
| 150 | |
| 151 | |
| 152 | void |
| 153 | diag(char const *fmt, ...) |
| 154 | { |
| 155 | va_list ap; |
| 156 | va_start(ap, fmt); |
| 157 | fprintf(tapout, "# " ); |
| 158 | vfprintf(tapout, fmt, ap); |
| 159 | emit_endl(); |
| 160 | va_end(ap); |
| 161 | } |
| 162 | |
| 163 | typedef struct signal_entry { |
| 164 | int signo; |
| 165 | void (*handler)(int); |
| 166 | } signal_entry; |
| 167 | |
| 168 | static signal_entry install_signal[]= { |
| 169 | { SIGINT, handle_core_signal }, |
| 170 | { SIGQUIT, handle_core_signal }, |
| 171 | { SIGILL, handle_core_signal }, |
| 172 | { SIGABRT, handle_core_signal }, |
| 173 | { SIGFPE, handle_core_signal }, |
| 174 | { SIGSEGV, handle_core_signal } |
| 175 | #ifdef SIGBUS |
| 176 | , { SIGBUS, handle_core_signal } |
| 177 | #endif |
| 178 | #ifdef SIGXCPU |
| 179 | , { SIGXCPU, handle_core_signal } |
| 180 | #endif |
| 181 | #ifdef SIGXCPU |
| 182 | , { SIGXFSZ, handle_core_signal } |
| 183 | #endif |
| 184 | #ifdef SIGXCPU |
| 185 | , { SIGSYS, handle_core_signal } |
| 186 | #endif |
| 187 | #ifdef SIGXCPU |
| 188 | , { SIGTRAP, handle_core_signal } |
| 189 | #endif |
| 190 | }; |
| 191 | |
| 192 | int skip_big_tests= 1; |
| 193 | ulong start_time= 0; |
| 194 | |
| 195 | void |
| 196 | plan(int count) |
| 197 | { |
| 198 | char *config= getenv("MYTAP_CONFIG" ); |
| 199 | size_t i; |
| 200 | |
| 201 | start_time= start_timer(); |
| 202 | |
| 203 | if (config) |
| 204 | skip_big_tests= strcmp(config, "big" ); |
| 205 | |
| 206 | setvbuf(tapout, 0, _IONBF, 0); /* provide output at once */ |
| 207 | /* |
| 208 | Install signal handler |
| 209 | */ |
| 210 | |
| 211 | for (i= 0; i < sizeof(install_signal)/sizeof(*install_signal); ++i) |
| 212 | signal(install_signal[i].signo, install_signal[i].handler); |
| 213 | |
| 214 | g_test.plan= count; |
| 215 | switch (count) |
| 216 | { |
| 217 | case NO_PLAN: |
| 218 | break; |
| 219 | default: |
| 220 | if (count > 0) |
| 221 | { |
| 222 | fprintf(tapout, "1..%d\n" , count); |
| 223 | fflush(tapout); |
| 224 | } |
| 225 | break; |
| 226 | } |
| 227 | } |
| 228 | |
| 229 | |
| 230 | void |
| 231 | skip_all(char const *reason, ...) |
| 232 | { |
| 233 | va_list ap; |
| 234 | va_start(ap, reason); |
| 235 | fprintf(tapout, "1..0 # skip " ); |
| 236 | vfprintf(tapout, reason, ap); |
| 237 | fflush(tapout); |
| 238 | va_end(ap); |
| 239 | exit(0); |
| 240 | } |
| 241 | |
| 242 | void |
| 243 | ok(int pass, char const *fmt, ...) |
| 244 | { |
| 245 | va_list ap; |
| 246 | va_start(ap, fmt); |
| 247 | |
| 248 | if (!pass && *g_test.todo == '\0') |
| 249 | ++g_test.failed; |
| 250 | |
| 251 | vemit_tap(pass, fmt, ap); |
| 252 | va_end(ap); |
| 253 | if (*g_test.todo != '\0') |
| 254 | emit_dir("todo" , g_test.todo); |
| 255 | emit_endl(); |
| 256 | } |
| 257 | |
| 258 | void |
| 259 | ok1(int const pass) |
| 260 | { |
| 261 | va_list ap; |
| 262 | |
| 263 | memset(&ap, 0, sizeof(ap)); |
| 264 | |
| 265 | if (!pass && *g_test.todo == '\0') |
| 266 | ++g_test.failed; |
| 267 | |
| 268 | vemit_tap(pass, NULL, ap); |
| 269 | |
| 270 | if (*g_test.todo != '\0') |
| 271 | emit_dir("todo" , g_test.todo); |
| 272 | |
| 273 | emit_endl(); |
| 274 | } |
| 275 | |
| 276 | void |
| 277 | skip(int how_many, char const * const fmt, ...) |
| 278 | { |
| 279 | char reason[80]; |
| 280 | if (fmt && *fmt) |
| 281 | { |
| 282 | va_list ap; |
| 283 | va_start(ap, fmt); |
| 284 | vsnprintf(reason, sizeof(reason), fmt, ap); |
| 285 | va_end(ap); |
| 286 | } |
| 287 | else |
| 288 | reason[0] = '\0'; |
| 289 | |
| 290 | while (how_many-- > 0) |
| 291 | { |
| 292 | va_list ap; |
| 293 | memset((char*) &ap, 0, sizeof(ap)); /* Keep compiler happy */ |
| 294 | vemit_tap(1, NULL, ap); |
| 295 | emit_dir("skip" , reason); |
| 296 | emit_endl(); |
| 297 | } |
| 298 | } |
| 299 | |
| 300 | |
| 301 | void |
| 302 | todo_start(char const *message, ...) |
| 303 | { |
| 304 | va_list ap; |
| 305 | va_start(ap, message); |
| 306 | vsnprintf(g_test.todo, sizeof(g_test.todo), message, ap); |
| 307 | va_end(ap); |
| 308 | } |
| 309 | |
| 310 | void |
| 311 | todo_end() |
| 312 | { |
| 313 | *g_test.todo = '\0'; |
| 314 | } |
| 315 | |
| 316 | int exit_status() |
| 317 | { |
| 318 | char buff[60]; |
| 319 | |
| 320 | /* |
| 321 | If there were no plan, we write one last instead. |
| 322 | */ |
| 323 | if (g_test.plan == NO_PLAN) |
| 324 | plan(g_test.last); |
| 325 | |
| 326 | if (g_test.plan != g_test.last) |
| 327 | { |
| 328 | diag("%d tests planned but%s %d executed" , |
| 329 | g_test.plan, (g_test.plan > g_test.last ? " only" : "" ), g_test.last); |
| 330 | return EXIT_FAILURE; |
| 331 | } |
| 332 | |
| 333 | if (g_test.failed > 0) |
| 334 | { |
| 335 | diag("Failed %d tests!" , g_test.failed); |
| 336 | return EXIT_FAILURE; |
| 337 | } |
| 338 | if (start_time) |
| 339 | { |
| 340 | end_timer(start_time, buff); |
| 341 | printf("Test took %s\n" , buff); |
| 342 | fflush(stdout); |
| 343 | } |
| 344 | |
| 345 | return EXIT_SUCCESS; |
| 346 | } |
| 347 | |
| 348 | #if defined(__WIN__) || defined(__NETWARE__) |
| 349 | #include <time.h> |
| 350 | #else |
| 351 | #include <sys/times.h> |
| 352 | #ifdef _SC_CLK_TCK // For mit-pthreads |
| 353 | #undef CLOCKS_PER_SEC |
| 354 | #define CLOCKS_PER_SEC (sysconf(_SC_CLK_TCK)) |
| 355 | #endif |
| 356 | #endif |
| 357 | |
| 358 | static ulong start_timer(void) |
| 359 | { |
| 360 | #if defined(__WIN__) || defined(__NETWARE__) |
| 361 | return clock(); |
| 362 | #else |
| 363 | struct tms tms_tmp; |
| 364 | return times(&tms_tmp); |
| 365 | #endif |
| 366 | } |
| 367 | |
| 368 | |
| 369 | /** |
| 370 | Write as many as 52+1 bytes to buff, in the form of a legible |
| 371 | duration of time. |
| 372 | |
| 373 | len("4294967296 days, 23 hours, 59 minutes, 60.00 seconds") -> 52 |
| 374 | */ |
| 375 | |
| 376 | static void nice_time(double sec,char *buff, my_bool part_second) |
| 377 | { |
| 378 | ulong tmp; |
| 379 | if (sec >= 3600.0*24) |
| 380 | { |
| 381 | tmp=(ulong) (sec/(3600.0*24)); |
| 382 | sec-=3600.0*24*tmp; |
| 383 | buff+= sprintf(buff, "%ld %s" , tmp, tmp > 1 ? " days " : " day " ); |
| 384 | } |
| 385 | if (sec >= 3600.0) |
| 386 | { |
| 387 | tmp=(ulong) (sec/3600.0); |
| 388 | sec-=3600.0*tmp; |
| 389 | buff+= sprintf(buff, "%ld %s" , tmp, tmp > 1 ? " hours " : " hour " ); |
| 390 | } |
| 391 | if (sec >= 60.0) |
| 392 | { |
| 393 | tmp=(ulong) (sec/60.0); |
| 394 | sec-=60.0*tmp; |
| 395 | buff+= sprintf(buff, "%ld min " , tmp); |
| 396 | } |
| 397 | if (part_second) |
| 398 | sprintf(buff,"%.2f sec" ,sec); |
| 399 | else |
| 400 | sprintf(buff,"%d sec" ,(int) sec); |
| 401 | } |
| 402 | |
| 403 | |
| 404 | static void end_timer(ulong start_time,char *buff) |
| 405 | { |
| 406 | nice_time((double) (start_timer() - start_time) / |
| 407 | CLOCKS_PER_SEC,buff,1); |
| 408 | } |
| 409 | |
| 410 | |
| 411 | /** |
| 412 | @mainpage Testing C and C++ using MyTAP |
| 413 | |
| 414 | @section IntroSec Introduction |
| 415 | |
| 416 | Unit tests are used to test individual components of a system. In |
| 417 | contrast, functional tests usually test the entire system. The |
| 418 | rationale is that each component should be correct if the system is |
| 419 | to be correct. Unit tests are usually small pieces of code that |
| 420 | tests an individual function, class, a module, or other unit of the |
| 421 | code. |
| 422 | |
| 423 | Observe that a correctly functioning system can be built from |
| 424 | "faulty" components. The problem with this approach is that as the |
| 425 | system evolves, the bugs surface in unexpected ways, making |
| 426 | maintenance harder. |
| 427 | |
| 428 | The advantages of using unit tests to test components of the system |
| 429 | are several: |
| 430 | |
| 431 | - The unit tests can make a more thorough testing than the |
| 432 | functional tests by testing correctness even for pathological use |
| 433 | (which shouldn't be present in the system). This increases the |
| 434 | overall robustness of the system and makes maintenance easier. |
| 435 | |
| 436 | - It is easier and faster to find problems with a malfunctioning |
| 437 | component than to find problems in a malfunctioning system. This |
| 438 | shortens the compile-run-edit cycle and therefore improves the |
| 439 | overall performance of development. |
| 440 | |
| 441 | - The component has to support at least two uses: in the system and |
| 442 | in a unit test. This leads to more generic and stable interfaces |
| 443 | and in addition promotes the development of reusable components. |
| 444 | |
| 445 | For example, the following are typical functional tests: |
| 446 | - Does transactions work according to specifications? |
| 447 | - Can we connect a client to the server and execute statements? |
| 448 | |
| 449 | In contrast, the following are typical unit tests: |
| 450 | |
| 451 | - Can the 'String' class handle a specified list of character sets? |
| 452 | - Does all operations for 'my_bitmap' produce the correct result? |
| 453 | - Does all the NIST test vectors for the AES implementation encrypt |
| 454 | correctly? |
| 455 | |
| 456 | |
| 457 | @section UnitTest Writing unit tests |
| 458 | |
| 459 | The purpose of writing unit tests is to use them to drive component |
| 460 | development towards a solution that passes the tests. This means that the |
| 461 | unit tests has to be as complete as possible, testing at least: |
| 462 | |
| 463 | - Normal input |
| 464 | - Borderline cases |
| 465 | - Faulty input |
| 466 | - Error handling |
| 467 | - Bad environment |
| 468 | |
| 469 | @subsection NormalSubSec Normal input |
| 470 | |
| 471 | This is to test that the component have the expected behaviour. |
| 472 | This is just plain simple: test that it works. For example, test |
| 473 | that you can unpack what you packed, adding gives the sum, pincing |
| 474 | the duck makes it quack. |
| 475 | |
| 476 | This is what everybody does when they write tests. |
| 477 | |
| 478 | |
| 479 | @subsection BorderlineTests Borderline cases |
| 480 | |
| 481 | If you have a size anywhere for your component, does it work for |
| 482 | size 1? Size 0? Sizes close to <code>UINT_MAX</code>? |
| 483 | |
| 484 | It might not be sensible to have a size 0, so in this case it is |
| 485 | not a borderline case, but rather a faulty input (see @ref |
| 486 | FaultyInputTests). |
| 487 | |
| 488 | |
| 489 | @subsection FaultyInputTests Faulty input |
| 490 | |
| 491 | Does your bitmap handle 0 bits size? Well, it might not be designed |
| 492 | for it, but is should <em>not</em> crash the application, but |
| 493 | rather produce an error. This is called defensive programming. |
| 494 | |
| 495 | Unfortunately, adding checks for values that should just not be |
| 496 | entered at all is not always practical: the checks cost cycles and |
| 497 | might cost more than it's worth. For example, some functions are |
| 498 | designed so that you may not give it a null pointer. In those |
| 499 | cases it's not sensible to pass it <code>NULL</code> just to see it |
| 500 | crash. |
| 501 | |
| 502 | Since every experienced programmer add an <code>assert()</code> to |
| 503 | ensure that you get a proper failure for the debug builds when a |
| 504 | null pointer passed (you add asserts too, right?), you will in this |
| 505 | case instead have a controlled (early) crash in the debug build. |
| 506 | |
| 507 | |
| 508 | @subsection ErrorHandlingTests Error handling |
| 509 | |
| 510 | This is testing that the errors your component is designed to give |
| 511 | actually are produced. For example, testing that trying to open a |
| 512 | non-existing file produces a sensible error code. |
| 513 | |
| 514 | |
| 515 | @subsection BadEnvironmentTests Environment |
| 516 | |
| 517 | Sometimes, modules has to behave well even when the environment |
| 518 | fails to work correctly. Typical examples are when the computer is |
| 519 | out of dynamic memory or when the disk is full. You can emulate |
| 520 | this by replacing, e.g., <code>malloc()</code> with your own |
| 521 | version that will work for a while, but then fail. Some things are |
| 522 | worth to keep in mind here: |
| 523 | |
| 524 | - Make sure to make the function fail deterministically, so that |
| 525 | you really can repeat the test. |
| 526 | |
| 527 | - Make sure that it doesn't just fail immediately. The unit might |
| 528 | have checks for the first case, but might actually fail some time |
| 529 | in the near future. |
| 530 | |
| 531 | |
| 532 | @section UnitTest How to structure a unit test |
| 533 | |
| 534 | In this section we will give some advice on how to structure the |
| 535 | unit tests to make the development run smoothly. The basic |
| 536 | structure of a test is: |
| 537 | |
| 538 | - Plan |
| 539 | - Test |
| 540 | - Report |
| 541 | |
| 542 | |
| 543 | @subsection TestPlanning Plan the test |
| 544 | |
| 545 | Planning the test means telling how many tests there are. In the |
| 546 | event that one of the tests causes a crash, it is then possible to |
| 547 | see that there are fewer tests than expected, and print a proper |
| 548 | error message. |
| 549 | |
| 550 | To plan a test, use the @c plan() function in the following manner: |
| 551 | |
| 552 | @code |
| 553 | int main(int argc, char *argv[]) |
| 554 | { |
| 555 | plan(5); |
| 556 | . |
| 557 | . |
| 558 | . |
| 559 | } |
| 560 | @endcode |
| 561 | |
| 562 | If you don't call the @c plan() function, the number of tests |
| 563 | executed will be printed at the end. This is intended to be used |
| 564 | while developing the unit and you are constantly adding tests. It |
| 565 | is not indented to be used after the unit has been released. |
| 566 | |
| 567 | |
| 568 | @subsection TestRunning Execute the test |
| 569 | |
| 570 | To report the status of a test, the @c ok() function is used in the |
| 571 | following manner: |
| 572 | |
| 573 | @code |
| 574 | int main(int argc, char *argv[]) |
| 575 | { |
| 576 | plan(5); |
| 577 | ok(ducks == paddling_ducks, |
| 578 | "%d ducks did not paddle", ducks - paddling_ducks); |
| 579 | . |
| 580 | . |
| 581 | . |
| 582 | } |
| 583 | @endcode |
| 584 | |
| 585 | This will print a test result line on the standard output in TAP |
| 586 | format, which allows TAP handling frameworks (like Test::Harness) |
| 587 | to parse the status of the test. |
| 588 | |
| 589 | @subsection TestReport Report the result of the test |
| 590 | |
| 591 | At the end, a complete test report should be written, with some |
| 592 | statistics. If the test returns EXIT_SUCCESS, all tests were |
| 593 | successfull, otherwise at least one test failed. |
| 594 | |
| 595 | To get a TAP complient output and exit status, report the exit |
| 596 | status in the following manner: |
| 597 | |
| 598 | @code |
| 599 | int main(int argc, char *argv[]) |
| 600 | { |
| 601 | plan(5); |
| 602 | ok(ducks == paddling_ducks, |
| 603 | "%d ducks did not paddle", ducks - paddling_ducks); |
| 604 | . |
| 605 | . |
| 606 | . |
| 607 | return exit_status(); |
| 608 | } |
| 609 | @endcode |
| 610 | |
| 611 | @section DontDoThis Ways to not do unit testing |
| 612 | |
| 613 | In this section, we'll go through some quite common ways to write |
| 614 | tests that are <em>not</em> a good idea. |
| 615 | |
| 616 | @subsection BreadthFirstTests Doing breadth-first testing |
| 617 | |
| 618 | If you're writing a library with several functions, don't test all |
| 619 | functions using size 1, then all functions using size 2, etc. If a |
| 620 | test for size 42 fails, you have no easy way of tracking down why |
| 621 | it failed. |
| 622 | |
| 623 | It is better to concentrate on getting one function to work at a |
| 624 | time, which means that you test each function for all sizes that |
| 625 | you think is reasonable. Then you continue with the next function, |
| 626 | doing the same. This is usually also the way that a library is |
| 627 | developed (one function at a time) so stick to testing that is |
| 628 | appropriate for now the unit is developed. |
| 629 | |
| 630 | @subsection JustToBeSafeTest Writing unnecessarily large tests |
| 631 | |
| 632 | Don't write tests that use parameters in the range 1-1024 unless |
| 633 | you have a very good reason to belive that the component will |
| 634 | succeed for 562 but fail for 564 (the numbers picked are just |
| 635 | examples). |
| 636 | |
| 637 | It is very common to write extensive tests "just to be safe." |
| 638 | Having a test suite with a lot of values might give you a warm |
| 639 | fuzzy feeling, but it doesn't really help you find the bugs. Good |
| 640 | tests fail; seriously, if you write a test that you expect to |
| 641 | succeed, you don't need to write it. If you think that it |
| 642 | <em>might</em> fail, <em>then</em> you should write it. |
| 643 | |
| 644 | Don't take this as an excuse to avoid writing any tests at all |
| 645 | "since I make no mistakes" (when it comes to this, there are two |
| 646 | kinds of people: those who admit they make mistakes, and those who |
| 647 | don't); rather, this means that there is no reason to test that |
| 648 | using a buffer with size 100 works when you have a test for buffer |
| 649 | size 96. |
| 650 | |
| 651 | The drawback is that the test suite takes longer to run, for little |
| 652 | or no benefit. It is acceptable to do a exhaustive test if it |
| 653 | doesn't take too long to run and it is quite common to do an |
| 654 | exhaustive test of a function for a small set of values. |
| 655 | Use your judgment to decide what is excessive: your milage may |
| 656 | vary. |
| 657 | */ |
| 658 | |
| 659 | /** |
| 660 | @example simple.t.c |
| 661 | |
| 662 | This is an simple example of how to write a test using the |
| 663 | library. The output of this program is: |
| 664 | |
| 665 | @code |
| 666 | 1..1 |
| 667 | # Testing basic functions |
| 668 | ok 1 - Testing gcs() |
| 669 | @endcode |
| 670 | |
| 671 | The basic structure is: plan the number of test points using the |
| 672 | plan() function, perform the test and write out the result of each |
| 673 | test point using the ok() function, print out a diagnostics message |
| 674 | using diag(), and report the result of the test by calling the |
| 675 | exit_status() function. Observe that this test does excessive |
| 676 | testing (see @ref JustToBeSafeTest), but the test point doesn't |
| 677 | take very long time. |
| 678 | */ |
| 679 | |
| 680 | /** |
| 681 | @example todo.t.c |
| 682 | |
| 683 | This example demonstrates how to use the <code>todo_start()</code> |
| 684 | and <code>todo_end()</code> function to mark a sequence of tests to |
| 685 | be done. Observe that the tests are assumed to fail: if any test |
| 686 | succeeds, it is considered a "bonus". |
| 687 | */ |
| 688 | |
| 689 | /** |
| 690 | @example skip.t.c |
| 691 | |
| 692 | This is an example of how the <code>SKIP_BLOCK_IF</code> can be |
| 693 | used to skip a predetermined number of tests. Observe that the |
| 694 | macro actually skips the following statement, but it's not sensible |
| 695 | to use anything than a block. |
| 696 | */ |
| 697 | |
| 698 | /** |
| 699 | @example skip_all.t.c |
| 700 | |
| 701 | Sometimes, you skip an entire test because it's testing a feature |
| 702 | that doesn't exist on the system that you're testing. To skip an |
| 703 | entire test, use the <code>skip_all()</code> function according to |
| 704 | this example. |
| 705 | */ |
| 706 | |