WO2007081979A2 - Method of and system for dynamic automated test case generation and execution - Google Patents

Method of and system for dynamic automated test case generation and execution Download PDF

Info

Publication number
WO2007081979A2
WO2007081979A2 PCT/US2007/000585 US2007000585W WO2007081979A2 WO 2007081979 A2 WO2007081979 A2 WO 2007081979A2 US 2007000585 W US2007000585 W US 2007000585W WO 2007081979 A2 WO2007081979 A2 WO 2007081979A2
Authority
WO
WIPO (PCT)
Prior art keywords
test
given
computer
implemented method
interchangeable
Prior art date
Application number
PCT/US2007/000585
Other languages
French (fr)
Other versions
WO2007081979A3 (en
Inventor
Jesse A. Noller
Robert S. Mason, Jr.
Original Assignee
Archivas, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Archivas, Inc. filed Critical Archivas, Inc.
Publication of WO2007081979A2 publication Critical patent/WO2007081979A2/en
Publication of WO2007081979A3 publication Critical patent/WO2007081979A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/36Preventing errors by testing or debugging software
    • G06F11/3668Software testing
    • G06F11/3672Test management
    • G06F11/3688Test management for test execution, e.g. scheduling of test suites
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/36Preventing errors by testing or debugging software
    • G06F11/3668Software testing
    • G06F11/3672Test management
    • G06F11/3684Test management for test design, e.g. generating new test cases

Definitions

  • the present invention relates generally to quality assurance testing of computer systems and software.
  • test case is typically defined as a single step, and its expected result, along with one or more other pieces of information such as a case identifier, test author, or the like.
  • a test case may be a series of steps with one expected result or outcome.
  • the one or more steps associated with a given test case often are stored in a test plan document, a spreadsheet, a database, or other common repository (which may also store past test results and similar information).
  • a set or collection of test cases is sometimes referred to a test suite.
  • test cases typically are derived from a well-defined series of simple discrete, atomic base actions (e.g., writing a set of files is a first action, reading the set of files is a second action, and so on), hi this manner, and given a library of base actions, it is possible to design and implement a set of well-defined test cases.
  • These test cases would then be useful to perform a given degree of test coverage for the system feature areas, but such testing is necessarily circumscribed by the nature and scope of the base library and the limited manner in which the actions may be associated with one another.
  • test case generation approaches can be useful, they have not adequately addressed these problems.
  • An object of the present invention is to provide an automated system that intelligently and randomly generates test cases for use in hardware or software quality assurance testing, wherein a given test case comprises a sequence (or "chain") of discrete, atomic steps (or "building blocks”).
  • a test case is a chain of building blocks, preferably with each building block linked to a next building block. Accordingly, a given building block therefore is sometimes referred to as a "link.”
  • each building block is executable in a single-threaded or multi-threaded execution environment.
  • one or more building blocks are executed concurrently in different threads, and individual building blocks may themselves be single or multithreaded.
  • test case is considered successful (i.e., a pass) if each building block in the chain executes successfully; if any building block fails, the test case, in its entirety, is considered a failure, but not necessarily discarded.
  • a particular test case is a sequence of potentially infinite size; thus, a given sequence has a variable number of building blocks.
  • a given link may also pass information to another link, and one link may, in turn, reference another during the course of execution.
  • the system randomly generates test cases (chains) based on a series of underlying building blocks (or other pre-existing test cases) that are then linked (chained) together into one or more applicable test cases, and then executes those tests. More generally, the system takes a set of test actions (or even test cases) and links them together in a relevant and useful manner to create a much larger library of test cases or "chains.”
  • the chains comprise a large number of random sequence tests that facilitate "chaos-like" or exploratory testing of the overall system under test.
  • the chains themselves potentially can then be chained to one another to create a new tier of tests.
  • each chain is a test case that is self-documenting, discrete and re-creatable as needed.
  • the system adapts and dynamically generates new test cases as underlying data changes (e.g., a building block is added, deleted, modified) or as new test cases themselves are generated.
  • the system is tunable to generate test sequences that have a given (e.g., higher) likelihood of finding bugs or generating errors from which the testing entity can then assess the system operation.
  • a tunable risk factor or level is associated with a given link in a chain.
  • a given link may have one of several values such as basic, intermediate or advanced.
  • a basic link has a first value
  • an intermediate link has a second value
  • an advanced link has a third value.
  • the number of levels and their associated values are tunable.
  • the system can generate a risk assessment (e.g., expressed as a percentage value or other metric) for the chain as a whole, or for some given portion of the chain.
  • a risk assessment e.g., expressed as a percentage value or other metric
  • An aggregate or composite risk assessment for the chain (or chain portion) preferably is then determined by counting the number of occurrences of each link and its associated value.
  • the testing entity increases the likelihood of finding or isolating bugs or errors in the system under test.
  • the testing entity can identify such sequences by configuring the system* e.g., "generate chains with an aggregate percentage greater than x%" or the like. Where multi-threaded execution is used, the system may tune the number of threads as a function of the designated risk assessment for the chain or chain portion.
  • an initial risk assessment is recalculated based on a number of threads and the likelihood of two actions from a same "pool" occurring at the same point in time instead, for example, linearly in the sequence.
  • each executor thread within the system maintains a journal of what action was executed at what time. Merging the journals, e.g., after each run, allows replayability of the chain in a multi-threaded execution environment, thereby recreating a previous run.
  • the system includes an ability to replay generated chains, thereby enabling given test sequences (no matter how complex) to be easily reproducible.
  • the ability to replay generated chains in facilitated by having each of a set of execution threads maintain discrete journals that can then be merged into an aggregate journal for a given test case.
  • Figure 1 is a simplified block diagram of how the test generation system of the present invention generates tests from a library of test actions
  • Figure 2 is a simplified representation of a redundant array of independent nodes each of which supports an archive cluster application for which the test cases are generated according to the present invention
  • Figure 3 is a high level representation of the various components of the archive cluster application executing on a given node in the system of Figure 2;
  • Figure 4 illustrates in more detail how test sequences are generated from a library of test actions according to the present invention.
  • Figure 5 is a block diagram of an automated test case generation system of the present invention.
  • the present invention is a test generation system 100 that is used to generate a set of tests 102 from a library of test actions 104.
  • the tests 102 are then executed (e.g., by some other system) against a system or product under test (SUT) for quality assurance (Q/A) purposes.
  • SUT system or product under test
  • Q/A quality assurance
  • a particular test 102 may be replayed at any time.
  • the particular system under test may be of any type, as the present invention is concerned primarily with the methodology for generating the tests themselves.
  • the invention may be used to generate test cases for any system under test (SUT) regardless of the machine, operating environment, application or networking environment.
  • the invention is used to generate the test cases for an "archive cluster application" that provides an archival storage management system.
  • the storage management system is defined by a set of nodes that potentially comprise different hardware and thus may be considered “heterogeneous.”
  • a node typically has access to one or more storage disks, which may be actual physical storage disks, or virtual storage disks, as in a storage area network (SAN).
  • the archive cluster application that is supported on each node may be the same or substantially the same.
  • enterprises can create permanent storage for many different types of fixed content information such as documents, e-mail, video, and the like, among others. As described in U.S. Publication No.
  • the archive cluster application is a distributed software application executed on each node that captures, preserves, manages, and retrieves digital assets.
  • a physical boundary of an individual archive is referred to as a cluster.
  • a cluster is not a single device, but rather a collection of devices. Devices may be homogeneous or heterogeneous.
  • a typical device is a computer or machine running an operating system such as Linux.
  • Clusters of Linux -based systems hosted on commodity hardware provide an archive that can be scaled from a few storage node servers to many nodes that store thousands of terabytes of data. Preferably, data is replicated across the cluster so that the archive is protected from device failure.
  • the illustrative cluster comprises nodes 202, a pair of network switches 204, power distribution units (PDUs) 206, and uninterruptible power supplies (UPSs) 208.
  • a node 202 typically comprises one or more commodity servers and contains a CPU (e.g., Intel x86, suitable random access memory (RAM), one or more hard drives (e.g., standard IDE/SAT A, SCSI, or the like), and network interface (NIC) cards.
  • the network switches 204 typically comprise an internal switch 205 that enables peer-to-peer communication between nodes, and an external switch 207 that allows extra-cluster access to each node.
  • a cluster is connectable to a network, such as the public Internet, an enterprise intranet, or other wide area or local area network. End users access the cluster using any conventional interface or access tool. Thus, for example, access to the cluster may be carried out over any IP-based protocol (HTTP, FTP, NFS, AFS, SMB, a Web service, or the like), via an API, or through any other known or later-developed access method, service, program or tool.
  • Client applications access the cluster through one or more types of external gateways such as standard UNIX file protocols, or HTTP APIs.
  • the archive preferably is exposed through a virtual file system that sits under any standard UNIX file protocol-oriented facility.
  • the archive cluster application runs on a redundant array of independent nodes (H-RAIN) that are networked together (e.g., via Ethernet) as a cluster.
  • H-RAIN redundant array of independent nodes
  • Each node runs an instance 300 of a distributed application comprised of several runtime components as illustrated in Figure 3.
  • the software components comprise a gateway protocol layer 302, an access layer 304, a file transaction and administration layer 306, and a core components layer 308.
  • the gateways provide native file services such as NFS 310 and SMB/CIFS 312, as well as a Web services API to build custom applications. HTTP support 314 is also provided.
  • the access layer 304 provides access to the archive.
  • a fixed content file system (FCFS) 316 emulates a native file system to provide full access to archive objects.
  • the access layer 304 preferably also includes a Web user interface (UI) 318 and an SNMP gateway 320.
  • the Web user interface 318 provides interactive access to an administration engine 322 in a file transaction and administration layer 306.
  • the SNMP gateway 320 provides storage management applications access to the administration engine 322, enabling them to securely monitor and control cluster activity.
  • the administration engine monitors cluster activity, including system and policy events.
  • the file transaction and administration layer 306 also includes a request manager process 324.
  • the request manager 324 manages requests from the external world
  • the components may include a metadata manager 328, and one or more instances of a storage manager 330.
  • the storage manager 330 on a given node is responsible for managing the physical storage devices.
  • the request manager 324 is responsible for executing the set of operations needed to perform archive actions by interacting with other components within the system.
  • a metadata manager 328 preferably is installed on each node.
  • the metadata managers in a cluster act as a distributed database, managing archive objects.
  • the storage manager 330 provides a file system layer available to all other components in the distributed application. Preferably, it stores the data objects in a node's local file system.
  • the storage manager 330 also provides system information, integrity checks on the data, and the ability to traverse local directly structures.
  • the cluster manages internal and external communication through a communications middleware layer 332 and a DNS manager 334.
  • the application instance executes on a base operating system 336.
  • a cluster's primary responsibility is to store an unlimited number of files on disk reliably.
  • the present invention is a method of and system for automated test case generation of a system under test (SUT) such as the archive cluster described above.
  • a test case generation system is used to self-generate test cases from a series of basic actions.
  • the actions are the basic building blocks of the test cases, and each building block or link is preferably a discrete, atomic operation.
  • building blocks may be as simple as "write N files into the archive,” “drop a given node of the archive,” “execute a given policy against the archive,” “write 10,000 files with 100 threads via multiple input gateways,” “disable a random storage point within the system,” or the like.
  • the above are merely illustrative.
  • test cases 400a-n for use in quality assurance testing, wherein a given test case 400b comprises a sequence (or "chain") of discrete, atomic steps (or “building blocks”) 402a-n.
  • a particular test case i.e., a given sequence
  • the system takes a set of test actions (or even test cases) and links them together in a relevant and useful manner to create a much larger library of test cases or "chains.”
  • the chains comprise a large number of intelligently random sequenced test building blocks that facilitate random (“chaos-like") or exploratory testing of the overall system under test 404.
  • the test case Upon execution in/against the system under test, the test case is considered successful (i.e., a pass) if each building block (or link) in the chain executes successfully; if any building block (or link) fails, the test case, in its entirety, is considered a failure.
  • the system adapts and dynamically generates new test cases as underlying data changes (e.g., a building block is added, deleted, modified) or new test cases themselves are generated. Also, the system is tunable to generate test sequences that have a given (e.g., higher) likelihood of finding bugs or generating errors from which the user can then assess the system operation.
  • Figure 5 illustrates a representative test system 500 of the present invention.
  • the system comprises a database 502 of potential test actions or links, which actions (as described below) are combinable into one or more action modules 504.
  • Action modules 504 may, in turn, be classified by the nature of the resulting tests, such as "basic" modules 501, intermediate modules 503, and advanced modules 505. Of course, the number of such classifications is configurable.
  • the modules 501, 503 and 505 maybe stored in or as a directory.
  • a test action (a "building block” or "link”) is some discrete function, method or class that, typically, cannot be broken down into a smaller component or action. A given test action thus may be considered immutable within the context of the particular system under test.
  • test action usually will depend on the nature and type of system under test.
  • a particular test action might be: “disable target machine,” “write 10,000 1KB files,” disable NIC device,” “read 100 files,” “drop a node,” “drop a node's backend Ethernet card,” “run a garbage collection,” “drop a node, bring the node back up, check node's availability,” and so on.
  • these are merely representative.
  • the test system 500 also comprises a set of functions that, in a representative embodiment, are implemented as a set of software classes written in an object-oriented compiled or interpreted programming language, such as Python.
  • the individual classes may be implemented in other languages, such as Java, and particular classes are illustrated and designated for convenience only. Particular classes shown in Figure 5 may be combined in any convenient manner.
  • a particular test action or link is implemented within an action module 504 (preferably implemented in Python), and, as indicated above, a given module may contain any number (e.g., ones, tens, hundreds, etc.) of discrete actions.
  • the test system preferably is executed on at least one machine having commodity hardware and operating system software (e.g., Linux).
  • the system also includes one or more input and output devices, as well as conventional applications and interfaces.
  • a user inputs options 507 to the system using a configuration file and/or a command line interface.
  • These input options are read by a configuration loader 506 or a command parser 508, depending on the nature of the input, and passed to a logging subsystem class 510 and a state verification class 512.
  • these options are a set of one or more tunable features, such as a number of execution threads, links, links in the chain, a number of chains, and the like.
  • the options 507 may also include application-specific configuration options depending on the type of system under test.
  • these options may include, for example, a cluster identifier, a domain name of the cluster, the number of nodes in the cluster, a number of tolerable points of failure (TPOF) for the cluster, a size of the files to be written, and the like.
  • a sequence generator class 514 receives the input (the set of one or more tunable options) from the state verification class 512 to obtain a list of valid test actions (i.e. the list of Python modules) from the directory.
  • a black box import class 516 with or without support from a module/directory class 517, may be used optionally to import a given library of actions.
  • the sequence generator class 514 parses the list and generates one or more link sequences 518, according to the input constraints. Thus, given a library of actions A, B and C, the sequence generator class might generate sequences: ABC, BCA, CAB, AABC, ABCB, and so on. Generalizing, the sequence generator receives the actions and mixes them to randomly generate permutations, each of which is a sequence or "chain.” The sequence generator class 514 passes each generated sequence to a risk assessor class 520. The sequences may be passed one at a time or in batch. According to a feature of the invention, preferably a tunable risk factor or level is associated with a given link in a generated chain.
  • a given link may have one of several values: basic, intermediate or advanced. Taking the example set forth above, a given link (such as link A) may be considered a basic test. Another link (such as link B) may be considered an intermediate test. Yet another link (such as link C) may be considered an advanced test.
  • a basic link is seeded with a first (e.g., 10%) value
  • an intermediate link is seeded with a second (e.g., 50%) value
  • an advanced link is seeded with a third (e.g., 90%) value.
  • the system can generate a risk assessment (e.g., expressed as a percentage value or metric) for the generated chain as a whole, or for some given portion of the chain.
  • An aggregate or composite risk assessment for the chain (or chain portion) preferably is determined (based on a 0-100% scale) by counting the number of occurrences of each link and its associated value.
  • a chain such as ABCBBACCCAB would have a composite risk factor of 98%, which is calculated as two (2) occurrences of A (at 5% each) + four (4) occurrences of B (at 50% each) + four (4) occurrences of C (at 90% each).
  • the risk assessor class 524 performs this risk assessment function.
  • the conflict hasher class checks the chain database to determine whether the sequence has been previously generated.
  • a chain is registered within the database, and each chain has an associated unique identifier (e.g., a serial number) that is preferably derived from the links within the chain.
  • Any convenient hashing algorithm may be used to generate the unique identifier. For example, an MD4, MD5 or SHA-I hash of the chain itself or the in-memory chain object may be used as the identifier.
  • the conflict hasher class 522 registers the chain with the database, using the unique identifier, and then stores the new sequence in the database.
  • the database may have fields to disable or otherwise modify chains. For example, it may be desirable to include override risk/relevance criteria that would control the system, e.g., to run a given chain in a lower priority, to never run a given chain, or the like.
  • sequence slicer class 524 breaks the sequence up into one or more chunks or portions 526 depending on a number of worker threads desired.
  • the slicer class 524 is not required if a single-threaded execution environment is used, as each link in the sequence (theoretically) will then need to be performed individually and in sequence.
  • a multi-threaded approach is preferred, however, in which case the sequence slicer class 524 breaks the sequences and generates associated work queues 530a-n for a set of worker threads. Once the queues are generated, a sequence dispatcher class 532 is called.
  • the dispatcher class 532 instantiates a set of worker threads 534a-n, which threads then execute the test sequence.
  • the worker threads each maintain a journal 540 of what action was executed at what time (with a value instantiated at runtime to be "0"); upon merger of the thread journals, the system can replay (i.e. re-run) an exact chain in a substantially exact fashion at a subsequent time.
  • the dispatcher class 532 does not have to have any knowledge as to the given library of actions, or how a particular test sequence is generated. As noted above, if any link in the chain fails, preferably the system records the result of the test as a failure.
  • the system 500 may also include a set of helper classes including, without limitation, the logging subsystem class, the state verification class, and a result parser class 540. These classes provide facilities to the system for information gathering and reporting. Thus, for example, the logging class records the results of each test, the state verification class records the ongoing state or health of the overall system or product under test, and the result parser class determines whether a particular test is successful.
  • each test action is a self-contained, discrete Python class/function/method that performs a relatively simple action.
  • a "write action” library might comprise a set of basic actions, such as write_ ⁇ les_n ⁇ , write Jilesjittp and write_files_smb.
  • Each action accepts some standard criteria or information; when the action called, the information is passed in, and the action is performed.
  • An example is set forth below:
  • This function accepts the incoming dictionary, parses it using a library that it self- imports, and then performs the action.
  • the retval variable is a return message that indicates the"pass/fail" status of the function.
  • the second line of the function is a string that is called by a helper class that documents the results. This facilitates a self- documenting system.
  • write libraries should have a default.
  • write libraries should have a default. o
  • the designated options cause the system to generate and run ten (10) chains ten (10) links deeps, with a maximum file range of 1024 bytes, with an initial count of 100,000.
  • this is merely a representative input data set. It is desirable to execute a test sequence from start to finish whenever possible. To this end, and before running a given test, the state verification class may be executed to determine whether the cluster is accessible, whether all nodes in the cluster are still available, whether given paths within the archive are reachable, whether there are unusual log entries, or the like. Although this may not prevent further actions from taking place, it ensures that the state of the SUT is known for debugging/investigation purposes.
  • the test generation system may include other helper classes.
  • a ratings class may be used to assign a rating to a given chain after that chain is generated and run. The chain ratings may then be used to order test cases, with test sequences that stress the system the most having a highest rating, whereas those that create less stress having lower ratings.
  • the present invention provides numerous advantages over the prior art.
  • the system exploits an automation paradigm by which a library of test actions is processed selectively to create larger test cases or chains.
  • the generated chains may, in turn, be recombined into larger, more complex chains themselves.
  • These chains make up for the deficiencies in the prior art (where there are a given number of static test cases) by automatically generating what are, in effect, "random" test sequences.
  • the inventive methodology can generate over 3,000,000 combinations of those ten actions, creating a much more robust, useful and extensive test suite.
  • the number and type of test sequences are constrained only by a set of user-tunable or system-tunable options.
  • An example of a system-tunable constraint in the context of the archive is a policy that might associate a given "weight” with a given link to facilitate creation of "weighted” random sequences, which allows for tunable amounts of actions constrained by risk assessment or "level” (e.g., "basic,” “intermediate” or “advanced”) to appear within a chain.
  • the user- or system-tunable constraints may also designate a maximum number of links in a chain, a maximum number of times a given link may occur in a chain, a maximum number of times that a given link may appear next to itself in a given chain, or the like.
  • each link is a standalone or discrete that may pass, or fail. It the link fails, the chain dies.
  • each link has a conduit through which given information may be passed through to a next or other link in the chain.
  • This conduit for example, is an in- memory data structure, a cached-on-disk data structure, a text file, a socket communication link, or the like.
  • the conduit is an in-memory data structure passed from the executor to the individual link, and this data structure comprises information about the SUT such as (in the archive cluster application) current file counts, current file locations, number of nodes, and the like.
  • the conduit is implemented on disk cache as a plaintext or Python-pickled data structure.
  • the information in the data structure is modifiable and/or expandable by each link, and preferably all links must either accept the in-memory data structure or assume the on-disk location, as the case must be.
  • the conduit is maintained by the executor, on the chain's behalf.
  • a given link in a chain may be either a self-contained entity (e.g., a set of self- contained libraries), or an entity that has associated dependencies. If a particular link has some dependency, preferably that dependency is specified by some criteria, which criteria can then be applied to the link using appropriate system-level or local processing or other resources. If a dependency is needed but is not met, " the link is "skipped" or its results ignored. In a preferred embodiment, however, any link dependencies are recognized and verified prior to inclusion of the link in a candidate chain. In this embodiment, a given link has associated therewith given criteria that identify required dependencies. When the system is initialized (e.g., at library import time), the system calls a given dependency verification function within each library to confirm that all identified dependencies are met. In such case, the verification of dependencies must occur before the given link can be used by the sequence generator.
  • a given dependency verification function within each library to confirm that all identified dependencies are met. In such case, the verification of dependencies must occur before the given link can be
  • the test generation system of the present invention is also advantageous in that it allows for dynamic test sequence generation as a function of new test actions or new test cases themselves. Indeed, a flexible quality assurance program must take into account regular additions, deletions or modifications of the building blocks and test cases.
  • the inventive system is fluid in the sense that it can at any time generate chains from new actions and tests as needed, easily deprecating irrelevant chains in the process. By registering each chain as it is created, the user or system need not cycle through pre- generated chains before performing a given test suite.
  • the present invention is a test case generation methodology by which a set of discrete test actions form a set of interchangeable building blocks that are permutated (intelligently random) to generate a set of large-scale tests against which a system under test is then exercised in a re-creatable manner.
  • Each test action is an atomic, preferably immutable step in the overall chain/test case generated.
  • test generation system may also create the building blocks themselves in an automated manner.
  • Such a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including an optical disk, a CD-ROM, and a magnetic-optical disk, a read-only memory (ROM), a random access memory (RAM), a magnetic or optical card, or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus.
  • a computer readable storage medium such as, but is not limited to, any type of disk including an optical disk, a CD-ROM, and a magnetic-optical disk, a read-only memory (ROM), a random access memory (RAM), a magnetic or optical card, or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus.
  • a given implementation of the present invention is software written in a given programming language (e.g., Python) that runs on a standard Intel hardware platform running an operating system such as Linux.

Abstract

An automated system that randomly and dynamically generates test cases for use in hardware or software quality assurance testing, wherein a given test case comprises a sequence of actions stored as number of building blocks. The system takes a set of test actions and links them together in a relevant and useful manner to create a much larger library of test cases comprising a number of random sequence tests that facilitate exploratory testing of the overall system under test. Upon execution in the system under test, the test case is considered successful if each building block in the chain executes successfully; if any building block fails, the test case, in its entirety, is considered a failure.

Description

METHOD OFAND SYSTEMFORDYNAMICAUTOMATED TEST CASE GENERATION ANDEXECUTION
This application is based on and claims priority to Serial No. 11/329,631, filed January 11, 2006.
BACKGROUND OF THE INVENTION Technical Field
The present invention relates generally to quality assurance testing of computer systems and software.
Description of the Related Art
There are many well-known approaches to testing software during the development of a complex system. During the feature/test development phase, for example, a quality assurance engineer may devise a series of static test cases against which a system under development may be exercised. A test case is typically defined as a single step, and its expected result, along with one or more other pieces of information such as a case identifier, test author, or the like. Occasionally, a test case may be a series of steps with one expected result or outcome. The one or more steps associated with a given test case often are stored in a test plan document, a spreadsheet, a database, or other common repository (which may also store past test results and similar information). A set or collection of test cases is sometimes referred to a test suite.
The process of generating test cases is usually a manual one, although automated techniques have been described, e.g., U.S. Patent No. 5,542,043. Regardless of the type of system involved, the test cases typically are derived from a well-defined series of simple discrete, atomic base actions (e.g., writing a set of files is a first action, reading the set of files is a second action, and so on), hi this manner, and given a library of base actions, it is possible to design and implement a set of well-defined test cases. These test cases would then be useful to perform a given degree of test coverage for the system feature areas, but such testing is necessarily circumscribed by the nature and scope of the base library and the limited manner in which the actions may be associated with one another.
There are inherent issues with the prior art with respect to the veracity of a given test cases (e.g., the subjective value of the test case vis-a-vis the features actually tested), as well as the number of test cases that can be implemented efficiently. The latter problem typically is caused by time and labor restraints, especially where manual testing is used. While automated test case generation approaches (such as described in the above-identified patent) can be useful, they have not adequately addressed these problems.
It would be highly desirable to provide an improved technique for automating the generation and managed execution of a much broader range of test cases than has been possible using the prior art. The present invention addresses this problem.
BRIEF SUMMARY OF THE INVENTION
An object of the present invention is to provide an automated system that intelligently and randomly generates test cases for use in hardware or software quality assurance testing, wherein a given test case comprises a sequence (or "chain") of discrete, atomic steps (or "building blocks"). A test case is a chain of building blocks, preferably with each building block linked to a next building block. Accordingly, a given building block therefore is sometimes referred to as a "link." Preferably, each building block is executable in a single-threaded or multi-threaded execution environment. Thus, in a multi-threaded environment, one or more building blocks are executed concurrently in different threads, and individual building blocks may themselves be single or multithreaded. The test case is considered successful (i.e., a pass) if each building block in the chain executes successfully; if any building block fails, the test case, in its entirety, is considered a failure, but not necessarily discarded. A particular test case is a sequence of potentially infinite size; thus, a given sequence has a variable number of building blocks. A given link may also pass information to another link, and one link may, in turn, reference another during the course of execution.
In one embodiment, the system randomly generates test cases (chains) based on a series of underlying building blocks (or other pre-existing test cases) that are then linked (chained) together into one or more applicable test cases, and then executes those tests. More generally, the system takes a set of test actions (or even test cases) and links them together in a relevant and useful manner to create a much larger library of test cases or "chains." The chains comprise a large number of random sequence tests that facilitate "chaos-like" or exploratory testing of the overall system under test. The chains themselves potentially can then be chained to one another to create a new tier of tests. Preferably, each chain is a test case that is self-documenting, discrete and re-creatable as needed. The system adapts and dynamically generates new test cases as underlying data changes (e.g., a building block is added, deleted, modified) or as new test cases themselves are generated.
According to another feature of the invention, the system is tunable to generate test sequences that have a given (e.g., higher) likelihood of finding bugs or generating errors from which the testing entity can then assess the system operation. To this end, preferably a tunable risk factor or level is associated with a given link in a chain. In an illustrative embodiment, a given link may have one of several values such as basic, intermediate or advanced. A basic link has a first value, an intermediate link has a second value, and an advanced link has a third value. Preferably, the number of levels and their associated values are tunable. Using these individual link risk factors, the system can generate a risk assessment (e.g., expressed as a percentage value or other metric) for the chain as a whole, or for some given portion of the chain. An aggregate or composite risk assessment for the chain (or chain portion) preferably is then determined by counting the number of occurrences of each link and its associated value. By generating high (or higher) risk sequences, the testing entity increases the likelihood of finding or isolating bugs or errors in the system under test. The testing entity can identify such sequences by configuring the system* e.g., "generate chains with an aggregate percentage greater than x%" or the like. Where multi-threaded execution is used, the system may tune the number of threads as a function of the designated risk assessment for the chain or chain portion.
In the case of multi-threaded execution, preferably an initial risk assessment is recalculated based on a number of threads and the likelihood of two actions from a same "pool" occurring at the same point in time instead, for example, linearly in the sequence. In this embodiment, preferably each executor thread within the system maintains a journal of what action was executed at what time. Merging the journals, e.g., after each run, allows replayability of the chain in a multi-threaded execution environment, thereby recreating a previous run. Thus, according to another important advantage of the invention, the system includes an ability to replay generated chains, thereby enabling given test sequences (no matter how complex) to be easily reproducible. As noted above, in one embodiment, the ability to replay generated chains in facilitated by having each of a set of execution threads maintain discrete journals that can then be merged into an aggregate journal for a given test case.
The foregoing has outlined some of the more pertinent features of the invention. These features should be construed to be merely illustrative. Many other beneficial results can be attained by applying the disclosed invention in a different manner or by modifying the invention as will be described.
BRIEF DESCRIPTION OF THE DRAWINGS
For a more complete understanding of the present invention and the advantages thereof, reference is now made to the following descriptions taken in conjunction with the accompanying drawings, in which: Figure 1 is a simplified block diagram of how the test generation system of the present invention generates tests from a library of test actions;
Figure 2 is a simplified representation of a redundant array of independent nodes each of which supports an archive cluster application for which the test cases are generated according to the present invention; Figure 3 is a high level representation of the various components of the archive cluster application executing on a given node in the system of Figure 2;
Figure 4 illustrates in more detail how test sequences are generated from a library of test actions according to the present invention; and
Figure 5 is a block diagram of an automated test case generation system of the present invention.
DETAILED DESCRIPTION OF AN EMBODIMENT As shown in Figure 1, the present invention is a test generation system 100 that is used to generate a set of tests 102 from a library of test actions 104. The tests 102 are then executed (e.g., by some other system) against a system or product under test (SUT) for quality assurance (Q/A) purposes. A particular test 102 may be replayed at any time. The particular system under test may be of any type, as the present invention is concerned primarily with the methodology for generating the tests themselves. The invention may be used to generate test cases for any system under test (SUT) regardless of the machine, operating environment, application or networking environment.
In this representative embodiment, the invention is used to generate the test cases for an "archive cluster application" that provides an archival storage management system. By way of brief background, it is assumed that the storage management system is defined by a set of nodes that potentially comprise different hardware and thus may be considered "heterogeneous." A node typically has access to one or more storage disks, which may be actual physical storage disks, or virtual storage disks, as in a storage area network (SAN). The archive cluster application that is supported on each node may be the same or substantially the same. Using the system, enterprises can create permanent storage for many different types of fixed content information such as documents, e-mail, video, and the like, among others. As described in U.S. Publication No. 2005/0120025, the archive cluster application is a distributed software application executed on each node that captures, preserves, manages, and retrieves digital assets. As illustrated in Figure 2, a physical boundary of an individual archive is referred to as a cluster. Typically, a cluster is not a single device, but rather a collection of devices. Devices may be homogeneous or heterogeneous. A typical device is a computer or machine running an operating system such as Linux. Clusters of Linux -based systems hosted on commodity hardware provide an archive that can be scaled from a few storage node servers to many nodes that store thousands of terabytes of data. Preferably, data is replicated across the cluster so that the archive is protected from device failure. If a disk or node fails, the cluster automatically fails over to other nodes in the cluster that maintain replicas of the same data. The illustrative cluster comprises nodes 202, a pair of network switches 204, power distribution units (PDUs) 206, and uninterruptible power supplies (UPSs) 208. A node 202 typically comprises one or more commodity servers and contains a CPU (e.g., Intel x86, suitable random access memory (RAM), one or more hard drives (e.g., standard IDE/SAT A, SCSI, or the like), and network interface (NIC) cards. The network switches 204 typically comprise an internal switch 205 that enables peer-to-peer communication between nodes, and an external switch 207 that allows extra-cluster access to each node. PDUs 206 are used to power all nodes and switches, and the UPSs 208 are used that protect all nodes and switches. Typically, a cluster is connectable to a network, such as the public Internet, an enterprise intranet, or other wide area or local area network. End users access the cluster using any conventional interface or access tool. Thus, for example, access to the cluster may be carried out over any IP-based protocol (HTTP, FTP, NFS, AFS, SMB, a Web service, or the like), via an API, or through any other known or later-developed access method, service, program or tool. Client applications access the cluster through one or more types of external gateways such as standard UNIX file protocols, or HTTP APIs. The archive preferably is exposed through a virtual file system that sits under any standard UNIX file protocol-oriented facility.
In this example testing environment, the archive cluster application runs on a redundant array of independent nodes (H-RAIN) that are networked together (e.g., via Ethernet) as a cluster. Each node runs an instance 300 of a distributed application comprised of several runtime components as illustrated in Figure 3. As will be described, the present invention generates test cases for this system under test. The software components comprise a gateway protocol layer 302, an access layer 304, a file transaction and administration layer 306, and a core components layer 308. The gateways provide native file services such as NFS 310 and SMB/CIFS 312, as well as a Web services API to build custom applications. HTTP support 314 is also provided. The access layer 304 provides access to the archive. A fixed content file system (FCFS) 316 emulates a native file system to provide full access to archive objects. The access layer 304 preferably also includes a Web user interface (UI) 318 and an SNMP gateway 320. The Web user interface 318 provides interactive access to an administration engine 322 in a file transaction and administration layer 306. The SNMP gateway 320 provides storage management applications access to the administration engine 322, enabling them to securely monitor and control cluster activity. The administration engine monitors cluster activity, including system and policy events.
The file transaction and administration layer 306 also includes a request manager process 324. The request manager 324 manages requests from the external world
(through the access layer 304), as well as internal requests from a policy manager 326 in the core components layer 308. In addition to the policy manager 326, the components may include a metadata manager 328, and one or more instances of a storage manager 330. The storage manager 330 on a given node is responsible for managing the physical storage devices. The request manager 324 is responsible for executing the set of operations needed to perform archive actions by interacting with other components within the system. A metadata manager 328 preferably is installed on each node. Collectively, the metadata managers in a cluster act as a distributed database, managing archive objects. The storage manager 330 provides a file system layer available to all other components in the distributed application. Preferably, it stores the data objects in a node's local file system. The storage manager 330 also provides system information, integrity checks on the data, and the ability to traverse local directly structures. As also illustrated in Figure 3, the cluster manages internal and external communication through a communications middleware layer 332 and a DNS manager 334. The application instance executes on a base operating system 336. A cluster's primary responsibility is to store an unlimited number of files on disk reliably.
The present invention is a method of and system for automated test case generation of a system under test (SUT) such as the archive cluster described above. In a representative embodiment, a test case generation system is used to self-generate test cases from a series of basic actions. The actions are the basic building blocks of the test cases, and each building block or link is preferably a discrete, atomic operation. In the context of the archival system, such building blocks may be as simple as "write N files into the archive," "drop a given node of the archive," "execute a given policy against the archive," "write 10,000 files with 100 threads via multiple input gateways," "disable a random storage point within the system," or the like. Of course, the above are merely illustrative.
As illustrated in Figure 4, the automated system generates test cases 400a-n for use in quality assurance testing, wherein a given test case 400b comprises a sequence (or "chain") of discrete, atomic steps (or "building blocks") 402a-n. A particular test case (i.e., a given sequence) has a variable number of building blocks. The system takes a set of test actions (or even test cases) and links them together in a relevant and useful manner to create a much larger library of test cases or "chains." The chains comprise a large number of intelligently random sequenced test building blocks that facilitate random ("chaos-like") or exploratory testing of the overall system under test 404. Upon execution in/against the system under test, the test case is considered successful (i.e., a pass) if each building block (or link) in the chain executes successfully; if any building block (or link) fails, the test case, in its entirety, is considered a failure. Preferably, and as will be seen, the system adapts and dynamically generates new test cases as underlying data changes (e.g., a building block is added, deleted, modified) or new test cases themselves are generated. Also, the system is tunable to generate test sequences that have a given (e.g., higher) likelihood of finding bugs or generating errors from which the user can then assess the system operation.
Figure 5 illustrates a representative test system 500 of the present invention. The system comprises a database 502 of potential test actions or links, which actions (as described below) are combinable into one or more action modules 504. Action modules 504 may, in turn, be classified by the nature of the resulting tests, such as "basic" modules 501, intermediate modules 503, and advanced modules 505. Of course, the number of such classifications is configurable. The modules 501, 503 and 505 maybe stored in or as a directory. According to the inventive methodology, a test action (a "building block" or "link") is some discrete function, method or class that, typically, cannot be broken down into a smaller component or action. A given test action thus may be considered immutable within the context of the particular system under test. Of course, the definition of a particular test action usually will depend on the nature and type of system under test. In the context of the archive system illustrated in Figures 1-3, a particular test action might be: "disable target machine," "write 10,000 1KB files," disable NIC device," "read 100 files," "drop a node," "drop a node's backend Ethernet card," "run a garbage collection," "drop a node, bring the node back up, check node's availability," and so on. As noted above, these are merely representative.
The test system 500 also comprises a set of functions that, in a representative embodiment, are implemented as a set of software classes written in an object-oriented compiled or interpreted programming language, such as Python. The individual classes may be implemented in other languages, such as Java, and particular classes are illustrated and designated for convenience only. Particular classes shown in Figure 5 may be combined in any convenient manner. A particular test action or link is implemented within an action module 504 (preferably implemented in Python), and, as indicated above, a given module may contain any number (e.g., ones, tens, hundreds, etc.) of discrete actions. The test system preferably is executed on at least one machine having commodity hardware and operating system software (e.g., Linux). The system also includes one or more input and output devices, as well as conventional applications and interfaces.
In operation, a user inputs options 507 to the system using a configuration file and/or a command line interface. These input options are read by a configuration loader 506 or a command parser 508, depending on the nature of the input, and passed to a logging subsystem class 510 and a state verification class 512. Preferably, these options are a set of one or more tunable features, such as a number of execution threads, links, links in the chain, a number of chains, and the like. The options 507 may also include application-specific configuration options depending on the type of system under test. In the example archive system described above, these options may include, for example, a cluster identifier, a domain name of the cluster, the number of nodes in the cluster, a number of tolerable points of failure (TPOF) for the cluster, a size of the files to be written, and the like. As seen in Figure 5, a sequence generator class 514 receives the input (the set of one or more tunable options) from the state verification class 512 to obtain a list of valid test actions (i.e. the list of Python modules) from the directory. A black box import class 516, with or without support from a module/directory class 517, may be used optionally to import a given library of actions. The sequence generator class 514 parses the list and generates one or more link sequences 518, according to the input constraints. Thus, given a library of actions A, B and C, the sequence generator class might generate sequences: ABC, BCA, CAB, AABC, ABCB, and so on. Generalizing, the sequence generator receives the actions and mixes them to randomly generate permutations, each of which is a sequence or "chain." The sequence generator class 514 passes each generated sequence to a risk assessor class 520. The sequences may be passed one at a time or in batch. According to a feature of the invention, preferably a tunable risk factor or level is associated with a given link in a generated chain. In an illustrative embodiment, a given link may have one of several values: basic, intermediate or advanced. Taking the example set forth above, a given link (such as link A) may be considered a basic test. Another link (such as link B) may be considered an intermediate test. Yet another link (such as link C) may be considered an advanced test. In the illustrative embodiment, a basic link is seeded with a first (e.g., 10%) value, an intermediate link is seeded with a second (e.g., 50%) value, and an advanced link is seeded with a third (e.g., 90%) value. These values are merely representative, as the number of levels as well as their associated values preferably are tunable. Using these individual link risk factors, the system can generate a risk assessment (e.g., expressed as a percentage value or metric) for the generated chain as a whole, or for some given portion of the chain. An aggregate or composite risk assessment for the chain (or chain portion) preferably is determined (based on a 0-100% scale) by counting the number of occurrences of each link and its associated value. Thus, to give a concrete example (using the representative percentages described above), a chain such as ABCBBACCCAB would have a composite risk factor of 98%, which is calculated as two (2) occurrences of A (at 5% each) + four (4) occurrences of B (at 50% each) + four (4) occurrences of C (at 90% each). The risk assessor class 524 performs this risk assessment function.
Once the risk assessor class 520 verifies the risk level associated with a given sequence, the sequence is passed to a conflict hasher class 522. The conflict hasher class checks the chain database to determine whether the sequence has been previously generated. Preferably, a chain is registered within the database, and each chain has an associated unique identifier (e.g., a serial number) that is preferably derived from the links within the chain. Any convenient hashing algorithm may be used to generate the unique identifier. For example, an MD4, MD5 or SHA-I hash of the chain itself or the in-memory chain object may be used as the identifier. If the chain has not been seen before, the conflict hasher class 522 registers the chain with the database, using the unique identifier, and then stores the new sequence in the database. The database may have fields to disable or otherwise modify chains. For example, it may be desirable to include override risk/relevance criteria that would control the system, e.g., to run a given chain in a lower priority, to never run a given chain, or the like.
Once the generation and verification of the sequence is complete, optionally the sequence is passed to a sequence slicer class 524, which breaks the sequence up into one or more chunks or portions 526 depending on a number of worker threads desired. The slicer class 524 is not required if a single-threaded execution environment is used, as each link in the sequence (theoretically) will then need to be performed individually and in sequence. A multi-threaded approach is preferred, however, in which case the sequence slicer class 524 breaks the sequences and generates associated work queues 530a-n for a set of worker threads. Once the queues are generated, a sequence dispatcher class 532 is called. The dispatcher class 532 instantiates a set of worker threads 534a-n, which threads then execute the test sequence. Preferably, the worker threads each maintain a journal 540 of what action was executed at what time (with a value instantiated at runtime to be "0"); upon merger of the thread journals, the system can replay (i.e. re-run) an exact chain in a substantially exact fashion at a subsequent time. The dispatcher class 532 does not have to have any knowledge as to the given library of actions, or how a particular test sequence is generated. As noted above, if any link in the chain fails, preferably the system records the result of the test as a failure. One or more links may need to be executed before other links in the chain, and a particular link may pass given output information to another (e.g., a next) link in the chain. Or, an individual link of the chain may have no functional relationship with any other link. The system 500 may also include a set of helper classes including, without limitation, the logging subsystem class, the state verification class, and a result parser class 540. These classes provide facilities to the system for information gathering and reporting. Thus, for example, the logging class records the results of each test, the state verification class records the ongoing state or health of the overall system or product under test, and the result parser class determines whether a particular test is successful.
The following provides additional implementation-specific details of a representative embodiment.
As noted above, preferably each test action is a self-contained, discrete Python class/function/method that performs a relatively simple action. Given the archival system example, a "write action" library might comprise a set of basic actions, such as write_βles_nβ, write Jilesjittp and write_files_smb. Each action accepts some standard criteria or information; when the action called, the information is passed in, and the action is performed. An example is set forth below:
W/usr/bin/python defwrite_http(infodict):
"""write N files to archive"""
import comtnonlib cname.finin.fmax, /count = commonlib.write_parse(infodict) status = ...write... retval = commonlib.checkstatus(status) return retval
This function accepts the incoming dictionary, parses it using a library that it self- imports, and then performs the action. The retval variable is a return message that indicates the"pass/fail" status of the function. The second line of the function is a string that is called by a helper class that documents the results. This facilitates a self- documenting system.
Using the archival system example described above, the following is a representative command line interface (CLI) for designating chain generation options: • Initial interface (CLI) o Basic CLI should accept: ■ Cluster DNS name (-c [name])
■ Cluster TPOF (-t ftpofl)
■ List of cluster nodes, (-n [node,node,node])
Number of Chains to run (-chains fnumj)
■ Number of Links within chain(s) (-links [num]. ) Number of allowable Link repetitions (-reps [num.])
Number of allowable "in sequence " reps to allow (-seq [num])
• Only exposed if -reps is added, defaults to 0
• Example: -reps 2, we do not want to have the same link appear in sequence (ie: AA) but it is valid if it appears in the chain twice (ABA) Size of files (range, bytes) to be used by any "write " actions (-ft [min,max])
• Optional, write libraries should have a default.
■ Number of files to be used by any "write " actions (-fit [num])
• Optional, write libraries should have a default. o Example: ./links.py-c foo.archivas.com -t 1, -n IP1IP1IP -chains 10 -links 10 -fr
10,1024 -fn 100000
In the above, example, the designated options cause the system to generate and run ten (10) chains ten (10) links deeps, with a maximum file range of 1024 bytes, with an initial count of 100,000. Of course, this is merely a representative input data set. It is desirable to execute a test sequence from start to finish whenever possible. To this end, and before running a given test, the state verification class may be executed to determine whether the cluster is accessible, whether all nodes in the cluster are still available, whether given paths within the archive are reachable, whether there are unusual log entries, or the like. Although this may not prevent further actions from taking place, it ensures that the state of the SUT is known for debugging/investigation purposes.
The test generation system may include other helper classes. As one example, a ratings class may be used to assign a rating to a given chain after that chain is generated and run. The chain ratings may then be used to order test cases, with test sequences that stress the system the most having a highest rating, whereas those that create less stress having lower ratings.
The present invention provides numerous advantages over the prior art. As has been described, the system exploits an automation paradigm by which a library of test actions is processed selectively to create larger test cases or chains. The generated chains may, in turn, be recombined into larger, more complex chains themselves. These chains make up for the deficiencies in the prior art (where there are a given number of static test cases) by automatically generating what are, in effect, "random" test sequences. Thus, where a given set often (10) actions might have a set of fifty (50) static test cases associated therewith, the inventive methodology can generate over 3,000,000 combinations of those ten actions, creating a much more robust, useful and extensive test suite. Preferably, the number and type of test sequences are constrained only by a set of user-tunable or system-tunable options. An example of a system-tunable constraint in the context of the archive is a policy that might associate a given "weight" with a given link to facilitate creation of "weighted" random sequences, which allows for tunable amounts of actions constrained by risk assessment or "level" (e.g., "basic," "intermediate" or "advanced") to appear within a chain. The user- or system-tunable constraints may also designate a maximum number of links in a chain, a maximum number of times a given link may occur in a chain, a maximum number of times that a given link may appear next to itself in a given chain, or the like. As noted above, in a preferred embodiment, each link is a standalone or discrete that may pass, or fail. It the link fails, the chain dies. Preferably, each link has a conduit through which given information may be passed through to a next or other link in the chain. This conduit, for example, is an in- memory data structure, a cached-on-disk data structure, a text file, a socket communication link, or the like. In particular, in one representative embodiment, the conduit is an in-memory data structure passed from the executor to the individual link, and this data structure comprises information about the SUT such as (in the archive cluster application) current file counts, current file locations, number of nodes, and the like. In another embodiment, the conduit is implemented on disk cache as a plaintext or Python-pickled data structure. In either case, preferably the information in the data structure is modifiable and/or expandable by each link, and preferably all links must either accept the in-memory data structure or assume the on-disk location, as the case must be. Preferably, the conduit is maintained by the executor, on the chain's behalf. A given link in a chain may be either a self-contained entity (e.g., a set of self- contained libraries), or an entity that has associated dependencies. If a particular link has some dependency, preferably that dependency is specified by some criteria, which criteria can then be applied to the link using appropriate system-level or local processing or other resources. If a dependency is needed but is not met," the link is "skipped" or its results ignored. In a preferred embodiment, however, any link dependencies are recognized and verified prior to inclusion of the link in a candidate chain. In this embodiment, a given link has associated therewith given criteria that identify required dependencies. When the system is initialized (e.g., at library import time), the system calls a given dependency verification function within each library to confirm that all identified dependencies are met. In such case, the verification of dependencies must occur before the given link can be used by the sequence generator.
The test generation system of the present invention is also advantageous in that it allows for dynamic test sequence generation as a function of new test actions or new test cases themselves. Indeed, a flexible quality assurance program must take into account regular additions, deletions or modifications of the building blocks and test cases. The inventive system is fluid in the sense that it can at any time generate chains from new actions and tests as needed, easily deprecating irrelevant chains in the process. By registering each chain as it is created, the user or system need not cycle through pre- generated chains before performing a given test suite. Generalizing, the present invention is a test case generation methodology by which a set of discrete test actions form a set of interchangeable building blocks that are permutated (intelligently random) to generate a set of large-scale tests against which a system under test is then exercised in a re-creatable manner. Each test action is an atomic, preferably immutable step in the overall chain/test case generated.
As a variation, the test generation system may also create the building blocks themselves in an automated manner.
While the above describes a particular order of operations performed by certain embodiments of the invention, it should be understood that such order is exemplary, as alternative embodiments may perform the operations in a different order, combine certain operations, overlap certain operations, or the like. References in the specification to a given embodiment indicate that the embodiment described may include a particular feature, structure, or characteristic, but every embodiment may not necessarily include the particular feature, structure, or characteristic. While the present invention has been described in the context of a method or process, the present invention also relates to apparatus for performing the operations herein. This apparatus may be specially constructed for the required purposes, or it may comprise a general-purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including an optical disk, a CD-ROM, and a magnetic-optical disk, a read-only memory (ROM), a random access memory (RAM), a magnetic or optical card, or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus. A given implementation of the present invention is software written in a given programming language (e.g., Python) that runs on a standard Intel hardware platform running an operating system such as Linux.
While given components of the system have been described separately, one of ordinary skill will appreciate that some of the functions may be combined or shared in given instructions, program sequences, code portions, and the like. Having described our invention, what we claim is as follows.

Claims

1. A computer-implemented method to generate replayable test cases for a system under test (SUT), comprising: defining a set of test actions, each of the test actions defining a discrete, atomic operation associated with the system under test; storing the set of discrete test actions as a set of interchangeable building blocks; and using the set of interchangeable building blocks and at least one input constraint to generate a set of test sequences against which the system under test may then be exercised.
2. The computer-implemented method as described in claim 1 further including the steps of: adding an interchangeable building block, deleting an interchangeable building block, or modifying an interchangeable building block; and generating a new set of test sequences.
3. . The computer-implemented method as described in claim 1 wherein the step of using the set of interchangeable building blocks generates a permutation of the interchangeable building blocks.
4. The computer-implemented method as described in claim 1 wherein the input constraint is a number of interchangeable building blocks that may be used to generate a test sequence.
5. The computer-implemented method as described in claim 1 wherein the input constraint is a number of times that a given interchangeable building block may be used in a test sequence.
6. The computer-implemented method as described in claim 1 wherein the input constraint is a number of times that a given interchangeable building block may be used in a given manner within a test sequence.
7. The computer-implemented method as described in claim 1 further including the step of determining whether a given test sequence in the set of test sequences has been generated in a prior test generation sequence.
8. The computer-implemented method as described in claim 1 further including the step of registering a given test sequence in the set of test sequences.
9. The computer-implemented method as described in claim 1 wherein a given interchangeable building block has associated therewith a risk assessment value.
10. The computer-implemented method as described in claim 9 wherein a given test sequence has associated therewith a risk assessment value generated as a function of the risk assessment values of the interchangeable building blocks from which the given test sequence is generated.
11. The computer-implemented method as described in claim 1 wherein a given interchangeable building block has an associated execution dependency.
12. The computer-implemented method as described in claim 11 further including the step of verifying the associated execution dependency before including the given interchangeable building block into the set of interchangeable building blocks.
13. A computer-implemented method to generate replayable test cases for a system under test (SUT), comprising: collecting a set of interchangeable test actions, each of the test actions defining a discrete, atomic operation associated with the system under test; collecting a set of test sequences, wherein each of the set of test sequences is generated as a random permutation from the set of interchangeable test actions; in response to a given input, generating at least one new test sequence; determining whether the new test sequence already exists in the set of test sequences; if the new test sequence does not already exist in the set of test sequences, adding the new test sequence to the set of test sequences.
14. The computer-implemented method as described in claim 13 wherein a given interchangeable test action has associated therewith a risk assessment value.
15. The computer-implemented method as described in claim 14 wherein a given test sequence has associated therewith a risk assessment value generated as a function of the risk assessment values of the interchangeable test actions from which the given test sequence is generated.
16. The computer-implemented method as described in claim 13 wherein the given input is a given constraint.
17. The computer-implemented method as described in claim 16 wherein the given constraint is one of: a number of interchangeable test actions that may be used to generate a test sequence, a number of times that a given interchangeable test action may be used to generate a test sequence, and a number of times that a given interchangeable test action may be used in a given manner within a test sequence.
18. The computer-implemented method as described in claim 13 further including the step of generating a unique identifier for the new test sequence.
19. The computer- implemented method as described in claim 13 wherein a given constraint is associated with the system under test.
20. A computer-implemented method for exercising a system under test (SUT), comprising: defining a set of test actions, each of the test actions defining a discrete, atomic operation associated with the system under test; using the set of test actions and at least one input constraint to generate a set of test sequences; exercising the system under test against the set of test sequences; and re-exercising the system under test at a subsequent time using the at least one of the set of test sequences.
PCT/US2007/000585 2006-01-11 2007-01-10 Method of and system for dynamic automated test case generation and execution WO2007081979A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US11/329,631 2006-01-11
US11/329,631 US8473913B2 (en) 2006-01-11 2006-01-11 Method of and system for dynamic automated test case generation and execution

Publications (2)

Publication Number Publication Date
WO2007081979A2 true WO2007081979A2 (en) 2007-07-19
WO2007081979A3 WO2007081979A3 (en) 2008-04-17

Family

ID=38234192

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2007/000585 WO2007081979A2 (en) 2006-01-11 2007-01-10 Method of and system for dynamic automated test case generation and execution

Country Status (2)

Country Link
US (1) US8473913B2 (en)
WO (1) WO2007081979A2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8140901B2 (en) 2009-11-30 2012-03-20 International Business Machines Corporation Validation of processors using a self-generating test case framework
US8327191B2 (en) 2007-10-19 2012-12-04 International Business Machines Corporation Automatically populating symptom databases for software applications

Families Citing this family (78)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7966599B1 (en) * 2006-08-29 2011-06-21 Adobe Systems Incorporated Runtime library including a virtual file system
US20080098361A1 (en) * 2006-10-20 2008-04-24 Ashish Kumar Method and apparatus for filtering software tests
US20080115114A1 (en) * 2006-11-10 2008-05-15 Sashank Palaparthi Automated software unit testing
US8561024B2 (en) * 2007-01-23 2013-10-15 International Business Machines Corporation Developing software components and capability testing procedures for testing coded software component
US7937621B2 (en) * 2007-05-07 2011-05-03 Intel Corporation Transient fault detection by integrating an SRMT code and a non SRMT code in a single application
US7937620B2 (en) * 2007-05-07 2011-05-03 Intel Corporation Transient fault detection by integrating an SRMT code and a non SRMT code in a single application
US8627287B2 (en) * 2007-11-29 2014-01-07 Microsoft Corporation Prioritizing quality improvements to source code
US7836343B2 (en) * 2008-03-03 2010-11-16 International Business Machines Corporation Method and apparatus for reducing test case generation time in processor testing
US8479173B2 (en) * 2008-07-09 2013-07-02 International Business Machines Corporation Efficient and self-balancing verification of multi-threaded microprocessors
US8463760B2 (en) * 2008-09-04 2013-06-11 At&T Intellectual Property I, L. P. Software development test case management
US9575878B2 (en) * 2009-03-16 2017-02-21 International Business Machines Corporation Data-driven testing without data configuration
US9141513B2 (en) 2009-10-01 2015-09-22 Kryterion, Inc. Maintaining a secure computing device in a test taking environment
US9280907B2 (en) * 2009-10-01 2016-03-08 Kryterion, Inc. Proctored performance analysis
US10672286B2 (en) 2010-03-14 2020-06-02 Kryterion, Inc. Cloud based test environment
US20110223576A1 (en) * 2010-03-14 2011-09-15 David Foster System for the Administration of a Secure, Online, Proctored Examination
US9009668B2 (en) 2010-05-27 2015-04-14 Red Hat Israel, Ltd. Software testing using test entity
US8850396B2 (en) * 2010-05-27 2014-09-30 Red Hat Israel, Ltd. Performing software testing based on grouping of tests using test list entity
US8578344B2 (en) 2010-06-17 2013-11-05 Microsoft Corporation Incremental compositional dynamic test generation
US8499286B2 (en) * 2010-07-27 2013-07-30 Salesforce.Com, Inc. Module testing adjustment and configuration
US9137163B2 (en) 2010-08-04 2015-09-15 Kryterion, Inc. Optimized data stream upload
US8713130B2 (en) 2010-08-04 2014-04-29 Kryterion, Inc. Peered proctoring
US8856749B2 (en) 2010-09-28 2014-10-07 Microsoft Corporation Multi-path brokered test automation execution
CN102063421A (en) * 2010-12-29 2011-05-18 东莞市创锐电子技术有限公司 System for editing test item document in test program
US9104803B2 (en) * 2011-01-03 2015-08-11 Paypal, Inc. On-demand software test environment generation
CN102622294B (en) 2011-01-28 2014-12-10 国际商业机器公司 Method and method for generating test cases for different test types
US9104811B1 (en) * 2011-05-08 2015-08-11 Panaya Ltd. Utilizing testing data collected from different organizations to generate test scenario templates that suit a user profile
US8627147B2 (en) 2011-05-17 2014-01-07 International Business Machines Corporation Method and computer program product for system tuning based on performance measurements and historical problem data and system thereof
AU2012203333A1 (en) 2011-06-15 2013-01-10 Agile Software Pty Limited Method and apparatus for testing data warehouses
US8839035B1 (en) 2011-09-14 2014-09-16 Amazon Technologies, Inc. Cloud-based test execution
US9317268B2 (en) 2012-02-02 2016-04-19 Sungard Availability Services Lp Recovery automation in heterogeneous environments
US9612814B2 (en) * 2012-02-02 2017-04-04 Sungard Availability Services, Lp Network topology-aware recovery automation
DE112012006107B4 (en) * 2012-03-26 2015-12-03 Mitsubishi Electric Corp. The sequence program debugging supporting device
US9971676B2 (en) * 2012-08-30 2018-05-15 Toyota Motor Engineering & Manufacturing North America, Inc. Systems and methods for state based test case generation for software validation
US9727446B2 (en) * 2012-12-05 2017-08-08 The Mathworks, Inc. Modifiers that customize presentation of tested values to constraints
US9613023B2 (en) * 2013-04-04 2017-04-04 Wayne M. Kennard System and method for generating ethnic and cultural emoticon language dictionaries
US9201763B1 (en) * 2013-05-31 2015-12-01 The Mathworks, Inc. Efficient sharing of test fixtures and ordering of test suites
US9424169B1 (en) * 2014-05-22 2016-08-23 Emc Corporation Method of integrating heterogeneous test automation frameworks
WO2016064930A1 (en) * 2014-10-21 2016-04-28 Proofpoint, Inc. Systems and methods for application security analysis
WO2016070135A1 (en) 2014-10-31 2016-05-06 Proofpoint, Inc. Systems and methods for privately performing application security analysis
US9489288B1 (en) * 2014-11-18 2016-11-08 Amdocs Software Systems Limited System, method, and computer program for automating activities associated with a testing project
US20160147646A1 (en) * 2014-11-21 2016-05-26 Wipro Limited Method and system for executing automated tests in an integrated test environment
US9619372B2 (en) * 2015-02-10 2017-04-11 Wipro Limited Method and system for hybrid testing
CN104598114A (en) * 2015-02-28 2015-05-06 苏州路之遥科技股份有限公司 Editing method for microcomputer test process of wiring harness intelligent testing machine
US10176085B2 (en) 2015-06-24 2019-01-08 Tata Consultancy Services Limited Method and system for generating functional test cases for software systems
US9892019B2 (en) * 2015-10-16 2018-02-13 Successfactors Inc. Use case driven stepping component automation framework
US9600401B1 (en) 2016-01-29 2017-03-21 International Business Machines Corporation Automated GUI testing
US9898392B2 (en) 2016-02-29 2018-02-20 Red Hat, Inc. Automated test planning using test case relevancy
US10268570B2 (en) * 2016-05-05 2019-04-23 Oracle International Corporation Methods, systems, and computer readable media for automated generation of test files and testing network equipment using same
US9977725B2 (en) * 2016-08-26 2018-05-22 Cisco Technology, Inc. Automatic classification and parallel processing of untested code in a protected runtime environment
US10061685B1 (en) 2016-08-31 2018-08-28 Amdocs Development Limited System, method, and computer program for high volume test automation (HVTA) utilizing recorded automation building blocks
US10007597B2 (en) * 2016-09-23 2018-06-26 American Express Travel Related Services Company, Inc. Software testing management
US10346290B2 (en) 2016-10-31 2019-07-09 International Business Machines Corporation Automatic creation of touring tests
CN110383253A (en) * 2017-02-10 2019-10-25 欧普菲有限公司 Method, integrated testing instrument and computer program product
WO2018162048A1 (en) * 2017-03-07 2018-09-13 Advantest Corporation Test apparatus and method for characterizing a device under test
US9703693B1 (en) * 2017-03-08 2017-07-11 Fmr Llc Regression testing system for software applications
CN109271308A (en) * 2017-07-17 2019-01-25 中国矿业大学 Class based on search integrates the generation method of initial population in cycle tests problem
US10769056B2 (en) 2018-02-26 2020-09-08 The Ultimate Software Group, Inc. System for autonomously testing a computer system
US11954461B2 (en) 2018-02-26 2024-04-09 Ukg Inc. Autonomously delivering software features
US11010284B1 (en) 2018-05-31 2021-05-18 The Ultimate Software Group, Inc. System for understanding navigational semantics via hypothesis generation and contextual analysis
US11113175B1 (en) 2018-05-31 2021-09-07 The Ultimate Software Group, Inc. System for discovering semantic relationships in computer programs
US10977155B1 (en) 2018-05-31 2021-04-13 The Ultimate Software Group, Inc. System for providing autonomous discovery of field or navigation constraints
US10599767B1 (en) 2018-05-31 2020-03-24 The Ultimate Software Group, Inc. System for providing intelligent part of speech processing of complex natural language
US10747651B1 (en) 2018-05-31 2020-08-18 The Ultimate Software Group, Inc. System for optimizing system resources and runtime during a testing procedure
US10585779B2 (en) * 2018-07-30 2020-03-10 General Electric Company Systems and methods of requirements chaining and applications thereof
CN109508295B (en) * 2018-11-14 2021-11-09 联动优势科技有限公司 Block chain consensus algorithm testing method and device, calculating device and storage medium
US10353804B1 (en) 2019-01-22 2019-07-16 Capital One Services, Llc Performance engineering platform and metric management
EP3696629A1 (en) * 2019-02-12 2020-08-19 Siemens Aktiengesellschaft Method for monitoring an industrial installation, computer program, computer-readable medium and system
US11561997B2 (en) 2019-03-13 2023-01-24 Oracle International Corporation Methods, systems, and computer readable media for data translation using a representational state transfer (REST) application programming interface (API)
CN110175115B (en) * 2019-04-30 2022-12-27 中国航空无线电电子研究所 Variable-based dynamic data operation and management system
US11516000B2 (en) 2019-05-29 2022-11-29 International Business Machines Corporation Approximate hash verification of unused blockchain output
US11539527B2 (en) 2019-05-29 2022-12-27 International Business Machines Corporation Peer node recovery via approximate hash verification
US11429738B2 (en) * 2019-05-29 2022-08-30 International Business Machines Corporation Blockchain endorsement with approximate hash verification
US11570002B2 (en) 2019-05-29 2023-01-31 International Business Machines Corporation Reduced-step blockchain verification of media file
US11711202B2 (en) 2019-05-29 2023-07-25 International Business Machines Corporation Committing data to blockchain based on approximate hash verification
CN110377504A (en) * 2019-06-20 2019-10-25 平安普惠企业管理有限公司 Test method, device, equipment and the storage medium of application program smoothness degree
CN111221725B (en) * 2019-11-27 2023-04-07 杭州涂鸦信息技术有限公司 Automatic testing method and system, readable storage medium and computer equipment
US11347492B2 (en) * 2020-02-26 2022-05-31 Sap Se Software deployment control using blockchain
CN112129547B (en) * 2020-09-29 2023-07-28 北京艾尔动力科技有限公司 Task generation method for automobile endurance unmanned test based on high-precision positioning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5659554A (en) * 1994-01-26 1997-08-19 Kabushiki Kaisha Toshiba Test case generating apparatus
US20040255275A1 (en) * 2003-06-11 2004-12-16 Microsoft Corporation Method and system for generating an efficient test suite from a domain description with given constraints
US20050246207A1 (en) * 2004-03-31 2005-11-03 Noonan Scott A Method for risk based testing

Family Cites Families (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5542043A (en) * 1994-10-11 1996-07-30 Bell Communications Research, Inc. Method and system for automatically generating efficient test cases for systems having interacting elements
US5671351A (en) * 1995-04-13 1997-09-23 Texas Instruments Incorporated System and method for automated testing and monitoring of software applications
US6002869A (en) * 1997-02-26 1999-12-14 Novell, Inc. System and method for automatically testing software programs
US7039912B1 (en) * 1998-05-12 2006-05-02 Apple Computer, Inc. Integrated computer testing and task management systems
US7020697B1 (en) * 1999-10-01 2006-03-28 Accenture Llp Architectures for netcentric computing systems
US6668340B1 (en) * 1999-12-10 2003-12-23 International Business Machines Corporation Method system and program for determining a test case selection for a software application
AU2001264713A1 (en) * 2000-05-19 2001-12-03 Wu-Hon Francis Leung Methods and apparatus for preventing software modifications from invalidating previously passed integration tests
US6601017B1 (en) * 2000-11-09 2003-07-29 Ge Financial Assurance Holdings, Inc. Process and system for quality assurance for software
US6889199B2 (en) * 2000-11-30 2005-05-03 Hewlett-Packard Development Company, L.P. Method and apparatus for encoding and generating transaction-based stimulus for simulation of VLSI circuits
US7917895B2 (en) * 2001-07-27 2011-03-29 Smartesoft, Inc. Automated software testing and validation system
US7055067B2 (en) * 2002-02-21 2006-05-30 Siemens Medical Solutions Health Services Corporation System for creating, storing, and using customizable software test procedures
US7007207B2 (en) * 2002-10-21 2006-02-28 International Business Machines Corporation Scheduling of transactions in system-level test program generation
US7165241B2 (en) * 2002-11-26 2007-01-16 Sun Microsystems, Inc. Mechanism for testing execution of applets with plug-ins and applications
US7685576B2 (en) * 2004-01-26 2010-03-23 Siemens Corporation System and method for model based system testing of interactive applications
US7334220B2 (en) * 2004-03-11 2008-02-19 Microsoft Corporation Data driven test automation of web sites and web services
US7779302B2 (en) * 2004-08-10 2010-08-17 International Business Machines Corporation Automated testing framework for event-driven systems
US7861226B1 (en) * 2006-03-16 2010-12-28 Avaya Inc. Constraint solver to code based test data generation for improving software reliability and security
WO2008039964A1 (en) * 2006-09-28 2008-04-03 Nec Laboratories America, Inc. Software testing using machine learning
US8074204B2 (en) * 2006-11-21 2011-12-06 Microsoft Corporation Test automation for business applications

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5659554A (en) * 1994-01-26 1997-08-19 Kabushiki Kaisha Toshiba Test case generating apparatus
US20040255275A1 (en) * 2003-06-11 2004-12-16 Microsoft Corporation Method and system for generating an efficient test suite from a domain description with given constraints
US20050246207A1 (en) * 2004-03-31 2005-11-03 Noonan Scott A Method for risk based testing

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8327191B2 (en) 2007-10-19 2012-12-04 International Business Machines Corporation Automatically populating symptom databases for software applications
US8140901B2 (en) 2009-11-30 2012-03-20 International Business Machines Corporation Validation of processors using a self-generating test case framework

Also Published As

Publication number Publication date
US8473913B2 (en) 2013-06-25
US20070162894A1 (en) 2007-07-12
WO2007081979A3 (en) 2008-04-17

Similar Documents

Publication Publication Date Title
US8473913B2 (en) Method of and system for dynamic automated test case generation and execution
JP5535484B2 (en) Automated software testing framework
Giménez-Alventosa et al. A framework and a performance assessment for serverless MapReduce on AWS Lambda
US11455217B2 (en) Transaction consistency query support for replicated data from recovery log to external data stores
US8458690B2 (en) Automated software deployment triggered by state differences in distributed systems
US9710367B1 (en) Method and system for dynamic test case creation and documentation to the test repository through automation
US7191364B2 (en) Automatic root cause analysis and diagnostics engine
US20140282400A1 (en) Systems and methods for managing software development environments
EP2779044A1 (en) System and method to provide management of test data at various lifecycle stages
CA2523279A1 (en) Method, system and article of manufacture for data preservation and automated electronic software distribution across an enterprise system
US10922213B2 (en) Embedded quality indication data for version control systems
Lu et al. An empirical case study on the temporary file smell in dockerfiles
US10552306B2 (en) Automated test generation for multi-interface and multi-platform enterprise virtualization management environment
Cotroneo et al. Profipy: Programmable software fault injection as-a-service
CN112835924A (en) Real-time computing task processing method, device, equipment and storage medium
US20120036496A1 (en) Plug-in based high availability application management framework (amf)
Moise et al. Optimizing intermediate data management in MapReduce computations
US8438417B2 (en) Method and apparatus to simplify HA solution configuration in deployment model
Merenstein et al. {CNSBench}: A cloud native storage benchmark
Cavalli et al. StoRM-GPFS-TSM: A new approach to hierarchical storage management for the LHC experiments
US11550816B1 (en) Variable replication levels for an object of a snapshot of a block storage volume
Tavares et al. An efficient and reliable scientific workflow system
Negru et al. Analysis of data integrity and storage quality of a distributed storage system
Ngo et al. Enterprise Application Backup and Restore in Cloud Computing
Markatchev et al. A framework for executing long running jobs in grid environments

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 07717854

Country of ref document: EP

Kind code of ref document: A2