|Numéro de publication||US6668340 B1|
|Type de publication||Octroi|
|Numéro de demande||US 09/458,753|
|Date de publication||23 déc. 2003|
|Date de dépôt||10 déc. 1999|
|Date de priorité||10 déc. 1999|
|État de paiement des frais||Caduc|
|Numéro de publication||09458753, 458753, US 6668340 B1, US 6668340B1, US-B1-6668340, US6668340 B1, US6668340B1|
|Inventeurs||Ronald Bruce Baker, David Bruce Kumhyr|
|Cessionnaire d'origine||International Business Machines Corporation|
|Exporter la citation||BiBTeX, EndNote, RefMan|
|Citations de brevets (9), Référencé par (57), Classifications (9), Événements juridiques (4)|
|Liens externes: USPTO, Cession USPTO, Espacenet|
1. Technical Field
The present invention relates to an improved method for testing a software application and in particular to a method, system and program for determining a test case selection for a software application. Still more particularly, the present invention relates to a method, system and program for determining which test cases should be tested for a software application based on risk factors determined for each function of a software application.
2. Description of the Related Art
As computers have become more prevalent, it has become desirable for software developers to be able to market their products to users of a variety of operating systems, operating under a variety of versions of the operating systems, in a variety of languages. Thereby, software applications continue to grow in complexity as support for multiple operating systems and multiple versions of operating systems, in addition to support for multiple language versions, is added.
As the complexity of a software application is increased, the complexity of testing the software application is also increased. Ideally, the software application should be tested for every function under every supported operating system and with every supported language. However, typically, the time allotted for a testing cycle prior to a software application release does not provide for such testing. For example, a software application with 100 functions to test for with five operating systems and seven languages would take an unacceptable amount of time to test. In addition, if such time were allotted for testing, a software application would typically be outdated before testing would be completed and the product released.
In previous testing methods, each portion of code for a software application was weighted equally, resulting in the need to test each portion of the code. However, for typical software applications, it can be assumed that the typical user will only utilize about 40% of the code controlling the software application. In addition, there are typical functions which need to perform in order for a user to be satisfied with a software application and/or to meet contractual obligations for the performance of a software application. For example, installation of the spelling check function for a word processor application is typically highly important, however installation of the about box is of lower importance.
Therefore, in view of the growing complexity of software applications leading to a growing complexity in test cases, it would be preferable to provide a method of determining which functions of a software applications are the most important to test and thereby determining a method for testing each software application. In addition, it would preferable to determine a risk associated with the software application not performing each of the function, and testing all functions which have an associated risk which is greater than an acceptable risk factor, such that a selection of functions of a software applications that are the most important to test may be determined and tested in a timely manner.
In view of the foregoing, it is therefore an object of the present invention to provide a software application.
It is another object of the present invention to provide a method, system and program for determining a test case selection for a software application.
It is yet another object of the present invention to provide a method, system and program for determining which test cases should be tested for a software application based on risk values determined for each risk factor and each function of a software application.
In accordance with the method, system and program of the present invention, multiple risk values are assigned to each of the test cases of a software application, wherein the multiple risk values represent multiple types of risks associated with the software application not performing each function. An acceptable level of risk is designated. Only a subset of the test cases of the software are selected to test, wherein the subset comprises the test cases assigned risk values that exceed the acceptable level of risk, such that a risk-based test case selection for the software application is determined.
All objects, features, and advantages of the present invention will become apparent in the following detailed written description.
The invention itself, as well as a preferred mode of use, further objects, and advantages thereof, will best be understood by reference to the following detailed description of an illustrative embodiment when read in conjunction with the accompanying drawings, wherein:
FIG. 1 depicts a block diagram of a typical computer system that may utilize a preferred embodiment of the present invention;
FIG. 2 illustrates a chart utilized for entering risk values for multiple test cases with multiple types of risk;
FIG. 3 depicts a graphical display of a chart of risk values for a particular software application;
FIGS. 4a-4 b illustrates a graphical display of a graph for illustrating cumulative risk values for the test cases illustrated in FIG. 3; and
FIG. 5 depicts a high level logic flowchart of a process for determining a test case selection for a software application.
The present invention may be executed in a variety of systems, including a variety of computers under a number of different operating systems. The computer may be, for example, a personal computer, a network computer, a midrange computer or a mainframe computer. In addition, the computer may be a stand-alone system or part of a network such as a local-area network (LAN) or a wide-area network (WAN). For the purposes of illustration, a preferred embodiment of the present invention, as described below, is implemented on a personal computer, such as the Aptiva series manufactured by International Business Machines Corporation.
Referring now to the drawings and in particular to FIG. 1, there is depicted a block diagram of a typical computer system that may utilize a preferred embodiment of the present invention. As shown, a processor (CPU) 12, a read-only memory (ROM) 13, and a Random-Access Memory (RAM) 14 are connected to a system bus 11 of a computer system 10. CPU 12, ROM 13, and RAM 14 are also coupled to a PCI local bus 20 of computer system 10 through a PCI host bridge 16. PCI host bridge 16 provides a low latency path through which processor 12 may directly access PCI devices mapped anywhere within bus memory and/or I/O address spaces. PCI host bridge 16 also provides a high bandwidth path for allowing PCI devices to directly access RAM 14.
Also attaching to PCI local bus 20 are communications adapter 15, small computer system interface (SCSI) 18, and expansion bus bridge 29. Communications adapter 15 is for connecting computer system 10 to a network 17. SCSI 18 is utilized to control high-speed SCSI disk drive 19. Expansion bus bridge 29, such as a PCI-to-ISA bus bridge, may be utilized for coupling ISA bus 25 to PCI local bus 20. As shown, a keyboard 26 and a mouse 28 are attached to ISA bus 25 for performing certain basic I/O functions. In addition, an audio adapter 23 is attached to PCI local bus 20 for controlling audio output through speaker 24. A graphics adapter 21 is also attached to PCI local bus 20 for controlling visual output through display monitor 22. In alternate embodiments of the present invention, additional peripheral components may be added.
Computer system 10 also preferably includes an interface such as a graphical user interface (GUI) and an operating system (OS) that reside within machine readable media to direct the operation of computer system 10. Any suitable machine-readable media may retain the GUI and OS, such as RAM 14, ROM 13, SCSI disk drive 19, and other disk and/or tape drive(e.g. magnetic diskette, magnetic tape, CD-ROM, optical disk, or other suitable storage media). Any suitable GUI and OS may direct CPU 12. For example, the AIX operating system is one of IBM's operating systems which may be implemented.
Further, computer system 10 preferably includes at least one software application(e.g. program product) that resides within machine readable media, for example as a test program 9 residing within RAM 14. A software application contains instructions that when executed on CPU 12 carry out the operations depicted in the flow chart of FIG. 5 and others described herein.
With reference now to FIG. 2, there is depicted a chart utilized for entering risk values for multiple test cases with multiple types of risk. Chart 30 illustrates multiple functions 32 a-32 n along a first axis of chart 30 and multiple types of risk 34 a-32 n along a second axis. In addition, column 36, for accumulating the total risk of sets of multiple types of risk 34 a-32 n, is depicted. For function 32 a, risk values have been entered for each of multiple types of risk 34 a-32 n.
Each of the risk values entered represents a risk value of a test case. Preferably, a selection of test cases to perform is determined based on the risk values assigned thereto.
In addition, an accumulated risk value is calculated for the intersection of column 36 and function 32 a. In the present example, the accumulated risk is additive, however in other embodiments of the present invention, the accumulated risk may be determined utilizing a weighted formula or other accumulation factor.
In particular, in determining functions 32 a-32 n, it is desirable to model the normal life cycle of a software product as closely as possible. In general, the normal life cycle of a software product includes installation, configuration, normal operation, backup and archiving, patching, and deleting or upgrading. Therefore, functions 32 a-32 n preferably encompass all the functions utilized in the normal life cycle of a software product.
In determining the types of risk 34 a-32 n, a first major type of risk is the business risk associated with a function not performing, regardless of the operating or language. For example, there is a greater business risk associated with a software product not installing properly than with an about box not functioning.
In addition, there is a risk associated with a function not performing correctly or returning corrupt results for multiple operating systems. For example, there is a greater risk associated with a function not performing with an operating system that 60% of the target market utilizes than with an operating system that only 5% of the target market utilizes. Therefore, analysis of the market share of different operating systems may be utilized to determine the risk associated with each operating system.
Further, there is a risk associated with a function not translating into other written languages. For example, Western European languages typically add accented characters in translation. However, Asian languages utilize multibyte encodings and may also utilize composed characters. In determining the risk, for example, there is a greater risk associated with a function of a software application not performing in a written language utilized by 50% of the target market than with a written language utilized by 2% of the target market. Therefore, analysis of the countries and languages in which the software application will be utilized may determine the risk associated with each written language.
Referring now to FIG. 3, there is illustrated a graphical display of a chart of risk values for a particular software application in accordance with the invention. Chart 40 includes multiple installation functions 42, multiple client deployment functions 44, and multiple firewall functions 46. In addition, chart 40 includes multiple risk factors 48, including business risk, multiple operating systems and versions thereof, and multiple languages. Among the operating systems listed, AIX is a trademark of IBM Corporation and Windows NT is a trademark of Microsoft Corporation.
Multiple risk values, such as a risk value 50, are illustrated wherein each of functions 42, 44 and 46 are assigned risk values for each of risk factors 48. For the present example, a numerical value between 0 and 3 is assigned where a numerical value of 0 represents minimal performance risk and a numerical value of 3 represents substantial performance risk.
Each of the multiple risk values, such as risk value 50, represents a test case. In order to determine which test cases will be tested, an acceptable risk value level must be determined. Any test cases with risk values greater than the acceptable risk value level should be tested. For example, if a risk value level of “2” is determined, then each of the test cases with risk levels greater than “2” should be tested. Alternatively, an acceptable cumulative risk level for the cumulative risk value may be utilized.
With reference now to FIGS. 4a-4 b, there is depicted a graphical display of a graph for illustrating cumulative risk values for the test cases illustrated in FIG. 3. Referring to FIG. 4a, chart 60 illustrates a three-dimensional plot of the test cases depicted in FIG. for installation functions 42. A first axis 62 depicts each of risk factors 48 of FIG. 3. A second axis 64 illustrates each of functions 42 of FIG. 3. A third axis 66 depicts a numerical value corresponding to the risk value for each test case.
Referring to FIG. 4b, chart 70 depicts a three-dimensional plot of the test cases depicted in FIG. 3 for client deployment functions 44. A first axis 72 illustrates each of risk factors 48 of FIG. 3. A second axis 74 depicts each of functions 44 of FIG. 3. A third axis 76 illustrates a numerical value corresponding to the risk value for each test case. While not depicted, a three-dimensional plot of the test cases in FIG. 3 for firewall functions 46 may also be determined. Further, all of the functions depicted in FIG. 3 may be included in a single graphical plot.
Utilizing charts 60 and 70, all test cases which have a risk value greater than an acceptable risk level can be determined. For example, in chart 60, for an acceptable risk level of the numerical value 2, there are multiple risk values displayed above the numerical value of 2. By selecting test cases with a risk value greater than an acceptable risk level the number of test cases that must be tested is reduced.
In general, from the graphical plot, several islands of risk may be determined where all the test cases in an area are assigned the same risk value. For example, the test cases for all of the installation functions for operating systems AIX 3.2.5, AIX 4.1.4, and AIX 4.2 form an island of risk. In addition, islands of risk are distinguished by barrier reef formations in the plot where the risk values are less than the risk value assigned for the island of risk. For example, a barrier reef is formed by all the test cases involving installation functions for the Solaris 2.7 Japanese operating system. From analyzing the various islands of risk formed from the barrier reefs, an acceptable risk level may be determined. Further, acceptable risk levels may be determined for each set of functions. In addition, alternative methods may be utilized in determining an acceptable risk level.
Referring now to FIG. 5, there is illustrated a high level logic flowchart of a process for determining a test case selection for a software application. As depicted, the process starts at block 80 and thereafter proceeds to block 82. Block 82 illustrates determining a business risk value associated with each function. The business risk may be determined from user input, or for example, from statistical analysis designating a monetary value associated with each function. Thereafter, block 84 depicts determining a risk value associated with each operating system for each function. The risk value associated with each operating system may be determined from user input, or for example, from statistical analysis designating the percentages of the target market that utilize each operating system.
Next, block 86 illustrates determining a risk value associated with each language for each function. The risk value associated with each language may be determined from user input, or for example, from statistical analysis designating the percentages of the target market utilizing each language that are potential customers. Thereafter, block 88 depicts calculating the accumulated risk value for each function. The accumulated risk value may be determined by adding the risk values assigned to each function. Alternatively, a formula may be utilized to calculate an accumulated risk value.
Block 90 illustrates displaying a three-dimensional plot of the test cases. Thereafter, block 92 depicts determining an acceptable level of risk. The acceptable level of risk may be determined from user input, or for example, from analysis of the islands of risk and barrier reefs in the plot. Next, block 94 illustrates designating a selection of test cases for testing. Thereafter, the process ends.
As has been described, the present invention provides a method and system for determining a test case selection for a software application. Although the determination of a test case selection can be made through entering risk values into a chart and displaying the results of a chart in a graphical format, other methods of entering risk values may be utilized and other methods of displaying the accumulated risk values may be provided. For example, a three-dimensional graph is utilized to display the results of accumulated risk values, however other types of graphical models may also be utilized. In addition other methods of displaying all risk values which exceed an acceptable risk value level may be displayed by alternate methods. Further, the implementation of the present invention may be carried out utilizing conventional software applications which provide for entering data in a chart format which can then be graphically plotted.
It important to note that, although the present invention has been described in the context of a fully functional computer system, those skilled in the art will appreciate that the mechanisms of the present invention are capable of being distributed as a program product in a variety of forms, and that the present invention applies equally regardless of the particular type of signal-bearing media utilized to actually carry out the distribution. Examples of signal-bearing media include, but are not limited to, recordable-type media such as floppy disks or CD ROMs and transmission-type media such as analogue or digital communications links.
While the invention has been particularly shown and described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made herein without departing from the spirit and scope of the invention.
|Brevet cité||Date de dépôt||Date de publication||Déposant||Titre|
|US5544308 *||2 août 1994||6 août 1996||Giordano Automation Corp.||Method for automating the development and execution of diagnostic reasoning software in products and processes|
|US5631857 *||15 juin 1994||20 mai 1997||International Business Machines Corporation||Measuring test tool effectiveness|
|US5754760 *||30 mai 1996||19 mai 1998||Integrity Qa Software, Inc.||Automatic software testing tool|
|US5892947 *||1 juil. 1996||6 avr. 1999||Sun Microsystems, Inc.||Test support tool system and method|
|US5913023 *||30 juin 1997||15 juin 1999||Siemens Corporate Research, Inc.||Method for automated generation of tests for software|
|US6167352 *||26 juin 1997||26 déc. 2000||Agilent Technologies, Inc.||Model-based diagnostic system with automated procedures for next test selection|
|US6219805 *||15 sept. 1998||17 avr. 2001||Nortel Networks Limited||Method and system for dynamic risk assessment of software systems|
|US6385741 *||26 mai 1999||7 mai 2002||Fujitsu Limited||Method and apparatus for selecting test sequences|
|US6415396 *||26 mars 1999||2 juil. 2002||Lucent Technologies Inc.||Automatic generation and maintenance of regression test cases from requirements|
|Brevet citant||Date de dépôt||Date de publication||Déposant||Titre|
|US6859894 *||30 nov. 2001||22 févr. 2005||Emc Corporation||System and method for risk management of errors in software program code|
|US7065677 *||10 avr. 2003||20 juin 2006||Microsoft Corporation||Method for testing operating system components|
|US7191435 *||7 juin 2002||13 mars 2007||Sun Microsystems, Inc.||Method and system for optimizing software upgrades|
|US7278059 *||7 févr. 2006||2 oct. 2007||Microsoft Corporation||Method for testing operating system components|
|US7506312 *||31 janv. 2008||17 mars 2009||International Business Machines Corporation||Method and system for automatically determining risk areas to retest|
|US7747987 *||5 août 2004||29 juin 2010||Cisco Technology, Inc.||System and method of analyzing risk in risk-based software testing|
|US7788057||23 janv. 2008||31 août 2010||International Business Machines Corporation||Test apparatus and methods thereof|
|US7937689||15 oct. 2007||3 mai 2011||International Business Machines Corporation||Method, apparatus and computer program product for determining a relative measure of build quality for a built system|
|US8056060||5 juin 2007||8 nov. 2011||International Business Machines Corporation||Software testing method and system|
|US8266700 *||16 mai 2005||11 sept. 2012||Hewlett-Packard Development Company, L. P.||Secure web application development environment|
|US8463760||12 déc. 2008||11 juin 2013||At&T Intellectual Property I, L. P.||Software development test case management|
|US8473913 *||11 janv. 2006||25 juin 2013||Hitachi Data Systems Corporation||Method of and system for dynamic automated test case generation and execution|
|US8689187 *||7 mai 2007||1 avr. 2014||Cadence Design Systems, Inc.||Predictive run testing|
|US8855801 *||17 oct. 2007||7 oct. 2014||International Business Machines Corporation||Automated integration of feedback from field failure to order configurator for dynamic optimization of manufacturing test processes|
|US8930908 *||20 déc. 2010||6 janv. 2015||Sap Se||Aspect and system landscape capability-driven automatic testing of software applications|
|US8997052 *||19 juin 2013||31 mars 2015||Successfactors, Inc.||Risk-based test plan construction|
|US9396160 *||28 févr. 2013||19 juil. 2016||Amazon Technologies, Inc.||Automated test generation service|
|US9436725 *||28 févr. 2013||6 sept. 2016||Amazon Technologies, Inc.||Live data center test framework|
|US9444717 *||28 févr. 2013||13 sept. 2016||Amazon Technologies, Inc.||Test generation service|
|US9485207 *||30 oct. 2013||1 nov. 2016||Intel Corporation||Processing of messages using theme and modality criteria|
|US9489290 *||5 nov. 2013||8 nov. 2016||The Mathworks, Inc.||Scheduling tests based on a valuation system|
|US9529700 *||10 juin 2014||27 déc. 2016||Wipro Limited||Method of optimizing execution of test cases and a system thereof|
|US9569205 *||10 juin 2013||14 févr. 2017||Symantec Corporation||Systems and methods for remotely configuring applications|
|US9571561 *||27 déc. 2013||14 févr. 2017||Samsung Sds Co., Ltd.||System and method for dynamically expanding virtual cluster and recording medium on which program for executing the method is recorded|
|US20030229890 *||7 juin 2002||11 déc. 2003||Michael Lau||Method and system for optimizing software upgrades|
|US20040015375 *||1 avr. 2002||22 janv. 2004||John Cogliandro||System and method for reducing risk|
|US20060150025 *||7 févr. 2006||6 juil. 2006||Microsoft Corporation||Method for testing operating system components|
|US20060259973 *||16 mai 2005||16 nov. 2006||S.P.I. Dynamics Incorporated||Secure web application development environment|
|US20070162894 *||11 janv. 2006||12 juil. 2007||Archivas, Inc.||Method of and system for dynamic automated test case generation and execution|
|US20080046791 *||5 juin 2007||21 févr. 2008||International Business Machines Corporation||Software Testing Method and System|
|US20080104577 *||15 oct. 2007||1 mai 2008||Holden Ian G||Method, Apparatus and Computer Program Product for Determining a Relative Measure of Build Quality for a Built System|
|US20080172580 *||15 janv. 2007||17 juil. 2008||Microsoft Corporation||Collecting and Reporting Code Coverage Data|
|US20080172651 *||15 janv. 2007||17 juil. 2008||Microsoft Corporation||Applying Function Level Ownership to Test Metrics|
|US20080172652 *||15 janv. 2007||17 juil. 2008||Microsoft Corporation||Identifying Redundant Test Cases|
|US20080172655 *||15 janv. 2007||17 juil. 2008||Microsoft Corporation||Saving Code Coverage Data for Analysis|
|US20080282124 *||7 mai 2007||13 nov. 2008||Cadence Design Systems, Inc.||Predictive run testing|
|US20080300888 *||30 mai 2007||4 déc. 2008||General Electric Company||Systems and Methods for Providing Risk Methodologies for Performing Supplier Design for Reliability|
|US20090105863 *||17 oct. 2007||23 avr. 2009||International Business Machines Corporation||Automated integration of feedback from field failure to order configurator for dynamic optimization of manufacturing test processes|
|US20100057693 *||12 déc. 2008||4 mars 2010||At&T Intellectual Property I, L.P.||Software development test case management|
|US20120042302 *||16 août 2010||16 févr. 2012||Bhava Sikandar||Selective regression testing|
|US20120159445 *||20 déc. 2010||21 juin 2012||Sap Ag||Aspect and system landscape capability-driven automatic testing of software applications|
|US20130041796 *||8 août 2011||14 févr. 2013||Bank Of America Corporation||Application governance process and tool|
|US20140181226 *||17 déc. 2013||26 juin 2014||Samsung Electronics Co., Ltd.||Content-centric network communication method and apparatus|
|US20140189109 *||27 déc. 2013||3 juil. 2014||Samsung Sds Co., Ltd.||System and method for dynamically expanding virtual cluster and recording medium on which program for executing the method is recorded|
|US20140380277 *||19 juin 2013||25 déc. 2014||Successfactors, Inc.||Risk-based Test Plan Construction|
|US20150067648 *||27 août 2014||5 mars 2015||Hcl Technologies Limited||Preparing an optimized test suite for testing an application under test in single or multiple environments|
|US20150309918 *||10 juin 2014||29 oct. 2015||Wipro Limited||Method of optimizing execution of test cases and a system thereof|
|US20160182298 *||18 déc. 2014||23 juin 2016||International Business Machines Corporation||Reliability improvement of distributed transaction processing optimizations based on connection status|
|CN1318976C *||16 juin 2004||30 mai 2007||华为技术有限公司||Software detection method and system|
|CN1326044C *||22 oct. 2004||11 juil. 2007||中国工商银行股份有限公司||Automation software testing system based on script explanatory tool|
|CN101859276B||14 mai 2010||22 févr. 2012||东南大学||Regression testing case selection method based on hierarchical slicing|
|CN101866317A *||29 juin 2010||20 oct. 2010||南京大学||Regression test case selection method based on cluster analysis|
|CN101866317B||29 juin 2010||25 juil. 2012||南京大学||Regression test case selection method based on cluster analysis|
|CN102629229A *||2 mars 2012||8 août 2012||南京大学||Co-diversified random testing (CAT) method|
|CN102629229B||2 mars 2012||10 sept. 2014||南京大学||Co-diversified random testing (CAT) method|
|CN103294457A *||27 févr. 2012||11 sept. 2013||百度在线网络技术（北京）有限公司||Method and device for dynamically substituting C/C + + function in main program|
|CN103294457B *||27 févr. 2012||1 févr. 2017||百度在线网络技术（北京）有限公司||动态替换主程序中c/c++函数的方法及装置|
|Classification aux États-Unis||714/38.1, 717/124, 714/26, 714/E11.208|
|Classification coopérative||G06F11/3676, G06F11/3688|
|Classification européenne||G06F11/36T2E, G06F11/36T2A|
|10 déc. 1999||AS||Assignment|
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BAKER, RONALD B.;KUMHYR, DAVID B.;REEL/FRAME:010484/0861
Effective date: 19991208
|5 juil. 2007||REMI||Maintenance fee reminder mailed|
|23 déc. 2007||LAPS||Lapse for failure to pay maintenance fees|
|12 févr. 2008||FP||Expired due to failure to pay maintenance fee|
Effective date: 20071223