US20090024874A1 - Generic template to autogenerate reports for software target testing - Google Patents
Generic template to autogenerate reports for software target testing Download PDFInfo
- Publication number
- US20090024874A1 US20090024874A1 US11/879,716 US87971607A US2009024874A1 US 20090024874 A1 US20090024874 A1 US 20090024874A1 US 87971607 A US87971607 A US 87971607A US 2009024874 A1 US2009024874 A1 US 2009024874A1
- Authority
- US
- United States
- Prior art keywords
- software
- reports
- cases
- class
- testing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/36—Preventing errors by testing or debugging software
- G06F11/3668—Software testing
- G06F11/3672—Test management
- G06F11/3688—Test management for test execution, e.g. scheduling of test suites
Definitions
- the present invention relates to computing environments involved in testing software products (e.g., software targets). Particularly, it relates to software testing applications able to autogenerate reports useful to users diagnosing individual test cases of software products.
- software products e.g., software targets
- Various features include utilizing existing Java-based JUnit testing frameworks and an API of Apache Ant. Executable code in the form of a template enables convenient usability with multiple cases to be tested. Still other features include computer program products. Overall architecture in a virtual machine is another noteworthy aspect.
- test cases of the software target are written as to examine whether features of the target are operating properly. For example, it may be the situation that a web service providing stock market quotes with a GUI front end has need of determining whether a user fill-in box for stock symbols actually produces a stock quote for the symbol upon submission. In such instance, the software testing application uses code for an individual test case to determine whether the web service works as advertised.
- test case construction may involve many classes and methods in a language such as C, C++, or Java and problems exist when trying to integrate the JUnit test case execution followed by a user report generation in an automated fashion.
- TestCase class Another problem with using the JUnit framework to execute test cases requires extending the TestCase class by design. In turn, this may lead to programming issues when trying to modifying a third party test tool class hierarchy. For instance, the third party test tool's class loader may generate, or “throw,” exceptions when trying to load the altered class hierarchy.
- JUnit typically generates reports in an XML format, which is non-visually recognized by users. Converting the XML to a visually useful format, such as HTML, requires several transformation classes. While this may be typically done using an Apache Ant script, which runs as a separate process, launching the Ant script requires still further programming code. In other words, it is not self-executing. Interoperability issues may also exist between Linux- and Windows-brand operating systems, for example.
- one or more software test cases of a feature of a software target are identified for execution that indicate a pass or failure, for example.
- the JUnit framework autogenerates a first report in a first computing language, e.g., XML, based on test results obtained from the execution of each of the software cases.
- a transformation into one or more second reports in a second computing language e.g., HTML, occurs so that users can visually understand the report.
- Retrofitting existing software testing applications contemplates inserting executable code, in the form of a template, to obtain auto-generated reports.
- Executable code is available as a computer program product in the form of a download or on a computer-readable medium.
- Overall architecture in a virtual machine is another noteworthy aspect.
- the invention embeds an inner class within a main test script class acting as a proxy mechanism.
- Each method within the inner class calls the corresponding method in the main test script class.
- Certain of the proxy class tasks include: identifying all test cases, such as obtaining a list of the methods that start with the prefix “test” (required by JUnit as an executable test); using the JUnit framework to execute each method; interfacing with the API of Junit to gather the test result information and create an output file in a first computing language, e.g., XML; interfacing with the Ant API to transform the output file in the first computing language to a second computing language, e.g., transforming the XML to HTML, without Ant scripting; and specifying a location for placing the generated and transformed reports/files.
- Certain advantages of the invention include, but are not limited to: unification of the traditional JUnit testing process with autogenerated XML to HTML report transformation; creating a portable class structure transferable into any class; applicability to any third party software testing tool, especially without having to modify class hierarchies; and “hiding” underlying complexities, from users, relating to accessing both the JUnit and Ant frameworks in order to generate useful, standardized HTML user reports. More complex reports would require enhancements to the XSL file provided by the JUnit framework.
- FIG. 1 is a diagrammatic view in accordance with the present invention of a representative computing environment for a generic template to autogenerate reports for software target testing;
- FIG. 2 is a combined diagrammatic architecture and flow chart in accordance with the present invention for auto-generating reports.
- a representative environment 10 for auto-generating reports consists of one or more computing devices 15 or 15 ′ available per single or multiple instances of a software testing application or a software target to be tested, and/or singularly or collectively arranged as physical and/or virtual machines for the auto-generating of reports.
- an exemplary computing device typifies a server 17 , such as a grid or blade server.
- it includes a general or special purpose computing device in the form of a conventional fixed or mobile computer 17 having an attendant monitor 19 and user interface 21 .
- the computer internally includes a processing unit for a resident operating system, such as DOS, WINDOWS, MACINTOSH, VISTA, UNIX and LINUX, to name a few, a memory, and a bus that couples various internal and external units, e.g., other 23, to one another.
- a processing unit for a resident operating system such as DOS, WINDOWS, MACINTOSH, VISTA, UNIX and LINUX, to name a few
- a memory and a bus that couples various internal and external units, e.g., other 23, to one another.
- Representative other items 23 include, but are not limited to, PDA's, cameras, scanners, printers, microphones, joy sticks, game pads, satellite dishes, hand-held devices, consumer electronics, minicomputers, computer clusters, main frame computers, a message queue, a peer machine, a broadcast antenna, a web server, an AJAX client, a grid-computing node, a peer, a virtual machine, a web service endpoint, a cellular phone or the like.
- the other items may also be stand alone computing devices 15 ′ in the environment 10 or the computing device itself, upon which the document editor is installed.
- storage devices are contemplated and may be remote or local. While the line is not well defined, local storage generally has a relatively quick access time and is used to store frequently accessed data, while remote storage has a much longer access time and is used to store data that is accessed less frequently. The capacity of remote storage is also typically an order of magnitude larger than the capacity of local storage.
- storage is representatively provided for aspects of the invention contemplative of computer executable instructions, e.g., software, as part of computer program products on readable media, e.g., disk 14 for insertion in a drive of computer 17 .
- Computer executable instructions may also be available as a download or reside in hardware, firmware or combinations in any or all of the depicted devices 15 or 15 ′.
- the computer product can be any available media, such as RAM, ROM, EEPROM, CD-ROM, DVD, or other optical disk storage devices, magnetic disk storage devices, floppy disks, or any other medium which can be used to store the items thereof and which can be assessed in the environment.
- the computing devices communicate with one another via wired, wireless or combined connections 12 that are either direct 12 a or indirect 12 b . If direct, they typify connections within physical or network proximity (e.g., intranet). If indirect, they typify connections such as those found with the internet, satellites, radio transmissions, or the like, and are given nebulously as element 13 .
- other contemplated items include servers, routers, peer devices, modems, T1 lines, satellites, microwave relays or the like.
- the connections may also be local area networks (LAN) and/or wide area networks (WAN) that are presented by way of example and not limitation.
- the topology is also any of a variety, such as ring, star, bridged, cascaded, meshed, or other known or hereinafter invented arrangement.
- FIG. 2 illustrates a high-level organization for auto-generating reports, such as in a virtual machine.
- the architecture includes a main test script class 42 and a test fixture class 44 .
- code in the form of a plurality of test cases are given as 46 - 1 , 46 - 2 and 46 - 3 .
- test cases relate to or are associated with the software target and are written to examine whether features of the target are operating properly.
- a stock market web service example was given.
- test cases may typify a response for displaying results upon a user clicking a certain icon or pushbutton of a program. Alternatively, it may represent ascertaining determinations about mouse locations on a computing monitor whereby users navigate or “mouse-over” items.
- test cases per software targets are possible and such is largely dictated by the functionality of the target and its interaction with users, other programs, other websites, etc.
- the precise embodiment of a test case therefore, is merely contextual.
- test fixture class 44 it is important to the user of the software testing application to know and record the results of the test case, such as whether the test passed, failed, met a predetermined metric or other, such as including whether to throw an error message or not. Users can then use the information for purposes of troubleshooting, providing quality assurance reports, developing marketing material, improving the product, or the like.
- the reports of the test cases are generated for the users at 52 .
- the reports are in an HTML computing language and result from: 1) executing the code of each of the test cases through a JUnit testing framework at 54 , which, by its nature, generates reports in an XML computing language at 55 ; and 2) transforming or converting the generated XML reports to HTML by way of the Apache Ant API at 56 .
- the transformation contemplates conversion without resorting to the cumbersomeness of Ant scripting and uses existing functionality of the JUnit framework in the form of its XSL template. All of which, will be described in more detail relative to exemplary executable code of one or more working prototypes of the invention.
- the inner class template 50 is established by the Test Main 58 .
- m_ScriptTest.setTheReportsDirectory m_ReportsDirectory
- the inner class template is executed.
- test cases to be tested 46 - x are executed by way of the JUnit testing framework and the Ant API, while avoiding any Ant scripting. Thereafter, the reports at 55 are generated for users.
- the test cases are found by all methods beginning with the prefix labeled “test.” Using the Java introspection API all test cases beginning with the prefix labeled “test” will be found.
- class variable 60 In that other class variables and various initialization(s) need to be present, such is illustrated generically as “class variable” 60 .
- the main method contains additional lines of executable code (the “m_ScriptTest” lines) that create a new instance of an inner class 50 .
- a parameter called the “this” variable is passed to a constructor of the inner class.
- This object reference instance to the main test script class 42 is saved within the inner class.
- a single method call is made to set the location of where the reports are to be written and then made (e.g., m_ScriptTest. setTheReportsDirectory (m_ReportsDirectory)).
- the last step is to execute a line to run the script (e.g., m_ScriptTest. RunScript ( )).
- the inner class 50 located within the main test script class 42 inherits methods from the TestCase, which allows execution of each test case method gathered from within the inner class.
- Upon method invocation of each test case which acts as the proxy method to the “real” test case code located within the main test script class is then executed. In other words, the a pass-through of sorts exists to the test case code 46 - x for the ultimate execution thereof.
- the test case is then run thru or executed by the JUnit framework at 54 to generate the XML results at 55 .
- This XML result file is transformed to HTML at 52 by calling the XSL transform methods located within the Ant API at 56 .
- the HTML result files or reports are stored in the location specified above when the main test script “main” method was invoked.
- TestJunitScript the inner class will be found by the name “TestJunitScriptTest.” Comments in the code will be delineated by // . . . // markings or/** . . . */ markings.
- RFT may sometimes be used to identify a third party software testing application.
- Section 1 Main Test Script Class (Item 42 , FIG. 2)
- a third party test tool e.g. a software testing application such as RFT
- RFT software testing application
- Section 4. Script Main Method (Entry Point; Item 58 , FIG. 2)
- the inner class is actually the class passed to JUnit framework in which the test cases were are to be run, a link is needed to the main test class where the actual code or “work” for the test cases is actually implemented. This is where the additional code in the TestMain method comes into play.
- the inner class is passed a reference using the “this” pointer to the defined software testing application (e.g., RFT) class instance. This reference is used to “call” the actual methods that do the work from within the proxy inner class being executed by the Junit framework.
- RFT software testing application
- test cases are not guaranteed to be executed in a // particular order since they are obtained by introspection. // If test case order of execution is required, initialize // a String [ ] array with the test case method names. // The runScript method is overloaded and will accept the // //string array.
- m_ScriptTest new TestJunitScriptTest(this); m_ScriptTest.setTheReportsDirectory(m_ReportsDirectory); m_ScriptTest.setTheBuildNumber(m_BuildNumber); m_ScriptTest.runScript( ); ⁇ catch(TestProgramException tpe) ⁇ // // Log the error in RFT log logError(tpe.getMessage( ) ); throw new TestProgramException(tpe.getMessage( )); ⁇ ⁇
- the inner class provides both a proxy and a link to the JUnit framework. Skilled artisans will note that this proxy class inserts itself into the hierarchy which eventually extends the TestCase class required by JUnit.
- TestJunitScriptTest extends RFTTestFixture ⁇ /** * The following JUnit test case a sample template.
- Test Tool Level Proxy Class (RFTTestFixture Class, Item 44 , FIG. 2)
- This class provides the last level of abstraction via the client side public API.
- This class provides the wrapper mechanism that integrates into the JUnit and Ant framework.
- This code exists in the provided library jar item 56 , FIG. 2 in which the user does not have access to from the client source (nor do they likely desire to have access to).
- the “public void runScript( ) throws TestProgramException” method begins the running of the testing of cases through the JUnit framework.
- the following template was developed for our third party test tool which embeds the necessary code within any named test script.
- a script is created via the third party test tool the name given to the script would be substituted where necessary. This will enable any user creating a test script with a sample test case to be generated.
- This template frees the user to add code to test the target via the sample test case or adding any new ones using the mySampleTestCase method as a guide.
- m_ScriptTest new %script:name%Test(this); m_ScriptTest.setTheReportsDirectory(m_ReportsDirectory); m_ScriptTest.setTheBuildNumber(m_BuildNumber); m_ScriptTest.runScript( ); ⁇ catch(TestProgramException tpe) ⁇ // // Log the error in RFT log logError(tpe.getMessage( ) ); throw new TestProgramException(tpe.getMessage( )); ⁇ ⁇ /** * Standard fixture class that uses JUnit to run each testcase * and create a report using the ANT API.
- TestFixure class contains all methods to run the * JUnit testcases, collect the results, write out the xml result * document, and transform the xml data file to Html reports.
- */ public static class %script:name%Test extends RFTTestFixture ⁇ /** * The following JUnit test case a sample template.
Abstract
Description
- Generally, the present invention relates to computing environments involved in testing software products (e.g., software targets). Particularly, it relates to software testing applications able to autogenerate reports useful to users diagnosing individual test cases of software products. Various features include utilizing existing Java-based JUnit testing frameworks and an API of Apache Ant. Executable code in the form of a template enables convenient usability with multiple cases to be tested. Still other features include computer program products. Overall architecture in a virtual machine is another noteworthy aspect.
- “Software Testing Applications,” such as IBM's Rational Robot as well as the latest product, Functional Tester (RFT) have been known for some time for quality testing software targets, such as web or server applications, stand-alone computing applications, web services, computer program products, GUI interfaces, etc., to name a few. During use, test cases of the software target are written as to examine whether features of the target are operating properly. For example, it may be the situation that a web service providing stock market quotes with a GUI front end has need of determining whether a user fill-in box for stock symbols actually produces a stock quote for the symbol upon submission. In such instance, the software testing application uses code for an individual test case to determine whether the web service works as advertised. Regardless of success, it is important to the user of the software testing application to know and record the results of the test, e.g., whether the test passed, failed, met a predetermined metric, or other. (Of course, an infinite number of test cases per software targets are possible and the foregoing is merely a contextual illustration that skilled artisans will readily understand.)
- One well known and open source testing framework, known commonly as JUnit, is another tool in the quality assurance of testing software targets. However, constructing test cases that integrate well with the JUnit framework and automatically generating HTML reports based upon test results can sometimes be cumbersome when using third party test tools, e.g., RFT. As is known, test case construction may involve many classes and methods in a language such as C, C++, or Java and problems exist when trying to integrate the JUnit test case execution followed by a user report generation in an automated fashion.
- Another problem with using the JUnit framework to execute test cases requires extending the TestCase class by design. In turn, this may lead to programming issues when trying to modifying a third party test tool class hierarchy. For instance, the third party test tool's class loader may generate, or “throw,” exceptions when trying to load the altered class hierarchy.
- Still another problem lies with report generation in a format useful for users. Namely, JUnit typically generates reports in an XML format, which is non-visually recognized by users. Converting the XML to a visually useful format, such as HTML, requires several transformation classes. While this may be typically done using an Apache Ant script, which runs as a separate process, launching the Ant script requires still further programming code. In other words, it is not self-executing. Interoperability issues may also exist between Linux- and Windows-brand operating systems, for example.
- Accordingly, there is need in the art of software testing applications to easily and conveniently generate user reports that are presented in useful format. There is further need to make the reporting “invisible” relative to the users and software testing applications to minimize inconvenience. It is also important to generate reports quickly and provide appropriate results. In that many users already own and/or use a software testing application, it is further desirable to retrofit existing applications to avoid the costs of providing wholly new products. Taking advantage of existing testing frameworks, such as JUnit, and transformation-of-languages programs, such as Ant, is another feature that would optimizes existing resources. Naturally, any improvements along such lines should further contemplate good engineering practices, such as relative inexpensiveness, stability, ease of implementation, low complexity, flexibility, etc.
- The above-mentioned and other problems become solved by applying the principles and teachings associated with the hereinafter-described generic template to autogenerate reports for software target testing. In a basic sense, users test one or more features of targets with specifically written test case code. A template, common to each of the test cases, utilizes a JUnit testing framework and an Apache Ant API to autogenerate useful user reports heretofore unavailable.
- During use, one or more software test cases of a feature of a software target are identified for execution that indicate a pass or failure, for example. The JUnit framework autogenerates a first report in a first computing language, e.g., XML, based on test results obtained from the execution of each of the software cases. By way of the Ant API, a transformation into one or more second reports in a second computing language, e.g., HTML, occurs so that users can visually understand the report.
- Retrofitting existing software testing applications contemplates inserting executable code, in the form of a template, to obtain auto-generated reports. Executable code is available as a computer program product in the form of a download or on a computer-readable medium. Overall architecture in a virtual machine is another noteworthy aspect.
- In the architecture, the invention embeds an inner class within a main test script class acting as a proxy mechanism. Each method within the inner class calls the corresponding method in the main test script class. Certain of the proxy class tasks include: identifying all test cases, such as obtaining a list of the methods that start with the prefix “test” (required by JUnit as an executable test); using the JUnit framework to execute each method; interfacing with the API of Junit to gather the test result information and create an output file in a first computing language, e.g., XML; interfacing with the Ant API to transform the output file in the first computing language to a second computing language, e.g., transforming the XML to HTML, without Ant scripting; and specifying a location for placing the generated and transformed reports/files.
- Certain advantages of the invention include, but are not limited to: unification of the traditional JUnit testing process with autogenerated XML to HTML report transformation; creating a portable class structure transferable into any class; applicability to any third party software testing tool, especially without having to modify class hierarchies; and “hiding” underlying complexities, from users, relating to accessing both the JUnit and Ant frameworks in order to generate useful, standardized HTML user reports. More complex reports would require enhancements to the XSL file provided by the JUnit framework.
- These and other embodiments of the present invention will be set forth in the description which follows, and in part will become apparent to those of ordinary skill in the art by reference to the following description of the invention and referenced drawings or by practice of the invention. The claims, however, indicate the particularities of the invention.
- The accompanying drawings incorporated in and forming a part of the specification, illustrate several aspects of the present invention, and together with the description serve to explain the principles of the invention. In the drawings:
-
FIG. 1 is a diagrammatic view in accordance with the present invention of a representative computing environment for a generic template to autogenerate reports for software target testing; and -
FIG. 2 is a combined diagrammatic architecture and flow chart in accordance with the present invention for auto-generating reports. - In the following detailed description of the illustrated embodiments, reference is made to the accompanying drawings that form a part hereof, and in which is shown by way of illustration, specific embodiments in which the invention may be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the invention and like numerals represent like details in the various figures. Also, it is to be understood that other embodiments may be utilized and that process, mechanical, electrical, arrangement, software and/or other changes may be made without departing from the scope of the present invention. In accordance with the present invention, methods and apparatus for a generic template to autogenerate reports for software target testing are hereinafter described.
- With reference to
FIG. 1 , arepresentative environment 10 for auto-generating reports consists of one ormore computing devices server 17, such as a grid or blade server. Alternatively, it includes a general or special purpose computing device in the form of a conventional fixed ormobile computer 17 having anattendant monitor 19 anduser interface 21. The computer internally includes a processing unit for a resident operating system, such as DOS, WINDOWS, MACINTOSH, VISTA, UNIX and LINUX, to name a few, a memory, and a bus that couples various internal and external units, e.g., other 23, to one another. Representativeother items 23 include, but are not limited to, PDA's, cameras, scanners, printers, microphones, joy sticks, game pads, satellite dishes, hand-held devices, consumer electronics, minicomputers, computer clusters, main frame computers, a message queue, a peer machine, a broadcast antenna, a web server, an AJAX client, a grid-computing node, a peer, a virtual machine, a web service endpoint, a cellular phone or the like. The other items may also be standalone computing devices 15′ in theenvironment 10 or the computing device itself, upon which the document editor is installed. - In either, storage devices are contemplated and may be remote or local. While the line is not well defined, local storage generally has a relatively quick access time and is used to store frequently accessed data, while remote storage has a much longer access time and is used to store data that is accessed less frequently. The capacity of remote storage is also typically an order of magnitude larger than the capacity of local storage. Regardless, storage is representatively provided for aspects of the invention contemplative of computer executable instructions, e.g., software, as part of computer program products on readable media, e.g.,
disk 14 for insertion in a drive ofcomputer 17. Computer executable instructions may also be available as a download or reside in hardware, firmware or combinations in any or all of the depicteddevices - When described in the context of computer program products, it is denoted that items thereof, such as modules, routines, programs, objects, components, data structures, etc., perform particular tasks or implement particular abstract data types within various structures of the computing system which cause a certain function or group of functions. In form, the computer product can be any available media, such as RAM, ROM, EEPROM, CD-ROM, DVD, or other optical disk storage devices, magnetic disk storage devices, floppy disks, or any other medium which can be used to store the items thereof and which can be assessed in the environment.
- In network, the computing devices communicate with one another via wired, wireless or combined connections 12 that are either direct 12 a or indirect 12 b. If direct, they typify connections within physical or network proximity (e.g., intranet). If indirect, they typify connections such as those found with the internet, satellites, radio transmissions, or the like, and are given nebulously as
element 13. In this regard, other contemplated items include servers, routers, peer devices, modems, T1 lines, satellites, microwave relays or the like. The connections may also be local area networks (LAN) and/or wide area networks (WAN) that are presented by way of example and not limitation. The topology is also any of a variety, such as ring, star, bridged, cascaded, meshed, or other known or hereinafter invented arrangement. - With the foregoing representative computing environment as backdrop,
FIG. 2 illustrates a high-level organization for auto-generating reports, such as in a virtual machine. Given generally aselement 40, the architecture includes a maintest script class 42 and atest fixture class 44. Within the main test script class, code in the form of a plurality of test cases (three cases shown, but any number are possible) are given as 46-1, 46-2 and 46-3. - As in the background of the invention section, Software Testing Applications are known for quality testing software targets. The targets are any of a variety of programs or applications and include, representatively, web or server applications, stand-alone computing applications, web services, computer program products, GUI interfaces, etc., to name a few. In turn, the test cases relate to or are associated with the software target and are written to examine whether features of the target are operating properly. In the background section, a stock market web service example was given. In other examples, test cases may typify a response for displaying results upon a user clicking a certain icon or pushbutton of a program. Alternatively, it may represent ascertaining determinations about mouse locations on a computing monitor whereby users navigate or “mouse-over” items. In still others, it may ascertain responses to user inputs. Naturally, an infinite number of test cases per software targets are possible and such is largely dictated by the functionality of the target and its interaction with users, other programs, other websites, etc. The precise embodiment of a test case, therefore, is merely contextual.
- In any event, it is important to the user of the software testing application to know and record the results of the test case, such as whether the test passed, failed, met a predetermined metric or other, such as including whether to throw an error message or not. Users can then use the information for purposes of troubleshooting, providing quality assurance reports, developing marketing material, improving the product, or the like. As part of the
test fixture class 44, and its interaction with the test case code 46-x (by way of an inner class template 50) the reports of the test cases are generated for the users at 52. Representatively, the reports are in an HTML computing language and result from: 1) executing the code of each of the test cases through a JUnit testing framework at 54, which, by its nature, generates reports in an XML computing language at 55; and 2) transforming or converting the generated XML reports to HTML by way of the Apache Ant API at 56. In this regard, the transformation contemplates conversion without resorting to the cumbersomeness of Ant scripting and uses existing functionality of the JUnit framework in the form of its XSL template. All of which, will be described in more detail relative to exemplary executable code of one or more working prototypes of the invention. - The interaction between the test cases 46-x and the
test fixture class 44 occurs by way of theinner class template 50. In turn, the inner class template is established by theTest Main 58. Namely, an object, such as “TestJunitScriptTest,” is named and created according to language, such as m_ScriptTest=new TestJunitScriptTest (this). It is invoked upon language comparable to m_ScriptTest. RunScript ( ). In order to know where to place the reports ultimately established at 52, and/or the reports at 55, executable code of the form m_ScriptTest.setTheReportsDirectory (m_ReportsDirectory) is provided. Thereafter, upon invocation by theTest Main 58, the inner class template is executed. All cases to be tested 46-x are executed by way of the JUnit testing framework and the Ant API, while avoiding any Ant scripting. Thereafter, the reports at 55 are generated for users. The test cases are found by all methods beginning with the prefix labeled “test.” Using the Java introspection API all test cases beginning with the prefix labeled “test” will be found. - In that other class variables and various initialization(s) need to be present, such is illustrated generically as “class variable” 60.
- At a high level, the main method (section 4 below) contains additional lines of executable code (the “m_ScriptTest” lines) that create a new instance of an
inner class 50. A parameter called the “this” variable is passed to a constructor of the inner class. This object reference instance to the maintest script class 42 is saved within the inner class. A single method call is made to set the location of where the reports are to be written and then made (e.g., m_ScriptTest. setTheReportsDirectory (m_ReportsDirectory)). The last step is to execute a line to run the script (e.g., m_ScriptTest. RunScript ( )). This causes a method to be invoked which performs introspection of the inner class instance and obtains a list of all the test case code 46-x or methods that begin with the prefix of “test.” Theinner class 50 located within the maintest script class 42 inherits methods from the TestCase, which allows execution of each test case method gathered from within the inner class. Upon method invocation of each test case which acts as the proxy method to the “real” test case code located within the main test script class is then executed. In other words, the a pass-through of sorts exists to the test case code 46-x for the ultimate execution thereof. The test case is then run thru or executed by the JUnit framework at 54 to generate the XML results at 55. This XML result file is transformed to HTML at 52 by calling the XSL transform methods located within the Ant API at 56. The HTML result files or reports are stored in the location specified above when the main test script “main” method was invoked. - Administratively, the main test script class will be found hereafter by the name “TestJunitScript” while the inner class will be found by the name “TestJunitScriptTest.” Comments in the code will be delineated by // . . . // markings or/** . . . */ markings. The term “RFT” may sometimes be used to identify a third party software testing application.
- This is the main class that encompasses the entire test script. This class is free to extend any classes required by either a third party test tool (e.g. a software testing application such as RFT) or user defined classes. No modifications are necessary to the declaration.
- Section 2. Main Test Script Variables (“Class Variables”
item 60,FIG. 2 ) - Additional variables are required to “hold” specialized data upon script initialization. Initially, they are set to null.
-
// // Variables used when creating JUnit reports private String m_ReportsDirectory = null; private String m_BuildNumber = null; private Object [ ] m_ScriptArgs = null; // // Object instance to class under test private TestJunitScriptTest m_ScriptTest = null; - This is the constructor executed when the Main Test Scrip is created or the class is loaded by the virtual machine. The “m_ReportsDirectory” is where the null setting of Section 2 is switched before executing the JUnit framework begins.
-
/** * Additional script constructor */ public TestJunitScript( ) { super( ); // // Set the build number from the specified utilities class m_BuildNumber = getDesignerUtilsRef( ).getBuildNumber( ); System.out.println(“Current build number = ” + m_BuildNumber); // // Initialize the target reports directory //String targetOS = getDesignerUtilsRef( ).getOS( ); m_ReportsDirectory = getReportsDirectory(targetOS); System.out.println(“Reports directory = ” + m_ReportsDirectory); } - Since the inner class is actually the class passed to JUnit framework in which the test cases were are to be run, a link is needed to the main test class where the actual code or “work” for the test cases is actually implemented. This is where the additional code in the TestMain method comes into play. The inner class is passed a reference using the “this” pointer to the defined software testing application (e.g., RFT) class instance. This reference is used to “call” the actual methods that do the work from within the proxy inner class being executed by the Junit framework. Once this test case is executed by JUnit a call to the method located within the RFT class is performed. This is where the “script code” should be located. Stated differently, this is the entry point for third party software testing applications.
-
public void testMain(Object[ ] args) throws TestProgramException { try { m_ScriptArgs = (Object [ ])args.clone( ); // // Run the test script thru the JUnit framework. // Test cases are not guaranteed to be executed in a // particular order since they are obtained by introspection. // If test case order of execution is required, initialize // a String [ ] array with the test case method names. // The runScript method is overloaded and will accept the // //string array. m_ScriptTest = new TestJunitScriptTest(this); m_ScriptTest.setTheReportsDirectory(m_ReportsDirectory); m_ScriptTest.setTheBuildNumber(m_BuildNumber); m_ScriptTest.runScript( ); } catch(TestProgramException tpe) { // // Log the error in RFT log logError(tpe.getMessage( ) ); throw new TestProgramException(tpe.getMessage( )); } } - The inner class provides both a proxy and a link to the JUnit framework. Skilled artisans will note that this proxy class inserts itself into the hierarchy which eventually extends the TestCase class required by JUnit.
-
/** * Standard fixture class that uses JUnit to run each testcase * and create a report using the ANT API. * The TestFixure class contains all methods to run the * JUnit testcases, collect the results, write out the xml result * document, and transform the xml data file to Html reports. */ public static class TestJunitScriptTest extends RFTTestFixture { /** * The following JUnit test case a sample template. */ public void testMySampleTestCase( ) { try { m_ScriptPtr.mySampleTestCase( ); } catch(TestProgramException tpe) { fail(tpe.getMessage( )); } } /** * Default Constructor * @param None */ public TestJunitScriptTest( ) { super( ); } /** * Constructor that has a object instance pointer from * the parent class. This ptr is used to access the parent * class methods. * @param Ojbect instance ptr to the parent class. */ public TestJunitScriptTest(TestJunitScript_scriptPtr) { super( ); m_ScriptPtr = _scriptPtr; } /** * Method executed by JUnit before the test cases are run. * Used to initialize any data that the test cases may require */ protected void setUp( ) throws Exception { } /** * Method executed by JUnit after the test cases are run. * Used to clean up any data or connections. */ protected void tearDown( ) throws Exception { } /** * Static JUnit method used to add the supplied class as * a parameter to the TestSuite constructor. */ public static Test suite( ) { return new TestSuite(TestJunitScriptTest.class); } // // Variable declarations private TestJunitScript m_ScriptPtr = null; } - This class provides the last level of abstraction via the client side public API. This class provides the wrapper mechanism that integrates into the JUnit and Ant framework. This code exists in the provided
library jar item 56,FIG. 2 in which the user does not have access to from the client source (nor do they likely desire to have access to). The “public void runScript( ) throws TestProgramException” method begins the running of the testing of cases through the JUnit framework. The “for” loop (e.g., “for (int index=0; index <m_TestCaseNames.length; index++)”) identifies and executes all the test cases through the Junit framework before generating the first of the XML reports. -
public class RFTTestFixture extends TestFixture { /** * Method to get the current directory string stored in the class * *@return Returns a string of the currently active reports directory. */ public String getTheReportsDirectory( ) { } /** * Method to set the directory string * *@param Base report directory used to write JUnit reports to. */ public void setTheReportsDirectory(String _reportsBaseDir) { } /** *Method to get the currently active build number * *@return String representing the stored build number */ public String getTheBuildNumber( ) { } /** *Method to set the build number * *@param String representing the build number */ public void setTheBuildNumber(String _buildNumber) { } /** * * Method that will execute JUnits test runner class. * The runner class will * run the class specified by the new TestSuite parameter. * * @param none * @return void * @throws TestProgramException */ public void runScript( ) throws TestProgramException { writeMessage(“Starting JUnit test(s).”); // // Set report output params to be used by the JUnit API super.setTheReportsDirectory(getTheReportsDirectory( )); super.setTheBuildNumber(getTheBuildNumber( )); super.setTestToolType(getTheAutomationToolType( )); // // Run thru all the junit test cases for (int index = 0; index < m_TestCaseNames.length; index++) { runTest(m_TestCaseNames[index]); // // If there is an error while JUnit is running the test, // try getting it from the TestFixture and log it. String testErrorMsg = getJUnitTestResultErrorMessage(index); if(testErrorMsg != null) { RationalTestScript.logError(“Step : ” + testErrorMsg); } } // // Generate JUnit xml results + reports only if there // are test cases to run. if (m_TestCaseNames.length == 0) writeMessage(“---No JUnit testcases found.---”); else { generateJUnitXMLResults( ); generateJUnitReports( ); }
((Note: this is where the Ant API is accessed to perform conversion from XML to HTML—the code will be provided after the EXAMPLE.)) -
writeMessage(“Ending JUnit test(s).”); } /** * Method that will execute JUnits test runner class. * The runner class will run the class specified *by the new TestSuite parameter. * * @param Array of test case names * @return void * @throws TestProgramException */ public void runScript(String[ ] _testCaseNames) throws TestProgramException { } } -
-
public class TestFixture extends TestCase implements XMLConstants { public String getTheBuildNumber( ) { return m_BuildNumber; } public void setTheBuildNumber(String _bldNum) { } public String getTheReportsDirectory( ) { return m_ReportsDirectory; } public void setTheReportsDirectory(String _targetReportsDir) throws TestProgramException { } public int getTestToolType( ) { return m_TestToolType; } public void setTestToolType(int _type) { m_TestToolType = _type; } public void runTest(String testName) throws TestProgramException { setTestName(testName); // // Start clock long startTime = System.currentTimeMillis( ); // // Start the test // Create and add XML Element for this test Element currentTest = doc.createElement(TESTCASE); currentTest.setAttribute(ATTR_NAME, testName); rootElement.appendChild(currentTest); // // Sets the name of the test case with Junit API setName(testName); // // Run the test thru JUnit framework TestResult result = new TestResult( ); this.run(result); } }
Section 8. Main Script Test Case Code (Item 46-x,FIG. 2 ) -
public class TestJunitScript { /** * Method is a sample test case template. */ public void mySampleTestCase( ) throws TestProgramException { // // TODO // Insert code here to do work against target application // This code can drive any type of target, gui, web, or server // based application. // This method will be called if defined within the inner test // fixture class and run thru the Junit framework. } } - Appreciating additional code is necessary to illustrate 1) the exposure to the Ant API to perform the actual conversion of XML to HTML, such as from 55 to 52, by way of 56 (
FIG. 2 ) and 2) theTest Fixture Class 44, the following is representative: -
-
// ------------------------------------------------------------------------ // // XMLToHTMLTransform.java // // Copyright ©) 2005, Novell, Inc., All Rights Reserved // Written by Phil Proto // // ------------------------------------------------------------------------ package com.novell.test.autoscript.xwb.junit.framework; import java.io.*; import java.util.StringTokenizer; import java.util.Vector; import org.apache.tools.ant.BuildException; import org.apache.tools.ant.util.DOMElementWriter; import org.apache.tools.ant.taskdefs.optional.junit.XMLResultAggregator; import org.apache.tools.ant.types.selectors.FilenameSelector; import org.apache.tools.ant.types.FileSet; import org.apache.tools.ant.Project; / / ====================================================== =============== public class XMLToHTMLTransform { public XMLToHTMLTransform( ) { } public static void main(String [ ] args) { // make sure there's an arg if (args.length < 2) { usage( ); System.exit(1); } // // Make sure directory exists if (!new File(args[0]).exists( )) { System.out.println(“The location, “ + args[0] + ”, does not exist”); usage( ); } // // Transform xml->html reports XMLToHTMLTransform transform = new XMLToHTMLTransform( ); try { transform.run(args[0], args[1]); } catch (Exception e) { e.printStackTrace( ); } } public void run(String targetDir, String sourceXMLFiles) throws Exception { File f = new File(targetDir); FileSet fs = new FileSet( ); fs.setDir(f); FilenameSelector fns = new FilenameSelector( ); fns.setName(sourceXMLFiles); fs.addFilename(fns); XMLResultAggregator xmlra = new XMLResultAggregator( ); Project prj = new Project( ); prj.setBaseDir(f); xmlra.setProject(prj); xmlra.addFileSet(fs); xmlra.setTodir(f); xmlra.createReport( ); try { xmlra.execute( ); } catch (BuildException be) { System.out.println(be.getMessage( )); System.exit(1); } } private static void usage( ) { System.out.println(“Usage: java ” + “com.novell.test.autoscript.xwb.junit.framework.- XMLToHTMLTransform” + “ <target directory> <xml source files>”); } } - The following template was developed for our third party test tool which embeds the necessary code within any named test script. When a script is created via the third party test tool the name given to the script would be substituted where necessary. This will enable any user creating a test script with a sample test case to be generated. This template frees the user to add code to test the target via the sample test case or adding any new ones using the mySampleTestCase method as a guide.
-
%script:packageDeclaration% import %helper:fullName%; import com.rational.test.ft.*; import com.rational.test.ft.object.interfaces.*; import com.rational.test.ft.script.*; import com.rational.test.ft.value.*; import com.rational.test.ft.vp.*; import junit.framework.Test; import junit.framework.TestSuite; import junit.framework.Assert; import com.novell.test.automator.framework.TestProgramException; import com.novell.test.autoscript.xwb.junit.framework.RFTTestFixture; import com.novell.qa.fw.*; /** * TODO:Brief description. This line becomes the description in the class list (like a table of contents). * * * <h2>Detailed Description:</h2> * * <p>TODO: Enter a detailed description of the test here. * * <h2>Manual Execution steps:</h2> * <ol> * <li>TODO:Do this first.</li> * <li>Now do this.</li> * <li>This is the time to do this.</li> *</ol> * * <h2>Verification Points:</h2> * * <ol> * <li>TODO: Comments concerning vps. Maybe what the point is looking for.</li> * </ol> * * <h2>Known issues</h2> * * <ul> * <li>TODO:Open script issue 1.</li> * <li>Open script issue 2.</li> * <li>Open script issue 3.</li> * </ul> * * <h2>Dependencies</h2> * * <ul> * <li>TODO:Dependency 1.</li> * </ul> * * <p><b>Product:</b> TODO:XYZWidget Calculator v0.3</b> * <br><b>PRD Reference: none</b> * <br><b>Created:</b> %date% %time% * < b r > < b > O r i g i n a l H o s t : < / b > %static:com.rational.test.ft.sys.OperatingSystem.getVersion% * * * @since %date:yyyy/MM/dd% * @author %system:user.name% * */ public class %script:name% extends %helper:name% { /** * Method is a sample test case template. */ public void mySampleTestCase( ) throws TestProgramException { %script:insertBefore% this entire line will be deleted during initial script generation } /** * Entry point for RFT class loader * @param Object array used to pass arguments from callScript method */ public void testMain(Object[ ] args) throws TestProgramException { try { m_ScriptArgs = (Object [ ])args.clone( ); // // Run the test script thru the JUnit framework. // Test cases are not guaranteed to be executed in a // particular order since they are obtained by introspection. // If test case order of execution is required, initialize // a String [ ] array with the test case method names. // The runScript method is overloaded and will accept the string array. m_ScriptTest = new %script:name%Test(this); m_ScriptTest.setTheReportsDirectory(m_ReportsDirectory); m_ScriptTest.setTheBuildNumber(m_BuildNumber); m_ScriptTest.runScript( ); } catch(TestProgramException tpe) { // // Log the error in RFT log logError(tpe.getMessage( ) ); throw new TestProgramException(tpe.getMessage( )); } } /** * Standard fixture class that uses JUnit to run each testcase * and create a report using the ANT API. * The TestFixure class contains all methods to run the * JUnit testcases, collect the results, write out the xml result * document, and transform the xml data file to Html reports. */ public static class %script:name%Test extends RFTTestFixture { /** * The following JUnit test case a sample template. */ public void testMySampleTestCase( ) { try { m_ScriptPtr.mySampleTestCase( ); } catch(TestProgramException tpe) { fail(tpe.getMessage( )); } } /** * Default Constructor * @param None */ public %script:name%Test( ) { super( ); } /** * Constructor that has a object instance pointer from * the parent class. This ptr is used to access the parent * class methods. * @param Ojbect instance ptr to the parent class. */ public %script:name%Test(%script:name% _scriptPtr) { super( ); m_ScriptPtr = _scriptPtr; } /** * Method executed by JUnit before the test cases are run. * Used to initialize any data that the test cases may require */ protected void setUp( ) throws Exception { } /** * Method executed by JUnit after the test cases are run. * Used to clean up any data or connections. */ protected void tearDown( ) throws Exception { } /** * Static JUnit method used to add the supplied class as * a parameter to the TestSuite constructor. */ public static Test suite( ) { return new TestSuite(%script:name%Test.class); } // // Variable declarations private %script:name% m_ScriptPtr = null; } /** * Default script constructor */ public %script:name%( ) { super( ); // // Set the build number from the specified utilities class m_BuildNumber = getDesignerUtilsRef( ).getBuildNumber( ); System.out.println(“Current build number = ” + m_BuildNumber); // // Initialize the reports directory String targetOS = getDesignerUtilsRef( ).getOS( ); m_ReportsDirectory = getReportsDirectory(targetOS); System.out.println(“Reports directory = ” + m_ReportsDirectory); } // // Variables used when creating JUnit reports private String m_ReportsDirectory = null; private String m_BuildNumber = null; private Object [ ] m_ScriptArgs = null; // // Object instance to class under test private %script:name%Test m_ScriptTest = null; } - Certain advantages of the invention over the prior art should now be readily apparent. For example, there is now provided a software testing application that easily and conveniently generates user reports in a useful format. There is also a hidden aspect of the functionality making the application convenient to users. Retrofitting to existing software testing applications also exists as does taking advantage of existing testing frameworks, such as JUnit and Ant. Of course, these are only a few of the many advantages of the invention and skilled artisans will immediately recognize others.
- Finally, one of ordinary skill in the art will recognize that additional embodiments are also possible without departing from the teachings of the present invention. This detailed description, and particularly the specific details of the exemplary embodiments disclosed herein, is given primarily for clarity of understanding, and no unnecessary limitations are to be implied, for modifications will become obvious to those skilled in the art upon reading this disclosure and may be made without departing from the spirit or scope of the invention. Relatively apparent modifications, of course, include combining the various features of one or more figures with the features of one or more of other figures or expanding the system to replicate the embodiments multiple times.
Claims (22)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/879,716 US7725772B2 (en) | 2007-07-18 | 2007-07-18 | Generic template to autogenerate reports for software target testing |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/879,716 US7725772B2 (en) | 2007-07-18 | 2007-07-18 | Generic template to autogenerate reports for software target testing |
Publications (2)
Publication Number | Publication Date |
---|---|
US20090024874A1 true US20090024874A1 (en) | 2009-01-22 |
US7725772B2 US7725772B2 (en) | 2010-05-25 |
Family
ID=40265834
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/879,716 Expired - Fee Related US7725772B2 (en) | 2007-07-18 | 2007-07-18 | Generic template to autogenerate reports for software target testing |
Country Status (1)
Country | Link |
---|---|
US (1) | US7725772B2 (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090300423A1 (en) * | 2008-05-28 | 2009-12-03 | James Michael Ferris | Systems and methods for software test management in cloud-based network |
US20110209121A1 (en) * | 2010-02-24 | 2011-08-25 | Salesforce.Com, Inc. | System, method and computer program product for providing automated testing by utilizing a preconfigured point of entry in a test or by converting a test to a predefined format |
CN102455913A (en) * | 2010-10-20 | 2012-05-16 | 微软公司 | Customization of display templates |
CN102855175A (en) * | 2011-06-28 | 2013-01-02 | 北京新媒传信科技有限公司 | Automated testing method and automated testing device |
US8667333B2 (en) | 2010-06-01 | 2014-03-04 | The United States Of America As Represented By The Secretary Of The Navy | Extensible testing system |
US20140245070A1 (en) * | 2013-02-27 | 2014-08-28 | International Business Machines Corporation | Automated execution of functional test scripts on a remote system within a unit testing framework |
US9201763B1 (en) * | 2013-05-31 | 2015-12-01 | The Mathworks, Inc. | Efficient sharing of test fixtures and ordering of test suites |
CN105117630A (en) * | 2015-08-21 | 2015-12-02 | 宇龙计算机通信科技(深圳)有限公司 | Fingerprint authentication method, fingerprint authentication apparatus, and terminal |
CN105260309A (en) * | 2015-10-21 | 2016-01-20 | 上海斐讯数据通信技术有限公司 | Automatic test platform unrelated to test tool |
CN105760298A (en) * | 2016-02-16 | 2016-07-13 | 上海斐讯数据通信技术有限公司 | Third party testing service platform |
CN106776344A (en) * | 2017-01-16 | 2017-05-31 | 上海聚宝网络科技股份有限公司 | A kind of automated testing method of the intelligent express delivery cabinet based on APPIUM |
CN108776642A (en) * | 2018-06-01 | 2018-11-09 | 平安普惠企业管理有限公司 | Test report generation method, device, computer equipment and storage medium |
CN110287115A (en) * | 2019-06-26 | 2019-09-27 | 北京金山云网络技术有限公司 | Generation method, device and the server of test report |
CN111008150A (en) * | 2019-12-23 | 2020-04-14 | 杭州迪普科技股份有限公司 | Test report generation method, device and equipment |
US11481295B2 (en) * | 2017-02-10 | 2022-10-25 | Optofidelity Oy | Method, an all-in-one tester and computer program product |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102339219B (en) * | 2010-07-20 | 2016-08-24 | 甲骨文国际公司 | For supporting the system and method for object-oriented wscript.exe |
US20120192158A1 (en) | 2010-11-22 | 2012-07-26 | Carlo Amalfitano | Model Based Verification Using Forward and Reverse Traversal of Variable Time Line |
US9514031B2 (en) | 2014-09-22 | 2016-12-06 | International Business Machines Corporation | Auto-deployment and testing of system application test cases in remote server environments |
CN108241580B (en) * | 2016-12-30 | 2021-11-19 | 深圳壹账通智能科技有限公司 | Client program testing method and terminal |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5737518A (en) * | 1996-07-31 | 1998-04-07 | Novell, Inc. | Method and apparatus for testing an object management system |
US5909544A (en) * | 1995-08-23 | 1999-06-01 | Novell Inc. | Automated test harness |
US6002869A (en) * | 1997-02-26 | 1999-12-14 | Novell, Inc. | System and method for automatically testing software programs |
US6263376B1 (en) * | 1997-02-24 | 2001-07-17 | Novell, Inc. | Generic run-time binding interpreter |
US6411974B1 (en) * | 1998-02-04 | 2002-06-25 | Novell, Inc. | Method to collate and extract desired contents from heterogeneous text-data streams |
US20050015675A1 (en) * | 2003-07-03 | 2005-01-20 | Kolawa Adam K. | Method and system for automatic error prevention for computer software |
US20050114834A1 (en) * | 2003-11-26 | 2005-05-26 | International Business Machines Corporation | Grid-enabled ANT compatible with both stand-alone and grid-based computing systems |
US20060048100A1 (en) * | 2004-07-16 | 2006-03-02 | International Business Machines Corporation | System and method for software product test modularization |
US20060085681A1 (en) * | 2004-10-15 | 2006-04-20 | Jeffrey Feldstein | Automatic model-based testing |
US20060230320A1 (en) * | 2005-04-07 | 2006-10-12 | Salvador Roman S | System and method for unit test generation |
US20060248405A1 (en) * | 2005-03-21 | 2006-11-02 | Ponczak Joseph M | Method for automating unit test development |
US20070061625A1 (en) * | 2005-09-15 | 2007-03-15 | Acosta Juan Jr | Automation structure for software verification testing |
US20070061113A1 (en) * | 2005-08-25 | 2007-03-15 | International Business Machines Corporation | Enabling Test Script Play Back in Different Locales |
US20070094541A1 (en) * | 2005-10-25 | 2007-04-26 | Hua Kang | Method and apparatus for generating test execution sequences automatically for a software testing process |
US7526681B2 (en) * | 2006-08-07 | 2009-04-28 | Sap Portals Israel Ltd. | Software testing framework |
-
2007
- 2007-07-18 US US11/879,716 patent/US7725772B2/en not_active Expired - Fee Related
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5909544A (en) * | 1995-08-23 | 1999-06-01 | Novell Inc. | Automated test harness |
US5737518A (en) * | 1996-07-31 | 1998-04-07 | Novell, Inc. | Method and apparatus for testing an object management system |
US6263376B1 (en) * | 1997-02-24 | 2001-07-17 | Novell, Inc. | Generic run-time binding interpreter |
US6002869A (en) * | 1997-02-26 | 1999-12-14 | Novell, Inc. | System and method for automatically testing software programs |
US6411974B1 (en) * | 1998-02-04 | 2002-06-25 | Novell, Inc. | Method to collate and extract desired contents from heterogeneous text-data streams |
US20050015675A1 (en) * | 2003-07-03 | 2005-01-20 | Kolawa Adam K. | Method and system for automatic error prevention for computer software |
US20050114834A1 (en) * | 2003-11-26 | 2005-05-26 | International Business Machines Corporation | Grid-enabled ANT compatible with both stand-alone and grid-based computing systems |
US20060048100A1 (en) * | 2004-07-16 | 2006-03-02 | International Business Machines Corporation | System and method for software product test modularization |
US20060085681A1 (en) * | 2004-10-15 | 2006-04-20 | Jeffrey Feldstein | Automatic model-based testing |
US20060248405A1 (en) * | 2005-03-21 | 2006-11-02 | Ponczak Joseph M | Method for automating unit test development |
US20060230320A1 (en) * | 2005-04-07 | 2006-10-12 | Salvador Roman S | System and method for unit test generation |
US20070061113A1 (en) * | 2005-08-25 | 2007-03-15 | International Business Machines Corporation | Enabling Test Script Play Back in Different Locales |
US20070061625A1 (en) * | 2005-09-15 | 2007-03-15 | Acosta Juan Jr | Automation structure for software verification testing |
US20070094541A1 (en) * | 2005-10-25 | 2007-04-26 | Hua Kang | Method and apparatus for generating test execution sequences automatically for a software testing process |
US7526681B2 (en) * | 2006-08-07 | 2009-04-28 | Sap Portals Israel Ltd. | Software testing framework |
Cited By (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090300423A1 (en) * | 2008-05-28 | 2009-12-03 | James Michael Ferris | Systems and methods for software test management in cloud-based network |
US8732663B2 (en) * | 2010-02-24 | 2014-05-20 | Salesforce.Com, Inc. | System, method and computer program product for providing automated testing by utilizing a preconfigured point of entry in a test or by converting a test to a predefined format |
US20110209121A1 (en) * | 2010-02-24 | 2011-08-25 | Salesforce.Com, Inc. | System, method and computer program product for providing automated testing by utilizing a preconfigured point of entry in a test or by converting a test to a predefined format |
US8667333B2 (en) | 2010-06-01 | 2014-03-04 | The United States Of America As Represented By The Secretary Of The Navy | Extensible testing system |
US8688795B2 (en) | 2010-06-01 | 2014-04-01 | The United States Of America As Represented By The Secretary Of The Navy | GPS embedded interactive network interface |
US8855961B2 (en) | 2010-06-01 | 2014-10-07 | United States Of America As Represented By The Secretary Of The Navy | Binary definition files |
US9322872B2 (en) | 2010-06-01 | 2016-04-26 | The United States Of America As Represented By The Secretary Of The Navy | Correlated testing system |
US10817516B2 (en) | 2010-10-20 | 2020-10-27 | Microsoft Technology Licensing, Llc | Result types for conditional data display |
US9652545B2 (en) | 2010-10-20 | 2017-05-16 | Microsoft Technology Licensing, Llc | Result types for conditional data display |
US9135358B2 (en) | 2010-10-20 | 2015-09-15 | Microsoft Technology Licensing, Llc | Result types for conditional data display |
US10210260B2 (en) | 2010-10-20 | 2019-02-19 | Microsoft Technology Licensing, Llc | Templates for displaying data |
CN102455913A (en) * | 2010-10-20 | 2012-05-16 | 微软公司 | Customization of display templates |
CN102855175A (en) * | 2011-06-28 | 2013-01-02 | 北京新媒传信科技有限公司 | Automated testing method and automated testing device |
US20140245070A1 (en) * | 2013-02-27 | 2014-08-28 | International Business Machines Corporation | Automated execution of functional test scripts on a remote system within a unit testing framework |
US20150324276A1 (en) * | 2013-02-27 | 2015-11-12 | International Business Machines Corporation | Automated execution of functional test scripts on a remote system within a unit testing framework |
US9135150B2 (en) * | 2013-02-27 | 2015-09-15 | International Business Machines Corporation | Automated execution of functional test scripts on a remote system within a unit testing framework |
US9886375B2 (en) * | 2013-02-27 | 2018-02-06 | International Business Machines Corporation | Automated execution of functional test scripts on a remote system within a unit testing framework |
US9201763B1 (en) * | 2013-05-31 | 2015-12-01 | The Mathworks, Inc. | Efficient sharing of test fixtures and ordering of test suites |
CN105117630A (en) * | 2015-08-21 | 2015-12-02 | 宇龙计算机通信科技(深圳)有限公司 | Fingerprint authentication method, fingerprint authentication apparatus, and terminal |
CN105260309A (en) * | 2015-10-21 | 2016-01-20 | 上海斐讯数据通信技术有限公司 | Automatic test platform unrelated to test tool |
CN105760298A (en) * | 2016-02-16 | 2016-07-13 | 上海斐讯数据通信技术有限公司 | Third party testing service platform |
CN106776344A (en) * | 2017-01-16 | 2017-05-31 | 上海聚宝网络科技股份有限公司 | A kind of automated testing method of the intelligent express delivery cabinet based on APPIUM |
US11481295B2 (en) * | 2017-02-10 | 2022-10-25 | Optofidelity Oy | Method, an all-in-one tester and computer program product |
CN108776642A (en) * | 2018-06-01 | 2018-11-09 | 平安普惠企业管理有限公司 | Test report generation method, device, computer equipment and storage medium |
CN110287115A (en) * | 2019-06-26 | 2019-09-27 | 北京金山云网络技术有限公司 | Generation method, device and the server of test report |
CN111008150A (en) * | 2019-12-23 | 2020-04-14 | 杭州迪普科技股份有限公司 | Test report generation method, device and equipment |
Also Published As
Publication number | Publication date |
---|---|
US7725772B2 (en) | 2010-05-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7725772B2 (en) | Generic template to autogenerate reports for software target testing | |
US8739190B2 (en) | Determining an extension to use to process an input object to a call in a program | |
US9361211B2 (en) | Automated generation of test cases for regression testing | |
US7526681B2 (en) | Software testing framework | |
US7222333B1 (en) | Techniques for generating software application build scripts based on tags in comments | |
US8640104B2 (en) | Computer method and apparatus for debugging in a dynamic computer language | |
US7287247B2 (en) | Instrumenting a software application that includes distributed object technology | |
Zhang et al. | Refactoring middleware with aspects | |
US7676806B2 (en) | Deployment, maintenance and configuration of complex hardware and software systems | |
CN106663002B (en) | REST service source code generation | |
US10209968B2 (en) | Application compiling | |
US8166347B2 (en) | Automatic testing for dynamic applications | |
Beaton et al. | Usability challenges for enterprise service-oriented architecture APIs | |
US20050039172A1 (en) | Synthesizing application response measurement (ARM) instrumentation | |
US8935705B2 (en) | Execution of highly concurrent processing tasks based on the updated dependency data structure at run-time | |
US11113050B2 (en) | Application architecture generation | |
Okanović et al. | Towards performance tooling interoperability: An open format for representing execution traces | |
US20080010545A1 (en) | Computer system and method for monitoring execution of application program | |
US20170322817A1 (en) | Object-oriented programming system and library | |
EP3447635A1 (en) | Application architecture generation | |
CN113568839A (en) | Method, device, equipment and medium for software testing and statistical test coverage rate | |
US8812556B2 (en) | Storing modification data for recreating modifications | |
US7836449B2 (en) | Extensible infrastructure for task display and launch | |
Ansaloni et al. | Rapid development of extensible profilers for the Java virtual machine with aspect-oriented programming | |
Lui et al. | A generalized approach to real-time, non-intrusive instrumentation and monitoring of standards-based distributed middleware |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: NOVELL, INC., UTAH Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PROTO, PHILIP J.;REEL/FRAME:019644/0118 Effective date: 20070717 Owner name: NOVELL, INC.,UTAH Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PROTO, PHILIP J.;REEL/FRAME:019644/0118 Effective date: 20070717 |
|
AS | Assignment |
Owner name: CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, NEW YORK Free format text: GRANT OF PATENT SECURITY INTEREST;ASSIGNOR:NOVELL, INC.;REEL/FRAME:026270/0001 Effective date: 20110427 |
|
AS | Assignment |
Owner name: CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, NEW YORK Free format text: GRANT OF PATENT SECURITY INTEREST (SECOND LIEN);ASSIGNOR:NOVELL, INC.;REEL/FRAME:026275/0018 Effective date: 20110427 |
|
AS | Assignment |
Owner name: NOVELL, INC., UTAH Free format text: RELEASE OF SECURITY IN PATENTS SECOND LIEN (RELEASES RF 026275/0018 AND 027290/0983);ASSIGNOR:CREDIT SUISSE AG, AS COLLATERAL AGENT;REEL/FRAME:028252/0154 Effective date: 20120522 Owner name: NOVELL, INC., UTAH Free format text: RELEASE OF SECURITY INTEREST IN PATENTS FIRST LIEN (RELEASES RF 026270/0001 AND 027289/0727);ASSIGNOR:CREDIT SUISSE AG, AS COLLATERAL AGENT;REEL/FRAME:028252/0077 Effective date: 20120522 |
|
AS | Assignment |
Owner name: CREDIT SUISSE AG, AS COLLATERAL AGENT, NEW YORK Free format text: GRANT OF PATENT SECURITY INTEREST SECOND LIEN;ASSIGNOR:NOVELL, INC.;REEL/FRAME:028252/0316 Effective date: 20120522 Owner name: CREDIT SUISSE AG, AS COLLATERAL AGENT, NEW YORK Free format text: GRANT OF PATENT SECURITY INTEREST FIRST LIEN;ASSIGNOR:NOVELL, INC.;REEL/FRAME:028252/0216 Effective date: 20120522 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: NOVELL, INC., UTAH Free format text: RELEASE OF SECURITY INTEREST RECORDED AT REEL/FRAME 028252/0316;ASSIGNOR:CREDIT SUISSE AG;REEL/FRAME:034469/0057 Effective date: 20141120 Owner name: NOVELL, INC., UTAH Free format text: RELEASE OF SECURITY INTEREST RECORDED AT REEL/FRAME 028252/0216;ASSIGNOR:CREDIT SUISSE AG;REEL/FRAME:034470/0680 Effective date: 20141120 |
|
AS | Assignment |
Owner name: BANK OF AMERICA, N.A., CALIFORNIA Free format text: SECURITY INTEREST;ASSIGNORS:MICRO FOCUS (US), INC.;BORLAND SOFTWARE CORPORATION;ATTACHMATE CORPORATION;AND OTHERS;REEL/FRAME:035656/0251 Effective date: 20141120 |
|
AS | Assignment |
Owner name: MICRO FOCUS SOFTWARE INC., DELAWARE Free format text: CHANGE OF NAME;ASSIGNOR:NOVELL, INC.;REEL/FRAME:040020/0703 Effective date: 20160718 |
|
AS | Assignment |
Owner name: JPMORGAN CHASE BANK, N.A., AS SUCCESSOR AGENT, NEW Free format text: NOTICE OF SUCCESSION OF AGENCY;ASSIGNOR:BANK OF AMERICA, N.A., AS PRIOR AGENT;REEL/FRAME:042388/0386 Effective date: 20170501 |
|
AS | Assignment |
Owner name: JPMORGAN CHASE BANK, N.A., DELAWARE Free format text: SECURITY INTEREST;ASSIGNORS:ATTACHMATE CORPORATION;BORLAND SOFTWARE CORPORATION;NETIQ CORPORATION;AND OTHERS;REEL/FRAME:044183/0718 Effective date: 20170901 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.) |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.) |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20180525 |
|
AS | Assignment |
Owner name: JPMORGAN CHASE BANK, N.A., AS SUCCESSOR AGENT, NEW Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE TO CORRECT TYPO IN APPLICATION NUMBER 10708121 WHICH SHOULD BE 10708021 PREVIOUSLY RECORDED ON REEL 042388 FRAME 0386. ASSIGNOR(S) HEREBY CONFIRMS THE NOTICE OF SUCCESSION OF AGENCY;ASSIGNOR:BANK OF AMERICA, N.A., AS PRIOR AGENT;REEL/FRAME:048793/0832 Effective date: 20170501 |
|
AS | Assignment |
Owner name: NETIQ CORPORATION, WASHINGTON Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: MICRO FOCUS SOFTWARE INC. (F/K/A NOVELL, INC.), WASHINGTON Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: ATTACHMATE CORPORATION, WASHINGTON Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: SERENA SOFTWARE, INC, CALIFORNIA Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: MICRO FOCUS (US), INC., MARYLAND Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: BORLAND SOFTWARE CORPORATION, MARYLAND Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: MICRO FOCUS LLC (F/K/A ENTIT SOFTWARE LLC), CALIFORNIA Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: MICRO FOCUS SOFTWARE INC. (F/K/A NOVELL, INC.), WASHINGTON Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 035656/0251;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062623/0009 Effective date: 20230131 Owner name: MICRO FOCUS (US), INC., MARYLAND Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 035656/0251;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062623/0009 Effective date: 20230131 Owner name: NETIQ CORPORATION, WASHINGTON Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 035656/0251;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062623/0009 Effective date: 20230131 Owner name: ATTACHMATE CORPORATION, WASHINGTON Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 035656/0251;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062623/0009 Effective date: 20230131 Owner name: BORLAND SOFTWARE CORPORATION, MARYLAND Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 035656/0251;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062623/0009 Effective date: 20230131 |