An airline's frequent flyer's club awards benefits depending on which tier a customer is in. The software must determine which tier to allocate a customer to based on an input value of total Credits Earned to date
Customers initially join the Silver Tier and remain in that tier for the first 400 Credits Earned The next 400 Credits Earned moves the customer into the Gold Tier. The next 600 Credits Earned moves the customer into the Platinum Tier Further Credits Earned moves the customer into the Concierge Tier.
Test Cases have been written with the following total Credits Earned input values:
TC1 -400 Credits
TC2 - 500 Credits
TC3 - 800 Credits
TC4-1500 Credits
Applying the Equivalence Partitioning test design technique, what percentage of valid Equivalence Partitions have these 4 test cases collectively achieved?
Equivalence Partitioning is a black-box test design technique that divides input data of a software module into partitions of equivalent data from which test cases can be derived. In this context, the valid equivalence partitions are:
Silver Tier: 0 to 400 Credits
Gold Tier: 401 to 800 Credits
Platinum Tier: 801 to 1400 Credits
Concierge Tier: 1401+ Credits
The test cases provided cover all these partitions:
TC1 covers the Silver Tier boundary at 400 Credits.
TC2 covers within the Gold Tier at 500 Credits.
TC3 covers the Gold Tier boundary at 800 Credits.
TC4 covers within the Concierge Tier at 1500 Credits.
Since all valid partitions are covered by the test cases, 100% of the valid Equivalence Partitions have been achieved.
Which of the types of test tools noted below BEST describes tools that support reviews?
Tools for document review and annotation
Tools for document comparison and version control
Tools for checklist management and defect tracking
Tools for recording user actions and feedback
Tools for analyzing user behavior and satisfaction
Tools for simulating different user interfaces and devices
Tools for data validation and verification
Tools for data cleansing and transformation
Tools for data profiling and analysis
Tools for security testing
Tools for performance testing
Tools for accessibility testing
What one of the following would be a typical objective of running a pilot project when introducing a new tool into an organisation?
Gaining in-depth knowledge about the tool, understanding both its strengths and weaknesses
Evaluating how the tool fits with existing processes and practices, and determining what would need to change
Assessing whether the benefits will be achieved at reasonable cost
Identifying and resolving any technical and organizational issues before deploying the tool on a larger scale
Providing training, coaching, and mentoring for users of the tool
Defining metrics to measure the success of the tool deployment
The other options are not typical objectives of running a pilot project, because they are either part of the tool selection process, or the outcomes of the pilot project. They are:
Which of the types of test tools noted below BEST describes tools that support reviews?
Tools for document review and annotation
Tools for document comparison and version control
Tools for checklist management and defect tracking
Tools for recording user actions and feedback
Tools for analyzing user behavior and satisfaction
Tools for simulating different user interfaces and devices
Tools for data validation and verification
Tools for data cleansing and transformation
Tools for data profiling and analysis
Tools for security testing
Tools for performance testing
Tools for accessibility testing
Currently there are no comments in this discussion, be the first to comment!