Module 46 - T203 Part 2 of 2

T203 Part 2 of 2: How to Develop Test Cases for an ITS Standards-Based Test Plan, Part 2 of 2

HTML of the PowerPoint Presentation

(Note: This document has been converted from a PowerPoint presentation to 508-compliant HTML. The formatting has been adjusted for 508 compliance, but all the original text content is included, plus additional text descriptions for the images, photos and/or diagrams have been provided below.)

 

Slide 1:

Welcome - Graphic image of introductory slide. Please see the Extended Text Description below.

(Extended Text Description: Welcome - Graphic image of introductory slide. A large dark blue rectangle with a wide, light grid pattern at the top half and bands of dark and lighter blue bands below. There is a white square ITS logo box with words "Standards ITS Training" in green and blue on the middle left side. The word "Welcome" in white is to the right of the logo. Under the logo box is the logo for the U.S. Department of Transpotation, Office of the Assistant Secretary for Research and Technology.)

 

Slide 2:

Welcome slide with Ken Leonard and screen capture of home webpage. Please see the Extended Text Description below.

(Extended Text Description: This slide, entitled "Welcome" has a photo of Ken Leonard, Director, ITS Joint Program Office, on the left hand side, with his email address, Ken.Leonard@dot.gov. A screen capture snapshot of the home webpage is found on the right hand side - for illustration only - from August 2014. Below this image is a link to the current website: www.pcb.its.dot.gov - this screen capture snapshot shows an example from the Office of the Assistant Secretary for Research and Development - Intelligent Transportation Systems Joint Program Office - ITS Professional Capacity Building Program/Advanced ITS Education. Below the main site banner, it shows the main navigation menu with the following items: About, ITS Training, Knowledge Exchange, Technology Transfer, ITS in Academics, and Media Library. Below the main navigation menu, the page shows various content of the website, including a graphic image of professionals seated in a room during a training program. A text overlay has the text Welcome to ITS Professional Capacity Building. Additional content on the page includes a box entitled What's New and a section labeled Free Training. Again, this image serves for illustration only. The current website link is: http://www.pcb.its.dot.gov.)

 

Slide 3:

Activity. A placeholder graphic with an image of hand over a computer keyboard to show that an activity is taking place.

 

Slide 4:

T203 Part 2 of 2: How to Develop Test Cases for an ITS Standards-Based Test Plan, Part 2 of 2

 

Slide 5:

Instructor

Headshot photo of Manny Insignares Vice President, Technology Consensus Systems Technologies New York, NY, US

Manny Insignares

Vice President, Technology Consensus Systems Technologies

New York, NY, USA

 

Slide 6:

Target Audience

 

Slide 7:

Recommended Prerequisite(s)

 

Slide 8:

Curriculum Path

Curriculum Path. Please see the Extended Text Description below.

(Extended Text Description: Curriculum Path: A graphical illustration indicating the sequence of training modules that lead up to and follow each course. Each module is represented by a box with the name of the module in it and an arrow showing the logical flow of the modules and the current module highlighted. This slide focuses on the modules that lead up to the current course. The first box is labeled "T101 Introduction to ITS Standards Testing" followed by a line that connects to a box labeled "T201 How to Write a Test Plan" followed by a line that connects to a box labeled "T202 Overview of Test Design Specifications, Test Cases and Test Procedures." The T202 Module is connected to a box labeled "T203 Part 1 of 2: How to Develop Test Cases for an ITS Standards-based Test Plan, Part 1 of 2." This box, in turn connects to a box, colored to indicate that it represents this module, labeled "T203 Part 2 of 2: How to Develop Test Cases for an ITS Standards-based Test Plan, Part 2 of 2.")

 

Slide 9:

Learning Objectives

Part 1 of 2:

  1. Review the role of test cases within the overall testing process.
  2. Discuss ITS data structures used in NTCIP and center-to-center standards (TMDD), and provide examples.
  3. Learn to find information needed to develop a test case.
  4. Understand test case development.

Part 2 of 2:

  1. Handle standards that are with and without test documentation.
  2. Develop a Requirements to Test Case Traceability Matrix (RTCTM).
  3. Identify types of testing.
  4. Recognize the purpose of the test log and test anomaly report.

 

Slide 10:

Brief Review of Part 1 Discussion

  1. We have discussed the role of test cases within the overall testing process and test plan.
  2. We have discussed ITS data structures used in NTCIP and center-to-center standards (TMDD).
  3. We have learned about the test case development process.
  4. We have learned where to find information needed for test case preparation.

Let's re-cap IEEE 829 and the relationship between test cases and test procedures, then continue with the remaining four learning objectives 5-8.

 

Slide 11:

What does IEEE Std 829 Provide?

This slide shows the cover page of the IEEE Standard for Software and System Test Documentation - 829.

 

Slide 12:

Testing Documentation Structure (IEEE Std 829)

This slide shows a diagram with test documentation in a layered diagram. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows a diagram with test documentation in a layered diagram. The layers from top to bottom are labeled: Test Plan, Test Design Specification, Test Case Specification, Test Procedure Specification, Test Execution, Test Reports. Lines connect to show a one-for-one relationship of test design specifications to test plans, multiple test cases to a test design, multiple test cases to a test procedure. A single test is executed based on the test documentation resulting in test log, test anomaly, and test reports. The slide shows the definitions of Test Plan - describes the overall approach to testing, Test Design Specification - describes which requirements are to be tested and associated test cases, and Test Case Specification - identifies objectives, inputs, outcomes, and conditions for execution of a test. The diagram to the right has the following content: Test documentation is shown in a layered diagram. Three boxes made of dashed lines run from top to bottom along the left side. The first box contains the text – Test Plan describes the overall approach to testing. The second box contains the text – Test Design Specification describes which requirements are to be tested and associated test cases. The third box contains the text – Text Case Specification identifies objectives and inputs, outcomes, and conditions for execution of a test. Each box points to its corresponding place on the diagram to the right. The layers from top to bottom are labeled: Test Plan, Test Design Specification, Test Case Specification, Test Procedure Specification, Test Execution, Test Reports. The Test Plan layer contains a large dashed rectangle surrounding four other rectangles. Each rectangle is labeled differently. They are, from left to right, Unit Test, Integration Test, System Acceptance Test, and Periodic Maintenance Test. The Test Design Specification Layer contains four boxes, each connecting to a corresponding box on the previous layer. They are labeled, from left to right, Test Design Unit Test, Test Design Integ. Test, Test Design Sys. Acceptance, and Test Design Periodic Mtce. The Test Case Specification layer contains five boxes that are labelled Test Case 001, Test Case 002, Test Case 003, Test Case 004, and Test Case N. Each Test Case is connected to multiple boxes in the previous layer. Test Case 001 is connected to Test Design Unit Test and Test Design Integ. Test. Test Case 002 is connected to Test Design Unit Test, Test Design Integ Test, and Test Design Sys. Acceptance. Test Case 003 is connected to Test Design Integ Test and Test Design Sys. Acceptance. Test Case 004 is connected to Test Design Sys. Acceptance and Test Design Periodic Mtce. Test Case N is connected to Test Design Periodic Mtce. The Test Procedure Specification layer contains two rectangles containing the text Test Procedure 001 and Test Procedure 002. Test Procedure 001 links back to Test Cases 001 – 003. Test procedure 002 links back to Test Case 004 and N. The Text Execution layer has a single dashed oval labeled Test Plan Execution (Process) and Is linked back to both Test Procedures. The final layer is Test Reports and contains three text boxes. The top two are labeled Test Logs and Test Incident Reports and they connect back to Test Plan Execution. These boxes also lead down to the final box via arrows and this final box is labeled Test Plan Execution Summary Report.)

 

Slide 13:

Testing Documentation Structure (cont.)

This slide shows the same diagram as in the previous slide. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows the same diagram as in the previous slide. The slide shows definitions of Test Procedure Specification - defines the steps to execute a test, Multiple test cases may reference a single test procedure, Test procedures may be more costly to develop than test cases, and Test Reports: Test Log, Test Anomoly Reports, Test Report.)

 

Slide 14:

Learning Objective #5

Learning Objective #5: Handle Standards That Are With and Without Test Documentation

 

Slide 15:

Learning Objective #5

About ITS Standards

 

Slide 16:

Learning Objective #5

Status of C2F and C2C Standards Content

Standard SE Content (ConOps, User Needs, Requirements, Design) Testing Content (Testing Plan, Test Cases)
NTCIP 1202 ASC No (being added) No
NTCIP 1203 DMS Yes Yes
NTCIP 1204 ESS Yes Yes
NTCIP 1205 CCTV No No
NTCIP 1206 DCM No No
NTCIP 1207 RM No No
NTCIP 1209 TSS Yes No
NTCIP 1210 FMS Yes No
NTCIP 1211 SCP Yes No
NTCIP 1213 ELMS Yes No
TMDD v03 (C2C) Yes No

 

Slide 17:

Learning Objective #5

Preparation for a Test Case

 

Slide 18:

Learning Objective #5

Step-by-Step Process to Follow Through Case Studies

  1. Identify user needs
  2. Identify requirements
  3. Develop test case objective
  4. Identify design content: dialogs, inputs, and outputs
  5. Document value constraints of inputs and outputs
  6. Complete test case

 

Slide 19:

Learning Objective #5

Overview of Case Studies

  1. 1. NTCIP 1205 CCTV (C2F)
  2. 2. NTCIP 1204 ESS (C2F)
  3. 3. TMDD (C2C)

This slide shows a graphic image with System Under Test (SUT). Please see the Extended Text Description below.

(Extended Text Description: This slide shows a graphic image with System Under Test (SUT), with a building connected with bidirectional arrows to the Communications Cloud, with bidirectional arrows connected to another building - External Center - and the Communications Cloud connected to a computer labeled Test Software.)

 

Slide 20:

Case Study 1 Slide: This slide contains a graphic with the words Case Study 1 in large letters. A placeholder graphic of a control center and staff at their stations indicating a Case Study follows.

 

Slide 21:

Learning Objective #5

Case Study 1: NTCIP 1205 CCTV

Background

  1. Characteristics of the standard: NTCIP 1205 CCTV
  2. Information sources for this test case

 

Slide 22:

Learning Objective #5

Case Study CCTV User Needs

User Needs and Requirements are the basis for a system interface design; test cases verify each requirement.

A graphic image shows elements of control for a CCTV System. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: A graphic image shows elements of control for a CCTV System, including: pan-tilt-zoom, focus control, iris control, preset position selection, and image labeling.)

This slide shows a Protocol Requirements List (PRL) developed in the A317a course module. Please see the Extended Text Description below.

(Extended Text Description: This slide shows a Protocol Requirements List (PRL) developed in the A317a course module. Step 1: Identify User Needs. The table content is shown below:

UN ID User Need Req ID Requirement Additional Specs.
3.0 Remote Monitoring 3.3.3 Status condition within the device  
3.3.3.2 Temperature  
3.3.3.2 Pressure  
3.3.3.2 Washer fluid  
3.3.3.3 ID Generator  

Below the table are arrows pointing to User Needs and Requirements.)

 

Slide 23:

Learning Objective #5

Case Study CCTV Requirements

Case Study CCTV Requirements. Please see the Extended Text Description below.

(Extended Text Description: The table content is shown below:

Req ID Requirement Dialog Object Reference and Title NTCIP 1205 Section 3
3.3.3 Status condition within the device D.1 Generic SNMP GET Interface
3.3.3.2 Temperature 3.7.5 alarmTemperatureCurrentValue
3.3.3.2 Pressure 3.7.6 alarmPressureHighLowThreshold
3.7.7 alarmPressureCurrentValue
3.3.3.2 Washer fluid 3.7.8 alarmWasherFluidHighLowThreshold
3.7.9 alarmWasherCurrentValue
3.3.3.3 ID Generator 3.11 cctv label Objects

Below the table are arrows pointing to Requirements and Design Content. The text "We need to look up value constraints for these objects" is to the right of the arrows.)

Step 2: Identify Requirements

 

Slide 24:

Dialog Example for 3.3 Status Condition within the Device

This slide includes a graphics display of a center-to-field communications dialog. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide includes a graphics display of a center-to-field communications dialog. The top of the graphic shows a management station at left and a CCTV camera at right. The image illustrates sending an object identifier (in this case cctv alarm 5) from the management station to the CCTV camera is a request. The camera responds by returning the object's value (in this case alarm temperature current value) to the management station.)

 

Slide 25:

Learning Objective #5

CCTV Test Case Objective and Purpose

Test Case
ID:TC001 Title: Request Status Condition within the Device Dialog Verification (Positive Test Case)
Objective: To verify system interface implements (positive test case) requirements for a sequence of OBJECT requests for:
  • 3.7.5 alarmTemperatureCurrentValue
  • 3.7.6 alarmPressureHighLowThreshold
  • 3.7.7 alarmPressureCurrentValue
  • 3.7.8 alarmWasherFluidHighLowThreshold
  • 3.7.9 alarmWasherCurrentValue
  • 3.11 cctv label Objects
The test case verifies that the data value of the OBJECTs requested are within specified ranges. The object identifier (OID) of each object requested is the only input required. An output specification is provided to show valid value constraints per the NTCIP 1205 v01 object definitions.
Inputs:  
Outcome(s):  
Environmental Needs:  
Tester/Reviewer  
Special Procedure Requirements:  
Intercase Dependencies:  

Objective

 

Slide 26:

Learning Objective #5

CCTV Test Case Output Specification: Data Concept ID, Data Concept Name, and Data Concept Type Filled in

Test Case Output Specification
ID: TCOS001 Title: Status Condition within the Device
Data Concept ID Data Concept Name (Variable) Data Concept Type Value Constraints
3.7.5 alarmTemperatureCurrentValue Data Element  
3.7.6 alarmPressureHighLowThreshold Data Element  
3.7.7 alarmPressureCurrentValue Data Element  
3.7.8 alarmWasherFluidHighLowThresh old Data Element  
3.7.9 alarmWasherCurrentValue Data Element  
3.11 cctv label Objects Data Element  

No dialogs are documented in NTCIP 1205 v01. These are simple GET/SET operations, consisting of an object request followed by the result that we will document in an output specification.

Step 4: Identify Dialogs, Inputs, Outputs

 

Slide 27:

Learning Objective #5

Object Definition for:

3.7.5 alarmTemperatureCurrentValue

This slide shows an object definition for the temperature alarm current value parameter. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows an object definition for the temperature alarm current value parameter. The image highlights key information contained in the object definition relevant to the development of the test case – namely, value constraints and object identifier.)

Step 4: Identify Dialogs, Inputs, Outputs

 

Slide 28:

Learning Objective #5

Object Definition for:

3.7.8 alarmWasherFluidHighLowThreshold

This slide shows an object definition for the washer fluid alarm high-low threshold parameter. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows an object definition for the washer fluid alarm high-low threshold parameter. The image highlights key information contained in the object definition relevant to the development of the test case – namely, value constraints and object identifier. .)

 

Step 4: Identify Dialogs, Inputs, Outputs

 

Slide 29:

Learning Objective #5

CCTV Test Case Output Specification

Test Case Output Specification
ID: TCOS001 Title: Status Condition within the Device
Data Concept ID Data Concept Name (Variable) Data Concept Type Value Domain
3.7.5 alarmTemperatureCurrentValue Data Element OCTET STRING (SIZE(1)) - Range: 0 to 255.
3.7.6 alarmPressureHighLowThreshold Data Element OCTET STRING (SIZE(2)) - Range: 0 to 255 each byte Note: Byte 1 Low Threshold denotes the value of minimum pressure within the camera enclosure measured in psig, Byte2 High Threshold denotes the value of maximum pressure within the camera enclosure measured in psig
3.7.7 alarmPressureCurrentValue Data Element OCTET STRING (SIZE(1)) - Range: 0 to 255
3.7.8 alarmWasherFluidHighLowThreshold Data Element OCTET STRING (SIZE(2)) - Range: 0 to 100 each byte. Note: Byte1 Low Threshold denotes the percentage of minimum filled capacity between zero (0) and 100 percent, Byte2 HighThreshold denotes the percentage of maximum filled capacity between zero (0) and 100 percent.
3.7.9 alarmWasherCurrentValue Data Element OCTET STRING (SIZE(1)) - Range: 0 to 100
3.11 cctv label Objects Data Element etc. 3.11 contains numerous object definition entries.

Step 5: Document Value Constraints for Inputs, Outputs

 

Slide 30:

Learning Objective #5

CCTV Test Case

Test Case
ID: TC001 Title: Request Status Condition within the Device Dialog Verification (Positive Test Case)
Objective: To verify system interface implements (positive test case) requirements for a sequence of OBJECT requests for:
  • 3.7.5 alarmTemperatureCurrentValue
  • 3.7.6 alarmPressureHighLowThreshold
  • 3.7.7 alarmPressureCurrentValue
  • 3.7.8 alarmWasherFluidHighLowThreshold
  • 3.7.9 alarmWasherCurrentValue
  • 3.11 cctv label Objects
The test case verifies that the data value of the OBJECTs requested are within specified ranges. The object identifier (OID) of each object requested is the only input required. An output specification is provided to show valid value constraints per the NTCIP 1205 v01 object definitions.
Inputs: The object identifier of each object requested is needed.
Outcome(s): All data are returned and verified as correct per the OBJECT constraints of NTCIP 1205 v01. See Test Case Output Specification TCOS001 - Status Condition within the Device (Positive Test Case)
Environmental Needs: When testing for alarm temperature current value, set up is needed to measure temperature.
Tester/Reviewer M.I.
Special Procedure Requirements: None
Intercase Dependencies: None

Step 6: Complete Test Case

 

Slide 31:

Learning Objective #5

Summary of CCTV Case Study

 

Slide 32:

Learning Objective #5

What have we learned from this case study?

We have learned that the NTCIP 1205 CCTV Standard does NOT have user needs (PRL) and requirements (RTM) and we must derive content from the PCB Module A317b CCTV and then proceed with test case development.

Using CCTV design objects, we have learned to identify and document value constraints in a test case.

MODULE 34: A317b: UNDERSTANDING REQUIREMENTS FOR CCTV SYSTEMS

BASED ON NTCIP 1205 STANDARD

(Source: http://www.pcb.its.dot.gov/StandardsTraining)

 

Slide 33:

Case Study 2 Slide: This slide contains a graphic with the words Case Study 2 in large letters. A placeholder graphic of a control center and staff at their stations indicating a Case Study follows.

 

Slide 34:

Learning Objective #5

Case Study 2: NTCIP 1204 v03 ESS

Background

 

Slide 35:

Learning Objective #5

Case Study ESS User Needs

Source: NTCIP 1204 v03 Protocol Requirement List (PRL).

This slide shows a portion of the Protocol Requirements List (PRL) from NTCIP 1204 v03 ESS standard. Please see the Extended Text Description below.

(Extended Text Description: This slide shows a portion of the Protocol Requirements List (PRL) from NTCIP 1204 v03 ESS standard. The image highlights that user needs are on the left-hand side, and that requirements are on the right-hand side of the table.

User Need ID User Need FR ID Functional Requirement Conformance Project Requirement Additional Project Requirements
2.5.2.1.2 (Wind) Monitor Winds O.3 (1..*) Yes / No / NA  
    3.5.2.3.2.2 Retrieve Wind Data M Yes / NA  
    3.6.2 Required Number of Wind Sensors M Yes / NA The ESS shall support at least ___ wind sensors.

)

Step 1: Identify User Needs

 

Slide 36:

Learning Objective #5

Case Study ESS Requirements

This slide shows a portion of the Requirements Traceability Matrix (RTM) from NTCIP 1204 v03 ESS standard. Please see the Extended Text Description below.

(Extended Text Description: This slide shows a portion of the Requirements Traceability Matrix (RTM) from NTCIP 1204 v03 ESS standard. The image highlights that requirements are on the left-hand side, and that design content is on the right-hand side of the table.

Req ID Dialog Requirement Object ID Add'l Requirements/Object
3.5.2.3.2.2 F.4.6 Retrieve Wind Data  
      5.6.8 windSensorTableNumSensors
      5.6.10.1 windSensorIndex
      5.6.10.4 windSensorAvgSpeed
      5.6.10.5 windSensorAvgDirection
      5.6.10.6 windSensorSpotSpeed
      5.6.10.7 windSensorSpotDirection
      5.6.10.8 windSensorGustSpeed
      5.6.10.9 windSensorGustDirection
      5.6.10.10 windSensorSituation

)

Step 2: Identify Requirements

 

Slide 37:

Learning Objective #5

Discussion: Case Study ESS RTCTM

This slide shows a portion of the Requirements to Test Case Traceability Matrix (RTM) from NTCIP 1204 v03 ESS standard. Please see the Extended Text Description below.

(Extended Text Description: This slide shows a portion of the Requirements to Test Case Traceability Matrix (RTM) from NTCIP 1204 v03 ESS standard. The image highlights that requirements are on the left-hand side, and that test cases used to verify the requirements are on the right-hand side of the table.

Requirement Test Case
ID Title ID Title
3.5.2.3.2 Monitor Weather Condition
3.5.2.3.2.1 Retrieve Atmospheric Pressure
    C.2.3.3.2 Retrieve Atmospheric Pressure
3.5.2.3.2.2 Retrieve Wind Data
    C.2.3.3.3 Retrieve Wind Data
3.5.2.3.2.3 Retrieve Temperature
    C.2.3.3.4 Retrieve Temperature
3.5.2.3.2.4 Retrieve Daily Minimum and Maximum Temperature
    C.2.3.3.5 Retrieve Daily Minimum and Maximum Temperature
3.5.2.3.2.5 Retrieve Humidity
    C.2.3.3.6 Retrieve Humidity

)

 

Discussion

 

Slide 38:

Learning Objective #5

Discussion: Case Study ESS Test Procedure

C.2.3.3.3 Retrieve Wind Data

Test Case: 3.3 Title: Retrieve Wind Data
Description: This test case verifies thai the ESS allows a management station to determine turner?? wind infonnaiion.
Variables: Required_Wind_Sensors PRL 3.6 2
Pass/Fail Criteria: The device under test (DUT) shall pass every verification step included within the Test Case in order to pass the Test Case.
Step Test Procedure Device
1 CONFIGURE: Determine the number of wind sensors required by the specificalion (PRL 3.6.2). RECORD this information as: Required_Wind_Sensors  
2 GET the fo1lowing object(s): windSensorTableNumSensors.0 Pass/Fail (Clause 3.5.2.3.2.2)
3 VERIFY that the RESPONSE VALUE for windSensorTableNumSensors.0 is greater than or equal to Required_Wind_Sensors. Pass/Fail (Clause 3.6.2)
4 Determine the RESPONSE VALUE for windSensorTableNumSensors.0. RECORD this information as: Supported_Wind_Sensors  
5 FOR EACH value, N, from 1 to Supported_Wind_Sensors, perform Steps 5.1 through 5.22.  
5.1 GET the following object(s):
  • windSensorAvgSpeed.N
  • windSensorAvgDirection.N
  • windSensorSpotSpeed.N
  • windSensorSpotDirection.N
  • windSensorGustSpeed.N
  • windSensorGustDirection.N
  • windSensorSituation.N
Pass/Fail (Clause 3.5.2.3.2.2)
5.2 VERIFY that the RESPONSE VALUE for wind SensorAvgSpeed.N is greater than or equal to O. Pass/Fail (Clause 5.6.10.4)

Discussion

 

Slide 39:

Learning Objective #5

Discussion: ESS Standard's SE and Test Content Makes the Implementers' Job Easier

Discussion

 

Slide 40:

Learning Objective #5

ESS Test Case Objective and Purpose

Test Case
ID:TC0012 Title: Retrieve Wind Data (Positive Test Case)
Objective: To verify system interface implements (positive test case) requirements for a sequence of OBJECT requests for:
  • 5.6.8 windSensorTableNumSensors
  • 5.6.10.1 windSensorIndex
  • 5.6.10.4 windSensorAvgSpeed
  • 5.6.10.5 windSensorAvgDirection
  • 5.6.10.6 windSensorSpotSpeed
  • 5.6.10.7 windSensorSpotDirection
  • 5.6.10.8 windSensorGustSpeed
  • 5.6.10.9 windSensorGustDirection
  • 5.6.10.10 windSensorSituation
The test case verifies that the data value of the OBJECTs requested are within specified ranges. The object identifier (OID) of each object requested is the only input required. An output specification is provided to show the valid value constraints per the NTCIP 1204 v03 object definitions.
Inputs:  
Outcome(s):  
Environmental Needs:  
Tester/Reviewer  
Special Procedure Requirements:  
Intercase Dependencies:  

Step 3: Develop Test Case Objective

 

Slide 41:

Learning Objective #5

ESS Test Case Output Specification: With Data Concept ID, Data Concept Name, and Data Concept Type Filled in

Test Case Output Specification
ID: TCOS022 Title: Wind Data
Data Concept ID Data Concept Name (Variable) Data Concept Type Value Domain
5.6.8 windSensorTableNumSensors Data Element  
5.6.10.1 windSensorIndex Data Element  
5.6.10.4 windSensorAvgSpeed Data Element  
5.6.10.5 windSensorAvgDirection Data Element  
5.6.10.6 windSensorSpotSpeed Data Element  
5.6.10.7 windSensorSpotDirection Data Element  
5.6.10.8 windSensorGustSpeed Data Element  
5.6.10.9 windSensorGustDirection Data Element  
5.6.10.10 windSensorSituation Data Element  

Step 4: Identify Dialogs, Inputs, Outputs

 

Slide 42:

Learning Objective #5

Object Definition for: 5.6.10.4 windSensorAvgSpeed

Object Definition for: 5.6.10.4 windSensorAvgSpeed. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows an object definition for wind sensor average speed. The image highlights key information contained in the object definition relevant to the development of the test case – namely, value constraints and object identifier.)

 

Slide 43:

Learning Objective #5

Object Definition for: 5.6.10.10 windSensorSituation

Object Definition for: 5.6.10.10 windSensorSituation. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows an object definition for wind sensor situation. The image highlights key information contained in the object definition relevant to the development of the test case – namely, value constraints and object identifier.)

 

Step 4: Identify Dialogs, Inputs, Outputs

 

Slide 44:

Learning Objective #5

ESS Test Case Output Specification (1 of 3)

Test Case Output Specification
ID: TCOS022 Title: Wind Data
Data Concept ID Data Concept Name (Variable) Data Concept Type Value Constraints
5.6.8 windSensorTableNumSensors Data Element INTEGER (0..255)
5.6.10.1 windSensorIndex Data Element INTEGER (1..255)
5.6.10.4 windSensorAvgSpeed Data Element INTEGER (0..65535) tenths of meters per second WMO Binary Code Form FM 94 BUFR Table B item 0 11 002
5.6.10.5 windSensorAvgDirection Data Element INTEGER (0..361) The value of zero (0) shall indicate 'calm', when the associated speed is zero (0), or 'light and variable,' when the associated speed is greater than zero (0). Normal observations, as defined by the WMO, shall report a wind direction in the range of 1 to 360 with 90 meaning from the east and 360 meaning from the north. The value of 361 shall indicate an error condition and shall always be reported if the associated speed indicates error.
5.6.10.6 windSensorSpotSpeed Data Element INTEGER (0..65535) tenths of meters per second The value of 65535 shall indicate an error condition or missing value.

Step 5: Document Value Constraints for Inputs, Outputs

 

Slide 45:

Learning Objective #5

ESS Test Case Output Specification (2 of 3)

Test Case Output Specification
ID: TCOS022 Title: Wind Data
Data Concept ID Data Concept Name (Variable) Data Concept Type Value Constraints
5.6.10.7 windSensorSpotDirection Data Element INTEGER (0..361) The value of zero (0) shall indicate 'calm', when the associated speed is zero (0), or 'light and variable,' when the associated speed is greater than zero (0). Normal observations, as defined by the WMO, shall report a wind direction in the range of 1 to 360 with 90 meaning from the east and 360 meaning from the north. The value of 361 shall indicate an error condition and shall always be reported if the associated speed indicates error.
5.6.10.8 windSensorGustSpeed Data Element INTEGER (0..65535) tenths of meters per second WMO Code Form FM 94 BUFR Table B item 0 11 041. The value of 65535 shall indicate an error condition or missing value.
5.6.10.9 windSensorGustDirection Data Element INTEGER (0..361) (See 5.6.10.7)

Step 5: Document Value Constraints for Inputs, Outputs

 

Slide 46:

Learning Objective #5

ESS Test Case Output Specification (3 of 3)

Test Case Output Specification
ID: TCOS022 Title: Wind Data
Data Concept ID Data Concept Name (Variable) Data Concept Type Value Domain
5.6.10.10 windSensorSituation Data Element INTEGER other (1), unknown (2), calm (3), lightBreeze (4), moderateBreeze (5), strongBreeze (6), gale (7), moderateGale (8), strongGale (9), stormWinds (10), hurricaneForceWinds (11), gustyWinds (12) (See object definition for additional detail.)

Step 5: Document Value Constraints for Inputs, Outputs

 

Slide 47:

Learning Objective #5

ESS Test Case

Test Case
ID: TC0012 Title: Retrieve Wind Data (Positive Test Case)
Objective: To verify system interface implements (positive test case) requirements for a sequence of OBJECT requests for:
  • 5.6.8 windSensorTableNumSensors
  • 5.6.10.1 windSensorIndex
  • 5.6.10.4 windSensorAvgSpeed
  • 5.6.10.5 windSensorAvgDirection
  • 5.6.10.6 windSensorSpotSpeed
  • 5.6.10.7 windSensorSpotDirection
  • 5.6.10.8 windSensorGustSpeed
  • 5.6.10.9 windSensorGustDirection
  • 5.6.10.10 windSensorSituation
The test case verifies that the data value of the OBJECTs requestedare within specified ranges. The object identifier (OID) of each object requested is the only input required. An output specification is provided to show the valid value constraints per the NTCIP 1204 v03 object definitions.
Inputs: The object identifier of each object requested.
Outcome(s): All data are returned and verified as correct per the OBJECT constraints of NTCIP 1204 v03. See Test Case Output Specification TCOS022 - Wind Data.
Environmental Needs: When testing for average wind speed, an artificial wind device is needed to provide the wind for the sensor to measure.
Tester/Reviewer M.I.
Special Procedure Requirements: Wind simulator set-up. (See test procedures.)
Intercase Dependencies: None

Step 6: Complete Test Case

 

Slide 48:

Learning Objective #5

Summary of the ESS Case Study

 

Slide 49:

Learning Objective #5

What Have We Learned from This Case Study?

We have learned that the NTCIP 1204 ESS standard has SEP content : user needs (PRL)_ and requirements (RTM) and testing documentation which we can review for the test cases development.

Using ESS design objects, we have learned to identify and document value constraints and develop a test case.

MODULE 18: T313: Applying your test plan to NTCIP 1204 v03 ESS standard (Source: http://www.pcb.its.dot.gov/StandardsTraining)

 

Slide 50:

Activity. A placeholder graphic with an image of hand over a computer keyboard to show that an activity is taking place.

 

Slide 51:

Learning Objective #5

Which of the following standards provides testing content?

Answer Choices

  1. TMDD v03 C2C Standard
  2. NTCIP 1204 v03 ESS Standard
  3. NTCIP 1205 v01 CCTV Standard
  4. All of the above

 

Slide 52:

Learning Objective #5

Review of Answers

A small graphical red and yellow X representing incorrect.a) TMDD v03 C2C Standard
Incorrect. This standard only provides NRTM, and no test documentation.

A small graphical green and yellow check mark representing correct.b) NTCIP 1204 v03 ESS Standard
Correct! The NTCIP ESS standard has SE and testing content.

A small graphical red and yellow X representing incorrect.c) NTCIP 1205 v01 CCTV Standard
Incorrect. This standard only provides design elements (objects) and no SE content and no test documentation. Our case study 1 explained this.

A small graphical red and yellow X representing incorrect.d) All of the above
Incorrect. Only certain ITS standards were developed using SE content and only two have test documentation-DMS and ESS.

 

Slide 53:

Case Study 3 Slide: This slide contains a graphic with the words Case Study 3 in large letters. A placeholder graphic of a control center and staff at their stations indicating a Case Study follows.

 

Slide 54:

Learning Objective #5

Case Study 3: TMDD (C2C)

Background

  1. Characteristics of the TMDD v3.03c Standard
  2. Information Sources
  3. User Needs and Requirements are contained in the NRTM in Volume I
  4. Relevant Dialogs, Input and Output Definitions (Data Concepts) and RTM are contained in Volume II

 

Slide 55:

Learning Objective #5

Background: Context Diagram

This slide includes an image of a center-to-center communications dialog. Please see the Extended Text Description below.

(Extended Text Description: This slide includes an image of a center-to-center communications dialog. The top of the graphic shows an external station at left and an owner center at right. This image illustrates sending a link status request message from the external center to owner center as a request. The owner center responds by returning a link status message to the external center.)

 

Slide 56:

Learning Objective #5

Case Study TMDD Needs

Case Study TMDD Needs. Please see the Extended Text Description below.

(Extended Text Description: This slide shows a portion of the Needs to Requirements Traceability Matrix (NRTM) from the TMDD standard. The image highlights that user needs are on the left-hand side, and that requirements are on the right-hand side of the table. The table contains the following content:

UN ID User Need   Requirement ID Requirement Conformance Support Other Requirements
2.3.4.2.2 Need to Share Link State Optional [Yes] / No  
    Dialogs
      3.3.4.3.2.1 Send Link Status Information Upon Request M [Yes] The owner center shall respond within ___ (100ms - 1 hour; Default = 1 minute) after receiving the request. See Section 3.4.2.
      Publication and Subscription Dialog not shown.      
    Request Message
      3.3.4.1.1 Contents of the Traffic Network Information Request M [Yes]  
      3.3.4.1.1.1 Required Traffic Network Information Request Content M [Yes]  
      3.3.4.1.1.2.1 Authentication - Network (AuthNetwork) O [Yes] / No  
      3.3.4.1.1.2.1.1 Operator Identifier -Network AuthNetwork:O [Yes] / No / NA  
      3.3.4.3.2.4 Contents of the Link Status Request M [Yes]  
    Response Message
      3.3.4.3.2.5 Contents of the Link Status Information M [Yes]  
      3.3.4.3.2.5.1 Required Link Status Information Content M [Yes]  
      3.3.4.3.2.5.2.1 Restrictions - Link Status O [Yes] / No  
      3.3.4.3.2.5.2.2 Link Name - Link Status O [Yes] / No  
      3.3.4.3.2.5.2.3 Link Direction - Link Status O [Yes] / No  
      3.3.4.3.2.5.2.4 Lanes Open O [Yes] / No  
      3.3.4.3.2.5.2.19 Roadway Event Source O Yes / [No]  
      3.3.4.3.2.5.2.37 Event Description Time -Link Status O Yes / [No]  
      3.3.4.3.2.5.2.38 Link Status Date and Time Change Information O [Yes] / No  
    Error Report Message (detail not shown)

)

 

Slide 57:

Learning Objective #5

Case Study TMDD Requirements

A portion of the RTM is shown, see student supplement for a full view.

This slide shows a portion of the Requirements Traceability Matrix (RTM) from the TMDD standard. Please see the Extended Text Description below.

(Extended Text Description: This slide shows a portion of the Requirements Traceability Matrix (RTM) from the TMDD standard. The image highlights that requirements are on the left-hand side, and that design content is on the right-hand side of the table. The table contains the following content:

Req ID (Vol. I) Requirement Dialog DC Type Definition Class Name DC ID (Vol. II) Data Concept Instance Name
3.3.4.3.2.1 Send Link Status Information Upon Request 2.4.1 dialog dlLinkStatusRequest 3.1.13.2 dlLinkStatusRequest
3.3.4.3.2.2 Publish Link Status Information 2.4.3 dialog dlLinkStatusUpdate 3.1.34.2 dlLinkStatusUpdate
3.3.4.3.2.3 Subscribe to Link Status Information 2.4.2 dialog dlTrafficNetworkInformationSubscription 3.1.19.1 dlTrafficNetworkInformationSubscription
3.3.4.3.2.4 Contents of the Link Status Request message trafficNetworkInformationRequestMsg 3.2.19.1 trafficNetworkInformationRequestMsg
3.3.4.3.2.5 Contents of the Link Status Information message linkStatusMsg 3.2.13.2 linkStatusMsg
3.3.4.3.2.5.1 Required Link Status Information Content data-frame organizationInformation 3.3.16.3 organization-information
3.3.4.3.2.5.1 Required Link Status Information Content data-element transportation-network-identifier 3.4.20.1 network-id
3.3.4.3.2.5.1 Required Link Status Information Content data-element transportation-network-identifier 3.4.20.1 link-id
3.3.4.3.2.5.1 Required Link Status Information Content data-element link-status 3.4.14.34 link-status
3.3.4.3.2.5.2.1 Restrictions - Link Status data-frame restrictions 3.3.16.5 restrictions
3.3.4.3.2.5.2.2 Link Name - Link Status data-element transportation-network-name 3.4.21.1 link-name
3.3.4.3.2.5.2.3 Link Direction - Link Status data-element link-direction 3.4.14.9 link-direction
3.3.4.3.2.5.2.4 Lanes Open data-element link-lanes-count 3.4.14.12 lanes-number-open
3.3.4.3.2.5.2.5 Link Priority data-element link-priority-type 3.4.14.21 priority-type
3.3.4.3.2.5.2.6 Link Restrictions - Axles data-element link-restriction-axle-count 3.4.14.22 restriction-axle-count
3.3.4.3.2.5.2.7 Link Restrictions - Height data-element link-restriction-height 3.4.14.23 restriction-height
3.3.4.3.2.5.2.8 Link Restrictions - Length data-element link-restriction-length 3.4.14.24 restriction-length

 

)

Step 2: Identify Requirements

 

Slide 58:

Learning Objective #5

TMDD Test Case Objective and Purpose

Test Case
ID:TC001 Title: Link Status Request-Response Dialog Verification (Positive Test Case)
Objective: To verify system interface implements (positive test case) requirements for:
  1. Link Status Request-Response Dialog message exchange
  2. Contents of the Link Status Request Message
  3. Contents of the Link Status Information Message
The test case verifies that the dialog, request message content, and response message content are correct by sending a request message (verified to be correct) across the system interface, and verification that the response message is correct. Input and output specifications are provided to verify the request and response message are correct per the requirements for the request and response message.
Inputs:  
Outcome(s):  
Environmental Needs:  
Tester/Reviewer  
Special Procedure Requirements:  
Intercase Dependencies:  

Step 3: Develop Test Case Objective

 

Slide 59:

Learning Objective #5

Data Concept Definition for:

3.3.16.3 organizationInformation (1 of 2)

3.3.16.3 organizationInformation (1 of 2). Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows a data concept definition for organization information. The image highlights key information contained in the data concept definition relevant to the development of the test case – namely, data concept type (data frame), and for data frames data structure of the data frame.)

Step 4: Identify Dialogs, Inputs, Outputs

 

Slide 60:

Learning Objective #5

Data Concept Definition for:

3.3.16.3 organizationInformation (2 of 2)

3.3.16.3 organizationInformation (2 of 2). Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide is a continuation of the previous slide showing valid sequence of data concepts that make up the data frame.)

Step 4: Identify Dialogs, Inputs, Outputs

 

Slide 61:

Learning Objective #5

Data Concept Definition for: 3.4.14.34 link-status

Data Concept Definition for: 3.4.14.34 link-status. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows a data concept definition for link status. This images highlights key information contained in the data concept definition relevant to the development of the test case – namely, data concept type (data element), and for data elements the value constraints for data.)

Step 4: Identify Dialogs, Inputs, Outputs

 

Slide 62:

Learning Objective #5

TMDD Test Case Input Specification

Test Case Input Specification
ID: TCIS001 Title: Input Specification for Link Status Information Request (Positive Test Case)
Data Concept ID Data Concept Name (Variable) Data Concept Type Value Constraints
3.2.19.1 trafficNetworklnformationReq uestMsg Message  
3.3.16.3 organization-requesting Data Frame  
3.4.16.8 organization-id Data Element IA5String (SIZE(1..32))
3.4.16.9 organization-name Data Element IA5String (SIZE(1..128))
3.4.20.2 network-information-type Data Element 1 = "node inventory" 2 = "node status" 3 = "link inventory" 4 = "link status" 5 = "route inventory" 6 = "route status" 7 = "network inventory"

Step 5: Document Value Constraints for Inputs, Outputs

 

Slide 63:

Learning Objective #5

TMDD Test Case Output Specification

Test Case Output Specification
ID: TCOS001 Title: Output Specification for Link Status Information Request (Positive Test Case)
Data Concept ID Data Concept Name (Variable) Data Concept Type Value Constraints
3.2.13.2 linkStatusMsg Message  
3.3.16.3 organization-information Data Frame  
3.4.16.8 organization-id Data Element IA5String (SIZE(1..32))
3.4.16.9 organization-name Data Element IA5String (SIZE(1..128))
3.4.20.1 network-id Data Element IA5String (SIZE(1..32))
3.4.20.1 link-id Data Element IA5String (SIZE(1..32))
3.4.21.1 link-name Data Element IA5String (SIZE(1..128))
3.4.14.34 link-status Data Element 1 = "no determination" 2 = "open" 3 = "restricted" 4 = "closed"
3.4.14.37 travel-time Data Element INTEGER (0..65535), units=seconds

Step 5: Document Value Constraints for Inputs. Outputs

 

Slide 64:

Learning Objective #5

TMDD Test Case

Test Case
ID: TC001 Title: Link Status Request-Response Dialog Verification (Positive Test Case)
Objective: To verify system interface implements (positive test case) requirements for:
  1. Link Status Request-Response Dialog message exchange
  2. Contents of the Link Status Request Message
  3. Contents of the Link Status Information Message
The test case verifies that the dialog, request message content, and response message content are correct by sending a request message (verified to be correct) across the system interface, and verification that the response message is correct. Input and output specifications are provided to verify the request and response message are correct per the requirements for the request and response message.
Inputs: Use the input file linkStatusRequest_pos.xml. See Test Case Input Specification TCIS001 - LinkStatusRequest (Positive Test Case). Set network-information-type to 4 or the text "link status".
Outcome(s): All data are returned and verified as correct: correct sequence of message exchanges, structure of data, and valid value of data content. See Test Case Output Specification TCOS001 -LinkStatusInformation (Positive Test Case)
Environmental Needs: No additional needs outside of those specified in the test plan.
Tester/Reviewer M.I.
Special Procedure Requirements: None
Intercase Dependencies: None

Step 6: Complete Test Case

 

Slide 65:

Learning Objective #5

Summary of the TMDD Case Study

 

Slide 66:

Learning Objective #5

Summary of Learning Objective #5

Handle Standards That Are With and Without Test Documentation

 

Slide 67:

Learning Objective #6

Learning Objective #6: Develop a Requirements to Test Case Traceability Matrix (RTCTM)

 

Slide 68:

Learning Objective #6

How Does RTCTM fit in a Test Coverage?

 

Slide 69:

Learning Objective #6

Format of an RTCTM

Format of an RTCTM. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows a portion of the Requirements to Test Case Traceability Matrix (RTM) from NTCIP 1204 v03 ESS standard. The image highlights that requirements (Requirement, ID and Title are taken from the project PRL (C2F) ro NRTM (C2C)) are on the left-hand side, and that test cases used to verify the requirements (Unique Test Case ID and Title are assigned by the user) are on the right-hand side of the table. Each Test Case verifies whether a stated requirement is implemented in a working system properly. The table contains the following content:

Requirement Test Case
ID Title ID Title
    C.2.3.2.8 Configure Passive Ice Detection Logic
3.5.2.1.9 Configure Snapshot Camera
    C.2.3.2.9 Configure Snapshot Camera

)

 

Slide 70:

Learning Objective #6

How Each Requirement is Handled

Example:

3.5.2.1.9 Configure Snapshot Camera (From ESS PRL)

Upon request, the ESS shall store a textual description of the location to which the camera points and the filename to be used when storing new snapshots.

Relevant Objects for Value constraints (from ESS RTM)

RTCTM connects a Requirement ID to a Test Case ID:

Configure Snapshot Camera (From ESS PRL). Please see the Extended Text Description below.

(Extended Text Description: The image contains a table with the following content:

3.5.2.1.9 Configure Snapshot Camera
    C.2.3.2.9 Configure Snapshot Camera

)

 

Slide 71:

Learning Objective #6

Example of RTCTM

Source: NTCIP 1204 v03 ESS Standard

Requirement Test Case
ID Title ID Title
    C.2.3.2.8 Configure Passive Ice Detection Logic
3.5.2.1.9 Configure Snapshot Camera
    C.2.3.2.9 Configure Snapshot Camera
3.5.2.3 Sensor Data Retrieval Requirements
3.5.2.3.1 Retrieve Weather Profile with Mobile Sources
    C.2.3.3.1 Retrieve Weather Profile with Mobile Sources
3.5.2.3.2 Monitor Weather Condition
3.5.2.3.2.1 Retrieve Atmospheric Pressure
    C.2.3.3.2 Retrieve Atmospheric Pressure
3.5.2.3.2.2 Retrieve Wind Data
    C.2.3.3.3 Retrieve Wind Data
3.5.2.3.2.3 Retrieve Temperature
    C.2.3.3.4 Retrieve Temperature
3.5.2.3.2.4 Retrieve Daily Minimum and Maximum Temperature
    C.2.3.3.5 Retrieve Daily Minimum and Maximum Temperature
3.5.2.3.2.5 Retrieve Humidity
    C.2.3.3.6 Retrieve Humidity
3.5.2.3.2.6 Monitor Precipitation
3.5.2.3.2.6.1 Retrieve Precipitation Presence

 

Slide 72:

Learning Objective #6

Let's Build an RTCTM for the CCTV Case Study

 

Slide 73:

Learning Objective #6

Build RTCTM for CCTV Case Study Requirements

This slide shows a portion of the RTM from the CCTV Case Study. Please see the Extended Text Description below.

(Extended Text Description: Author's relevant notes: This slide shows a portion of the RTM from the CCTV Case Study. The image highlights that the requirements we'll use to develop the RTCTM are on the left-hand side, and that design content at right may be ignored. The table contains the following content:

Req ID Requirement Dialog Object Reference and Title NTCIP 1205 Section 3
3.3.3 Status condition within the device D.1 Generic SNMP GET Interface
3.3.3.2 Temperature 3.7.5 alarmTemperatureCurrentValue
3.3.3.2 Pressure 3.7.6 alarmPressureHighLowThreshold
3.7.7 alarmPressureCurrentValue
3.3.3.2 Washer fluid 3.7.8 alarmWasherFluidHighLowThreshold
3.7.9 alarmWasherCurrentValue
3.3.3.3 ID Generator 3.11 cctv label Objects

)

 

Slide 74:

Learning Objective #6

Build RTCTM for CCTV Case Study Test Case

Test Case
ID:TC001 Title: Request Status Condition within the Device Dialog Verification (Positive Test Case)
Objective: To verify system interface implements (positive test case) requirements for a sequence of OBJECT requests for:
  • 3.7.5 alarmTemperatureCurrentValue
  • 3.7.6 alarmPressureHighLowThreshold
  • 3.7.7 alarmPressureCurrentValue
  • 3.7.8 alarmWasherFluidHighLowThreshold
  • 3.7.9 alarmWasherCurrentValue
  • 3.11 cctv label Objects
The test case verifies that the data value of the OBJECTs requested are within specified ranges. The object identifier (OID) of each object requested is the only input required. An output specification is provided to show valid value constraints per the NTCIP 1205 v01 object definitions.
Inputs: The object identifier of each object requested is needed.
Outcome(s): All data are returned and verified as correct per the OBJECT constraints of NTCIP 1205 v01. See Test Case Output Specification TCOS001 - Status Condition within the Device (Positive Test Case)
Environmental Needs: When testing for alarm temperature current value, set up is needed to measure temperature.
Tester/Reviewer M.I.
Special Procedure Requirements: None
Intercase Dependencies: None

 

Slide 75:

Learning Objective #6

Build RTCTM for CCTV Case Study

Requirements to Test Case Traceability Matrix
Requirement Test Case
ID Title ID Title
3.3.3 Status Condition within the Device
    TC001 Title: Request Status Condition within the Device Dialog Verification (Positive Test Case)

 

Slide 76:

Learning Objective #6

Importance of RTCTM

 

Slide 77:

Activity. A placeholder graphic with an image of hand over a computer keyboard to show that an activity is taking place.

 

Slide 78:

Learning Objective #6

The Requirements to Test Case Traceability Matrix relates to which of the following?

Answer Choices

  1. Requirements and design
  2. Test cases and requirements
  3. Needs and requirements
  4. None of the above

 

Slide 79:

Learning Objective #6

Review of Answers

A small graphical red and yellow X representing incorrect.a) Requirements and design
Incorrect. The RTM relates requirements and design content

A small graphical green and yellow check mark representing correct.b) Test cases and requirements
Correct! The RTCTM relates test cases and the requirements the test cases verify.

A small graphical red and yellow X representing incorrect.c) Needs and requirements
Incorrect. The PRL or NRTM relates needs and requirements.

A small graphical red and yellow X representing incorrect.d) None of the above
Incorrect. The correct answer is b) above.

 

Slide 80:

Learning Objective #6

Summary of Learning Objective #6

Develop a Requirements to Test Case Traceability Table (RTCTM)

 

Slide 81:

Learning Objective #7

Learning Objective #7: Identify Types of Testing

 

Slide 82:

Learning Objective #7

Types of Testing

Function Test

 

Slide 83:

Learning Objective #7

Types of Testing (cont.)

Performance Test

 

Slide 84:

Learning Objective #7

Types of Testing (cont.)

Load Test

 

Slide 85:

Learning Objective #7

Types of Testing (cont.)

Stress Test

 

Slide 86:

Learning Objective #7

Types of Testing (cont.)

Other Testing

 

Slide 87:

Learning Objective #7

Types of Testing (cont.)

Other Testing: Integration Tests

 

Slide 88:

Learning Objective #7

Types of Testing (cont.)

Other Testing: System Acceptance Tests

 

Slide 89:

Activity. A placeholder graphic with an image of hand over a computer keyboard to show that an activity is taking place.

 

Slide 90:

Learning Objective #7

Which of the following is used to test error handling of a system?

Answer Choices

  1. System acceptance test
  2. Negative test
  3. Periodic maintenance test
  4. Unit test

 

Slide 91:

Learning Objective #7

Review of Answers

A small graphical red and yellow X representing incorrect.a) System acceptance test
Incorrect. This test is not intended specifically to test for error handling.

A small graphical green and yellow check mark representing correct.b) Negative test
Correct! A negative test is designed to test error handling by the system.

A small graphical red and yellow X representing incorrect.c) Periodic maintenance test
Incorrect. This test is not intended specifically to test for error handling.

A small graphical red and yellow X representing incorrect.d) Unit test
Incorrect. This is a function test to a subsystem or unit of the system.

 

Slide 92:

Learning Objective #7

Summary of Learning Objective #7

Identify Types of Testing

 

Slide 93:

Learning Objective #8

Learning Objective #8: Recognize the Purpose of the Test Log and Test Anomaly Reports

 

Slide 94:

Learning Objective #8

Impacts of Failure of a Test Case

 

Slide 95:

Learning Objective #8

Test Log

 

Slide 96:

Learning Objective #8

Test Anomaly Report

 

Slide 97:

Learning Objective #8

Summary of Learning Objective #8

Understand the Purpose of the Test Log and Test Anomaly Reports

 

Slide 98:

What We Have Learned

  1. Showed that some standards have, test documentation, some do not, and how to deal with the gaps.
  2. Showed elements of a requirements to test case traceability matrix and how it can be structured to account for all requirements and all test cases.
  3. Learned how to document and handle impact of test case failures.
    1. What to document in the test log.
    2. What to document in the test anomaly report.
  4. Learned about types of testing.

 

Slide 99:

What We Have Learned (cont.)

5. Process to gather information and develop a test case:

  1. Identify user needs
  2. Identify requirements
  3. Develop test case objective
  4. Identify design content
  5. Document value constraints
  6. Complete test case

 

Slide 100:

What We Have Learned (cont.)

6. Test reports for documenting test case failure

  1. Test log
  2. Test anomaly report

 

Slide 101:

What We Have Learned (cont.)

7. Test cases can be re-used in different types of testing:

  1. Function test
  2. Performance test
  3. Load test
  4. Stress test
  5. Benchmark test
  6. Integration test
  7. System acceptance test

 

Slide 102:

Resources

  1. IEEE Std 829-2008 IEEE Standard for Software and System Test Documentation
  2. IEEE Std 610-1990 Standard Glossary of Software Engineering Terminology
  3. NTCIP 8007 Testing and Conformity Assessment Documentation within NTCIP Standards Publications
  4. A317a: Understanding User Needs for CCTV Systems Based on NTCIP 1205 Standard
  5. A317b: Understanding Requirements for CCTV Systems Based on NTCIP 1205 Standard
  6. NTCIP 1205 v01 CCTV Standard
  7. NTCIP 1204 v03 Environmental Sensor Station Interface Standard
  8. Traffic Management Data Dictionary Version 3.03
  9. PCB Testing Modules: T311 DMS; T313 ESS; and T321 TMDD
  10. Student Supplement (Combined for Part 1 and 2).

 

Slide 103:

Questions? A placeholder graphic image with word Questions? at the top, and an image of a lit light bulb on the lower right side.