3. Data, Analysis Methods, and Performance Measures

TTI logo

This chapter provides documentation on the archived traffic detector data as well as the analysis methods used to process and summarize the data. The definition and calculation procedures for the performance measures reported in the Mobility Monitoring Program are also included. The chapter is organized as follows:

  • Participating cities and their archived data — presents information on the cities that participated by submitting archived data, including the type of sensor technology used to collect the data, the level of detail of the archived data, and the data elements that were submitted.
  • Overview of data processing — provides an overview of the data processing steps used to prepare the data for analysis, including pre-processing, data quality checking, and aggregation to a common data standard, and finally the mobility and reliability analysis.
  • Data quality procedures — describes the data quality checks used in preparing the archived data for analysis.
  • Congestion and reliability measures — provides definitions and calculation procedures for the performance measures reported in this report.

Participating Cities and Their Archived Data

Public and private agencies from a total of 29 cities participated in the Mobility Monitoring Program by providing archived traffic detector data from 2003 (Table 2). The format and organization of the archived data varies considerably among cities. To provide more consistency and reduce ambiguity, the project team developed some basic guidelines on preferred data formats (Figure 1). These preferred data formats were developed to clarify exactly what data was needed, as well as to “standardize” the archived data (with some minor variation) being submitted to the Program. Some of the preferred data formats and organization arose from how the data were to be processed in the SAS application software. Other details of organization were included because they were already present and consistent in the majority of cities submitting archived data. Note that the preferred data formats reference data elements contained in national ITS standards like the Traffic Management Data Dictionary.7

7 See http://www.ite.org/tmdd/index.asp for more details.

In future years, the project team will encourage use of these preferred data formats to reduce our pre-processing burden and standardize our initial data processing software for all cities. Because participation and data submission is strictly voluntary, however, we are in a position to accept the data format and organization that is most convenient for cities to provide.

Table 2. Participating Cities and Agencies for 2003 Archived Data
City # Participating City Contact Agency/Group Historical Data Since
1 Albany, NY New York State DOT 2001
2 Atlanta, GA Georgia DOT 2000
3 Austin, TX Texas DOT 2000
4 Baltimore, MD Maryland SHA & Univ. of Maryland 2003
5 Charlotte, NC North Carolina DOT 2001
6 Cincinnati, OH-KY ARTIMIS & Kentucky Transportation Cabinet 2000
7 Dallas, TX Texas DOT & TTI 2003
8 Detroit, MI Michigan DOT 2000
9 El Paso, TX Texas DOT 2003
10 Hampton Roads, VA Virginia DOT & ADMS Virginia 2000
11 Houston, TX Texas DOT & TTI 2000
12 Los Angeles, CA Caltrans & PeMS 2000
13 Louisville, KY Kentucky Transportation Cabinet 2001
14 Milwaukee, WI Wisconsin DOT 2001
15 Minneapolis-St. Paul, MN Minnesota DOT & Univ. of Minnesota-Duluth 2000
16 Orange County, CA Caltrans & PeMS 2003
17 Orlando, FL Florida DOT 2001
18 Philadelphia, PA Mobility Technologies 2001
19 Phoenix, AZ Arizona DOT 2000
20 Pittsburgh, PA Mobility Technologies 2001
21 Portland, OR Oregon DOT 2001
22 Riverside-San Bernardino, CA Caltrans & PeMS 2003
23 Sacramento, CA Caltrans & PeMS 2002
24 Salt Lake City, UT Utah DOT 2002
25 San Antonio, TX Texas DOT 2000
26 San Diego, CA Caltrans & PeMS 2001
27 San Francisco, CA Caltrans & PeMS 2003
28 Seattle, WA Washington State DOT 2000
29 Washington, DC Maryland SHA & Univ. of Maryland, Virginia DOT & ADMS Virginia 2002
Figure 1. Preferred Archived Data Formats for Mobility Monitoring Program

PREFERRED DATA FORMATS FOR FHWA’S MOBILITY MONITORING PROGRAM

The document summarizes the preferred formats for submitting data to FHWA’s Mobility Monitoring Program. While other formats are acceptable, the following formats are encouraged for unambiguous and efficient data exchange. The required data submissions should include two distinct datasets: 1) actual traffic data records; and 2) traffic sensor location information. Many of the data elements have already been defined by national ITS standards (e.g., Traffic Management Data Dictionary, TMDD) and are indicated as such.

File Formats And Organization

  • The traffic data records should be submitted in delimited ASCII-text files. Acceptable delimiting characters include commas, tabs, or spaces.
  • Missing or null values should be indicated by providing a blank space or other null value code in the respective field. Metadata should document particular error codes (e.g., “-1” or “255”) and meaning if these error codes are contained in the dataset. A separate delimited text file should be submitted for each day for each city, with data from all sensor locations being included in this single daily file. The file should be named to include a location or agency code and a date stamp (YYYYMMDD format). For example, “phx_20040101.txt” contains data for Jan. 1, 2004 for Phoenix, Arizona.
  • The text files should be compressed for transmission using industry standard PC (*.zip) or UNIX (*.z or *.gz) compression software.
  • The archived traffic data should be submitted by CD-ROM or FTP.

Data Elements

  • The data should be aggregated to 5-minute time periods for each travel lane. Even 5-minute time periods should be used (e.g., 12:00 am, 12:05 am, 12:10 am, etc.).
  • Each row of the delimited text file should contain the following data elements:
    1. Time (HH:MM with 24-hour clock) and date (MM/DD/YYYY) stamp. Documentation should indicate whether this is a start or ending time;
    2. Detector identifier (DETECTOR_Identifier_identifier in TMDD);
    3. Vehicle traffic volume count (DETECTOR_VehicleCount_quantity);
    4. Average lane occupancy, if available, (DETECTOR_Occupancy_percent); and
    5. Average speed or travel time (LINK_SpeedAverage_rate or LINK_TravelTime_quantity).
  • If the data have been aggregated from a shorter time period (e.g., 20 seconds or 1 minute), each 5-minute record should indicate how many detailed records were used in the summary statistic calculation. This “completeness” value is reported as the percentage of the total possible records that are included in the summary statistic.

Sensor Location Information

  • Location information should be provided for each unique traffic sensor. The location information can be provided in delimited text files, spreadsheets, or databases.
  • The location information should include the following for each traffic sensor that reports data at any time during the year:
    1. Detector identifier (same as used in the traffic data records, DETECTOR_Identifier_identifier);
    2. Lane designation or code (DETECTOR_LaneNumber_code);
    3. Number of directional through travel lanes at that location (LINK_LaneCount_quantity);
    4. Roadway name and/or designation (LINK_RoadDesignator_number);
    5. Roadway direction (DETECTOR_Direction_code);
    6. Roadway facility type, such as mainlane, HOV, entrance ramp, etc. (LINK_Type_code);
    7. A linear distance reference such as roadway milepost (in miles or kilometers);
    8. Sensor activation date which indicates when the sensor began providing valid data; and
    9. Sensor de-activation date which indicates when the sensor stopped providing valid data. If the sensor is still active, this field could be blank or null values.

Additional Documentation
Additional documentation on the archived data is encouraged and appreciated. This documentation could include information on the following aspects:

  • Data collection technology and source;
  • Data quality control checks and summary results;
  • Data transformation or estimation processes (e.g., equations used to estimate speeds from single loops); and
  • Other information that would help analysts better interpret the quality and content of the ITS data archives.

The mobility and reliability measure calculations in the Mobility Monitoring Program are built around estimated travel times for directional freeway routes. However, as Table 3 indicates, nearly all of the participating cities have traffic management centers that collect traffic speeds and volumes at specific points along the freeway route. For 28 of the 29 total cities (all except Houston), the data were collected at point locations using a variety of traffic sensor technologies including single and double inductance loops, microwave radar, passive acoustic, and video image processing. For Houston, link travel times are collected by a toll tag-based system, and these link travel times are supplemented with volume trend data from a limited number of double inductance loops. In many cities, multiple sensor technologies were used to collect the traffic speed and volume data. All of these technologies use a small, fixed zone of detection, and the traffic speed and volume measurements are taken as vehicles pass through this zone. The last section in this chapter describes how these point speeds and volumes are transformed to travel time estimates for directional freeway routes.

Table 3 also indicates the level of detail at which the archived data is submitted to the Mobility Monitoring Program. The time aggregation level varies widely, from 20 seconds in San Antonio to 15 minutes in several areas. In some cases, the data are collected in smaller time intervals (e.g., 20 seconds to 2 minutes) but aggregated to larger time intervals for archiving purposes. Most of the archived data are provided on a per lane basis.

The extent of freeway monitoring coverage is shown in Table 4, which reports the percentage of each city’s freeway system for which archived data are available. The freeway coverage ranges from 6 percent in Dallas to 100+ percent in Milwaukee and Salt Lake City (both cities monitor freeway miles outside of the urban area). The average coverage of freeway centerline miles is 46 percent, whereas the average coverage of freeway lane-miles is slightly higher at 53 percent. Generally speaking, only half of the urban freeway mileage in the 29 participating cities is covered in this report’s analysis. For all 29 cities, there are nearly 3,000 total freeway miles on which real-time traffic data is being collected and archived.

Readers should note that the participating cities were not chosen based on their monitoring coverage, but on their ability to provide archived data. In many cities, this freeway coverage includes the most congested freeways as well as some lightly congested freeway routes. In several cities, the monitoring coverage does not include very congested routes for a variety of reasons (e.g., reconstruction, upcoming deployment, etc.).

Table 3. Summary of Data Collection Technologies and Data Level of Detail in 2003
Participating City Traffic Sensor Technology Data Level of Detail
Time Space
Albany, NY loop detectors 15 minutes by lane
Atlanta, GA video imaging, microwave radar 15 minutes by lane
Austin, TX loop detectors 1 minute by lane
Baltimore, MD microwave radar, loop detectors 5 minutes by lane
Charlotte, NC microwave radar 30 seconds by lane
Cincinnati, OH-KY loop detectors, microwave radar, video imaging 15 minutes by direction
Dallas, TX video imaging, loop detectors, microwave radar 5 minutes by lane
Detroit, MI loop detectors 1 minute by lane
El Paso, TX loop detectors, microwave radar 1 minute by direction
Hampton Roads, VA loop detectors, microwave radar 5 minutes by direction
Houston, TX probe vehicle (toll tags); also loop detectors, video imaging, and microwave radar vehicle-based
link travel times
Los Angeles, CA loop detectors 5 minutes by direction
Louisville, KY microwave radar, video imaging 15 minutes by direction
Milwaukee, WI loop detectors, microwave radar, video imaging 5 minutes by lane
Minneapolis-St. Paul, MN loop detectors 30 seconds by lane
Orange County, CA loop detectors 5 minutes by direction
Orlando, FL loop detectors 1 minute by lane
Philadelphia, PA microwave radar, passive acoustic detectors 5 minutes by lane
Phoenix, AZ loop detectors, passive acoustic detectors 5 minutes by lane
Pittsburgh, PA microwave radar, passive acoustic sensors 5 minutes by lane
Portland, OR loop detectors 15 minutes by lane
Riverside-San Bernardino, CA loop detectors, microwave radar 5 minutes by direction
Sacramento, CA loop detectors, microwave radar 5 minutes by direction
Salt Lake City, UT loop detectors, acoustic detectors, microloops 15 minutes by lane
San Antonio, TX loop detectors, acoustic detectors 20 seconds by lane
San Diego, CA loop detectors 5 minutes by direction
San Francisco, CA loop detectors, microwave radar 5 minutes by direction
Seattle, WA loop detectors 5 minutes by lane
Washington, DC
– Maryland
– Virginia
microwave radar, loop detectors (MD)
loop detectors (VA)
5 minutes
1 minute
by lane
by lane

Source: Conversations with agency representatives and the 2002 ITS Deployment Tracking Survey (http://itsdeployment2.ed.ornl.gov/its2002/default.asp).

Table 4. Summary of 2003 Freeway Archived Data Coverage
Participating City Centerline miles Lane-miles
Instrumented Total Coverage (%) Instrumented Total Coverage (%)
Albany, NY 17 104 16% 159 550 29%
Atlanta, GA 91 302 30% 862 2,300 37%
Austin, TX 23 107 21% 146 590 25%
Baltimore, MD 28 290 10% 167 1,543 11%
Charlotte, NC 15 93 16% 89 488 18%
Cincinnati, OH-KY 54 184 29% 367 1,035 35%
Dallas, TX 19 302 6% 151 1,904 8%
Detroit, MI 110 282 39% 751 1,810 41%
El Paso, TX 31 52 60% 193 288 67%
Hampton Roads, VA 20 191 10% 158 970 16%
Houston, TX 236 383 62% 1,764 2,548 69%
Los Angeles, CA 411 463 89% 3,697 4,253 87%
Louisville, KY 26 137 19% 147 670 22%
Milwaukee, WI 118 111 106% 698 605 115%
Minneapolis-St. Paul, MN 219 317 69% 1,194 1,595 75%
Orange County, CA 178 195 91% 1,557 1,530 102%
Orlando, FL 32 166 19% 230 775 30%
Philadelphia, PA 126 352 36% 687 2,095 33%
Phoenix, AZ 83 214 39% 729 1,360 54%
Pittsburgh, PA 101 292 35% 541 1,228 44%
Portland, OR 54 139 39% 245 725 34%
Riverside-San Bernardino, CA 57 143 40% 469 918 51%
Sacramento, CA 52 109 48% 370 715 52%
Salt Lake City, UT 95 84 113% 656 545 120%
San Antonio, TX 90 212 42% 576 1,073 54%
San Diego, CA 134 254 53% 1,065 1,848 58%
San Francisco, CA 322 342 94% 2,349 2,455 96%
Seattle, WA 120 249 48% 889 1,763 50%
Washington, DC 73 309 24% 468 1,975 24%
TOTAL 2,935 6,382 46% 21,374 40,149 53%

Source: The estimates for total freeway centerline miles and total freeway lane-miles are from FHWA’s Highway Performance Monitoring System and TTI’s Urban Mobility Study (http://mobility.tamu.edu/ums/).

Real-time traffic data collection and archiving processes have been developed independently in most of the cities and the details of these processes vary among the cities. As a general rule, traffic centers at least have the capability to archive data from their traffic detector systems. In a few cases, this capability is not used because of priorities elsewhere in the traffic center, but it is clear that traffic center software is being constructed with archiving as a function. However, the state of the practice in traffic center archiving is still fairly primitive. The most common practice is to transfer the data to a storage device where they reside in simple text file formats without an active information management system. Quality control is seldom performed at this level and access to the data is provided on a case-by-case basis without the benefit of a query or reporting structure — data are simply provided in whatever file formats are used to store them.

The typical steps from field data collection to data archiving as are follows:

  1. Traffic data are collected by detectors and accumulated in roadside controllers. These field measurements are typically collected for each individual lane of traffic. At 20-second to 2-minute intervals, the roadside controllers transmit the data to a central location, typically a traffic center.
  2. Some cities perform quality control on field-collected data, but this checking is simple and based on minimum and maximum range value thresholds.
  3. Cities that use single inductance loop detectors can measure only volumes and lane occupancies directly. In these cases, speed estimation algorithms are used to compute spot speeds from volumes and lane occupancies. These speed estimation algorithms vary among cities.
  4. Internal processes at the traffic center aggregate the traffic data to specified time intervals for archival purposes. These time intervals vary from 20 seconds (no aggregation) to 15 minutes. In some cases, the data are also aggregated across all lanes in a given direction at a sensor location.
  5. The aggregated data are then stored in text files or databases unique to each traffic center. CDs or DVDs are routinely created at the traffic centers to offload some of the storage burden and to satisfy outside requests for the data.

Calibration and maintenance of field equipment and communications are nearly universal problems. The main impediment is lack of resources to devote to these tasks; traffic center budgets are limited and must be used to address a multitude of issues. Calibration—at least to very tight tolerances—is not seen as a priority, given that operators focus on a broad range of operating conditions rather than precise volume and speed measurements. Or in some cases traffic managers may be willing to accept a certain level of data quality to satisfy only their current operations applications. This philosophy may be changing as a result of more stringent data requirements for traveler information purposes (e.g., travel time messages on variable message signs). However, we have found the current data quality requirements used by traffic centers to be quite coarse for supporting their traditional operations activities, such as incident detection and ramp meter control.

Maintenance is a problem (due primarily to funding limitations) even when detectors are known to be producing erroneous or no data. The problem is exacerbated where loop or other pavement-based detectors are used because most agencies are reluctant to shut down traffic on heavily traveled freeways just for detector repair. This is not to say that faulty detectors are never repaired, but maintenance is often postponed to coincide with other roadway activities, which helps spread the cost burden as well.

Field checking of detectors is done periodically but no standardized procedures are used across all cities. If a detector is producing values that are clearly out of range, inspection and maintenance are usually performed. However, calibration to a known standard is rarely, if ever, performed. This means that more subtle errors may go undetected. As most traffic centers typically do not require highly accurate data for most of their operations, this approach is reasonable and practical. Work zones exacerbate these problems and often contractors unknowingly sever communication lines or pave over inductance loops.

Overview of Data Processing

This section presents a brief overview of the data processing steps used to transform the archived data into congestion and reliability statistics. The relatively mundane topic of data processing is included here because of its departure from traditional traffic data monitoring practices. In analyzing the archived traffic data from the 29 participating cities, the project team processed about 6 billion data records, with a total computer processing time best measured in days or weeks.

Figure 2 shows an overview of the data processing steps used to prepare and analyze the archived data. Perhaps the greatest challenge in the data processing was “standardizing” the archived datasets from 29 different cities, or numerous different legacy systems. In many cases, the lack of adequate metadata (i.e., descriptive information about the archived data) complicates the process of properly interpreting and analyzing the archived data. For example, each city’s dataset may use different data error codes to indicate various hardware or software failures. Or similar data error codes could be used by several cities to mean different types of data errors. In other cases, various flaws, nuances, or characteristics in the archived data may be known by the data collector but undocumented, and potentially go undetected by the project team unless careful study was initiated. The experience of the project team indicates that dealing with legacy system data is much more manageable when metadata is used to describe the origin, lineage, characteristics, and subtle nuances of the archived data.

The data processing for the Mobility Monitoring Program is primarily accomplished using SAS software on a Microsoft Windows platform for two reasons: 1) the project team’s previous software programming experience with SAS; and 2) ability and flexibility of SAS to handle a wide range of complex computations on very large datasets. Many other relational database management systems (RDBMS) could also be used to accomplish the same data processing tasks as was performed in SAS.

The data processing flows shown in Figure 2 have been optimized for the use of SAS in generating annual mobility and reliability reports. Some of the data processing steps, however, may be similar for other data archiving and analysis activities. For example, the first step that includes the base code is known as extraction, transformation and loading (ETL) in the data warehouse industry and is a common function for most data warehouse projects. The project team has attempted to standardize the software code as much as possible for ease and automation of data processing. However, the software code is custom-tailored (mostly in the base code) to meet the different formats and organization of submitted archived data.

Figure 2. Overview of Data Processing within Mobility Monitoring Program
This figure illustrates the following data processing steps (in order): 1) Non-standard is processed to daily ASCII-text files. 2) The daily ASCII-text files are processed to a standardized dataset using a base code; 3) The standardized datasets are processed to summary report datasets using a summary code; 4) Production graphics are created by using an OLE copy from SAS to Excel.

The data processing as shown in Figure 2 would ideally start with daily ASCII-text files that meet the preferred data formats indicated in Figure 1. However, many cities submit data in a form that requires pre-processing (e.g., binary file formats or thousands of separate files per city per day). Pre-processing this non-standard data requires extra steps and time at the beginning to prepare the archived data to be processed using the base code.

Once the submitted archived data meets basic formatting and organization requirements, it is processed using the base code. This software code: 1) imports the data to SAS; 2) performs data quality checking; 3) aggregates detailed data to a common standard (currently 5-minute lane-by-lane); and 4) generates summary statistics on the data quality checking and processing steps. Some of these steps, such as the data quality checks, have been standardized for all cities. Other steps are unique to each city based on the aggregation level and other data characteristics. This step involves the longest amount of processing time, sometimes taking up to 24 hours for the largest cities with the most detailed data (e.g., 20-seconds, lane-by-lane).

The standardized datasets are produced as a result of the base code. The data elements and table structure of these datasets are very similar with a few exceptions (e.g., some cities are 5-minute by-lane, others may be 15-minute by direction). Thus the summary code, which contains the mobility and reliability measure calculations described later in this chapter, has largely been standardized for all cities. The standardized datasets are analogous to the database tables that would be kept on-line in an RDBMS environment.

The summary code performs all mobility and reliability measure calculations, and produces relatively small datasets (less than 1 megabyte total) that are then used to produce the charts and tables shown throughout this report and the city report appendices. Microsoft Excel was selected for the ease of producing report-ready graphics.

In summary, the data processing steps and software code used to analyze the archived data has developed in this way as a result of: 1) previous project team experience; and 2) the specific application of creating annual mobility and reliability reports. Different approaches are very likely given different implementation scenarios and development teams. Several of the data processing steps conducted in the Mobility Monitoring Program may be relevant to other data archiving or data warehouse activities. In particular, the base code contains data quality checking procedures and other steps that are most likely required in other data warehouse efforts. The summary code contains congestion and reliability measure calculations that are described in later in this chapter and may be useful to others developing performance measure programs.

Data Quality Checking

The topic of data quality is included here because of its overall importance in checking and evaluating the validity of archived data. Readers should note that the project team has not been able to systematically assess data accuracy. This means that the traffic speeds and volumes in the archived data could be systematically higher or lower (e.g., ± 10 to 20 percent) than true speeds and still be within the range of possible data values that pass quality control.

Table 5 presents the data quality checks that were used in processing the 2003 archived data. The data quality checks have been developed from these sources:

  • Current practices in other traffic centers or data archiving systems;
  • Suggested practices recommended in the literature; and
  • Practices found to be necessary from project team analysis of the archived data.

These data quality checks can be characterized as basic validity checks and should detect major problems with data errors. More subtle erroneous or suspect data could potentially go undetected with these basic rules. The project team is reviewing the use of more sophisticated data quality checking, and we will continue to balance the sophistication of the data quality checking with the amount of available data processing time. The data quality checks shown in Table 5 may evolve as the project team accumulates more experience with the archived data. More sophisticated quality checks could include tests like these:

  • Rapid fluctuations in values across successive time periods;
  • Detectors in adjacent lanes at the same location reporting significantly different values or trends;
  • Detectors in adjacent upstream or downstream locations reporting significantly different values or trends;
  • Detectors from multiple locations reporting the same values (indicative of a system problem);
  • Reported values that are significantly different from the location’s history for similar days of the calendar.

The results of these quality control checks are shown in Table 6, which reports the percent of the submitted data that passed the quality control checks. The table presents traffic volume and speed data quality separately, as some of the validity checks could have rejected one of the data values but not the other. Also note that Table 6 only evaluates the validity of the data that was archived and submitted. This table does not reflect data that are missing and were never reported because of various hardware or software failures.

Table 7 summarizes information on data completeness or availability, another dimension of data quality. The data completeness measures the number of available data values to the number of total possible values that one could expect (given the number of sensors and a polling rate). For example, if the data are reported by 5-minute time interval, 288 data values or records per day per detector are to be expected (i.e., 1,440 minutes per day divided by 5-minute periods equals 288 records). Table 7 reports data completeness for the original dataset as submitted by participating cities, as well as the analysis dataset (after quality control and imputation) that is used for mobility and reliability performance measure calculations.

Interested readers should refer to the September 2004 FHWA report on traffic data quality measurement for more details on calculating the validity and completeness measures, as well as several other data quality measures.8

8 Battelle, Cambridge Systematics, and Texas Transportation Institute. Traffic Data Quality Measurement: Final Report. Federal Highway Administration, September 2004.

Table 5. 2003 Data Validity Checks in the Mobility Monitoring Program
Quality Control Test and Description Sample Code with Threshold Values Action
Controller error codes

  • Special numeric codes that indicate that controller or system software has detected an error or a function has been disabled.
If VOLUME={code} or OCC={code} or SPEED={code} where {code} typically equals
“-1” or “255”
  • Set values with error codes to missing/null, assign missing value flag/code.
No vehicles present

  • Speed values of zero when no vehicles present
  • Indicates that no vehicles passed the detection zone during the detection time period.
If SPEED=0 and VOLUME=0 (and OCC=0)
  • Set SPEED to missing/null, assign missing value code
  • No vehicles passed the detection zone during the time period.
Consistency of elapsed time between records

  • Polling period length may drift or controllers may accumulate data if polling cycle is missed.
  • Data collection server may not have stable or fixed communication time with field controllers.
Elapsed time between consecutive records exceeds a predefined limit or is not consistent
  • Action varies. If polling period length is inconsistent, volume-based QC rules should use a volume flow rate, not absolute counts.
Duplicate records

  • Caused by errors in data archiving logic or software process.
Detector and date/time stamp combination are identical.
  • Remove/delete duplicate records.
QC1-QC3: Logical consistency tests

  • Typically used for date, time and location.
  • Caused by various types of failures.
If VOLUME > 17 (20 sec.)
If VOLUME > 25 (30 sec.)
If VOLUME > 250 (5 min.)
If VPHPL > 3000 (any time period length)
  • Assign QC flag to VOLUME, write failed record to off-line database, set VOLUME to missing/null.
QC5: Maximum occupancy

  • Empirical evidence suggests that all data values at high occupancy levels are suspect.
  • Caused by detectors that may be “stuck on.”
If OCC > 95% (20 to 30 sec.)
If OCC > 80% (1 to 5 min.)
  • Assign QC flag to VOLUME, OCCUPANCY and SPEED; write failed record to off-line database; set VOLUME, OCCUPANCY and SPEED to missing/null
QC6: Minimum speed

  • Empirical evidence suggests that actual speed values at low speed levels are inaccurate.
If SPEED < 5 mph
  • Assign QC flag to SPEED, write failed record to off-line database, set SPEED value to missing/null
QC7: Maximum speed

  • Empirical evidence suggests that actual speed values at high speed levels are suspect.
If SPEED > 100 mph (20 to 30 sec.)
If SPEED > 80 mph (1 to 5 min.)
  • Assign QC flag to SPEED, write failed record to off-line database, set SPEED value to missing/null
Maximum reduction in speed

  • Empirical evidence suggests that speed reductions greater than some maximum value are suspect.
  • Used only for AVI probe vehicle data that reports space mean speeds.
If SPEEDn+1 < (0.45 x SPEEDn)
  • Assign QC flag to SPEED, write failed record to off-line database, set SPEED value to missing/null
QC8: Multi-variate consistency

  • Zero speed values when volume (and occupancy) are non-zero
  • Speed trap not functioning properly
If SPEED = 0 and VOLUME > 0 (and OCC > 0)
  • Assign QC flag to SPEED, write failed record to off-line database, set SPEED value to missing/null
QC9: Multi-variate consistency

  • Zero volume values when speed is non-zero.
  • Unknown cause.
If VOLUME = 0 and SPEED > 0
  • Assign QC flag to VOLUME, write failed record to off-line database, set VOLUME to missing/null
QC10: Multi-variate consistency

  • Zero speed and volume values when occupancy is non-zero.
  • Unknown cause.
If SPEED = 0 and VOLUME = 0 and OCC > 0
  • Assign QC flag to VOLUME, OCCUPANCY and SPEED; write failed record to off-line database; set VOLUME, OCCUPANCY and SPEED to missing/null
QC11: Truncated occupancy values of zero

  • Caused when software truncates or rounds to integer value
  • Calculate maximum possible volume (MAXVOL) for an occupancy value of “1”:
If OCC = 0 and VOLUME > MAXVOL where MAXVOL=(2.932*ELAPTIME*SPEED)/600
  • Assign QC flag to VOLUME, OCCUPANCY and SPEED; write failed record to off-line database; set VOLUME, OCCUPANCY and SPEED to missing/null
QC12: Maximum estimated density

  • Caused by improbable combinations of volume and speed.
  • Traffic flow theory suggests that vehicle density rarely exceeds 220 vehicles per lane per mile.
IF ((VOLUME*(3600/NOM_POLL))/SPEED) > 220 where NOM_POLL is the nominal polling cycle length in seconds.
  • Assign QC flag to VOLUME, OCCUPANCY and SPEED; write failed record to off-line database; set VOLUME, OCCUPANCY and SPEED to missing/null
QC13: Consecutive identical volume-occupancy-speed values

  • Research and statistical probability indicates that consecutive runs of identical data values are suspect.
  • Typically caused by hardware failures.
No more than 8 consecutive identical volume-occupancy-speed values. That is, the volume AND occupancy AND speed values have more than 8 consecutive identical values, respectively. Zero (“0”) values are included in this check.
  • Assign QC flag to VOLUME, OCCUPANCY and SPEED; write failed record to off-line database; set VOLUME, OCCUPANCY and SPEED to missing/null
Table 6. Summary of 2003 Freeway Archived Data Validity
Participating City Validity (%)
Volume data Speed data
Albany, NY 55% 54%
Atlanta, GA 94% 89%
Austin, TX 80% 56%
Baltimore, MD 99% 73%
Charlotte, NC 92% 93%
Cincinnati, OH-KY 76% 74%
Dallas, TX 100% 96%
Detroit, MI 100% 99%
El Paso, TX 55% 49%
Hampton Roads, VA 75% 58%
Houston, TX n.a. 98%
Los Angeles, CA 98% 99%
Louisville, KY 93% 93%
Milwaukee, WI 80% 82%
Minneapolis-St. Paul, MN 83% 78%
Orange County, CA 98% 99%
Orlando, FL n.a. n.a.
Philadelphia, PA 100% 99%
Phoenix, AZ 72% 68%
Pittsburgh, PA 100% 99%
Portland, OR 62% 85%
Riverside-San Bernardino, CA 96% 100%
Sacramento, CA 100% 100%
Salt Lake City, UT 62% 53%
San Antonio, TX 96% 84%
San Diego, CA 100% 99%
San Francisco, CA 100% 100%
Seattle, WA 85% 100%
Washington, DC 100% 98%

Notes: Validity is reported as the percentage of submitted data values that passed the quality control rules specified in Table 5. See Traffic Data Quality Measurement: Final Report (footnote 8 on page 24) for more details on data quality measure calculation.

Table 7. Summary of 2003 Freeway Archived Data Completeness
Participating City Completeness (%) Original Data Completeness (%) Analysis Data
Volume data Speed data Volume data Speed data
Albany, NY 69% 69% 38% 37%
Atlanta, GA 17% 17% 57% 54%
Austin, TX 97% 97% 77% 59%
Baltimore, MD 77% 77% 63% 57%
Charlotte, NC 60% 61% 55% 57%
Cincinnati, OH-KY 47% 47% 44% 41%
Dallas, TX 76% 76% 46% 44%
Detroit, MI 61% 61% 61% 62%
El Paso, TX 98% 98% 33% 33%
Hampton Roads, VA 75% 75% 49% 39%
Houston, TX n.a. n.a. n.a. 56%
Los Angeles, CA 100% 99% 98% 98%
Louisville, KY 82% 82% 82% 76%
Milwaukee, WI 100% 100% 80% 77%
Minneapolis-St. Paul, MN 100% 100% 83% 79%
Orange County, CA 99% 94% 97% 93%
Orlando, FL n.a. n.a. n.a. n.a.
Philadelphia, PA 92% 92% 92% 91%
Phoenix, AZ 89% 89% 63% 60%
Pittsburgh, PA 94% 94% 94% 93%
Portland, OR 98% 98% 84% 83%
Riverside-San Bernardino, CA 71% 67% 70% 67%
Sacramento, CA 88% 84% 88% 83%
Salt Lake City, UT 95% 95% 44% 38%
San Antonio, TX 63% 63% 67% 66%
San Diego, CA 95% 93% 95% 92%
San Francisco, CA 98% 92% 97% 92%
Seattle, WA 79% 79% 80% 81%
Washington, DC 52% 52% 33% 33%

Note: Completeness is reported as the percentage of data values available for use. It is calculated as the ratio of total available data values to total expected data values. See Traffic Data Quality Measurement: Final Report (footnote 8 on page 24) for more details on data quality measure calculation.

Congestion and Reliability Measure Calculations

With the exception of Houston, which reported travel times collected with their AVI system, archived data from the participating cities consisted of traffic speeds and volumes collected at various points along the freeway routes. Because the mobility and reliability performance measures are based on travel time, the project team estimated freeway route travel times from the spot speeds. Figure 3 illustrates the process whereby lane-by-lane volumes and speeds are used as the basis for estimating freeway route travel times and vehicle-miles of travel (VMT). The steps are as follows:

  1. If data are reported by lane, the lane-by-lane data are combined into a “station” (e.g., all lanes in a direction). Traffic volumes are summed across all lanes, and traffic speeds are a weighted average, with weighting based on respective traffic volumes.
  2. Link properties were estimated from “station” data by assuming that each detector had a zone of influence equal to half the distance to the detectors immediately upstream and downstream from it. The measured speeds were then assumed to be constant within each zone of influence, and travel times were calculated using the equivalent link lengths. VMT were also computed in this way using traffic volume.
  3. Freeway links were then grouped with other similar adjacent link into analysis sections, which were typically 5 to 10 miles in length. The beginning and end points of analysis sections were typically selected to coincide with major highway interchanges or other locations where traffic conditions were expected to change because of traffic or roadway characteristics.

Travel times for these analysis sections then served as the basis for all subsequent mobility and reliability measure calculations. The specifics of these performance measure calculations are contained later in this section. Readers should note that equations using travel time refer to the analysis section travel times as described above.

Several other aspects and definitions used in preparing the archived data for analysis were:

  • Holidays were excluded from the weekday peak period analysis, as holidays were considered to be atypical of normal travel patterns. Holidays are included in several daily total statistics, which also include weekend days. The holidays that are excluded from weekday analyses include:
    1. New Year’s Day
    2. Martin Luther King, Jr. Day
    3. President’s Day/Washington’s Birthday
    4. Memorial Day
    5. Independence Day
    6. Labor Day
    7. Thanksgiving Day (and the day after)
    8. Christmas (and day before or after, depending on day of week)
    9. New Year’s Eve Day
  • Fixed and consistent time periods were defined for all cities. These were:
    1. 12:00 am to 6:00 am – early morning
    2. 6:00 am to 9:00 am – morning peak
    3. 9:00 am to 4:00 pm – mid-day
    4. 4:00 pm to 7:00 pm – afternoon peak
    5. 7:00 pm to 12:00 am – late evening
  • Only mainline freeway detectors were included. Some cities reported ramp data, but these were dropped to maintain consistency across the cities.

Figure 3. Estimating Directional Route Travel Times and VMT From Spot Speeds and Volumes
This figure shows the process whereby individual lane-by-lane sensor data are aggregated to a section-level travel and VMT estimate. The process starts with traffic sensors that typically collect in each lane at a 1/2-mile nominal spacing. Summary statistics are then computed across all lanes in a given direction. Point-based properties (traffic volume and speed) are extrapolated to roadways links that are typically ½- to 3-miles  in length. Link properties (travel time and vehicle-miles of travel) are summed to analysis sections that are typically 5 to 10 miles in length.

Congestion Measures

The Mobility Monitoring Program tracks traffic congestion using the three measures below. For most applications, these measures are reported for the peak periods (6 to 9 a.m. and 4 to 7 p.m.):

  • Travel time index (measures congestion intensity, also congestion duration when shown by time of day)
  • Percent of congested travel (measures congestion extent, also congestion duration when shown by time of day)
  • Total delay (measures congestion intensity)

The travel time index is the ratio of average peak travel time to a free-flow travel time (Equation 1). In this report, the free-flow conditions are travel times at a speed of 60 mph. Index values can be related to the general public as an indicator of the length of extra travel time spent during a trip. For example, a value of 1.20 means that average peak travel times are 20 percent longer than free-flow travel times. In this report, the travel time index is calculated for directional freeway sections (as shown in Figure 3), then combined into an areawide average by weighting each freeway section by the respective VMT (Equation 2).

Equation 1
For a specific road section and time period, the travel time index (which has no units) is computed as the average travel time (in minutes) divided by the free-flow travel time (in minutes).

Equation 2
For several road sections and time periods, the average travel time index is computed as the weighted average of all road sections and time periods. The weighting factor is vehicle-miles of travel (VMT).

The travel time index can be applied to various transportation system elements with different free-flow speeds, although only freeways were analyzed in the Mobility Monitoring Program. The index can be averaged for streets, freeways, bus and carpool lanes, bus and rail transit, bicycle facilities, and sidewalks. All of these system elements have a free-flow travel time and when crowded, the travel time increases. In a multi-modal context, the VMT in Equation 2 is replaced with person-miles of travel (PMT). An average corridor value can be developed with the number of persons using each facility or mode to calculate the weighted average of the conditions on adjacent streets, freeways, HOV lanes, bus routes and/or rail transit lines. The corridor values can be computed for certain time periods and weighted by the number of travelers to estimate peak-period or daily index values.

The percent of congested travel is calculated as the ratio of congested VMT to total VMT (Equation 3). In this report, a free-flow speed of 60 mph is used as the value below which VMT is considered to be congested.

Equation 3
The percent of congested travel is calculated as the congested VMT divided by the total VMT.

Our experience indicates that the use of a 60 mph threshold in the percent congested travel measure may over-represent the magnitude of congestion. In several cities, the spot speeds collected by point-based detectors are less than 60 mph even in light traffic conditions. These point-based detectors are also more likely to record lower speeds than longer distance travel time measurements, due to their common location near entrance ramps and the much greater variation in speed over short sections than long sections. These considerations suggest that a lower speed may be more appropriate for the congestion threshold in this measure when using point-based sensors. Unlike the other congestion measures, congested travel is a binary attribute—travel is either congested or it is not congested, no matter how close the speed is to the congestion threshold. Thus, for a given time period, the VMT is assigned as either congested or not congested, even if the average speeds are just below the congestion threshold. For example, if the nighttime speed limit on an urban freeway system is 55 mph, a significant portion of travel could be categorized as congested without heavy traffic being the cause.

Delay is calculated as the additional travel time that is incurred when actual travel times are greater than free-flow travel times (Equations 4 and 5), expressed in this report as vehicle-hours as well as vehicle-hours per 1,000 VMT. The delay measure can also be expressed in person-hours in a multi-modal context where person travel quantities are known.

Equation 4
The delay (in vehicle-hours) is computed by calculating the difference between average travel time (minutes) and free-flow travel time (minutes), multiplying by the traffic volume (vehicles), the dividing by 60 (a minutes-to-hours conversion factor).

Equation 5
For several road sections and time periods, total delay is calculated as the sum of delay for all road sections and time periods.

Reliability Measures

The congestion measures in the previous section represent the average and total levels of congestion. In addition to average and total statistics, there is a growing recognition of the need to track variability of congestion and the reliability of travel. The Mobility Monitoring Program tracks these measures for travel reliability:

  • Planning time index
  • Buffer index

The planning time index is statistically defined as the 95th percentile travel time index (Equation 6) and also represents the extra time most travelers add to a free-flow travel time when planning trips. For example, a planning time index of 1.60 means that travelers should plan for an additional 60 percent travel time above the free-flow travel time to ensure on-time arrival most of the time (95 percent in this report).

Equation 6
For a specific road section and time period, the planning time index (no units) is calculated as the 95th percentile travel time index.

The planning time index is useful because it can be directly compared to the travel time index on similar numeric scales. For example, assume that the peak period travel time index for a particular road section is 1.20, which means that average travel times are 20 percent longer in the peak period than during free-flow conditions. Now assume that the planning time index for that same road and time period is 1.60, which means that 95 percent of all travel times are less than 60 percent longer than during free-flow conditions. In other terms, the planning time index marks the upper limit for the nearly worst (95 percent of the time) travel conditions.

The buffer index represents the extra time (buffer) most travelers add to their average travel time when planning trips (Equation 7). The buffer index is differentiated from the planning time index in these two important ways:

  • The buffer index is expressed as a percentage;
  • The buffer index represents the extra time between the average travel time and near-worst case travel time (95th percentile), whereas the planning time index represents the extra time between the free-flow travel time and the near-worst case travel time (95th percentile).

Equation 7
For a specific road section and time period, the buffer index (percent) is calculated as the difference between the 95th percentile travel time and the average travel time, divided by the average travel time.

For example, a buffer index of 40 percent means that a traveler should budget an additional 8-minute buffer for a 20-minute average peak travel time to ensure on-time arrival most of the time (95 percent in this report). The 95th percentile travel time was chosen for these reliability measures to represent a near-worst case scenario. For example, the 95th percentile travel time corresponds to a 95 percent on-time arrival rate, which can be simply explained in non-technical terms as “being late for work one day per month.” Other percentiles, such as the 85th or 90th percentile, could be used in this or other applications. Ultimately, the application of the reliability measure will determine the percentile used in its calculation.

Equations 6 and 7 show the reliability measure calculations for a specific road section and time period. For these reliability measures, the road section and time period should be chosen in a way that accurately represents the reliability of interest. For example, an analysis of urban commuting reliability would likely consider freeway sections 5 to 10 miles in length whose endpoints correspond to major freeway or major arterial interchanges. Alternatively, an analysis of intercity travel reliability would consider much longer freeway sections whose endpoints correspond to popular city origins and destinations. The time period(s) should be selected to include conditions of a similar nature and interest to travelers. For example, a buffer index for a typical commuter audience will likely focus on periods throughout the day in which commute travel is made, and should not mix travel times from these different periods. That is, travel times from the evening peak period should not be combined into the same distribution as the morning peak travel times when calculating a 95th percentile.

The average planning time or buffer index values (across several road sections, time periods, etc.) can be calculated by using the VMT as a weighting factor (Equation 8).

Equation 8
For several road sections and time periods, the average index value for planning time index or buffer index is calculated as the VMT-weighted average for all road sections and time periods.

The Program also tracks throughput using peak-period and total daily VMT, which is calculated as shown in Figure 3.

Other Considerations for Performance Measure Calculations

The performance measure analysis uses data in a standard format, which currently consists of 5-minute data (all times of the day and days of the year) for 5- to 10-mile freeway sections. This standard format corresponds with the bottom part of the diagram in Figure 3. Combining the estimated travel time values or performance measures from each 5-minute time period is accomplished using VMT as a weighting factor for each time period.

Measures that do not use specific origins and destinations generally provide easier comparisons because these measures are length-neutral and can be applied to a wider variety of situations. If trip-based measures are desired as examples for specific origins and destinations, the performance measures described here can be used with the estimated travel time for a specific trip. This combination of generalized, length-neutral measures as well as specific examples should provide statistics with which most audiences can relate.

There is no single best performance measure and users should resist the urge to select a single measure or index for all situations. Each performance measure reported here addresses different dimensions of traffic congestion or different aspects of reliability. The “dashboard” concept of using a “few good measures” is appropriate, and performance monitoring programs should consider selecting a few (for example, two or three of the five presented here) measures for an executive summary or dashboard report.

This analysis defines fixed-length time periods in which to compute average peak period measures. No single time period will be correct for all analyses, but there are several considerations as follows:

  • Peak hour or peak period — Transportation engineers have traditionally used a peak hour to describe congestion, but major urban areas now experience slow speeds for multiple hours in both the morning and the afternoon. In many areas, congestion growth occurs in the hours before or after the traditional peak hour. Use of a single peak hour misses the congestion that occurs during other times, prompting many areas to define a multi-hour peak period.
  • Urban area size — Using a 3- to 4-hour peak period for all area sizes may mask congestion for the smaller urban areas. Smaller areas can probably develop useful statistics with only peak hour analyses.
  • City-to-city comparison — A consistent peak-period length is necessary for any type of comparison between cities. Comparative studies between urban areas should probably use peak period analyses, rather than only a peak hour.
  • Daily or peak comparisons — For national comparisons of reliability trends, a day-to-day comparison is appropriate. For local purposes, where individual trip planning is also an issue, it may be useful to also include travel reliability within an hour or for several segments of a multi-hour peak period.