1. Trang chủ
  2. » Kỹ Thuật - Công Nghệ

Transportation Systems Planning Methods and Applications 12

23 75 0

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 23
Dung lượng 688,44 KB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

Transportation Systems Planning Methods and Applications 12 Transportation engineering and transportation planning are two sides of the same coin aiming at the design of an efficient infrastructure and service to meet the growing needs for accessibility and mobility. Many well-designed transport systems that meet these needs are based on a solid understanding of human behavior. Since transportation systems are the backbone connecting the vital parts of a city, in-depth understanding of human nature is essential to the planning, design, and operational analysis of transportation systems. With contributions by transportation experts from around the world, Transportation Systems Planning: Methods and Applications compiles engineering data and methods for solving problems in the planning, design, construction, and operation of various transportation modes into one source. It is the first methodological transportation planning reference that illustrates analytical simulation methods that depict human behavior in a realistic way, and many of its chapters emphasize newly developed and previously unpublished simulation methods. The handbook demonstrates how urban and regional planning, geography, demography, economics, sociology, ecology, psychology, business, operations management, and engineering come together to help us plan for better futures that are human-centered.

Trang 1

III

Systems Simulation and Applications

0273_book Page 1 Friday, October 25, 2002 8:33 AM

© 2003 CRC Press LLC

Trang 2

Microsimulation

CONTENTS

12.1 Introduction12.2 What Is Microsimulation?

Further Reading

12.1 Introduction

The purpose of this chapter is to provide an overview of microsimulation concepts and methods that may be used in travel-related forecasting applications Including this very brief introductory section, the chapter is divided into eight sections Section 12.2 defines the term microsimulation Section 12.3 discusses the reasons why microsimulation may prove useful or even necessary for at least some types of activity-based travel forecasting applications Section 12.4 defines the important concepts of objects, agents, and cellular automata, which represent fundamental organizing constructs in modern microsimulation mod-els Section 12.5 discusses a key step in the microsimulation process — synthesizing and updating the attributes of the population or sample of individuals whose behavior is being simulated Section 12.6

then discusses some of the major issues associated with the development and application of operational microsimulation methods, while Section 12.7 provides representative examples of transportation-related microsimulation applications Finally, Section 12.8 presents several microsimulation models drawn from

a range of applications, including activity-based travel forecasting

12.2 What Is Microsimulation?

While many current modeling efforts are microsimulation based, the term itself is rarely defined and tends to mean different things to different people Perhaps due to the recent proliferation of network microsimulators, many people tend to think of microsimulation specifically in terms of network route choice and performance models On the other hand, given the use of microsimulation methods to generate the disaggregate inputs required by their models, many activity modelers think of microsimu-Eric J Miller

University of Toronto

Trang 3

lation as the procedures used in this input data synthesis and updating process In this chapter we adopt

a comprehensive definition of microsimulation as a method or approach (rather than a model per se) for exercising a disaggregate model over time

Simulation generally refers to an approach to modeling systems that possess one or both of the following two key characteristics

1 The system is a dynamic one, whose behavior must be explicitly modeled over time

2 The system’s behavior is complex In addition to the dynamic nature of the system (which generally

in itself introduces complexity), this complexity typically has many possible sources, including:

a Complex decision rules for the individual actors within the system

b Many different types of actors interacting in complex ways

c System processes that are path dependent (i.e., the future system state depends both on the current system state and explicitly on how the system evolves from this current state over time)

d A generally open system on which exogenous forces operate over time, thereby affecting the internal behavior of the system

e Significant probabilistic elements (uncertainties) that exist in the system, with respect to random variations in exogenous inputs to the system or the stochastic nature of endogenous processes at work within the system

Note that in speaking of complexity, we are not merely referring to the difficulty in dealing with very large models with large data sets defined over many attributes for hundreds if not thousands of zones Rather, we are referring to the more fundamental notion of the difficulty in estimating likely future system states given the inherently complex nature of the system’s behavioral processes

Given the system’s complexity, closed-form analytical representations of the system are generally not possible, in which case numerical, computer-based algorithms are the only feasible method for generating estimates of future system states Similarly, given the system’s path dependencies and openness to time-varying exogenous factors, system equilibrium often is not achieved, rendering equilibrium-based models inappropriate in such cases.1 In the absence of explicit equilibrium conditions, the future state of the system again generally can only be estimated by explicitly tracing the evolutionary path of the system over time, beginning with current known conditions Such numerical, computer-based models that trace

a system’s evolution over time are what we generally refer to as simulation models

Note that conventional four-stage travel demand models most clearly are not simulation models under this definition Conventional four-stage models are static equilibrium models that predict a path-inde-pendent future year end state without concern for either the initial (current) system state or the path traveled by the system from the current to the future year state

The prefix micro simply indicates that the simulation model is formulated at the disaggregate or micro level of individual decision-making agents (or other relevant units), such as individual persons, house-holds, and vehicles A full discussion of the relative merits of disaggregate vs more traditional aggregate modeling methods is beyond the scope of this chapter It is fair to say that a broad consensus exists within the activity–travel demand modeling community that disaggregate modeling methods possess consider-able advantages over more aggregate approaches (including minimization of model bias, maximization

of model statistical efficiency, improved policy sensitivity, and improved model transferability — and hence usability within forecasting applications), and that they will continue to be the preferred modeling approach for the foreseeable future With respect to microsimulation, the relevant question is to what extent does microsimulation represent a feasible and useful mechanism for using disaggregate models within various forecasting applications

To answer this question, first consider the well-known short-run policy analysis or forecasting dure known as sample enumeration In this procedure, a disaggregate behavioral model of some form has been developed (say, for sake of illustration, a logit work trip mode choice model) A representative

Trang 4

sample of decision makers (in this case workers) typically exists, since such a sample is generally required for model development This sample defines all relevant inputs to the model with respect to the attributes

of all the individuals in the sample The short-run impact of various policies that might be expected to affect work trip mode choice can then be tested by “implementing” a given policy, and then using the model to compute the response of each individual to this policy Summing up the responses of the individuals provides an unbiased estimate of the aggregate system response to the policy in question.Figure 12.1 summarizes this procedure This figure can be taken as a very generic representation of a microsimulation process for the case of a short-run forecast, in which all model inputs except those relating to the policy tests of interest are fixed, and hence all that needs to be simulated are the behavioral responses of the sampled decision makers to the given policy stimuli Thus, in such cases, sample enumeration and microsimulation are essentially synonymous, and use of the latter term simply empha-sizes the disaggregate, dynamic2 nature of the model The majority of activity-based microsimulation models developed to date basically fall into this category of short-run sample enumeration-based models.Sample enumeration is a very efficient and effective forecasting method providing that:

1 A representative sample is available

2 One is undertaking a short-run forecast (so that the sample can be assumed to remain tative over the time frame of the forecast)

represen-3 The sample is appropriate for testing the policy of interest (i.e., the policy applies in a useful way

to the sample in question)

Many forecasting situations, however, violate one or more of these conditions Perhaps most monly, one is often interested in forecasting over medium to long time periods, during which time the available sample will clearly become unrepresentative (people will age and even die; workers will change jobs or residential locations; new workers with different combinations of attributes will join the labor force; etc.) The question then becomes how to properly update the sample in order to maintain its representativeness In other cases, the sample may not be adequate to test a given policy (e.g., it contains too few observations of a particularly important subpopulation for the given policy test) If this is the case, how does one extend the sample so that a statistically reliable test of the policy can be performed? Finally, there may be cases in which a suitable sample simply does not exist (e.g., perhaps the model has been transferred from another urban area) In such a case, how does one generate or synthesize a representative sample?

com-In all of these cases, microsimulation provides a means of overcoming the limitations of the available sample In the case of the sample becoming less and less representative over time, Figure 12.2 presents

a simple microsimulation framework in which the sample is explicitly updated over time Updating may involve changing attributes of the households or persons in the sample (age of population, changes

FIGURE 12.1 Forecasting with sample enumeration (short-run microsimulation).

demographic and socioeconomic dynamics that are discussed immediately below.

Disaggregate Behavioral Model

Predicted Behavior

Exogenous Inputs

Trang 5

in household structure, changes in income or employment, etc.), deleting households or persons from the sample (due to out-migration from the study area, death, etc.), or adding new households or persons (due to in-migration, births, etc.) The result is that the sample (hopefully) remains repre-sentative for each point in simulated time and so provides a valid basis for predicting behavior at each such point in time.

If the original sample is either inadequate or missing altogether, then, as shown in Figure 12.3, an additional step must be inserted into the model, involving synthesizing a representative sample from other available (typically more aggregate) data, such as census data Procedures for doing this are discussed

in greater detail below

Note that these figures assume that the disaggregate behavioral model is itself a dynamic one that must

be stepped through time (and hence its inclusion within the time loop) Many current models, however, are fairly static in nature In such cases, the behavioral model can be removed from the time loop and executed only once, using the desired future year sample that has been estimated through the microsim-ulation procedure Thus, one may distinguish between static microsimulation, in which a fixed, repre-

FIGURE 12.2 Typical microsimulation model design using an available sample.

FIGURE 12.3 Typical microsimulation model design using a synthesized sample.

Base Sample for t = t 0

Endogenous Changes to Sample during This ∆t

Disaggregate Behavioral Model

Behavior/System State

at (t + ∆t)

Exogenous Inputs This ∆t

Disaggregate Behavioral Model

Behavior/System State

at (t + ∆t)

Exogenous Inputs This ∆t

t = t + ∆t

Trang 6

sentative sample is used to test various policy alternatives within a microanalytic framework, and dynamic microsimulation, wherein the representative sample changes or evolves over time as a function of endog-enous or exogenous processes.

In summary, key features of the full microsimulation process include the following:

1 The model must have as its primary input a disaggregate list of actors (or entities or behavioral units) upon which it operates This list often consists of a representative sample of individuals drawn from the relevant population Alternatively, it is becoming more common to use a 100% sample, i.e., a list of the entire set of actors in the population In general, two sources exist for this input list or base sample: a sample of actual individuals drawn from the population, obtained through conventional survey methods, or a list of synthetic individuals, statistically constructed from more aggregate data concerning the population being modeled (e.g., census data)

2 Over time, the demographic, social, and economic characteristics of the population being analyzed will change Thus, the list of actors must be updated over time within the model so that it remains representative of the population at each point in time Processes affecting the evolution of the population over time can be both endogenous (aging, births, deaths, etc.) and exogenous (in-migration, etc.)

3 Once the attributes of the list of actors are known for a given point in time, the behavior of these actors can be simulated using a relevant behavioral model Depending on the application, this model may deal with a single process (e.g., activity–travel choices in response to travel demand management (TDM) measures) or many nested processes (residential location, employ-ment location, activity–travel, etc.), involving a complex set of interconnected submodels In general, this behavior will depend on past and current system states (endogenous factors) as well as exogenous factors

4 Primary outputs from the microsimulation include both the attributes and behaviors of the actors over time These outputs are generally expressible both in aggregate terms (link volumes, average modal splits, etc.) and in terms of disaggregate trajectories of the individuals being simulated (i.e., the historical record of the behavior of each individual over time)

Different microsimulation applications, of course, will involve different implementations of

Figure 12.3 So-called static microsimulations, in particular, do not require updating of the base sample

or population, and they do not require iterating the model through time Similarly, microsimulations that are based on an observed sample of actors do not require a synthetic sample to be constructed

12.3 Why Microsimulate?

As briefly discussed in the previous section, a primary motivation for adopting a microsimulation modeling approach is that it may well be the best (and in some cases perhaps the only) way to generate the detailed inputs required by disaggregate models The strength of the disaggregate modeling approach

is in being able to fix decision makers within explicit choice contexts with respect to:

1 Salient characteristics of the actors involved

2 Salient characteristics of the choice context (in terms of the options involved, the constraints faced

by the actors, etc.)

3 Any context-specific rules of behavior that may apply

This inherent strength of the disaggregate approach is clearly compromised if one cannot provide adequately detailed inputs to the model Such compromises occur in at least two forms One involves using overly aggregate forecast inputs, resulting in likely aggregation biases in the forecasts The other involves developing more aggregate models in the first place, so as to reduce the need for disaggregate forecast input data, thereby building the aggregation bias into the model itself A strong case can be made that a primary reason for the relatively slow diffusion of disaggregate modeling methods into travel demand forecasting practice is due to the difficulty practitioners have in generating the disag-

Trang 7

gregate forecast inputs required by these methods As described in the previous section, lation in principle eliminates this problem by explicitly generating the detailed inputs required for each actor simulated.

microsimu-A second driving force for using microsimulation relates to the outputs required from the travel demand model As discussed further in Chapter 4, many emerging road network assignment procedures are themselves microsimulation based, and hence require quite microlevel inputs from the travel forecasting model

In addition, the disaggregate nature of the model outputs, in which behavior is explicitly attached to individual actors with known attributes, permits very detailed analysis of model results For example, the impacts of a given policy on specific subgroups (the elderly, the poor, suburban vs central area inhabitants, etc.) can be readily identified, since the disaggregate model outputs can, in principle, be aggregated in almost any user-specified fashion Thus, questions of equity, distribution of cost and benefits (both spatially and socioeconomically), details of the nature of the behavioral responses (e.g., who travels less or more, who changes modes, etc.), etc., can all be explored within a microsimulation framework

in ways that are generally infeasible with more aggregate models

A third point is that, despite the obviously large computational requirements of a large microsimulation model, in many cases microsimulation is computationally more efficient than conventional methods for dealing with large-scale forecasting problems In particular, a micro list-based approach to storing large spatial databases is far more efficient than aggregate matrix-based approaches To illustrate this, consider

a simple example in which one might want to keep track of the number of workers by their place of residence, place of work, number of household automobiles, and total number of household members Further assume that there are 1000 traffic zones, 3 auto ownership levels (e.g., 0, 1, 2+), and 5 household size categories (e.g., 1, 2, 3, 4, 5+) To save this information in matrix format would require a four-dimensional matrix with a total of 1000 × 1000 × 3 × 5 = 15 × 106 data items Also note that a large number of the cells in this matrix will have the value zero, either because they are infeasible (or at least extremely unlikely; e.g., 2+ autos in a one-person household) or because one simply does not observe nonzero values for many cells (as will be the case for many origin–destination (O-D) pairs)

In a list-based approach, one record is created for each worker, with each record containing the worker’s residence zone, employment zone, number of household autos, and household size Thus, four data storage locations are required per worker, meaning that as long as there are less than (15 × 106)/4 = 3.75 × 106

workers in this particular urban area, the list-based approach will require less memory (or disk space) than the matrix-based approach to store the same information Obviously, as the number of worker attributes that need to be stored increase, the relative superiority of the list-based approach increases

The advantages of list-based data structures for large-scale spatial applications have been recognized for at least 30 years Aggregate urban simulation models such as NBER, for example, developed in the 1970s, used list-based data structures (see Ingram et al., 1972; Wilson and Pownall, 1976) The key point

to be made here with respect to microsimulation is that once one begins to think in list-based terms, the conceptual leap to microsimulation model designs is a relatively small one Or, turning it around, if one takes a microsimulation approach to model design, efficient list-based data structures quickly emerge as the natural way for storing information

Whether microsimulation possesses other inherent computational advantages relative to more gate methods is less clear Certainly one can advance the proposition that by working at the microlevel

aggre-of the individual decision maker, relatively simple, clear, and computationally efficient models aggre-of process can generally be developed Whether this efficiency in computing each actor’s activities translates into overall computation time savings relative to other approaches given the large number of actors being simulated remains to be seen It is, however, important to note that Harvey and Deakin (1996) report a major advantage of the STEP microsimulation model (discussed further below): it can provide model results in many applications much more quickly (both in terms of getting ready to do the model runs and in terms of the computational effort involved in running the model) than conventional four-stage modeling systems Vovsha etþal (2002) similarly argue that microsimulation is far more efficient than conventional methods for modeling very large urban areas such as the New York metropolitan area

Trang 8

A fourth argument in favor of microsimulation is that it raises the possibility of emergent behavior, that is, of predicting outcomes that are not “hardwired” into the model Simple examples of emergent behavior of relevance to this discussion might include the generation of single-parent households by

a demographic simulator as a result of more fundamental processes dealing with fertility and household formation and dissolution, or the prediction of unexpected activity–travel patterns by an activity-based model as a result of the occurrence within the simulation of certain combinations of household needs, constraints, etc

The importance of emergent behavior within travel demand forecasting is at least twofold First, it offers the potential for the development of parsimonious models in the sense that relatively simple (but fundamental) rules of behavior can generate very complex behavior This is an attractive property of any model for two reasons In practical terms, it implies computational efficiency At a more theoretical level, parsimony is an important criteria in the evaluation of any model: it is generally assumed that a model that can satisfactorily explain behavior with fewer variables, parameters, rules, etc., is preferred, all else being equal, over more complicated formulations Second, while all models are to at least some degree captive to past behavior through use of historical data to estimate model parameters, the potential for emergent behavior increases the likelihood of the model generating unanticipated outcomes, and hence for departures from the trend to occur

Finally, it may well be the case that microsimulation models will ultimately prove easier to explain or

“sell” to decision makers than more aggregate models Since microsimulation models are formulated at the level of individual actors (workers, homeowners, parents, etc.), relatively clear and simple “stories” can be told concerning what the model is trying to accomplish (e.g., the model estimates the out-of-home activities that a given household will undertake on a typical weekday, and when and where these activities will occur) to which laypeople can readily relate The technical details of the model’s imple-mentation typically will be very complex, but the fundamental conceptual design is, in most cases, surprisingly simple to convey to others

12.4 Object, Agents, and Cellular Automata

The purpose of microsimulation is to model the behavior of actors, or objects, in the real world In a travel-related microsimulation, these objects may include persons, households, vehicles, jobs, firms, dwelling units, etc The real world consists of these objects evolving and interacting over time; micro-simulation models attempt to emulate this evolution and interaction with as high a degree of fidelity as

is required or feasible in the given application

Object-oriented software systems employ object-oriented analysis, design, and programming niques These techniques were specifically developed to handle complex problems such as large-scale microsimulation In an object-oriented system, the program consists of many objects that have their own states and behaviors There is a one-to-one mapping between objects in the real world and objects

tech-in the simulated world The conceptual benefits of the one-to-one mapptech-ing between the real world and objects in an object-oriented system should be clear Every object in the real world is represented by a similar object in the simulated world — these objects are known as abstractions of their real-world counterparts The behavior of objects in an object-oriented program is given by a set of methods or member functions There is a member function corresponding to every behavior that the object exhibits For example, the decision to change jobs is a behavior that is part of the person object

Traditional programming methods (such as functional or procedural programming) are built on data flow diagrams and are not well suited for microsimulation applications These methods use an approach known as top-down design with stepwise refinement and require that the program begin with a single problem statement that can be refined over time Many complex real-world systems, however, do not consist of a single abstract problem; rather, they are comprised of a set of objects that interact in complex ways over time

Object-oriented techniques provide a solution that is conceptually cleaner and easier for noncomputer scientists to understand and validate By matching real-world objects with their synthetic counterparts,

Trang 9

researchers can focus on understanding the problem rather than programming the solution When dealing with a system as complex as a microsimulation of events in the real world, any technique that will help

to reduce this inherent complexity will surely benefit the project Object-oriented systems are generally better equipped than procedural systems in solving problems where there is significant complexity in the problem domain

To give an example, we can think of the individual persons in the system These persons have certain behaviors that can easily be described In addition, there is information about each person that must be stored — this information might include age, gender, education level, marital status, job, etc Object-oriented systems allow us to encapsulate behavior and data together in the object The object’s behavior

is responsible for changing its state For example, a person will have the behavior that he or she ages over time This aging process is a behavior that is programmed into the person object that updates the data corresponding to the individual’s age Other behaviors are responsible for moving, finding jobs, making travel decisions, etc

Agents are objects of particular interest in microsimulation models in that they exhibit autonomous behavior that is typically the primary focus of the microsimulation Trip makers making travel decisions, households making residential location or auto ownership decisions, and firms hiring or firing workers are all examples of agents who independently make decisions as a function of their own attributes and

of the state of the system that they find themselves within These actions of the agents, in turn, change the system state over time (congestion levels, housing prices and vacancy rates, unemployment rates, etc.) Multiagent simulation models are simply microsimulation models that focus on modeling auton-omous agent objects, often using specialized programming languages (e.g., SWARM) that have be spe-cifically developed to facilitate this type of modeling

Cellular automata are a very special form of object or agent, in which the agents exist within a regular spatial pattern Traffic zones, roadway links, and grid cells are all examples of objects that exist in a regular spatial pattern (i.e., in which one object’s relationship or interaction with another object is determined

by their spatial relationship).3 In such cases, each object or cell can be modeled as an autonomous agent that interacts with other cells in a highly localized way (e.g., usually only with its immediate neighbors)

In such cases, very efficient (typically integer-based) algorithms can be developed to model cell (and hence system) behavior The TRANSIMS network microsimulation model is an example of the cellular automata approach In this model roads are divided into small segments, where each segment is approx-imately one car length in size Each road segment is a cell At each instant in time each cell is either occupied by a vehicle or not, and its interactions with other cells simply consist of receiving or not receiving a vehicle from the upstream cell, and of sending or not sending a vehicle to the downstream cell

12.5 Agent Attribute Synthesis and Updating

Microsimulation models by definition operate on a set of individual actors whose combined simulated behaviors define the system state over time As previously discussed, in short-run forecasting applications,

a representative sample may often exist that can define the set of actors whose behavior is to be simulated (Figure 12.1) In medium- and long-term forecasting applications, however, even if such a sample exists for the base year of the simulation, this sample cannot generally be assumed to remain representative over the forecast time period In such cases the microsimulation model must be extended to include methods for updating the attributes of the set of actors so that they continue to be representative at each point of time within the simulation (Figure 12.2) In addition, in many applications (particularly larger-scale, general-purpose regional modeling applications), the base year sample of actors either may not be available or may not be suitable for the task at hand In such cases, the microsimulation model must also include a procedure for synthesizing a suitable base year set of actors as input to the dynamic

Trang 10

behavioral simulation portion of the model (Figure 12.3) Each of these two processes — synthesis and updating — are discussed in the following two subsections.

Before discussing synthesis and updating methods, however, one other important model design issue needs to be addressed The discussion to this point has assumed that the set of actors being simulated

is a sample drawn in an appropriate way from the overall population Situations exist, however, in which

it may be useful or even necessary to work with the entire population of actors within the microsimulation, rather than a representative sample At least two major reasons exist for why one might prefer to work

at the population level rather than with a sample

First, situations exist in which computing population totals based on weighted sample results can be difficult to do properly Consider, for example, the problem of simulating residential mobility Assume that one is working with a 5% sample of households Then, on average, each household in the sample will carry

a weight of 20 in terms of its contribution to the calculation of population totals If it is determined within the simulation that a given sample household will move from its current zone of residence i to another zone j, does this imply that 20 identical households make the same move? The answer is probably not More complex weighting schemes can undoubtedly be devised, but it may prove to be conceptually simpler, more accurate, and perhaps even computationally more efficient to deal directly with the residential mobility decisions of every household and thereby avoid the weighting problem entirely

All sample-based models inherently represent a form of aggregation in that each observation in the sample stands for or represents n actual population members (where, as illustrated above, 1/n is the average sample rate) These n population members will possess at least some heterogeneity and hence variability

in behavior In many applications (microsimulation or otherwise) this aggregation problem is negligible, and the efficiency in working with a (small) sample of actors rather than the entire population is obvious

In many other applications, such as the one described above, however, use of a sample may introduce aggregation bias into the forecast unless considerable care (and associated additional computational effort)

is taken In such cases, the relative advantages of the two approaches are far less clear

Second, as one moves from short-run, small-scale, problem-specific applications to longer-run, scale, general-purpose applications (e.g., testing a wide range of policies within a regional planning context — presumably an eventual goal of at least some modeling efforts), the definition of what constitutes a representative sample becomes more ambiguous A sample that is well suited to one policy test or application may not be suitable for another This is particularly the case when one requires adequate representation spatially (typically by place of residence and place of work) as well as socioeconomically

larger-In such cases, a sufficiently generalized sample may be so large or sufficiently complex to generate that

it might be just as easy to work with the entire population

In trying to build a case for population-based microsimulations, one certainly cannot ignore the computational implications (in terms of processing time, memory, and data storage requirements) of such an approach Nevertheless, it is important to note that the conceptual case for population-based microsimulation does exist, in at least some applications; computing capabilities and costs are continu-ously improving, and several population-based models are currently operational or under development.The synthesis and updating methods discussed in the following subsections do not depend in any significant conceptual way on whether they are operating on a sample or the entire population For simplicity of discussion, however, the presentations in these sections assume that it is a disaggregated representation of the entire population that is being either synthesized or updated

Trang 11

across the entire set of attributes, however, is not known, although it may be the case that sample data (which provide specific attributes for the observed individuals) obtained from sources such as activ-ity–travel surveys, public use microdata sample (PUMS) files, etc., are available In such cases, these data provide an estimate of the joint population attribute distribution The synthesis task, as shown in Figure 12.4, is to generate a list of individual population units (in the case of Figure 12.4, households) that is statistically consistent with the available aggregate data.

All synthesis procedures developed to date use some form of Monte Carlo simulation to draw a realization of the disaggregate population from the aggregate data At least two general procedures for doing this currently exist The first appears to have been originally proposed by Wilson and Pownall (1976) In this method, the marginal and two-way aggregate distributions for a given zone (or census tract) are used sequentially to construct the specific attribute values for a given person (or household, etc.) living in this zone For example, assume that we are synthesizing households with three attributes,

X1, X2, and X3 Also assume that we have the marginal distribution for X1, which defines the marginal probabilities P(X1 = x1) for the various valid values x1 for this attribute We also have the joint distributions for X1 and X2 and for X2 and X3, which can be used to define the conditional probabilities P(X2 = x2|x1) and P(X3 = x3|x2) An algorithm for generating specific values (x1h, x2h, x3h) for household h is then:

1 Generate a uniform random number u1h on the range (0, 1) Given u1h, determine x1h from the distribution P(X1 = x1h)

2 Generate a uniform random number uh2 Given x1h and u2h, determine x2h from the distribution P(X2 = x2h|x1h)

3 Generate a uniform random number u3h Given x2h and u3h, determine x3h from the distribution P(X3 = x3h|x2h)

FIGURE 12.4 Population synthesis.

1 2 3 4 5+

1 2 3 4 5+

0 1 2 3+

1 2 3+

0 1 2+

No of Households by Household Size

of workers, and no of autos.

HH Size No of No of

1 4 2 2

2 5+ 1 1

3 2 2 1 .

.

The synthesis process converts the aggregate matrix-based data, combined with any small sample data available into a list of synthetic (but representative) households with specific attribute values.

Autos

Ngày đăng: 05/05/2018, 09:29