That is,when a user submits a query, the user does not need to be aware that multiple search engines may be used to process this query, and when the user receives thesearch result from t
Trang 1Yu WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:52 Char Count= 0
W EB S EARCH T ECHNOLOGY
750
document selector is to utilize the fact that most search
engines return retrieved results in groups Usually, only
the top 10 to 20 results are returned in the first result page
but the user can make additional requests for more result
pages and more results Hence, a document selector may
ask each search engine to return the first few result pages
This method tends to return the same number of pages
from each selected search engine Since different search
engines may contain different numbers of useful pages for
a given query, retrieving the same number of pages from
each search engine is likely to cause over-retrieval from
less useful databases and under-retrieval from highly
use-ful databases
More elaborate document selection methods try to tie
the number of pages to retrieve from a search engine to the
ranking score (or the rank) of the search engine relative
to the ranking scores (or ranks) of other search engines
This can lead to proportionally more pages to be retrieved
from search engines that are ranked higher or have higher
ranking scores This type of approach is referred to as a
weighted allocation approach in (Meng et al., 2002).
For each user query, the database selector of the
metasearch engine computes a rank (i.e., 1st, 2nd, .)
and a ranking score for each local search engine Both
the rank information and the ranking score information
can be used to determine the number of pages to retrieve
from different local search engines For example, in the
D-WISE system (Yuwono & Lee, 1997), the ranking score
information is used Suppose for a given query q, r i
de-notes the ranking score of the local database D i , i= 1, ,
k, where k is the number of selected local databases for
the query, andα =k
j=1r j denotes the total ranking score
for all selected local databases D-WISE uses the ratio r i/α
to determine how many pages should be retrieved from
D i More precisely, if m pages across these k databases
are to be retrieved, then D-WISE retrieves m ∗ r i/α pages
from database D i An example system that uses the rank
information to select documents is CORI Net (Callan
et al., 1995) Specifically, if m is the total number of pages
to be retrieved from k selected local search engines, then
for u < v, more pages will be retrieved from the uth ranked
database than from the vth ranked database Because
exactly m pages will be retrieved from the k top-ranked
databases In practice, it may be wise to retrieve slightly
more than mpages from local databases in order to reduce
the likelihood of missing useful pages
It is possible to combine document selection and
database selection into a single integrated process In
Database Selection, we described a method for ranking
databases in descending order of the estimated ity of the most similar document in each database for
similar-a given query A combined dsimilar-atsimilar-absimilar-ase selection similar-and
doc-ument selection method for finding the m most similar
pages based on these ranked databases was proposed in
Yu et al (1999) This method is sketched below First, for
some small positive integer s (e.g., s can be 2), each of the
stop-ranked databases are searched to obtain the actual
global similarity of its most similar page This may quire some locally top-ranked pages to be retrieved from
re-each of these databases Let min sim be the minimum of these s similarities Next, from these s databases, retrieve
all pages whose actual global similarities are greater than
or equal to min sim If m or more pages have been
re-trieved, then sort them in descending order of similarities,
return the top m pages to the user, and terminate this
pro-cess Otherwise, the next top ranked database (i.e., the
(s+ 1)th ranked database) is considered and its most ilar page is retrieved The actual global similarity of this
sim-page is then compared with the current min sim and the
minimum of these two similarities will be used as the
new min sim Then retrieve from these s+ 1 databasesall pages whose actual global similarities are greater than
or equal to the new min sim This process is repeated til m or more pages are retrieved and the m pages with
un-the largest similarities are returned to un-the user A ing problem with this combined method is that the samedatabase may be searched multiple times In practice, thisproblem can be avoided by retrieving and caching an ap-propriate number of pages when a database is searchedfor the first time In this way, all subsequent “interactions”with the database would be carried out using the cachedresults This method has the following property (Yu et al.,
seem-1999) If the databases containing the m desired pages are
ranked higher than other databases and the similarity (or
desirability) of the mth most similar (desirable) page is distinct, then all of the m desired pages will be retrieved
while searching at most one database that does not
con-tain any of the m desired pages.
Result Merging
Ideally, a metasearch engine should provide local systemtransparency to its users From a user’s point of view,such a transparency means that a metasearch searchshould behave like a regular search engine That is,when a user submits a query, the user does not need
to be aware that multiple search engines may be used
to process this query, and when the user receives thesearch result from the metasearch engine, he/she should
be hidden from the fact that the results are retrievedfrom multiple search engines Result merging is a nec-essary task in providing the above transparency Whenmerging the results returned from multiple search en-gines into a single result, pages in the merged resultshould be ranked in descending order of global similari-ties (or global desirabilities) However, the heterogeneitiesthat exist among local search engines and between themetasearch engine and local search engine make resultmerging a challenging problem Usually, pages returnedfrom a local search engine are ranked based on thesepages’ local similarities Some local search engines makethe local similarities of returned pages available to the
Trang 2Yu WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:52 Char Count= 0
user (as a result, the metasearch engine can also
ob-tain the local similarities) while other search engines
do not make them available For example, Google and
AltaVista do not provide local similarities while Northern
Light and FirstGov do To make things worse, local
simi-larities returned from different local search engines, even
when made available, may be incomparable due to the
use of different similarity functions and term-weighting
schemes by different local search engines Furthermore,
the local similarities and the global similarity of the same
page may be quite different still as the metasearch engine
may use a similarity function different from those used in
local systems In fact, even when the same similarity
func-tion were used by all local systems and the metasearch
engine, local and global similarities of the same page may
still be very different This is because some statistics used
to compute term weights, for example the document
fre-quency of a term, are likely to be different in different
systems
The challenge here is how to merge the pages returnedfrom multiple local search engines into a single ranked list
in a reasonable manner in the absence of local similarities
and/or in the presence of incomparable similarities An
additional complication is that retrieved pages may be
returned by different numbers of local search engines For
example, one page could be returned by one of the selected
local search engines and another may be returned by all of
them The question is whether and how this should affect
the ranking of these pages
Note that when we say that a page is returned by asearch engine, we really mean that the URL of the page
is returned One simple approach that can solve all of the
above problems is to actually fetch/download all returned
pages from their local servers and compute their global
similarities in the metasearch engine One metasearch
engine that employs this approach for result merging
is the Inquirus system (http://www.neci.nec.com/∼
lawrence/inquirus.html) Inquirus ranks pages returned
from local search engines based on analyzing the
con-tents of downloaded pages, and it employs a ranking
formula that combines similarity and proximity matches
(Lawrence & Lee Giles, 1998) In addition to being able
to rank results based on desired global similarities, this
approach also has some other advantages (Lawrence
& Lee Giles, 1998) For example, when attempting to
download pages, obsolete URLs can be discovered This
helps to remove pages with dead URLs from the final
result list In addition, downloading pages on the fly
ensures that pages will be ranked based on their current
contents In contrast, similarities computed by local
search engines may be based on obsolete versions of Web
pages The biggest drawback of this approach is its slow
speed as fetching pages and analyzing them on the fly
can be time consuming
Most result merging methods utilize the local ties or local ranks of returned pages to perform merging
similari-The following cases can be identified:
Selected Databases for a Given Query Do Not Share
Pages, and All Returned Pages Have Local Similarities
Attached. In this case, each result page will be returned
from just one search engine Even though all returned
pages have local similarities, these similarities may be malized using different ranges by different local search en-gines For example, one search engine may normalize itssimilarities between 0 and 1 and another between 0 and
nor-1000 In this case, all local similarities should be malized based on a common range, say [0, 1], to improvethe comparability of these local similarities (Dreilinger &Howe, 1997; Selberg & Etzioni, 1997)
renor-Renormalized similarities can be further adjustedbased on the usefulness of different databases for thequery Recall that when database selection is performedfor a given query, the usefulness of each database is esti-mated and is represented as a score The database scorescan be used to adjust renormalized similarities The idea
is to give preference to pages retrieved from highly rankeddatabases In CORI Net (Callan et al., 1995), the adjust-
ment works as follows Let s be the ranking score of cal database D and s be the average of the scores of all
lo-searched databases for a given query Then the following
weight is assigned to D : w = 1 + k * (s − s)/s, where k
is the number of databases searched for the given query
It is easy to see from this formula that databases with
higher scores will have higher weights Let x be the malized similarity of page p retrieved from D Then CORI Net computes the adjusted similarity of p by w * x The re-
renor-sult merger lists returned pages in descending order of justed similarities A similar method is used in ProFusion(Gauch et al., 1996) For a given query, the adjusted sim-
ad-ilarity of a page p from a database D is the product of the renormalized similarity of p and the ranking score of
D.
Selected Databases for a Given Query Do Not Share Pages, but Some Returned Pages Do Not Have Local Similarities Attached. Again, each result page will be re-turned by one local search engine In general, there aretwo types of approaches for tackling the result-mergingproblem in this case The first type uses the local rankinformation of returned pages directly to perform themerge Note that in this case, local similarities that may
be available for some returned pages would be ignored.The second type first converts local ranks to local simi-larities and then applies techniques described for the firstcase to perform the merge
One simple way to use rank information only for resultmerging is as follows (Meng et al., 2002) First, arrangethe searched databases in descending order of usefulnessscores Next, a round-robin method based on the databaseorder and the local page rank order is used to produce
an overall rank for all returned pages Specifically, inthe first round, the top-ranked page from each searcheddatabase is taken and these pages are ordered based on thedatabase order such that the page order and the databaseorder are consistent; if not enough pages have been ob-tained, the second round starts, which takes the secondhighest-ranked page from each searched database, ordersthese pages again based on the database order, and placesthem behind those pages selected earlier This process isrepeated until the desired number of pages is obtained
In the D-WISE system (Yuwono & Lee, 1997), the lowing method for converting ranks into similarities is
fol-employed For a given query, let r be the ranking score of
Trang 3Yu WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:52 Char Count= 0
W EB S EARCH T ECHNOLOGY
752
database D i , r min be the smallest database ranking score, r
be the local rank of a page from Di , and g be the converted
similarity of the page The conversion function is g= 1 −
(r − 1) * F i , where F i = r min /(m * r i ) and m is the number
of documents desired across all searched databases This
conversion has the following properties First, all locally
top-ranked pages have the same converted similarity, i.e.,
1 Second, F iis the difference between the converted
sim-ilarities of the jth and the ( j+ 1)th ranked pages from
database D i , for any j= 1, 2, Note that the distance is
larger for databases with smaller ranking scores
Conse-quently, if the rank of a page p in a higher rank database is
the same as the rank of a page pin a lower rank database
and neither p nor pis top-ranked, then the converted
sim-ilarity of p will be higher than that of p This property can
lead to the selection of more pages from databases with
higher scores into the merged result As an example,
con-sider two databases D1and D2 Suppose r1= 0.2, r2= 0.5,
and m = 4 Then r min = 0.2, F1= 0.25, and F2= 0.1 Thus,
the three top-ranked pages from D1will have converted
similarities 1, 0.75, and 0.5, respectively, and the three
top-ranked pages from D2will have converted similarities 1,
0.9, and 0.8, respectively As a result, the merged list will
contain three pages from D2and one page from D1
Selected Databases for a Given Query Share Pages. In
this case, the same page may be returned by multiple local
search engines Result merging in this situation is usually
carried out in two steps In the first step, techniques
dis-cussed in the first two cases can be applied to all pages,
regardless of whether they are returned by one or more
search engines, to compute their similarities for merging
In the second step, for each page p returned by
multi-ple search engines, the similarities of p due to multimulti-ple
search engines are combined in a certain way to
gener-ate a final similarity for p Many combination functions
have been proposed and studied (Croft, 2000), and some of
these functions have been used in metasearch engines For
example, the max function is used in ProFusion (Gauch
et al., 1996), and the sum function is used in MetaCrawler
(Selberg & Etzioni, 1997)
CONCLUSION
In the past decade, we have all witnessed the explosion
of the Web Up to now, the Web has become the largest
digital library used by millions of people Search engines
and metasearch engines have become indispensable tools
for Web users to find desired information
While most Web users probably have used search
en-gines and metasearch enen-gines, few know the technologies
behind these wonderful tools This chapter has provided
an overview of these technologies, from basic ideas to
more advanced algorithms As can be seen from this
chap-ter, Web-based search technology has its roots from text
retrieval techniques, but it also has many unique features
Some efforts to compare the quality of different search
engines have been reported (for example, see (Hawking,
Craswell, Bailey, & Griffiths, 2001)) An interesting issue is
how to evaluate and compare the effectiveness of different
techniques Since most search engines employ multiple
techniques, it is difficult to isolate the effect of a particular
technique on effectiveness even when the effectiveness ofsearch engines can be obtained
Web-based search is still a pretty young discipline, and
it still has a lot of room to grow The upcoming transition
of the Web from mostly HTML pages to XML pages willprobably have a significant impact on Web-based searchtechnology
ACKNOWLEDGMENT
This work is supported in part by NSF GrantsIIS-9902872, IIS-9902792, EIA-9911099, IIS-0208574,IIS-0208434 and ARO-2-5-30267
GLOSSARYAuthority page A Web page that is linked from hubpages in a group of pages related to the same topic
Collection fusion A technique that determines how
to retrieve documents from multiple collections andmerge them into a single ranked list
Database selection The process of selecting potentiallyuseful data sources (databases, search engines, etc.) foreach user query
Hub page A Web page with links to important ity) Web pages all related to the same topic
(author-Metasearch engine A Web-based search tool that lizes other search engines to retrieve information forits user
uti-PageRank A measure of Web page importance based onhow Web pages are linked to each other on the Web
Search engine A Web-based tool that retrieves tially useful results (Web pages, products, etc.) for eachuser query
poten-Result merging The process of merging documents trieved from multiple sources into a single ranked list
re-Text retrieval A discipline that studies techniques toretrieve relevant text documents from a documentcollection for each query
Web (World Wide Web) Hyperlinked documents ing on networked computers, allowing users to navi-gate from one document to any linked document
hid-Bruce Croft (Ed.), Advances in information retrieval:
Re-cent research from the Center for Intelligent Information Retrieval (pp 127–150) Dordrecht, The Netherlands:
Kluwer Academic
Callan, J., Connell, M., & Du, A (1999) Automatic
dis-covery of language models for text databases In ACM
SIGMOD Conference (pp 479–490) New York: ACM
Press
Trang 4Yu WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:52 Char Count= 0
Callan, J., Croft, W., & Harding, S (1992) The INQUERY
retrieval system In Third DEXA Conference, Valencia,
Spain (pp 78–83) Wien, Austria: Springer-Verlag.
Callan, J., Lu, Z., & Croft, W (1995) Searching
dis-tributed collections with inference networks In ACM
SIGIR Conference, Seattle (pp 21–28) New York: ACM
Press
Chakrabarti, S., Dom, B., Raghavan, P., Rajagopalan, S.,
Gibson, D., Kleinberg, J (1998) Automatic resourcecompilation by analyzing hyperlink structure and asso-
ciated text In 7th International World Wide Web
Confer-ence, Brisbane, Australia (pp 65–74) Amsterdam, The
Netherlands: Elsevier
Chakrabarti, S., Dom, B., Kumar, R., Raghavan, P.,
Rajagopalan, S., et al (1999) Mining the Web’s link
structure IEEE Computer, 32, 60–67.
Croft, W (2000) Combining approaches to information
retrieval In W Bruce Croft (Ed.), Advances in
infor-mation retrieval: Recent research from the Center for Intelligent Information Retrieval (pp 1–36) Dordrecht:
Kluwer Academic
Cutler, M., Deng, H., Manicaan, S., & Meng, W (1999)
A new study on using HTML structures to improve
retrieval In Eleventh IEEE Conference on Tools with
Artificial Intelligence, Chicago (pp 406–409)
Washing-ton, DC: IEEE Computer Society
Dreilinger, D., & Howe, A (1997) Experiences with
selecting search engines using metasearch ACM
Transactions on Information Systems, 15, 195–222.
Fan, Y., & Gauch, S (1999) Adaptive agents for
infor-mation gathering from multiple, distributed
informa-tion sources In AAAI Symposium on Intelligent Agents
in Cyberspace, Stanford University (pp 40–46) Menlo
Park, CA: AAAI Press
Gauch, S., Wang, G., & Gomez, M (1996) ProFusion:
Intelligent fusion from multiple, distributed search
engines Journal of Universal Computer Science, 2, 637–
649
Gravano, L., Chang, C., Garcia-Molina, H., & Paepcke,
A (1997) Starts: Stanford proposal for Internet
meta-searching In ACM SIGMOD Conference, Tucson,
AZ (pp 207–218) New York: ACM Press.
Hawking, D., Craswell, N., Bailey, P., & Griffiths, K (2001)
Measuring search engine quality Journal of
Informa-tion Retrieval, 4, 33–59.
Hearst, M., & Pedersen, J (1996) Reexamining the
clus-ter hypothesis: Scatclus-ter/gather on retrieval results In
ACM SIGIR Conference (pp 76–84) New York: ACM
Press
Kahle, B., & Medlar, A (1991) An information system for
corporate users: Wide area information servers (Tech.
Rep TMC199) Thinking Machine Corporation.Kirsch, S (1998) The future of Internet search: Infoseek’s
experiences searching the Internet ACM SIGIR Forum,
32, 3–7 New York: ACM Press.
Kleinberg, J (1998) Authoritative sources in a
hyper-linked environment In Ninth ACM-SIAM Symposium
on Discrete Algorithms (pp 668–677) Washington, DC:
ACM–SIAM
Koster, M (1994) ALIWEB: Archie-like indexing in the
Web Computer Networks and ISDN Systems, 27, 175–
182
Lawrence, S., & Lee Giles, C (1998) Inquirus, the NECi
meta search engine In Seventh International World
Wide Web Conference (pp 95–105) Amsterdam, The
Netherlands: Elsevier
Manber, U., & Bigot, P (1997) The search broker
In USENIX Symposium on Internet Technologies and
Systems, Monterey, CA (pp 231–239) Berkeley, CA:
USENIX
Meng, W., Yu, C., & Liu, K (2002) Building efficient and
effective metasearch engines ACM Computing Surveys,
34, 48–84.
Page, L., Brin, S., Motwani, R., & Winograd, T (1998)
The PageRank citation ranking: Bring order to the Web
(Technical Report) Stanford, CA: Stanford University.Pratt, W., Hearst, H., & Fagan, L (1999) A knowledge-based approach to organizing retrieved documents In
Sixteenth National Conference on Artificial Intelligence
(pp 80–85) Menlo Park, CA: AAAI Press and bridge, MA: MIT Press
Cam-Salton, G., & McGill, M (1983) Introduction to modern
information retrieval New York: McCraw-Hill.
Selberg, E., & Etzioni, O (1997) The MetaCrawler
ar-chitecture for resource aggregation on the Web IEEE
Expert, 12, 8–14.
Wu, Z., Meng, W., Yu, C., & Li, Z (2001) Towards ahighly scalable and effective metasearch engine In
Tenth World Wide Web Conference (pp 386–395) New
York: ACM Press
Yu, C., Meng, W., Liu, L., Wu, W., & Rishe, N (1999).Efficient and effective metasearch for a large number
of text databases In Eighth ACM International
Con-ference on Information and Knowledge Management
(pp 217–214) New York: ACM Press
Yuwono, B., & Lee, D (1997) Server ranking for tributed text resource systems on the Internet In
dis-Fifth International Conference on Database Systems for Advanced Applications (pp 391–400) Singapore:
World Scientific
Trang 5Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
Web Services
Akhil Sahai, Hewlett-Packard Laboratories Sven Graupner, Hewlett-Packard Laboratories Wooyoung Kim, University of Illinois at Urbana-Champaign
The Genesis of Web Services 754
Tightly Coupled Distributed Software
Convergence of the Two Independent Trends 755
Single Sign-On and Digital Passports 760Payment Systems for Web Services 762
The Future of Web Services 763Dynamic Web Services Composition and
End-to-End Web Service Interactions 764Future Web Services Infrastructures 765
There were two predominant trends in computing over
the past decade—(i) a movement from monolithic
soft-ware to distributed objects and components and (ii) an
increasing focus on software for the Internet Web
ser-vices (or e-serser-vices) are a result of these two trends
Web services are defined as distributed services that are
identified by Uniform Resource Identifiers (URI’s), whose
interfaces and binding can be defined, described, and
dis-covered by eXtensible Markup Language (XML) artifacts,
and that support direct XML message-based interactions
with other software applications over the Internet Web
services that perform useful tasks would often exhibit the
following properties:
Discoverable—The foremost requirement for a Web
ser-vice to be useful in commercial scenarios is that it be
discovered by clients (humans or other Web services)
Communicable—Web services adopt a message-driven
operational model where they interact with each other
and perform specified operations by exchanging XML
messages The operational model is thus referred to
as the Document Object Model (DOM) Some of
pre-eminent communication patterns that are being used
between Web services are synchronous, asynchronous,
and transactional communication
Conversational—Sending a document or invoking a
met-hod, and getting a reply are the basic communication
primitives in Web services A sequence of the
primi-tives that are related to each other (thus, conversation)
forms a complex interaction between Web services
Secure and Manageable—Properties such as security,
re-liability, availability, and fault tolerance are critical for
commercial Web services as well as manageability and
quality of service
As the Web services gain critical mass in the informationtechnology (IT) industry as well as academia, a dominantcomputing paradigm of that of software as a monolithicobject-oriented application is gradually giving way to soft-ware as a service accessible via the Internet
THE GENESIS OF WEB SERVICES
Contrary to general public perception, the development ofWeb services followed a rather modest evolutionary path.The underpinning technologies of Web services borrowheavily from object-based distributed computing and de-velopment of the World Wide Web (Berners-Lee, 1996)
In the chapter, we review related technologies that helpshape the notion of Web services
Tightly Coupled Distributed Software Architectures
The study of various aspects of distributed computing can
be dated back as early as the invention of time-shared tiprocessing Despite the early start, distributed comput-ing remained impractical until the introduction of ObjectManagement Group’s (OMG) Common Object RequestBroker Architecture (CORBA) and Microsoft’s DistributedComponent Object Model (DCOM), a distributed ex-tension to the Component Object Model (COM) BothCORBA and DCOM create an illusion of a single machineover a network of (heterogeneous) computers and allowobjects to invoke remote objects as if they were on thesame machine, thereby vastly simplifying object sharingamong applications They do so by building their abstrac-tions on more or less OS- and platform-independent mid-dleware layers In these software architectures, objects de-fine a number of interfaces and advertise their services
mul-by registering the interfaces Objects are assigned fiers at the time of creation The identifiers are used for
identi-754
Trang 6Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
discovering their interfaces and their implementations In
addition, CORBA supports discovery of objects using
de-scriptions of the services they provide Sun Microsystems’
Java Remote Method Invocation (Java RMI) provides a
similar functionality, where a network of platform-neutral
Java virtual machines provides the illusion of a single
ma-chine Java RMI is a language-dependent solution, though
the Java Native Interface (JNI) provides language
inde-pendence to some extent
The software architectures supported by CORBA and
DCOM are said tightly coupled because they define their
own binary message encoding, and thus objects are
inter-operable only with objects defined in the same software
architecture; for example, CORBA objects cannot invoke
methods on DCOM objects Also, it is worth noting that
security was a secondary concern in these software
archi-tectures—although some form of access control is highly
desirable—partly because method-level/object-level
ac-cess control is too fine-grained and incurs too much
over-head, and partly because these software architectures
were developed for use within the boundary of a single
administrative domain, typically a local area network
Loosely Coupled Distributed
Software Architectures
Proliferation and increased accessibility of diverse
intel-ligent devices in today’s IT market have transformed the
World Wide Web to a more dynamic, pervasive
environ-ment The fundamental changes in computing landscape
from a static client-server model to a dynamic peer-to-peer
model encourage reasoning about interaction with these
devices in terms of more abstract notion of service rather
than a traditional notion of object For example, printing
can be viewed as a service that a printer provides;
print-ing a document is to invoke the print service on a printer
rather than to invoke a method on a proxy object for a
printer
Such services tend to be dispersed over a wide area,often crossing administrative boundaries, for better re-
source utilization This physical distribution calls for
more loosely coupled software architectures where
scal-able advertising and discovery are a must and low-latency,
high-bandwidth interprocessor communication is highly
desirable As a direct consequence, a number of
service-centric middleware developments have come to light
We note three distinctive systems from computer
in-dustry’s research laboratories, namely, HP’s client utility
(e-Speak), Sun Microsystems’ Jini, and IBM’s TSpaces
(here listed in the alphabetic order) These have been
im-plemented in Java for platform independence
Client Utility
HP’s client utility is a somewhat underpublicized system
that became the launching pad for HP’s e-Speak (Karp,
2001) Its architecture represents one of the earlier forms
of peer-to-peer system, which is suitable for Web service
registration, discovery, and invocation (Kim, Graupner, &
Sahai, 2002) The fundamental idea is to abstractly
repre-sent every element in computing as a uniform entity called
“service (or resource).” Using the abstraction as a building
block, it provides facilities for advertising and discovery,
dynamic service composition, mediation and ment, and capability-based fine-grain security What dis-tinguishes client utility most from the other systems is thefact that it makes advertisement and discovery visible toclients Clients can describe their services using vocabu-laries and can specifically state what services they want todiscover
manage-Jini
The Jini technology at Sun Microsystems is a set of tocol specifications that allows services to announce theirpresence and discover other services in their vicinity It ad-vocates a network-centric view of computing However,
pro-it relies on the availabilpro-ity of multicast capabilpro-ity, tically limiting its applicability to services/devices con-nected with a local area network (such as home network).Jini exploits Java’s code mobility and allows a service to ex-port stub code which implements a communication proto-col using Java RMI Joining, advertisement, and discoveryare done transparently from other services It has been de-veloped mainly for collaboration within a small, trustedworkgroup and offers limited security and scalability sup-ports
prac-TSpaces
IBM’s TSpaces (TSpaces, 1999) is network middlewarethat aims to enable communication between applicationsand devices in a network of heterogeneous computers andoperating systems It is a network communication bufferwith database capabilities, which extends Linda’s Tuplespace communication model with asynchrony TSpacessupports hierarchical access control on the Tuple spacelevel Advertisement and discovery are implicit in TSpacesand provided indirectly through shared Tuple spaces
Convergence of the Two Independent Trends
Web services are defined at the cross point of the evolutionpaths of service-centric computing and the World WideWeb The idea is to provide service-centric computing byusing the Internet as platform; services are delivered overthe Internet (or intranet) Since its inception, the WorldWide Web has strived to become a distributed, decentra-lized, all pervasive infrastructure where information is putout for other users to retrieve It is this decentralized,distributed paradigm of information dissemination thatupon meeting the concept of service-centric computinghas led to the germination of the concept of Web services.The Web services paradigm has caught the fancy of theresearch and development community Many computerscientists and researchers from IT companies as well asuniversities are working together to define concepts, plat-forms, and standards that will determine how Web ser-vices are created, deployed, registered, discovered, andcomposed as well as how Web services will interact witheach other
WEB SERVICES TODAY
Web services are appearing on the Internet in theform of e-business sites and portal sites For example,
Trang 7Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
W EB S ERVICES
756
priceline.com (http://www.priceline.com) and Expedia
com (http://www.expedia.com) act as a broker for airlines,
hotels, and car rental companies They offer through their
portal sites statically composed Web services that have
prenegotiated an understanding with certain airlines and
hotels These are mostly a business-to-consumer (B2C)
kind of Web services A large number of technologies
and platforms have appeared and been standardized so
as to enable the paradigm of Web services to support
business-to-business (B2B) and B2C scenarios alike in a
uniform manner These standards enable creation and
de-ployment, description, and discovery of Web services, as
well as communication amongst them We describe some
preeminent standards below
The Web Services Description Language (WSDL) is a
standard to describe service interfaces and publish them
together with services’ access points (i.e., bindings) and
supported interfaces Once described in WSDL, Web
ser-vices can be registered and discovered using the
Univer-sal Description, Discovery, and Integration (UDDI)
Af-ter having discovered its partners, Web services use the
Simple Object Access Protocol (SOAP), which is in fact
an incarnation of the Remote Procedure Call (RPC) in
XML, over the HyperText Transfer Protocol (HTTP) to
ex-change XML messages and invoke the partners’ services
Though most services are implemented using
platform-independent languages such as Java and C#, development
and deployment platforms are also being standardized;
J2EE and NET are two well known ones Web services
and their users often expect different levels of security
depending on their security requirements and
assump-tion The primary means for enforcing security are
dig-ital signature and strong encryption using the Public
Key Infrastructure (PKI) SAML, XKMS, and XACML are
some of recently proposed security standards Also, many
secure payment mechanisms have been defined (See
Figure 1)
Web Services Description
In traditional distributed software architectures,
devel-opers use an interface definition language (IDL) to
de-fine component interfaces A component interface
typi-cally describes the operations the component supports by
specifying their inputs and expected outputs This enables
developers to decouple interfaces from actual
implemen-tations As Web services are envisaged as software
acces-sible through the Web by other Web services and users,
Figure 1: Web services.
Web services need to be described so that their interfacesare decoupled from their implementations WSDL serves
as an IDL for Web services
WSDL enables description of Web services dently of the message formats and network protocolsused For example, in WSDL a service is described as a set
indepen-of endpoints An endpoint is in turn a set indepen-of operations
An operation is defined in terms of messages received orsent out by the Web service:
Message—An abstract definition of data being cated consisting of message parts
communi-Operation—An abstract definition of an action supported
by the service Operations are of the following types:one-way, request–response, solicit–response, and noti-fication
Port type—An abstract set of operations supported by one
Service—A collection of related endpoints
As the implementation of the service changes or evolvesover time, the WSDL definitions must be continuouslyupdated and versioning the descriptions done
Web Services Discovery
When navigating the Web for information, we use keywords to find Web sites of interest through search engines.Often times, useful links in search results are mixed with
a lot of unnecessary ones that need to be sifted through.Similarly, Web services need to discover compatibleWeb services before they undertake business with them.The need for efficient service discovery necessitates somesort of Web services clearing house with which Webservices register themselves UDDI (http://www.uddi.org)supported by Ariba, IBM, Microsoft, and HP, is an ini-tiative to build such a Web service repository; it is nowunder the auspice of OASIS (http://www.oasis-open.org).These companies maintain public Web-based registries(operator sites) consistent with each other that makeavailable information about businesses and their techni-cal interfaces and application program interfaces (APIs)
A core component of the UDDI technology is tion, an XML document defining a business and the Webservices it provides There are three parts to the regis-
registra-tration, namely a white page for name, address, contact information, and other identifiers; a yellow page for clas-
sification of a business under standard taxonomies; and
a green page that contains technical information about
the Web services being described UDDI also lists a set ofAPIs for publication and inquiry The inquiry APIs are forbrowsing information in a repository (e.g., find business,get businessDetail) The publication APIs are for businessentities to put their information on a repository
E-marketplaces have been an important development
in the business transaction arena on the Internet Theyare a virtual meeting place for market participants(i.e., Web services) In addition to the basic registration
Trang 8Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
and discovery, e-marketplaces offer their participants a
number of value-added services, including the following:
Enabling inter-Web service interaction after the discovery
(the actual interaction may happen with or without thedirect participation of the e-marketplace);
Enabling supply and demand mechanisms through
tradi-tional catalogue purchasing and request for purchase(RFP), or through more dynamic auctions and ex-changes;
Enabling supply-chain management through
collabora-tive planning and inventory handling; andOther value-added services, such as rating, secured pay-
ment, financial handling, certification services, and tification services
no-Thus, e-marketplaces can be developed as an entity that
uses public UDDI registries The e-marketplaces are
cat-egorized as vertical and horizontal depending on their
target market The vertical e-marketplaces, such as
Ver-ticalNet, GlobalNetXChange, and Retailer Market
Ex-change, target a specific industry sector where
partici-pants perform B2B transactions In particular, Chemdex,
E-Steel, DirectAg.com, and many more have been
success-ful in their respective markets By contrast, horizontal
ex-changes, such as eBay, are directed at a broad range of
clients and businesses
Web Services Orchestration
By specifying a set of operations in their WSDL document,
Web services make visible to the external world a certain
subset of internal business processes and activities
There-fore, the internal business processes must be defined and
some of their activities linked to the operations before
publication of the document This in turn requires
mod-eling a Web service’s back-end business processes as well
as interactions between them On the other hand, Web
ser-vices are developed to serve and utilize other Web serser-vices
This kind of interaction usually takes a form of a sequence
of message exchanges and operation executions, termed
conversation Although conversations are described
inde-pendently of the internal flows of the Web services, they
result in executions of a set of backend processes A Web
service and its ensuing internal processes together form
what is called a global process.
Intra-Web Service Modeling and Interaction
The Web Services Flow Language (WSFL) (Leymann,
2001), the Web Services Conversation Language (WSFL)
(W3C, 2002), the Web Service Choreography Interface
(WSCI) (BEA, 2002) and XLANG (Thatte, 2001) are some
of many business process specification languages for Web
services
WSFL introduces the notion of activities and flowswhich are useful for describing both local business pro-
cess flows and global message flows between multiple Web
services WSFL models business processes as a set of
ac-tivities and links An activity is a unit of useful work while
a link connects two activities A link can be a control link
where a decision of what activity to follow is made, or a
data link specifying that a certain datum flows from an
activity to another These activities may be made visiblethrough one or more operations grouped as endpoints As
in WSDL, a set of endpoints defines a service WSFL fines global message flows in a similar way A global flowconsists of plug links that link up operations of two ser-vice providers Complex services involving more than twoservice providers are created by recursively defining pluglinks
de-XLANG developed by Microsoft extends the XMLSchema Definition Language (XSDL) to provide a mecha-nism for process definition and global flow coordination.The extension elements describe the behavioral aspects
of a service A behavior may span multiple operations.Action is an atomic component of a behavior definition
An action element can be an operation, a delay element,
or a raise element A delay element can be of type
de-layFor or delayUntil dede-layFor and delayUntil introducedelays in execution for a process to wait for something
to happen (for example, a timeout) and to wait till anabsolute date-time has been reached, respectively Raiseelements are used to specify exception handling Excep-tions are handled by invoking the corresponding handlerregistered with a raise definition Finally, processes com-bine actions in different ways: some of them are sequence,switch, while, all, pick, and empty
Inter-Web Service Modeling and Interaction
Web services must negotiate and agree on a protocol inorder to engage in a business transaction on the Web.X-EDI, ebXML, BTP, TPA-ML, cXML, and CBL have beenproposed as an inter-Web service interaction protocol Wefocus on ebXML as it is by far the most successful one.(See Figure 2.)
In ebXML (http://www.ebxml.org/) parties to engage in
a transaction have Collaboration Protocol Profiles (CPP’s)that they register at ebXML registries A CPP contains thefollowing:
Process Specification Layer—Details the business tions that form the collaboration It also specifies theorder of business transactions
transac-Delivery Channels—Describes a party’s message receivingand sending characteristics A specification can con-tain more than one delivery channel
A
C B
X
Y
Z
Pt o2 o3
Pt o5
o7
C B
Trang 9Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
W EB S ERVICES
758
Document Exchange Layer—Deals with processing of the
business documents like digital signatures, encryption,
and reliable delivery
Transport Layer—Identifies the transport protocols to be
used with the endpoint addresses, along with other
properties of the transport layer The transport
proto-cols could be SMTP, HTTP, and FTP
When a party discovers another party’s CPP they
ne-gotiate certain agreement and form a Collaboration
Pro-tocol Agreement (CPA) The intent of the CPA is not to
expose the business process internals of the parties but
to make visible only the processes that are involved in
interactions between the parties Message exchange
be-tween the parties can be facilitated with the ebXML
Mes-saging Service (ebMS) A CPA and the business process
specification document it references define a conversation
between parties A typical conversation consists of
mul-tiple business transactions which in turn may involve a
sequence of message exchanges for requests and replies
Although a CPA may refer to multiple business process
specification documents, any conversation is allowed to
involve only a single process specification document
Con-ceptually, the B2B servers of parties involved are
respon-sible for managing CPAs and for keeping track of the
conversations They also interface the operations defined
in a CPA with the corresponding internal business
pro-cesses
Web Services Platforms
Web services platforms are the technologies, means, and
methods available to build and operate Web services
Plat-forms have been developed and changed over the course
of time A classification into four generations of platform
technology should help to structure the space:
First Generation: HTML and CGI—Characterized by Web
servers, static HTML pages, HTML FORMS for simple
dialogs, and the Common Gateway Interface (CGI) to
connect Web servers to application programs, mostly
Perl or Shell scripts (See Figure 3.)
Second Generation: Java—Server-side dynamic
genera-tion of HTML pages and user session support; the Java
servlet interface became popular for connecting to
ap-plication programs
Third Generation: Application server as Richer
develop-ment and run-time environdevelop-ments—J2EE as foundation
for application servers that later evolved towards the
fourth generation
Service
Service B
CPP
CPA
ebXML registry
Figure 3: ebXML service-to-service interaction.
front-end web server app server back-end
DB
DB
Figure 4: Basic four-tier architecture for Web services.
Fourth Generation: Web services—Characterized by theintroduction of XML and WSDL interfaces for Webservices with SOAP-based messaging A global serviceinfrastructure for service registration and discoveryemerged: UDDI Dynamic Web services aggregation—Characterized by flow systems, business negotiations,agent technology, etc
Technically, Web services have been built according to a
pattern of an n-tier architecture that consists of a
front-end tier, firewall (FW), load balancer (LB), a Web-servertier (WS), an application (server) (AS) tier, and a back-end tier for persistent data, or the database tier (DB) (SeeFigure 4.)
First Generation: HTML and CGI
The emergence of the World Wide Web facilitated theeasy access and decent appearance of linked HTML mark-
up pages in a user’s browser In the early days, it wasmostly static HTML content Passive information servicesthat provided users with the only capability of naviga-ting though static pages could be built However, HTMLsupported from the very beginning FORMS that allowedusers to enter text or select from multiple-choice menus.FORMS were treated specially by Web servers They werepassed onto CGI, behind which small applications, mostlyPerl or Shell scripts, could read the user’s input, performrespective actions, and return a HTML page that couldthen be displayed in the user’s browser This primitivemechanism enabled a first generation of services on theWeb beyond pure navigation through static contents
Second Generation: Java
With the growth of the Web and the desire for richer vices such as online shopping and booking, the initialmeans to build Web services quickly became too primi-tive Java applets also brought graphical interactiveness tothe browser side Java appeared as the language of choicefor Web services Servlets provided a better interface be-tween the Web server and the application Technology tosupport dynamic generation of HTML pages at the serverside was introduced: JSP (Java Server Pages) by Sun Mi-crosystems, ASP (Active Server Pages) by Microsoft, orPHP pages in the Linux world enabled separation of pre-sentation, the appearance of pages in browsers, from con-tent data Templates and content were then merged onthe fly at the server in order to generate the final page re-turned to the browser Since user identification was crit-ical for business services, user log-in and user sessionswere introduced Applications were becoming more com-plex, and it turned out that there was a significant overlap
ser-in common functions needed for many services such assession support, connectivity to persistent databases, andsecurity functions
Trang 10Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
Figure 5: The J2EE platform.
Third Generation: Application Server
The observation that many functions were shared and
common among Web services drove the development
toward richer development environments based on the
Java language and Java libraries A cornerstone of these
environments became J2EE (Java 2 Platform, Enterprise
Edition), which is a Java platform designed for
enterprise-scale computing Sun Microsystems (together with
in-dustry partners such as IBM) designed J2EE (Figure 5)
to simplify application development for Web services by
decreasing the need for programming through reusable
modular components and by providing standard
func-tions such as session support and database
connecti-vity
J2EE primarily manifests in a set of libraries used byapplication programs performing the various functions
Web service developers still had to assemble all the pieces,
link them together, connect them to the Web server, and
manage the various configurations This led to the
emer-gence of software packages that could be deployed
eas-ier on a variety of machines These packages later
be-came application servers They significantly reduced the
amount of configuration work during service deployment
such that service developers could spend more time on
business logic and the actual function of the service Most
application server are based on J2EE technology
Exam-ples are IBM’s WebSphere suite, BEA’s WebLogic
environ-ment, the Sun ONE Application Framework, and Oracle’s
9i application server (See Figure 5.)
Fourth Generation: Web Services
Prior generations of Web services mostly focused on
end-users, people accessing services from Web browsers
How-ever, accessing services from services other than browsers
turned out to be difficult This circumstance has prevented
the occurrence of Web service aggregation for a long time
Web service aggregation meant that users would only have
to contact one Web service, and this service then would
resolve the user’s requests with further requests to other
Web services
HTML is a language defined for rendering and senting content in Web browsers It does not allow per se
pre-separating content from presentation information With
the advent of XML, XML became the language of choice
for Web services for providing interfaces that could not
only be accessed by users through Web browsers but also
by other services XML is now pervasively being used
in Web services messaging (mainly using SOAP) and forWeb service interface descriptions (WSDL) In regard toplatforms, XML enhancements were added to J2EE andapplication servers The introduction of XML is the majordifferentiator between Web services platforms of the thirdand the fourth generation in this classification
A major step toward the service-to-service integrationwas the introduction of the UDDI service (see the abovesection Web Services Discovery)
Three major platforms for further Web services teraction and integration are: Sun Microsystems’ SunONE (Open Net Environment), IBM WebSphere, and Mi-crosoft’s NET
in-Sun ONE—in-Sun’s standards-based software architectureand platform for building and deploying services ondemand Sun ONE’s architecture is built around exis-ting business assets: Data, applications, reports, andtransactions, referred to as the DART model Majorstandards are supported: XML, SOAP, J2EE, UDDI,LDAP, and ebXML The architecture is composed ofseveral product lines: the iPlanet Application Frame-work (JATO), Sun’s J2EE application framework forenterprise Web services development, application ser-ver, portal server, integration server, directory server,e-commerce components, the Solaris Operating Envi-ronment, and development tools
IBM WebSphere—IBM’s platform to build, deploy, andintegrate your e-business, including components such
as foundation and tools, reach and user experience,business integration, and transaction servers andtools
Microsoft NET—Microsoft’s NET platform for ing lead technology for future distributed applicationsinherently seen as Web services With Microsoft NET,Web services’ application code is built in discrete units,XML Web services, which handle a specified set oftasks Because standard interfaces based on XML sim-plify communication among software, XML Web ser-vices can be linked together into highly specific applica-tions and experiences The vision is that the best XMLWeb services from any provider around the globe can
provid-be used to create a needed solution quickly and easily
Trang 11Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
W EB S ERVICES
760
Microsoft will provide a core set of XML Web services,
called Microsoft NET My Services, to provide
func-tions such as user identification and calendar access
Security and Web Services
Due to their public nature, security is vital for Web
ser-vices Security attacks can be classified as threats of
infor-mation disclosure, unauthorized alteration of data,
de-nial of use, misuse or abuse of services, and, more rarely
considered, repudiation of access Since Web services
link networks together with businesses, further attacks,
such as masquerading, stealing or duplicating identity
and conducting business under false identity, or accessing
or transferring funds from or to unauthorized accounts,
need to be considered
Security is vital for establishing the legal basis for
businesses done over the Web Identification and
authen-tication of business partners are the basic security
re-quirements Others include integrity and authenticity of
electronic documents Electronic contracts must have the
same binding legal status as conventional contracts
Re-fusal and repudiation of electronic contracts must be
provable in order to be legally valid Finally, payment and
transferring funds between accounts must be safe and
se-cure
Security architectures in networks are typically
com-posed of several layers:
Secure data communication—IPsec (Internet Protocol
Security), SSL (Secure Socket Layer), TLS (Transport
Layer Security);
Secured networks—VPNs (Virtual Private Networks);
Authenticity of electronic documents and issuing
individuals—digital signatures;
Secure and authenticated access—digital certificates;
Secure authentication and certification—PKI (Public Key
Infrastructure); and
Single sign-on and digital passports
Single Sign-On and Digital Passports
Digital passport emerged from the desire to provide an
in-dividual’s identity information from a trusted and secure
centralized place rather then repeatedly establishing this
information with each collaborating partner and
main-taining separate access credentials for each pair of
collab-orations Individuals only need one such credential, the
passport, in order to provide collaborating partners with
certain parts of an individual’s identity information This
consolidates the need for maintaining separate identities
with different partners into a single identification
mech-anism Digital passports provide an authenticated access
to a centralized place where individuals have registered
their identity information such as phone numbers, social
security numbers, addresses, credit records, and payment
information Participating individuals, both people and
businesses, will access the same authenticated
informa-tion assuming trust to the authority providing the
pass-port service Two initiatives have emerged: Microsoft’s
.NET Passport and the Liberty Alliance Project, initiated
by Sun Microsystems
Microsoft NET Passport (Microsoft NET, 2002) is asingle sign-on mechanism for users on the Internet In-stead of creating separate accounts and passwords withevery e-commerce site, users only need to authenticatewith a single Passport server Then, through a series ofauthentications and encrypted cookie certificates, the user
is able to purchase items at any participating e-commercesite without verifying the user’s identity again .NET Pass-port is an online service that enables use of an e-mail ad-dress and a single (Passport server) password to securelysign in to any NET Passport participating Web site orservice It allows users to easily move among participat-ing sites without the need to verify their identity again.The Microsoft NET Passport had initially been plannedfor signing into Microsoft’s own services Expanding it to-ward broader use in the Web has been seen as critical.This concern gave reason for the Liberty Alliance Projectinitiative that is now widely supported in industry andpublic
The Liberty Alliance Project (Liberty Alliance Project,2002) is an organization being formed to create an open,federated, single sign-on identity solution for the digi-tal economy via any device connected to the Internet.Membership is open to all commercial and noncommer-cial organizations The Alliance has three main objec-tives:
1 To enable consumers and businesses to maintain sonal information securely
per-2 To provide a universal, open standard for single sign-onwith decentralized authentication and open authoriza-tion from multiple providers
3 To provide an open standard for network identity ning all network-connected devices
span-With the emergence of Web services, specific rity technology is emerging Two major security techno-logy classes are Java-based security technology and XML-based security technology
secu-Both classes basically provide mappings of securitytechnologies, such as authentication and authorization,encryption, and signatures, into respective environments
Java-Based Security Technology for Web Services
Java-based security technology is primarily availablethrough the Java 2 SDK and J2EE environments in theform of sets of libraries:
Encryption—JSSE (Java Secure Socket Extension); theJCE (Java Cryptography Extension) provides a frame-work and implementations for encryption, key gener-ation and key agreement, and Message AuthenticationCode (MAC) algorithms Support for encryption in-cludes symmetric, asymmetric, block, and stream ci-phers The software also supports secure streams andsealed objects
Secure messaging—Java GSS-API is used for securelyexchanging messages between communicating appli-cations The Java GSS-API contains the Java bindingsfor the Generic Security Services Application ProgramInterface (GSS-API) defined in RFC 2853 GSS-API
Trang 12Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
offers application programmers uniform access tosecurity services atop a variety of underlying securitymechanisms, including Kerberos
Authentication and Authorization—JAAS (Java
Authenti-cation and Authorization Service) for authentiAuthenti-cation
of users, to reliably and securely determine who is rently executing Java code, and for authorization ofusers to ensure they have the access rights (permis-sions) required to do security-sensitive operations
cur-Certification—Java Certification Path API
X.509 Certificates and Certificate Revocation Lists (CRLs)
and Security Managers
These libraries are available for use when Web services
are built using Java They are usually used when building
individual Web services with application servers
For Web services interaction, XML technology inates the tied binding to Java Consequently, a similar
elim-set of XML-based security technologies enabling
cross-service interactions is emerging
XML-Based Security Technology for Web Services
The Organization for the Advancement of Structured
Information Standards (OASIS) merges security into
Web services at a higher level than the common
Inter-net security mechanisms and practices described above
Proposals are primarily directed toward providing XML
specifications for documents and protocols suitable for
cross-organizational Web services interactions
XML-based security technology can be classified into the
fol-lowing:
XML Document-Level Security—encryption and digitally
signing XML documents;
Protocol-Level Security for XML Document Exchanges—
exchanging XML documents for authentication andauthorization of peers; and
XML-Based Security Frameworks—infrastructures for
establishing secure relationships among parties
XML Document-Level Security: Encryption and
Signature. The (preliminary) XML encryption
specifi-cation (Reagle, 2000) details requirements on how to
digitally encrypt a Web resource in general, and an XML
document in particular XML encryption can be applied
to a part of or complete XML document The granularity
of encryption can be reduced to an element, attributes,
or text content Encryption can be recursive The
specifi-cation does not address confidence or trust relationships
and key establishment The specification addresses both
key-encrypting-keys and data keys The specification will
not address the expression of access control policies
asso-ciated with portions of the XML document This will be
addressed by XACML
XML signature defines the XML schema and ing rules for creating and representing digital signatures
process-in any digital content (data object), process-includprocess-ing XML An
XML signature may be applied to the content of one
or more documents Enveloped or enveloping signatures
are over data within the same XML document as the
signature; detached signatures are over data external tothe signature element More specifically, this specificationdefines an XML signature element type and an XML sig-nature application; conformance requirements for eachare specified by way of schema definitions and prose re-spectively This specification also includes other usefultypes that identify methods for referencing collections ofresources, algorithms, and keying and management infor-mation
The XML Signature (Bartel, Boyer, Fox, LaMacchia,
& Simon, 2002) is a method of associating a key withreferenced data (octets); it does not normatively specifyhow keys are associated with persons or institutions, northe meaning of the data being referenced and signed.Consequently, while this specification is an importantcomponent of secure XML applications, it itself is not suf-ficient to address all application security/trust concerns,particularly with respect to using signed XML (or otherdata formats) as a basis of human-to-human communi-cation and agreement Such an application must specifyadditional key, algorithm, processing, and rendering re-quirements The SOAP Digital Signature Extensions de-fines how specifically SOAP messages can be digitallysigned
Protocol-Level Security for XML Document Exchanges. Protocol-level security defines documentexchanges with the purpose of establishing secure rela-tionships among parties, typically providing well-definedinterfaces and XML bindings to an existing public key in-frastructure Protocol-level security can be built upon thedocument-level security
The XML Key Management Specification (Ford et al.,2001) defines protocols for validating and registering pub-lic keys, suitable for use in conjunction with the pro-posed standard for XML signature developed by the WorldWide Web Consortium (W3C) and the Internet Engineer-ing Task Force (IETF) and an anticipated companion stan-dard for XML encryption The XML Key ManagementSpecification (XKMS) comprises two parts: the XML KeyInformation Service Specification (X-KISS) and the XMLKey Registration Service Specification (X-KRSS).The X-KISS specification defines a protocol for a trustservice that resolves public key information contained inXML-SIG document elements The X-KISS protocol al-lows a client of such a service to delegate part or all of thetasks required to process <ds:KeyInfo> elements embed-ded in a document A key objective of the protocol design
is to minimize the complexity of application tations by allowing them to become clients and therebyshielded from the complexity and syntax of the underlyingPublic Key Infrastructure (OASIS PKI Member Section,2002) used to establish trust relationships-based specifi-cations such as X.509/PKIX, or SPKI (Simple Public KeyInfrastructure, 1999)
implemen-The X-KRSS specification defines a protocol for a webservice that accepts registration of public key information.Once registered, the public key may be used in conjunc-tion with other web services including X-KISS
XML-Based Security Frameworks. XML-based curity frameworks go one step further than the above
Trang 13se-Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
W EB S ERVICES
762
The Security Assertion Markup Language (SAML),
de-veloped under the guidance of OASIS (OASIS, 2002), is an
XML-based framework for exchanging security
informa-tion with established, SAML-compliant security services
This security information is expressed in the form of
as-sertions about subjects, where a subject is an entity (either
human or program) that has an identity in some security
domain A typical example of a subject is a person,
iden-tified by his or her e-mail address in a particular Internet
DNS domain
Assertions can convey information about
authentica-tion acts performed by subjects, attributes of subjects,
and authorization decisions about whether subjects are
allowed to access certain resources Assertions are
repre-sented as XML constructs and have a nested structure,
whereby a single assertion might contain several
differ-ent internal statemdiffer-ents about authdiffer-entication,
authoriza-tion, and attributes Assertions containing authentication
statements merely describe acts of authentication that
happened previously
Assertions are issued by SAML authorities, namely,
au-thentication authorities, attribute authorities, and policy
decision points SAML defines a protocol by which
rely-ing parties can request assertions from SAML
authori-ties and get a response from them This protocol,
consist-ing of XML-based request-and-response message formats,
can be bound to many different underlying
communica-tions and transport protocols Currently it defines only
one binding, namely SOAP over HTTP
SAML authorities can use various sources of
informa-tion, such as external policy stores and assertions that
were received as input in requests, in creating their
re-sponses Thus, while clients always consume assertions,
SAML authorities can be both producers and consumers
of assertions
Payment Systems for Web Services
Effective payment systems are a prerequisite for business
with Web services This section introduces and classifies
different approaches for payment systems that have been
developed over the passed years However, payments in
the Internet are mostly conducted through the existing
payment infrastructure that was developed before the
In-ternet became pervasive End-consumer retail business on
the Internet primarily relies on credit card transactions
Other traditional payment methods are offered as well:
personal checks, money orders, or invoice billing In the
business-to-business segment, traditional invoice billing
is still the major payment method An overview is given
in (Weber, 1998) W3C has adopted payment standards
(Micropayment Overview, 2002)
Payments by Credit Cards
The reason why credit card payments are well accepted is
that credit card providers act as intermediaries between
payers and recipients of payments (payees) They do also
guarantee payments up to a limit (important to the payee),
and they carry the risk of misuse All parties must
regis-ter accounts before transfers can be conducted Another
important service is the verification of creditability of a
person or a business before opening an account
SET—The Secure Electronic Transaction Standard
SET (Secure Electronic Transaction, 2002) is an opentechnical standard for the commerce industry initiallydeveloped by two major credit card providers, Visa andMasterCard, as a way to facilitate secure payment cardtransactions over the Internet Digital certificates (DigitalCertificates, 1988) create a trust chain throughout thetransaction, verifying cardholders’ and merchants’ iden-tity SET is a system for ensuring the security of finan-cial transactions of credit card providers or bank acco-unts Its main objective is to provide a higher securitystandard for credit card payments on the Internet A ma-jor enhancement compared to traditional credit card pay-ments is that neither credit card credentials nor payers’identity are revealed to merchants With SET, a user isgiven an electronic wallet (digital certificate) A transac-tion is conducted and verified using a combination of digi-tal certificates and digital signatures among the purchaser,
a merchant, and the purchaser’s bank in a way that sures privacy and confidentiality
en-Not all payments required by Web services can be ducted through credit card transactions First, credit cardtransactions are typically directed from an end-customer,
con-a person, to con-a business thcon-at ccon-an receive such pcon-ayments.Second, the amounts transferred through a credit cardtransaction are limited to a range between currency equiv-alents of > $0.10 up to several thousand dollars depending
on an individual’s credit limits Micropayments <$0.10, aswell as macropayments> $10,000, are typically not pro-vided The lower payment bound is also caused by the costper transaction model credit card providers use Third,payments among persons, as for instance required forauctions among people or for buying and selling usedgoods, cannot be conducted through credit card accounts.Traditional payment methods are used here: personalchecks, money orders, or cash settlement Fourth, onlyindividuals with registered accounts can participate incredit card payments Individuals that do not qualify areexcluded This restriction is also a major barrier for Webservice business in developing countries
Micropayments
The purpose of micropayments is primarily for per-use” models where the usage is measured and im-mediately charged to customers in very small amounts.Transaction costs for micropayment systems need to besignificantly lower, and the number of transactions may
“pay-be significantly higher than that of credit card payments.Accurate, fine-grained charging is enabled These are thetwo major differentiators of micropayment systems W3Cproposes the Common Markup for Micropayment “per-fee-links.”
Micropayments involve a buyer or customer, a vendor
or merchant, and potentially one or more additional ties that keep accounts in order to aggregate micro pay-ments for final charge These mediators are called brokers(in Millicent), billing servers (in IBM MicroPayments),
par-or intermediaries (in France Telecom Micropayments), toname a few
Millicent. One micropayment system is Millicent(Glassman, 2000) The MilliCent Microcommerce
Trang 14Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
Network provides new pay-per-click/earn-per-click
func-tionality for Internet users It allows buying and selling
digital products costing from 1/10th of a cent to up to
$10.00 or more MilliCent can be used by Web services to
build any number of parallel revenue streams through the
simultaneous use of pay-per-click purchases,
subscrip-tions, and advertising It can also be used to make direct
monetary payments to users MilliCent is optimized for
buying and selling digital products or services over the
Internet such as articles, newsletters, real-time data,
strea-ming audio, electronic postage, video streams, maps,
financial data, multimedia objects, interactive games,
software, and hyperlinks to other sites
NetBill. NetBill is a Carnegie Mellon University
Inter-net billing server project, which is used as a payment
method for buying information goods and services via the
Internet It aims at secure payment for and delivery of
information goods, e.g., library services, journal articles,
and CPU cycles The NetBill system charges for
transtions and requires customers to have a prepaid NetBill
ac-count from which all payments are deducted The NetBill
payment system uses both symmetric key and public key
cryptography It relies on Kerberos for authentication An
account server, called NetBill server, maintains accounts
for both customers and merchants NetBill acts as an
ag-gregator to combine many small transactions into larger
conventional transactions, thus amortizing conventional
overhead fees Customers and merchants have to trust the
NetBill server
Digital Money and Digital Coins
In contrast to account-based payment systems, such as
credit card-based systems, where amounts are
trans-ferred between accounts inside or between credit card
or bank providers, digital money represents a value
amount flowing from a payer to a payee across the
network Establishing accounts with providers before
ser-vices can actually be used is unnecessary Advantages
are the same as for cash money: no mutual accounts
need to be established before a payment can be
con-ducted No mutual authentication is needed for
improv-ing convenience for both parties In addition, as with
cash money, the payer does not need to reveal any
identity credentials to the payee or someone else
Pay-ments are anonymous and nontraceable A major
hur-dle for this approach is the prevention of duplication
and forging of digital money since no physical security
marks such as watermarks can be applied to digitized
bit strings
The basic idea behind digital money is that a sumer purchases “digital coins” from an issuer using a
con-regular payment method such as a credit card The issuer
generates an account for that customer and deposits the
amount into it It then hands out a set of digital coins to
the customer that he or she can use for payments For a
payment, the customer transfers coins to the merchant or
service provider The provider then transfers coins to the
issuer and deposits them into his account The merchant,
however, may also use these coins to pay its suppliers
Dig-ital coins will thus flow among participants similarly like
cash money flows among people
The following requirements need to be met by digitalmoney systems:
digital money must be protected from duplication or ing; and
forg-digital money should neither contain nor reveal identitycredentials of any involved party in order to be anony-mous
The first requirement is achieved by not actually senting an amount by a digital coin, but rather a reference
repre-to an allocated amount in the possessor’s account withthe issuer When digital coins are copied, the reference
is copied, not the amount itself However, the first vidual redeeming a coin with the issuer will receive theamount Identity at redemption cannot be verified sincedigital coins do not carry identifying credentials of thepossessor The only term the issuer can verify is whether
indi-or not a coin has already been redeemed By thus, theft ofdigital money is possible, and parties have an interest inkeeping their coins protected
Achieving complete anonymity between an issuer andsubsequent receivers of digital money is a key characteris-tic of digital money It is basically achieved by blinded sig-natures (Chaum, 1985) that guarantee to uniquely assigncoins with allocated amounts within the issuer’s accountsystem and without revealing any identification informa-tion of the holder of that account
E-cash. E-cash (CrytoLogic Ecash FAQ, 2002) stands for
“electronic cash,” a system developed by DigiCash thatunderwent field tests in the late 1990s E-cash is a legalform of computer-generated currency This currency can
be securely purchased with conventional means: creditcards, checks, money orders, or wire transfers
MicroMint. MicroMint is a proposal by Rivest andShamir about coins that can only efficiently be produced
in very large quantities and are hard to produce in smallquantities The validity of a coin is easily checked Mi-croMint is optimized for unrelated low-value payments
It uses no public key operations However, the scheme isvery complex and would require a lot of initial and opera-tional efforts Therefore, it is unlikely that it ever will gainany practical importance
A broker will issue new coins at the beginning of aperiod and will revoke those of the prior period Coinsconsist of multiple hash collisions, i.e., different valuesthat all hash to the same value The broker mints coins bycomputing such hash collisions For that process manycomputations are required, but more and more hash col-lisions are detected with continued computation The bro-ker sells these MicroMint coins in batches to customers.Unused coins can be returned to the broker at the end of aperiod, e.g., a month Customers render MicroMint coins
as payment to merchants
THE FUTURE OF WEB SERVICES
In future we will see the unleashing of a Web servicesphenomenon This will involve the fulfillment of dynamicWeb service composition and orchestration vision, the ap-pearance of personalized Web services, concepts of Web
Trang 15Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
W EB S ERVICES
764
service management, and the development of Web service
infrastructure as a reusable, reconfigurable, self-healing,
self-managing, large-scale system
Dynamic Web Services Composition
and Orchestration
The vision of Web services intelligently interacting with
one another and performing useful tasks automatically
and seamlessly remains to become reality Major
mile-stones have been achieved: XML as a syntactic framework
and data representation language for Web services
inter-action; the Web infrastructure itself providing ubiquitous
access to Web services; the emergence of global
registra-tion and discovery services; and the technology to
sup-port the creation and maintenance of Web services, just
to name a few However, major pieces such as the
forma-lization and description of service semantic are yet to be
developed The effort of creating a semantic Web
(Se-mantic Web, 2001) is an extension of the current Web
in which information is given well-defined meaning,
bet-ter enabling compubet-ters and people to work in
coopera-tion Ontologies define the structure, relationships, and
meaning of terms appearing in service descriptions The
semantic Web vision is that these ontologies can be
reg-istered, discovered, and used for reasoning about Web
service selection before undertaking business Languages
like DAML+OIL (DAML, 2001) have been developed in
this context
In addition, sending a document or invoking a method
and getting a reply are the basic communication
prim-itives However, complex interactions between Web
ser-vices will involve multiple steps of communication that
are related to each other A conversation definition is a
sequencing of document exchanges (method invocations
in the network object model) that together accomplish
some business functionality In addition to agreeing upon
vocabularies and document formats, conversational Web
services also agree upon conversation definitions before
communicating with each other A conversation
defini-tion consists of descripdefini-tions of interacdefini-tions and
transi-tions Interactions define the atomic units of information
interchange between Web services Essentially, each
ser-vice describes each interaction in terms of the documents
that it will accept as input or will produce as output The
interactions are the building blocks of the conversation
definition Transitions specify the ordering amongst the
interactions Web services need to introspect other Web
services and obtain each other’s descriptions before they
start communicating and collaborating (Banerji et al.,
2002)
RosettaNet (RosettaNet, 2002) is a nonprofit
consor-tium of major information technology, electronic
com-ponents, and semiconductor manufacturing companies
working to create and implement industry-wide, open
e-business process standards, particularly targeting
busi-ness-to-business market places, workflow, and
supply-chain management solutions These standards form a
common e-business language, aligning processes between
supply-chain partners on a global basis Several examples
exist The centerpiece of the RosettaNet model is the
part-ner interface process (PIP) The PIP defines the activities,
decisions, and interactions that each e-business tradingparticipant is responsible for Although the RosettaNetmodel has been in development, it will be a while untilWeb services start using them to undertake business onthe Web
Once these hurdles are overcome, the basis and form for true Web services that will enable agent technolo-gies merging into Web services to provide the envisioneddynamic Web service aggregation on demand according
plat-to users’ specifications will emerge
Personalized Web Services
As Web service technology evolves, we anticipate thatthey will become increasingly sophisticated, and that thechallenges the Web service community will face will alsoevolve to meet their new capabilities One of the mostimportant of these challenges is the question of what itmeans to personalize Web services Personalization can
be achieved by using user profiles, i.e., monitoring userbehavior, devices, and context to customize Web services(Kuno & Sahai, 2002) for achieving metrics like quality ofexperience (QoE) (van Moorsel, 2001) This would involveproviding and meeting guarantees of service performance
on the user’s side Personalization could also result in thecreation of third-party rating agencies that will registeruser experiences, which could be informative for otherfirst-time users These rating mechanisms already exist in
an ad hoc manner, e.g., eBay and Amazon allow users torate sellers and commodities (books), respectively Salcen-tral.com and bizrate.com are third-party rating agenciesthat rate businesses These services could be also devel-oped as extended UDDI services These mechanisms willalso render Web services more “customer-friendly.”
End-to-End Web Service Interactions
Web services are federated in nature as they interactacross management domains and enterprise networks.Their implementations can be vastly different in nature.When two Web services connect to each other, they mustagree on a document exchange protocol and the appro-priate document formats (Austin, Barbir, & Garg 2002).From then on they can interoperate with each other,exchanging documents SOAP defines a common layerfor document exchange Services can define their ownservice-specific protocol on top of SOAP Often, these Webservice transactions will span multiple Web services A re-quest originating at a particular Web service can lead totransactions on a set of Web services For example, a pur-chase order transaction that begins when an employeeorders supplies and ends when he or she receives a con-firmation could result in 10 messages being exchangedbetween various services as shown in Figure 6
The exchange of messages between Web services could
be asynchronous Services sending a request messageneed not be blocked waiting for a response message Insome cases, all the participating services are like peers, inwhich case there is no notion of a request or a response.Some of the message flow patterns that result from thisasynchrony are shown in Figure 7 The first example inFigure 7 shows a single request resulting in multipleresponses The second example shows a broker-scenario,
Trang 16Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
7
6 9
8
1 purchase order
2 part of purchase order
3 the other part of the purchase order
10 purchase order confirmation
Figure 6: SOAP messages exchanged between Web services.
in which a request is sent to a broker but responses are
received directly from a set of suppliers
These Web services also interact with a complex web ofbusiness processes at their back-ends Some of these busi-
ness processes are exposed as Web service operations A
business process comprises a sequence of activities and
links as defined by WSFL and XLANG These business
processes must be managed so as to manage Web
ser-vice interactions Management of Web serser-vices thus is
a challenging task because of their heterogeneity,
asyn-chrony, and federation Managing Web services involves
managing business transactions by correlation of
mes-sages across enterprises (Sahai, Machiraju, & Wurster,
2001) and managing the business processes
Also, in order to manage business on the Web, userswill need to specify, agree, and monitor service level agree-
ments (SLAs) with each other Thus, Web services will
invariably have a large number of SLAs As less human
intervention is more desirable, the large number of SLAs
would necessitate automating the process as much as
pos-sible (Sahai, Machiraju, Sayal, Jin, & Casati, 2002)
Web service to Web service interaction managementcan also be done through mediation (Machiraju, Sahai,
& van Moorsel, 2002) Web service networks’ vision is to
mediate Web service interactions, so as to make it secure,
manageable, and reliable Such networks enable
version-ing management, reliable messagversion-ing, and monitorversion-ing of
message flows (e.g., Flamenco Networks, GrandCentral,
Transact Plus, Talking Blocks)
Future Web Services Infrastructures
Deployment and operational costs are determinants in
the balance sheets for Web service providers Web
of Web services (installation and configuration of ware and content data), the virtual wiring of machinesinto application environments independently of the physi-cal wiring in a data center They allow rearrangements ofWeb services’ applications among machines, the dynamicsizing of service capacities according to fluctuations in de-mands, and the isolation of service environments hosted
soft-in the same data center
HP’s Utility Data Center (HP Utility Data Center, 2001)
is such a platform The HP Utility Data Center with itsUtility Controller Software creates and runs virtual IT en-vironments as a highly automated service optimizing assetutilization and reducing staffing loads Resource virtual-ization is invisible to applications, sitting underneath theabstractions of operating systems
Two types of resources are virtualized:
Virtualized network resources, permitting the rewiring ofservers and related assets to create entire virtual IT en-vironments; and
Virtualized storage resources, for secure, effective age partitioning, and with disk images containing per-sistent states of application environments such as filesystems, bootable operating system images, and appli-cation software
stor-Figure 8 shows the basic building blocks of such a utilitydata center with two fabrics for network virtualizationand storage virtualization
The storage virtualization fabric with the storage areanetwork attaches storage elements (disks) to processingelements (machines) The network virtualization fabricthen allows linking processing elements together in a vir-tual LAN
Two major benefits for Web services management can
be achieved on top of the infrastructure:
Automated Web services deployment—By entirely taining persistent Web services’ states in the storagesystem and conducting programmatic control over
Trang 17main-Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
Network Virtualization
Dynamic capacity sizing of Web services—By the ability to
automatically launch additional service instances
ab-sorbing additional load occurring to the service
Ser-vice instances are launched by first allocating spare
machines from the pool maintained in the data center,
wiring them into the specific environment of the Web
service, attaching appropriate storage to those
ma-chines, and launching the applications obtained from
that storage Web server farms are a good example for
such a “breathing” (meaning dynamically adjustable)
configuration (Andrzejak, Graupner, Kotov, & Trinks,
2002; Graupner, Kotov, & Trinks, 2002)
IBM’s Autonomic Computing vision is to provide for
self-managing systems The intent is to create systems
that respond to capacity demands and system glitches
without human intervention These systems intend to
be configuring, healing, protecting, and
self-optimizing (IBM Autonomic Computing, 2002)
CONCLUSION
The Web services paradigm has evolved substantially
be-cause of concerted efforts by the software community The
genesis of Web services can be traced back to projects like
e-speak, Jini, and TSpaces Although progress has been
made in Web service standardization, the full potential of
Web services remains unrealized The future will see the
realization of Web services as a means of doing business
on the Web, the vision of dynamic composition of Web
services, personalized Web services, end-to-end
manage-ment of Web service interactions, and a dynamically
reusable service infrastructure that will adapt to
varia-tions in resource consumption
GLOSSARYBusiness process execution language for Web services (BPEL4WS) A standard business process descrip-tion language that combines features from WSFL andXLANG
Composition Creating composite Web services whenWeb services outsource their functionalities to otherWeb services
Conversation A set of message exchanges that can belogically grouped together
Description Describing Web services in terms of the erations and messages they support, so that they can
op-be registered and discovered at UDDI operator sites or
by using WS-Inspection
End-to-end management Protocol required to trackand manage Web service composition leading to atransaction being subdivided amongst multiple Webservices
Orchestration Web service to Web service interactionthat leads to the coupling of internal business pro-cesses
Personalization Personalizing or customizing Web vices to user/client profiles and requirements
ser-Platform One or more execution engines over which aWeb service implementation is executed
Service level agreement (SLA) An agreement thatspecifies quality-of-service guarantees between parties
Simple object access protocol (SOAP) A standard formessaging between Web services
Web service conversation language (WSCL) A guage to describe Web service conversations
lan-Web services flow language (WSFL) A language to scribe business processes
de-CROSS REFERENCES
See Client/Server Computing; Common Gateway Interface
(CGI) Scripts; Electronic Payment; Java; Perl; ization and Customization Technologies; Secure Electronic Transmissions (SET).
Personal-REFERENCES
Andrzejak, A., Graupner, S., Kotov, V., & Trinks, H (2002).Self-organizing control in planetary-scale computing
In IEEE International Symposium on Cluster
Comput-ing and the Grid (CCGrid), 2nd Workshop on based Cluster and Grid Computing (ACGC) New York:
Agent-IEEE
Austin, D., Barbir, A., & Garg, S (2002, 29 April) Web
services architecture requirements Retrieved November
2002 from 20020429
http://www.w3.org/TR/2002/WD-wsa-reqs-Banerji, A., Bartolini, C., Beringer, D., Chopella, V.,
Govin-darajan, K., Karp, A., et al (2002, March 14) WSCL
Web services conversation language Retrieved
Novem-ber 2002 from http://www.w3.org/TR/wscl10Bartel, M., Boyer, J., Fox, B., LaMacchia, B., & Si-
mon, E (2002, February 12) XML signature syntax
and processing Retrieved November 2002 from http://
www.w3.org/TR/2002/REC-xmldsig-core-20020212
Trang 18Sahai WL040/Bidgolio-Vol I WL040-Sample.cls July 16, 2003 18:35 Char Count= 0
BEA Systems, Intalio, SAP AG, and Sun Microsystems
(2002) Web Service Choreography Interface (WSCI)1.0 Specification Retrieved November 2002 fromhttp://wwws.sun.com/software/xml/developers/wsci
Berners-Lee, T (1996, August) The World Wide Web: Past,
present and future Retrieved November 2002 from
http://www.w3.org/People/Berners-Lee/1996/ppf.htmlChaum, D (1985) Security without identification: Trans-
action systems to make Big Brother obsolete
Commu-nications of the ACM, 28.
CryptoLogic Ecash FAQ (2002) Retrieved November 2002
from http://www.cryptologic.com/faq/faq-ecash.htmlDAML: The DARPA Agent Markup Language Home-
page (2001) Retrieved November 2002 from http://
www.daml.orgDigital Certificates, CCITT (1988) Recommendation
X.509: The Directory—Authentication Framework
ebXML: Enabling a global electronic market (2001)
Re-trieved November 2002 from http://www.ebxml.orgFord, W., Hallam-Baker, P., Fox, B., Dillaway, B., LaMac-
chia, B., Epstein, J., & Lapp, J (2001, March 30)
XML key management specification (XKMS) Retrieved
November 2002 from http://www.w3.org/TR/xkmsGlassman S., Manasse, M., Abadi, M., Gauthier P., Sobal-
varo, P (2000) The Millicent Protocol for InexpensiveElectronic Commerce Retrieved November 2002 fromhttp://www.w3.org/Conferences/WWW4/Papers/246/
Graupner, S., Kotov, V., & Trinks, H (2002)
Resource-sharing and service deployment in virtual data centers
In IEEE Workshop on Resource Sharing in Massively
Distributed Systems (RESH’02) New York: IEEE.
Hallam-Baker, P., & Maler, E (Eds.) (2002, March 29)
Assertions and protocol for the OASIS Security sertion Markup Language Retrieved November 2002
As-from http://www.oasis-open.org/committees/security/
docs/draft-sstc-core-29.pdfKarp, A., Gupta, R., Rozas, G., Banerji, A (2001) The
Client Utility Architecture: The Precursor to E-speak,
HP Technical Report Retrieved November 2002from http://lib.hpl.hp.com/techpubs/2001/HPL-2001-136.html
HP Utility Data Center: Enabling the adaptive
infrastruc-ture (2002, November) Retrieved November 2002
from http://www.hp.com/go/hpudcKim, W., Graupner, S., & Sahai, A (2002, January 7–
10) A secure platform for peer-to-peer computing inthe Internet Paper presented at 35th Hawaii Inter-national Conference on System Science (HICSS-35),Hawaii
Kuno, H., & Sahai, A (2002) My agent wants to talk to your
service: Personalizing Web services through agents
Re-trieved November 2002 from http://www.hpl.hp.com/
techreports/2002/HPL-2002-114
IBM Autonomic Computing (n.d.) Retrieved from http://
www.research.ibm.com/autonomic/
Leymann, F (Ed.) (2001) WSFL Web services flow
language (WSFL 1.0) Retrieved July 2003 from
http://www.ibm.com/software/solutions/webservices/pdf/WSFL.pdf
Liberty Alliance Project (2002) Retrieved November 2002from http://www.projectliberty.org/
Machiraju, V., Sahai, A., & van Moorsel, A (2002) Webservice management network: An overlay network forfederated service management Retrieved November
2002 from http://www.hpl.hp.com/techreports/2002/HPL-2002-234.html
Micropayments overview (2002) Retrieved ber 2002 from http://www.w3.org/ECommerce/Micropayments/
Novem-Microsoft NET Passport (2002) Retrieved November
Reagle, J (Ed.) (2000, October 6) XML encryption
require-ments Retrieved November 2002 from http://lists.w3.
org/Archives/Public/xml-encryption/2000Oct/att-0003/01-06-xml-encryption-req.html
RosettaNet (2002) Retrieved November 2002 fromhttp://www.rosettanet.org
Sahai, A., Machiraju, V., Sayal, M., Jin, L J., & Casati,
F (2002) Automated SLA monitoring for Web vices Retrieved November 2002 from http://www.hpl.hp.com/techreports/2002/HPL-2002-191.htmlSahai, A., Machiraju, V., & Wurster, K (2001, July).Monitoring and controlling Internet based services
ser-In Second IEEE Workshop on ser-Internet Applications
(WIAPP’01) New York: IEEE [Also as HP Tech Rep.
TSpaces: Intelligent Connectionware (1999) Retrieved
November 2002 from http://www.almaden.ibm.com/cs/TSpaces/
Van Moorsel, A (2001) Metrics for the Internet Age—
Quality of experience and quality of business
Re-trieved November 2002 from http://www.hpl.hp.com/techreports/2001/HPL-2001-179.html
Weber, R (1998) Chablis—Market analysis of digital
pay-ment systems Retrieved November 2002 from
Univer-sity of Munich Web site: muenchen.de/MStudy/x-a-marketpay.html
Trang 19http://chablis.informatik.tu-Irie WL040/Bidgoli-Vol III-Ch-62 June 23, 2003 16:43 Char Count= 0
Web Site Design
Robert E Irie, SPAWAR Systems Center San Diego
Consistent Navigation Mechanism 773
Miscellaneous Interface Issues 773
Designing and implementing a Web site is increasingly
be-coming a complex task that requires knowledge not only
of software programming principles but of graphical and
user interface design techniques as well While good
de-sign is important in regular software engineering and
ap-plication development, nowhere is it more essential than
in Web site development, due to the diverse and dynamic
nature of Web content and the larger intended audience
This chapter will cover some of the issues involved with
the two major components of a Web site, its design and
implementation The scope of this chapter is necessarily
limited, as Web development is a rich and heterogeneous
field A broad overview of techniques and technology is
given, with references to other chapters The reader is
di-rected to consult other chapters in this encyclopedia for
more detailed information about the relevant
technolo-gies and concepts mentioned below Occasionally links to
Web sites will be given They are either representative
ex-amples or suggestions for further reference, and should
not be construed as an endorsement
WEB SITE COMPONENTS
A Web site is an integration of three components, the
con-tent to be published on the Web, its presentation to the
user, and the underlying programming logic Each
com-ponent has its own particular representation and role in
shaping the overall user experience
Content
The content consists of all relevant data that are to be
pub-lished, or shown to the user It usually constitutes the bulk
of a Web site’s storage requirements and can be in the form
of text, images, binary and multimedia data, etc Statictextual and graphic content can be stored as HTML pages,whereas multimedia files like videos and sound recordingsare usually stored in large databases and served, in whole
or in parts, by dedicated servers Most of the discussion
in this chapter will focus on the former type
Presentation
The presentation component involves the user interface tothe Web site and the manner in which content is displayed.Typical elements include the graphical and structural lay-out of a Web document or page, text and graphic styles tohighlight particular content portions, and a mechanismfor the user to navigate the Web site Originally, files withHTML markups were used to store both content and in-formation regarding its presentation It is now commonpractice to store neither exclusively in HTML HTML isprimarily used to describe the structure of a Web docu-ment, by breaking down the page into distinct elementslike paragraphs, headings, tables, etc The actual textualcontent of the document can be stored separately in adatabase, to be dynamically inserted into the HTML pageusing programming logic A separate file, called the stylesheet, can be associated with the HTML page, and con-sequently the content, to affect the presentation A stylesheet file can describe how each structural element in anHTML file is displayed; sizes, colors, positions of fonts,blocks, backgrounds, etc., are all specified in a hierarchi-cal organization, using the standard Web-based style sheetlanguage, cascading style sheets (CSS) (Lie & Bos, 1999)
Logic
The programming logic determines which content todisplay, processes information entered by the user, and
768
Trang 20Irie WL040/Bidgoli-Vol III-Ch-62 June 23, 2003 16:43 Char Count= 0
generates new data It drives the interaction between the
Web site and the user and is the glue that binds the content
and its presentation To be useful, it needs to access the
content as well as its presentation information and
han-dle user input accordingly Logic is usually implemented
as small programs, or scripts, that are executed on the
Web server or the user’s browser These scripts can be
stored within the HTML page, along with the presentation
and content, or separately as distinct program files that
are associated with the content There are several
stan-dard programming languages that can be used in writing
scripts
Separation of Components
With the existence of a variety of technologies, protocols,
and standards, Web development is remarkably flexible,
and there are often multiple ways of accomplishing the
same task This is both an asset and a liability, as while
developers are free to choose their own techniques, it is
very easy to create sloppy or undisciplined documents and
code In regular application development, it is important
to adhere to sound software engineering techniques to
manage a code base for future enhancements and
simul-taneous development efforts The flexibility of Web
deve-lopment makes such good techniques even more critical
Until very recently, there was a great deal of overlap,
in terms of storage and implementation, of the three Web
site components mentioned above This led, for example,
to Web pages that contained all three components in a
single, often unmanageable, HTML file As Web site
de-velopment has matured, the principle of Web site
com-ponent separation has become widely encouraged, if not
accepted, and it is the central theme of this chapter
IMPLEMENTATION ISSUES
The World Wide Web (WWW) is a series of client/server
interactions, where the client is the user’s Web browser,
and the server is a particular Web site The WWW
Con-sortium (W3C) defines the hypertext markup language
(HTML) and the hypertext transfer protocol (HTTP) as
the standard mechanisms by which content is published
and delivered on the Web, respectively
In essence, the local Web browser initiates HTTP quests to the remote Web server, based on user input
re-The Web server retrieves the particular content specified
by the requests and transmits it back to the browser as
an HTTP response The Web browser then interprets the
response and renders the received HTML content into a
user-viewable Web page
Web site implementations can be classified by the level
of interactivity and the way content is stored, retrieved,
and displayed
Static Sites
Static sites are the simplest type of Web sites, with the
content statically stored in HTML files, which are simple
text files Updating the Web site requires manually
chang-ing individual HTML text files While this type of site was
prevalent in the beginning, most sites, especially
commer-cial ones, have come to incorporate at least some degree
Figure 1: Block diagram of a client/server
archi-tecture with a static Web site
of dynamic behavior, and users have come to expect someinteractivity
Figure 1 shows the basic client–server interaction for astatic Web site The client browser makes an HTTP request
to a Web server The URL specifies the particular Webserver and page The Web server retrieves the requestedWeb page, which is an HTML file, from the file system andsends it back to the client through the HTTP response.This very basic interaction between browser and server
is the basis for more complex, dynamic interactions Thistype is static because the Web page content is straightHTML, statically stored on disk, with no mechanism tochange the contents The Web server here serves solely as
a file transfer program
Developing static sites requires very few tools All that
is required, besides the Web server and browser, is a texteditor application The simplest text editor can be used tomanually create HTML files Complex, graphical HTMLeditors can make the task almost trivial by automaticallygenerating HTML files and behaving similarly to wordprocessors, with WYSIWYG (what you see is what youget) interfaces Creating graphics and images for staticWeb sites is also straightforward, requiring any typicalpaint or drawing program
DYNAMIC SITES
Dynamic sites share the same basic architecture as staticones, but with the addition of programming logic The twomajor types of dynamic sites reflect the place of execution
of the scripts Client-side scripting involves embedding tual source code in HTML, which is executed by the clientbrowser in the context of the user’s computer Server-sidescripts, on the other hand, are executed on the Web server.While the following discussion examines both types sep-arately, in an actual Web site both types can and often doexist simultaneously
ac-Client Side
Figure 2 shows the basic architecture for a dynamic sitewith client-side scripting Scripts are embedded withinHTML documents with the <script> </script> tags orstored in separate documents on the server’s file system.Scripts are transmitted, without execution, to the clientbrowser along with the rest of the HTML page Whenthe client browser renders the HTML page, it also in-terprets and executes the client script An example of aclient-side script is the functionality that causes a userinterface element, such as a menu, to provide visual feed-back when the user moves the mouse pointer over a menuoption
Trang 21Irie WL040/Bidgoli-Vol III-Ch-62 June 23, 2003 16:43 Char Count= 0
W EB S ITE D ESIGN
770
Figure 2: Block diagram of a Web site interaction
with client-side scripting
There are several client-side scripting languages, the
most common one being JavaScript, an object-oriented
language originally developed by Netscape It is now a
sta-ndardized language, defined by the international industry
group European Computer Manufacturers Association,
and called ECMAScript (European Computer
Manufac-turers Association, 1999) Netscape continues to use the
term JavaScript, however, and Microsoft calls its
imple-mentation of ECMAScript for Windows browsers JScript
The other major scripting language is Microsoft’s
VB-Script, short for Visual Basic Scripting Edition, which is
available only for Windows platforms (Champeon, 2001)
Regardless of the language, client-side scripts rely on a
standard programming interface, defined by the W3C and
called the Document Object Model (DOM), to dynamically
access and update the content, structure, and style of Web
documents (World Wide Web Consortium, 1998)
Cascading style sheets (CSS) is another W3C language
standard that allows styles (e.g., fonts, colors, and
spac-ing) to be associated with HTML documents Any specific
HTML tag or group of HTML tags can be modified It is
a language, separate from HTML, that expresses style in
common desktop publishing terminology The
combina-tion of HTML, CSS, and DOM client-side scripts is often
referred to as dynamic HTML (Lie & Bos, 1999)
Client-side scripting is used primarily for dynamic user
interface elements, such as pull-down menus and
ani-mated buttons The advantage of using client-side scripts
instead of server-side scripts for such elements is that
the execution is more immediate Since the script, once
loaded from the server, is being executed by the browser
directly on the user’s computer, there are no delays
asso-ciated with the network or the server load This makes the
user interface responsive and similar to standard platform
applications
One of the disadvantages is that client-side scripting
languages are usually more limited in functionality than
server-side languages, so that complex processing is not
possible Such limitations are by design, for security
rea-sons, and are not usually apparent for simple user
inter-face programming
Users may also specifically choose not to allow
client-side scripts to execute on their computers, resulting in a
partial or complete reduction in functionality and
usabil-ity of a Web site In general, it is recommended that a site
incorporate user interface scripting only sparingly, and
always with clear and functional alternatives
Finally, because client-side programs, whether
embed-ded or stored separately, must necessarily be accessible
and viewable by the Web browser, they are also
ulti-mately viewable by the user This may not be desirable for
Figure 3: Block diagram of a Web site interaction with
Server Side
Figure 3 shows the basic architecture for a server-side namic site Scripts are still stored in HTML documents onthe server’s file system, but are now executed on the server,with only the program results and output being sent to theclient browser, along with the rest of the HTML page Tothe client browser, the HTTP response is a normal staticHTML Web page Scripts are embedded in HTML docu-ments using special HTML-like tags, or templates, whosesyntax depends on the particular server-side scripting lan-guage (Weissinger, 2000)
dy-There are several common server-side scripting guages, including PHP, Active Server Pages (ASP), andJava Server Pages (JSP) The common gateway interface(CGI) is also a server-side scripting mechanism, wherebyneither the Web content nor the programming logic isstored in an HTML file A separate program, stored in thefile system, dynamically generates the content The Webserver forwards HTTP request information from the clientbrowser to the program using the CGI interface Theprogram processes any relevant user input, generates anHTML Web document and returns the dynamic content
lan-to the browser via the Web server and the CGI interface.This process is illustrated in Figure 4
Server-side scripting is used primarily for complex andtime-consuming programming logic tasks, where imme-diacy of response is not as critical as with user interfaceelements The advantage of using server-side scripts is thefreedom and computational power that is available on the
Figure 4: Block diagram of a Web site interaction
with common gateway interface scripting
Trang 22Irie WL040/Bidgoli-Vol III-Ch-62 June 23, 2003 16:43 Char Count= 0
server; server-side scripts do not have the same security
constraints as client-side scripts, and often have full
ac-cess to the server machine’s file system and resources The
user may not disable execution of such scripts, so that the
Web developer can reasonably expect that the Web site
will behave exactly the same regardless of user
configu-ration Finally, any proprietary server-side source code is
safely hidden from user view, as the client browser
re-ceives only the output of the script
Server-side scripts have the disadvantage of requiring
a request–response round trip between the client browser
and the server, which leads to slower response times
Server-side scripting languages normally interact sely with the Web server, which imposes some compatibil-
clo-ity constraints The choice of a Web server, particularly a
proprietary system, usually limits the selection of
server-side scripting languages, and vice versa
WEB APPLICATIONS
As a Web site becomes more complex, a robust and
effi-cient mechanism for the separation of content,
presenta-tion, and logic is necessary Web application servers are
Web sites that are more interactive, access large amounts
of data, and provide a rich functionality similar to that of
desktop applications Unlike desktop applications, where
all components are stored and executed on the same
com-puter, Web applications usually follow a three-tier
client/-server architecture (see Figure 5) consisting of the Web
browser, the Web server, and a database All content and
logic are stored in the database and are retrieved and
pro-cessed as necessary on the Web server The presentation
information can be embedded with the content or stored
as a separate style sheet on the database or the server
Usually a Web application server interfaces with a lational database, which stores data in rows of tables
re-Table 1 URLs of Various Web Resources
World Wide Web Consortium http://www.w3.org
Web Application Servers
BEA WebLogic http://www.beasys.com/products/weblogicIBM WebSphere http://www.ibm.com/software/webservers/appservMacromedia ColdFusion http://www.macromedia.com/software/coldfusionApache Jakarta http://jakarta.apache.org
The major disadvantage of developing with Web plication servers, besides the inherent complexity, is thenecessity of learning a nonstandard or proprietary server-side programming interface or language There are severalmajor Web application servers that support standard pro-gramming languages such as Java and C++, but each hasits own application programming interface (API) Table 1lists some of the popular commercial and open source ap-plication servers (see Web Resources)
ap-DESIGN ISSUES
Unlike implementation issues, which usually are forward to specify and quantify, design issues are muchmore subjective and are dependent on several factors, in-cluding the particular type of Web site and its purpose.Web site development efforts are often driven by con-flicting objectives and considerations, and a balance must
straight-be maintained straight-between business and financial concerns,
Trang 23Irie WL040/Bidgoli-Vol III-Ch-62 June 23, 2003 16:43 Char Count= 0
W EB S ITE D ESIGN
772
which often stress the commercial viability and
revenue-generating aspects of a Web site, and more user-centric
design concerns, which usually deal with usability issues
(Murray & Costanzo, 1999) Since the former are very
domain-specific, only the latter will be discussed in this
chapter In the discussion that follows, references to
sam-ple Web sites will be given
USABILITY ISSUES
The goal of designing a Web site with usability issues in
mind is to ensure that the users of the site find it usable
and useful Specifically, a Web site should be accessible,
appealing, consistent, clear, simple, navigable, and
forgiv-ing of user errors (Murray & Costanzo, 1999)
The first step in designing any Web site should be the
determination of the purpose of the site Too often the
rush to incorporate the latest Web technology or standard
prevents a thorough examination and determination of
the most important factor of the Web site, its intention or
purpose Most Web sites in essence are information
dis-semination mechanisms; their purpose is to publish
use-ful content to as wide an audience as possible Others also
have a commercial component, with the buying and
sell-ing of goods or services Still others foster a community
or group activity and are used as collaboration devices
The Web site’s purpose should drive the design and
im-plementation efforts A Web site advertising or describing
a company’s products will most likely need eye-catching
graphical designs and images A commerce site will need
to consider inventory mechanisms and secure
transac-tions A community site will need to solve problems
in-volving simultaneous collaboration of a distributed group
of users
It is also important to consider the intended audience
of a Web site There is a wide range in browser capabilities
and user technical competencies that must be taken into
account A Web site geared toward a younger, more
tech-nically inclined audience may contain highly interactive
and colorful designs, whereas a corporate site might want
to have a more professional, businesslike appearance It
is generally a good practice, if not essential, to consider
accessibility issues for all users, including those who do
not have access to high-end graphics-capable browsers
BASIC WEB SITE TYPES
Just as there are several implementation classifications
for Web sites, we can also classify them based on their
purpose Each type will lead to different choices in the
content, presentation, and logic components and require
emphasis on different usability issues A single Web site
may incorporate features of more than one basic type
News/Information Dissemination
This type of Web site design is geared toward providing
informational content to the Web user The content is
usu-ally textual in form, with some graphics or images The
presentation of the content and its navigation are kept as
clear and consistent as possible, so that the user will be
able to quickly access the desired information Not
sur-prisingly, newspaper companies usually have Web sites
with online news content (e.g., http:/./www.nytimes.com)
to incorporate more community-like features to promptusers to return to their sites (e.g., http://www.yahoo.com)
Community
Community sites foster interaction among their usersand provide basic collaboration or discussion capabili-ties Message boards, online chats, and file sharing areall typical functionalities of community sites The opensource software movement has promoted numerous Websites based on this type (e.g., http://www.sourceforge.net)
Search
There is a lot of overlap between this type of Web sites andportals Like portals, search sites provide a mechanism bywhich users discover other Web sites to explore Some so-phisticated programming logic, the search engine, formsthe foundation of this type of Web site Search sitesoften emphasize simple, almost minimalist interfaces(e.g., http://www.google.com)
E-commerce
This type of site is often a component of other Web sitetypes and allows users to purchase or sell goods and ser-vices in a secure manner Since potentially large amounts
of currency are involved, security is an important eration, as well as an interface that is tolerant of potentialuser errors An example of a successful commerce sitewith elements of a community is http://www.ebay.com
consid-Company/Product Information
With widespread Web use, having an official Web ence is almost a requirement for corporations Such sitesusually serve purposes similar to those of informationaland e-commerce sites, but with a more focused interface,reflecting the corporate image or logo (e.g., http://www.microsoft.com)
pres-Entertainment
This type of site is usually highly interactive and stressesappealing, eye-catching interfaces and designs Typicalapplications include online gaming sites, where users mayplay games with each other through the site, and sportingevent sites, where users may view streaming content inthe form of video or audio broadcasts of live events (e.g.,http://play.games.com)
BASIC DESIGN ELEMENTS
There is obviously no single best design for a Web site,even if one works within a single type There are, however,some guidelines that have gained common acceptance.Like any creative process, Web site design is a matter oftradeoffs A typical usability tradeoff is between making
Trang 24Irie WL040/Bidgoli-Vol III-Ch-62 June 23, 2003 16:43 Char Count= 0
an interface appealing and interactive and making it clear
and simple The former usually involves graphical designs
with animations and client-side scripting, whereas the
lat-ter favors minimal text-based inlat-terfaces Where a
particu-lar Web site belongs on the continuous spectrum between
the two extremes depends on its intended purpose and
au-dience, and should be a subjective, yet conscious decision
The safest design decision is to always offer tives, usually divided into high- and low-bandwidth ver-
alterna-sions of the same Web pages, so that the user experience
can be tailored to suit different preferences The major
disadvantage of this is the increase in development time
and management requirements
Accessibility/Connectivity
The two major factors affecting accessibility and
connec-tivity issues are the bandwidth of the user’s network
con-nection, and the particular graphical capabilities of the
user browser Low-bandwidth connections to the Internet
are still very common in homes By some measures, dialup
modems are still used in 90% of all homes that regularly
access the Internet (Marcus, 2001) This requires Web site
designers either to incorporate only small, compressed
images on their sites, or to provide alternative versions of
pages, for both high- and low-bandwidth users
Some user browsers do not have any graphics ity at all, for accessibility reasons or user preference For
capabil-example, visually impaired users and PDA (personal
digi-tal assistant) users most often require accessibility
consid-eration Estimates of the number of disabled users range
from 4 to 17% of the total online population (Solomon,
2000) PDA and mobile phone Internet usage is relatively
new in the United States, but is already approaching
10 million users (comScore Networks, 2002) For such
users, designing a separate text-only version of the Web
site is a possibility What would be better is to design a Web
site that contains automatic browser-specific
functiona-lity degradation An example is to associate relevant
tex-tual content to graphical images; graphical browsers may
display the images, while text browsers may display the
descriptions
Consistent Page Layout
One of the most important design for a Web site is a
con-sistent page layout While every single page does not need
to have the same layout, the more consistent each page
looks, the more straightforward it is for the user to
nav-igate through the site and the more distinctive the Web
site appears A typical Web page layout utilizes parts or
all of an artificially defined border around the content (see
Figure 6)
Originally, HTML frames or tables were the standardway of laying out a page, and they are still the preferred
method for most developers However, the W3C clearly is
favoring the use of cascading style sheets (CSS) for page
layout (World Wide Web Consortium, 2002) CSS also
pro-vides a mechanism for associating styles, such as color,
font type and size, and positioning, with Web content,
without actually embedding them in it This is in keeping
with the principle of separating content from its
presen-tation
Figure 6: A typical layout scheme for
a Web page
Consistent Navigation Mechanism
Web site navigation is an important component of the sign, and a consistent navigation mechanism supplements
de-a pde-age lde-ayout de-and mde-akes the user experience much pler and more enjoyable
sim-One of the best ways of providing a consistent gation mechanism is to have a menu bar or panel that
navi-is consnavi-istent across all pages of the site Such a menucan be a static collection of links, or a dynamic, interac-tive component similar to that of a desktop application.Figure 7 is an example of a simple and consistent naviga-tion scheme that utilizes two menu panels The top panel(with menu items A, B, C) is similar to a desktop appli-cation’s menu bar and is a global menu that is consistentthroughout the site and refers to all top-level pages of asite The left side panel is a context-dependent menu thatprovides further options for each top-level page This type
of navigation scheme can be seen on several public Websites (e.g., http://www.python.org)
While there are no absolute rules or guidelines for goodnavigation elements, they usually provide visual feedback(e.g., mouse rollover effects), have alternate text displays(for nongraphical or reduced capability browsers), andare designed to be easy to use as well as learn
MISCELLANEOUS INTERFACE ISSUES
The following are miscellaneous interface design issues.Again, only suggestions for possible design choices are of-fered, and they will not be applicable in all circumstances
Figure 7: An example of a consistent navigation scheme for
a Web site
Trang 25Irie WL040/Bidgoli-Vol III-Ch-62 June 23, 2003 16:43 Char Count= 0
W EB S ITE D ESIGN
774
Graphics
The two major interface issues concerning graphics are
size and color As the majority of Web users still access
Web sites over slow modem links, it is important to use
graphic images that are of reasonable size, to prevent
excessive download delays For photograph images, the
JPEG format offers a good compromise between lossy
compression size and image quality, with an adjustable
tradeoff point For line art and solid color images, lossless
compression is preferred, and the proprietary GIF format
is common, although the open standard PNG format is
gaining in acceptance (Roelofs, 2000)
The issue of colors is a complex one and depends on
many factors In general, Web graphic designers work
with 24-bit colors, with 256 possible values for each of
three color channels, red, green, and blue (RGB) Until
re-cently, the majority of users’ computers and Web browsers
could only support a palette, or set, of 256 colors
simul-taneously To ensure that colors appear uniformly across
platforms and browsers, a “Web-safe palette” of 216 colors
was established, consisting of combinations of six
pos-sible values, or points, for each of three color channels
(6 possible reds× 6 possible greens × 6 possible blues =
216 possible colors) (Niederst, 2001)
Recently, browsers and systems with 24-bit and 16-bit
support have drastically increased and now account for
about 94% of all users (Lehn & Stern, 2000)
Twenty-four-bit color support results in the full display of the designer’s
color scheme Sixteen-bit color systems are sometimes
problematic, as they nonuniformly sample the three color
channels (5 bits for red, 6 bits for green, and 5 bits for blue)
and provide a nonpalettized approximation of 24-bit color
Layout Styles
A comprehensive guide to layout styles is beyond the scope
of this chapter The major design decision is between
hav-ing page layouts of fixed or variable size (Niederst, 2001)
By default, HTML documents are variable-sized, in that
text and graphics positioning and line breaks are not
deter-mined by the user’s monitor resolution and browser
win-dow size Since a wide variety of sizes and resolutions is
almost a given, having a variable-sized page layout allows
flexible designs that scale to the capabilities and
prefer-ences of each user The disadvantage is that because each
user experience is different, and elements can be resized
or repositioned at will, it is difficult to design a consistent
and coherent interface; there is the possibility that some
configurations lead to poor or unusable interfaces
The alternative to the default variable-sized page
lay-out is to explicitly design the size and position of some
or all of the elements of a Web document An example of
this would be to limit the width of all content in a page
to fit within a certain horizontal screen resolution, such
as 640 pixels All text and graphics will remain stationary
even if the user resizes the browser window to greater
than 640 horizontal pixels The advantage of this method
is that designing an interface is much more deterministic,
so the Web designer will have some degree of control over
the overall presentation and is reasonably certain that all
users will have the same experience accessing the site
The disadvantage is that the designer must pick constants
that may not be pleasant or valid for all users For ample, a Web page designed for a 640× 480 resolutionscreen will look small and limited on a 1280× 1024 screen,whereas a Web page designed for an 800× 600 screenwould be clipped or unusable for users with only a 640×
ex-480 screen
Actually implementing either type of page layout can
be done with HTML frames, tables, or CSS style sheets,
or some combination of the three Although using stylesheets is the currently preferred method for page lay-out, browser support is still poor, and many sites still useframes or tables (Niederst, 2001)
Search Engines
A search engine is a useful tool to help users quicklyfind particular content or page as the content of a Website increases, or the navigation scheme becomes com-plicated The search engine is a server-side software pro-gram, often integrated with the Web server, that indexes
a site’s Web content for efficient and quick retrieval based
on a keyword or phrase Search engines are available with
a variety of configurations, interfaces, and capabilities
A good resource that summarizes the major commercialand open source engines is the Search Tools Web site(http://www.searchtools.com)
Cross-Browser Support
Designing a Web site that is consistent across multiplebrowsers and platforms is one of the most challengingtasks a developer faces Even different versions of thesame browser are sometimes incompatible At the mini-mum, the three major browsers to consider are InternetExplorer (IE), Netscape Navigator (NN), and text-basedbrowsers such as Lynx
For the most part, browser development and ties have preceded the establishment of formal standards
capabili-by the W3C, leading to numerous incompatibilities andnonuniform feature support The latest versions of the twocommon browsers (IE 6, NN 6.2) offer complete supportfor the current W3C standard HTML 4.01 However, themore common, earlier versions of the browsers (versions
4+ and 5+) had only incomplete support
Even more troublesome was their support of the W3Cstandard Document Object Model (DOM) Level 1, as eachhas historically taken a different track and implementedits own incompatible DOM features (Ginsburg, 1999)
In general, NN’s DOM support is much closer to the
“official” W3C DOM Level 1 specification, whereas IE hasseveral extensions that are more powerful, but are avail-able only on Windows platforms The latest versions of thetwo browsers have alleviated some of this problem by sup-porting, as a baseline, the complete Level 1 specification
WEB RESOURCES
Table 1 summarizes some useful online resources for Website development They are only suggestions and shouldnot be considered comprehensive or authoritative
CONCLUSION
This chapter has given an overview of Web site ment, including the design and implementation aspects
Trang 26develop-Irie WL040/Bidgoli-Vol III-Ch-62 June 23, 2003 16:43 Char Count= 0
This field is very dynamic, and technologies and practices
are constantly changing More complex object-oriented
programming paradigms and generalized markup
lan-guages are gaining widespread acceptance and use
XML (extensible markup language), XHTML (extensible
HTML), XML-RPC (XML remote procedure call), SOAP
(simple object access protocol), and SVG (scalable vector
graphics) are examples of such new standards However,
the basic principles of clarity, consistency, and
concise-ness are still applicable to the design of all sites regardless
of type or technology
The Web development profession is also rapidly ing field No longer is it feasible to have one person per-
chang-form all design and implementation duties A team of Web
developers, graphic designers, and database
administra-tors is usually required, with each member responsible for
the three components of Web site development: content
management, content presentation, and programming
logic However, it is still important to be aware of all
issues in order to work effectively in a Web development
team
GLOSSARY
Client/server architecture A process by which multiple
computers communicate The client initiates all munication with the server in the form of a request andreceives the results in the form of a response For Websites, the user’s browser is the client requesting content
com-or services from the Web server
Database A repository of information The data are
stored in a structured way to be easily and efficientlyretrieved Two popular types of databases are the rela-tional database and the object-oriented database Eachhas advantages and disadvantages with respect to effi-ciency, rich associations between information, etc
Hypertext A mechanism by which related content (text,
graphic, multimedia, etc.) is associated using links Ahypertext document allows the user to easily accessrelevant content in a seamless, integrated context, asopposed to traditional, sequentially viewed documents
Hypertext markup language (HTML) A standard
lan-guage for publishing content on the World Wide Web
HTML defines a set of markups, or tags, that are bedded in a Web document and provide structural,stylistic, and content information
em-Uniform resource locator (URL) The explicit format
for a reference to a hypertext document It is in the form
protocol://server:port/path The protocol can be any of
several standard Internet communications protocols,with HTTP being the most common for Web pages
By default, Web servers communicate using a standardport number, 80 In such cases the URL can be short-
ened to protocol://server/path.
User Anyone accessing the Web site, using a Web
bro-wser A related term, user interface, refers to the tire environment (text, graphics, and user input andresponse) that builds the experience for the user inter-acting with the site
en-Web site The integration of hypertext content,
presen-tation information, and controlling logic, that formsthe user experience Implemented on a Web server, its
purpose is usually to disseminate information, fostercollaboration, or obtain user input It is the basic unit
of discussion in this chapter and will refer to both theuser experience and the actual implementation
World Wide Web (WWW) A network of hypertext uments, existing on Web servers and accessible viathe Internet using computer programs called Webbrowsers
doc-CROSS REFERENCES
See Client/Server Computing; Databases on the Web;
HTML/XHTML (HyperText Markup Language/Extensible HyperText Markup Language); Usability Testing: An Eval- uation Process for Internet Communications.
REFERENCES
Champeon, S (2001) JavaScript: How did we get here?Retrieved April 16, 2002, from http://www.oreillynet.com/pub/a/javascript/2001/04/06/js history.htmlcomScore Networks (2002) Ten million Internet users
go online via a cellphone or PDA, reports ComscoreMedia Metrix Press Release Retrieved August 30, 2002,from http://www.comscore.com/news/cell pda 082802.htm
European Computer Manufacturers Association (1999).Standard ECMA-262: ECMAScript language specifica-tion Retrieved April 2, 2002, from ftp://ftp.ecma.ch/ecma-st/Ecma-262.pdf
Ginsburg, P E (1999) Building for 5.0 browsers trieved May 10, 2002, from http://builder.cnet.com/webbuilding/pages/Authoring/Browsers50
Re-Lehn, D., & Stern, H (2000) Death of the websafe colorpalette? Retrieved April 20, 2002, from http://hotwired.lycos.com/webmonkey/00/37/index2a.html?tw=designLie, H W., & Bos, B (1999) Cascading style sheets, level
1 W3C recommendation Retrieved April 2, 2002, fromhttp://www.w3.org/TR/REC-CSS1
Marcus, B (2001) Wireless, broadband penetrationcontinues Retrieved May 1, 2002, from http://www.digitrends.net/nwna/index 15935.html
Murray, G., & Costanzo, T (1999) Usability and the Web:
An overview Retrieved April 16, 2002, from http://www.nlc-bnc.ca/9/1/p1-260-e.html
Niederst, J (2001) Web design in a nutshell (2nd ed.)
Se-bastopol, CA: O’Reilly & Associates
Roelofs, G (2000) PNG, MNG, JNG, and Mozilla M17 trieved August 30, 2002, from http://www.libpng.org/pub/png/slashpng-2000.html
Solomon, K (2000) Smart biz: Enabling the disabled trieved May 1, 2002, from http://www.wired.com/news/print/0,1294,39563,00.html
Re-Weissinger, A K (2000) ASP in a nutshell (2nd ed.)
Se-bastopol, CA: O’Reilly & Associates
World Wide Web Consortium (1998) Document ject model (DOM) level 1 specification RetrievedApril 2, 2002, from http://www.w3.org/TR/REC-DOM-Level-1
ob-World Wide Web Consortium (2002) Hypertext markuplanguage (HTML) home page Retrieved August 28,
2002, from http://www.w3.org/MarkUp
Trang 27DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
Wide Area and Metropolitan Area Networks
Lynn A DeNoia, Rensselaer Polytechnic Institute
Facilities and Infrastructure 778
Differences around the World 780
Switching, Routing, and Signaling 781
Carriers and Service Providers 788Class of Service, Quality of Service 789
In today’s social, political, and economic environment,
individuals and organizations communicate and operate
over ever-increasing geographic distances This means
that access to and sharing of information and resources
must extend beyond the “local” office, building, or
cam-pus out across cities, states, regions, nations, continents,
and even beyond the planet Bridging this diversity of
dis-tances in ways that satisfy application requirements for
speed, capacity, quality, timeliness, etc at reasonable cost
is no simple challenge, from either a technical or a
busi-ness perspective In this chapter we concentrate on the
main elements required to meet such a challenge in wide
area and metropolitan area networks
HISTORY AND CONTEXT
Definitions
The public networking arena has typically been divided
into two segments with the following characteristics:
Metropolitan area networks (MANs) are built and
op-erated by service providers (SPs) who offer network
ser-vices to subscribers for a fee, covering distances up to
tens of miles, often within or surrounding a major city
MANs are often built by telecommunication companies
such as local exchange carriers (LECs) or by utility
com-panies A recent alternative using Ethernet for the MAN
has spawned a new category of companies called Ethernet
LECs or ELECs
Wide area networks (WANs) are built and operated
by SPs who offer network services to subscribers for a
fee, covering distances up to hundreds or thousands of
miles, such as between cities, across or between
coun-tries, across oceans, etc WANs designed for voice are
usu-ally built by telecommunication companies known in the
United States as interexchange carriers (IXCs) WANs for
data are also called public data networks (PDNs)
By contrast, local area networks (LANs) are typically
built and operated as private networks, by individuals
or enterprises, for their own use In addition, landlordsoperating as building LECs (BLECs) may offer LAN ser-vices to tenants In either case, the geographic scope of
a LAN is usually limited to a building or campus ronment where all rights of way for cabling purposes be-long to the individual/enterprise/landlord The boundariesbetween LANs and MANs and WANs began to blur as ge-ographic limitations of networking technologies were ex-tended with increasingly capable implementations overfiber-optic cabling Even the distinctions between privateand public networks became more difficult to draw withthe advent of “virtual private network” equipment andservices
envi-Challenges
The number of options and choices available to networkdesigners in both the subscriber and provider communi-ties continues to grow for both MANs and WANs Multi-ple technologies and standards, increasing numbers andtypes of applications, higher expectations for MAN andWAN performance comparable to (or at least approach-ing) that found in a LAN environment, and pressure tokeep unit costs low, all combine to create enormous chal-lenges for MAN and WAN builders Infrastructure choicesmust last long enough, not just for cost recovery, but
to achieve return on investment Service providers mustmarry new technologies to their existing installed base,create smooth transitions (e.g., for network upgrades, newservice roll-outs) with minimal disruption to customerservices, and add or enhance services to meet advancingcustomer expectations, all in an environment of increas-ing economic and competitive pressure Many providershave begun to recognize that their long-term survival de-pends on a strategy of simplification—reducing the com-plexity (to have fewer technologies, fewer equipment ven-dors, fewer equipment types, fewer management systems,etc.) of their infrastructure while maintaining the flexibil-ity to adapt to changing application, user, and competitiverequirements
776
Trang 28DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
The pressure to simplify is constantly at odds with thedifficulties of predicting the future:
Which technology will provide the best flexibility and
scal-ability at an acceptable investment cost?
How fast and in what ways will application needs and user
expectations develop?
Which services or enhancements will provide competitive
advantage?
How can value be added to those elements moving
down-ward into the commodity market?
The ability to develop shrewd answers to such tions is likely to determine which companies will thrive
ques-in the networkques-ing services busques-iness
Functional Requirements
The basic function that subscribers seek from MAN and
WAN service providers is the ability to deliver traffic from
one place to another (point-to-point) or to multiple
oth-ers (multipoint) This begins with connectivity For the
network in Figure 1, traffic can flow from A to C and/or D,
but not to B Once connectivity is established, the network
must have sufficient capacity in bandwidth and
switch-ing to get the traffic from the source to its intended
des-tination Subscribers want services that are reliable, as
measured by the percentage of time network resources
are available when needed and by the amount of traffic
(preferably none) that gets lost Subscribers also want
ser-vices that perform well enough so that their traffic gets
delivered in a timely fashion, with minimal delay (low
latency is particularly important for delay-sensitive
traf-fic such as voice or video) Providers, on the other hand,
want an infrastructure that is cost-effective, manageable,
and capable of supporting revenue generation and profits
Evolution and Coexistence
The first WANs were built from circuit-switched
connec-tions in the telephone system because that’s what was
available to cover the distances involved Circuit
switch-ing continues to be useful, particularly when the computer
MAN or WAN
D
CB
s
q
r
Figure 2: Connections, a)
point-to-point and b) multipoint-to-point
devices being connected need to exchange messages inreal time or with guaranteed delivery For occasional traf-
fic, dial-up connections similar to an individual telephonecall are used For continuous traffic or when applicationscannot tolerate the delay involved in call setup, circuitsare leased from a telephone company and “nailed up” intopermanent connections For two connected locations the
leased line is called a point-to-point connection (Figure 2a) More than two locations can be connected with a mul-
tipoint link (Figure 2b) if a sharing discipline is imposed
to prevent traffic from one source interfering with trafficsent from another at the same time In either case, theresources required to carry traffic across the leased lineare dedicated to the particular subscriber, creating an ef-fectively private connection through the service provider’spublic network resources
Two devices connected by a leased line may or maynot send traffic continuously, wasting capacity when theline is idle If there are multiple devices in one location
to be connected to one or more devices in a destinationlocation, a single leased line may be shared using a de-vice at each end of the line called a multiplexer Statisticalmultiplexing allows more devices to be connected thanthe capacity of the line could support in real time if allwere to transmit simultaneously This is called oversub-scription On the average, it is quite likely that only somedevices will be active, and the line is shared effectivelywith little traffic delay and less wasted capacity However,when many devices are active, performance can be de-graded The sending multiplexer adds a label to each unit
of traffic transmitted; the receiver reads (and removes) thelabel to figure out which device is the intended recipientand switches the traffic onto the appropriate output link.Packet switching is a form of statistical multiplexing.Originally circuit switching was designed to carry ana-log voice traffic and packet switching was designed for dig-ital data Today, however, public networks convert all types
of traffic into digital form for cost-effective transport Wecould say that “bits are bits,” whether they belong to voice,data, video, or some other application The same networkmight well be used to deliver multiple types of bits, instead
of having distinct networks dedicated for voice, data, etc
This is the concept of convergence, where a single network
carries various types of traffic In the context of gence, the important question shifts from whether circuit
conver-or packet switching is better, to what suppconver-ort a netwconver-ork
Trang 29DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
W IDE A REA AND M ETROPOLITAN A REA N ETWORKS
778
must provide so that traffic delivery meets user
expecta-tions and application requirements Convergence is
cer-tainly not new, because in early WANs, digital data were
transformed into analog signals and carried over public
networks that had been designed for voice Today
conver-gence is available through many more options for what
traffic to combine and how to do it
FACILITIES AND INFRASTRUCTURE
Digital Transmission
The heritage of digital WANs dates from the early 1960s,
when the Bell System first introduced the T-carrier
sys-tem of physical components to support transport of
digi-tal signals in the United States The accompanying
time-division multiplexed (TDM) digital signal scheme, called a
digital hierarchy, was based on a standard 64-kilobits per
second (Kbps) signal designed to carry one analog voice
signal transformed by pulse-code modulation (PCM) into
digital form This basic unit is known as DS0 The
Inter-national Telecommunication Union (ITU) now supports
an entire set of digital signaling standards (Table 1),
in-corporating elements from the North American (United
States/Canada), European, and Japanese standard
hierar-chies
The traditional U.S multiplexing hierarchy began with
combining 24 DS0-level signals into one DS1 It is
com-monly called a T1 stream, and consists of a sequence of 24
channels combined to create one frame Each channel is
filled with 8 bits (an octet or byte) representing one PCM
sample A particular challenge of the time was to ensure
synchronization between transmitter and receiver, which
can be accomplished in several ways For example, each
frame could be introduced by a unique starting sequence
of 12 bits to allow receiver synchronization to be renewed
on a frame by frame basis The U.S designers decided
in-stead to distribute the 12 bits over 12 frames, reducing
transmission overhead at the expense of receiver
com-plexity The 12-frame sequence was called a superframe
With improved hardware, synchronization is more easily
maintained over longer periods, and an extended
super-frame (ESF) has replaced the supersuper-frame ESF comprises
24 frames but only needs 6 bits for synchronization,
free-Table 1 Digital Signal Hierarchy
Capacity Number Designation (Mbps) of DS0s
manage-In the European scheme (also used by other countries
such as Mexico), the basic E1 stream aggregates 32 PCM
channels Rather than adding synchronization bits, E1dedicates the first PCM channel for synchronization andthe 17th for management and control signaling
Optical Fiber Systems
Service providers first used digital multiplexing withintheir own networks (e.g., trunking between Central Of-fices), to improve the return on and extend the life of theircopper cable infrastructure investments By the 1980s,however, interest had shifted to fiber optics for longer dis-tance, higher speed communications Standards were de-fined for the Synchronous Optical Network (SONET in theUnited States, equivalent to the Synchronous Digital Hier-archy, SDH, in Europe and elsewhere) to carry TDM trafficcost-effectively and reliably over metropolitan and widearea distances Today SONET specifies both a standardoptical interface signal and a digital signaling hierarchytailored to the fiber transmission environment The hier-archy is based on an 810-octet frame transmitted every
125 microseconds (µs) to create synchronous transport
signal-level 1 (STS-1) for electrical signals Each octet is
equivalent to a 64-Kbps PCM channel For fiber
transmis-sion, the STS-1 equivalent is optical carrier-level 1 (OC-1).
Higher level signals are formed from specific multiples
of OC-1 (Table 2) Each SONET frame is structured intotransport overhead and a synchronous payload envelope(SPE), which consists of both path overhead and payload
It is only the payload portion that carries subscriber traffic
to be routed and delivered through the SONET network.The major building blocks for SONET networks are thepoint-to-point multiplexer, and for point-to-multipointconfigurations, the add-drop multiplexer (ADM) In par-ticular, the ADM allows traffic to be dropped off and theresultant free capacity to be reused to carry traffic enter-ing the network at that point SONET ADMs can also beemployed to create highly survivable networks that max-imize availability using diverse routing and self-healing,survivable ring structures Figure 3a shows a dual-ringstructure where the network accommodates loss of a link
Table 2 Basic SONET Levels Designation Line rate SDH equivalent
Trang 30DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
a) Lost link
b) Lost nodeX
X
Figure 3: SONET ring survivability, a) lost link
and b) lost node
by looping traffic back on each side of the break, and
Fig-ure 3b shows how loss of a network node can be handled
similarly SONET has been deployed extensively by
ser-vice providers in metropolitan areas to create highly
re-liable and scalable transport capabilities Once the fiber
and switching equipment are in place, transport
capac-ity can be increased by installing higher-speed signaling
interfaces
Another approach to increasing the capacity of fibersystems has become available with advances in optical
component technology Rather than using the entire range
of wavelengths that can be carried over fiber as a
sin-gle transmission channel, newer equipment allows us to
divide the range into multiple channels for
simultane-ous transmission using wavelength-division multiplexing
(WDM) This is quite similar to sending multiple television
channels over a coaxial cable Channels must be spaced
far enough apart to limit the interference between
adja-cent signals that would degrade signal quality In coarse
WDM (CWDM) the channels are widely spaced; for dense
WDM (DWDM), they are very close together (spacing≤
25–50 GHz) By combining WDM and high-speed
sig-naling, transmission capacities of OC-192, OC-768, and
greater become possible, limited primarily by the quality
of existing fiber installations
Access Technologies
In order to get traffic in and out of a MAN or WAN,
sub-scribers must have physical connections, or access, to the
appropriate service provider’s network resources In the
regulated telecommunications environment of the United
States, this typically means acquiring connectivity from a
LEC to tie the subscriber’s physical premises to a WANservice provider’s (i.e., IXC’s) equipment as physically lo-cated in a point of presence (POP) In a metropolitanarea, a single company may be allowed to provide bothlocal exchange connections and MAN services The pri-mary means of accessing MAN and WAN service providernetworks are described below
Dial-up
Dial-up access is appropriate for occasional connections
of limited duration, as for making a telephone call Wherethe physical facilities used for dial-up were designed andinstalled to support analog voice traffic, two characteris-tics are particularly important for data networking:
Digital data must be converted to analog using a modem
at the subscriber end and reconverted to digital by amodem at the provider end of the connection.Data rates are limited by the analog frequency range ac-cepted at provider receiving equipment and by the sig-nal modulation techniques of the modem devices Themost widely accepted standards today support maxi-mum data rates of 56 Kbps
Leased Line
Leased-line access is more appropriate for connectionsthat need to be continuous and/or of better quality forhigher-speed transmission Such facilities are dedicated
to the use of a specific subscriber For example, a businessmay lease a T1 access line as its basic unit of connectioncapacity (1.544 Mbps), particularly for access to Internetservice providers Fractional-T1 and multiple-T1 lines arealso available in some areas A newer technology designed
to be digital from end to end over copper cabling, calleddigital subscriber line (DSL), is beginning to be offered
as a lower-cost alternative to the traditional T-carrier.Leased-line access requires matching equipment at eachend of the line (subscriber and service provider) to ensuretransmission quality suitable to the desired data rates
Wireless
Wireless access is growing in popularity among mobileindividuals who do not work from a fixed desktop in a sin-gle building location (e.g., salespeople, customer servicerepresentatives, and travelers) Rather than having to find
a suitable “land-line” telephone connection with an log data port to connect the modem, wireless users haveeither wireless network interface cards or data interfacecables that connect their modems to cellular telephones.Both approaches require proximity to a wireless receivingstation of matching technology that is then connected tothe wired resources making up the remainder of the MAN
ana-or WAN
Cable Modem
Cable modem access is provided by cable television panies who have expanded their business into data net-working A modem designed to transmit data signals overcoaxial, broadband television cable is connected, usuallyvia Ethernet technology, to the subscriber’s internal net-work or computer equipment In residential applications,
Trang 31com-DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
W IDE A REA AND M ETROPOLITAN A REA N ETWORKS
780
subscribers in a neighborhood typically share data
net-working capacity on the aggregate cable that carries traffic
back to the provider’s central service location This is
dif-ferent from a leased-line approach, where access capacity
is dedicated from each subscriber location all the way to
the provider’s POP In the United States, cable providers
are regulated differently from other public
telecommu-nications providers, and may not suffer the same
conse-quences for unavailable service
Management
Management for MANs and WANs typically began with
proprietary systems sold to service providers by each
man-ufacturer of telecommunications switching equipment
Networks composed of equipment from multiple vendors
thus contained multiple management systems
Equip-ment manageEquip-ment and service manageEquip-ment functions
are often tied together by an Operations Support System
(OSS) in order to automate operations (e.g., performance
monitoring), administration (e.g., ordering and billing),
maintenance (e.g., diagnostics, fault detection and
iso-lation), and provisioning (OAM&P) functions Many
ser-vice providers tailored what they could acquire as a basic
OSS in order to accommodate their own specific sets of
equipment and services, making it difficult to share
infor-mation, provide consistent management data in a
mul-tiprovider environment, and keep up to date with new
functional requirements This often leaves customers who
need services from multiple providers without a single,
coherent view of their enterprise WAN resources
Beginning in 1988, the Telecommunication
Standard-ization sector of the International Telecommunication
Union (ITU-T, formerly the Consultative Committee on
In-ternational Telephony and Telegraphy, CCITT) set about
establishing the precepts for a standard
Telecommunica-tions Management Network (TMN) While the concept of
a TMN encompasses the entire set of OAM&P
applica-tions in the network, what they do, and how they
com-municate, ITU-T standards focus on the information
re-quired and how it should be communicated rather than
how it is processed (M.3000 recommendation series) Two
types of telecommunications resources are encompassed:
managed systems (such as a switch), which are called
network elements (NE), and management systems,
usu-ally implemented as operations systems (OS) TMN
stan-dards are organized into interface specifications that
de-fine the interconnection relationships possible between
resources Figure 4 shows the relationship between the
TMN and the telecommunication network for which it is
responsible
TMN is based on the Open Systems
Interconnec-tion (OSI) management framework, using object-oriented
principles and standard interfaces to define
communica-tion for purposes of managing the network The primary
interface specification, Q3, allows direct communication
with an OS Any network component that does not
im-plement Q3 may not access an OS directly, but must go
through a mediation device (MD) instead Legacy
equip-ment and systems that rely on proprietary ASCII messages
for communication are accommodated by means of a
Q-adapter (QA) that can translate between messages
NE
NE
NE
OSData
network
OS
OS
Telecommunicationnetwork
TMNQ3
Figure 4: TMN and the network it manages.
representing the legacy information model and theobject-oriented representation expected in today’s TMN.TMN defines a layered architecture (ITU-T standardM.3010) as a logical model for the functions involved
in managing a telecommunication network effectively(Table 3) The object is to create a framework for inter-operability across heterogeneous operation systems andtelecommunication networks that is flexible, scalable, re-liable, easy to enhance, and ultimately, inexpensive to op-erate Standard management services have been definedfor alarm surveillance (Q.821), performance management(Q.822), traffic management (Q.823), ISDN service profilemanagement (Q.824), call detail recording (Q.825), androuting management (Q.826)
Differences around the World
Creating and operating WANs or MANs in different tries may present challenges well beyond identifying a ser-vice provider and getting connections established A par-ticular type of service may not be available in the desiredlocation, or a single provider may not offer services in ev-ery location, or the capacity required may not be available.Such differences may be due to telecommunication infras-tructure of varying ages and technologies, or to differentregulations on service offerings in various countries Forexample, T1 service is readily available in most U.S cities.Mexico, however, employs the European standard hier-archy Thus E1 service would need to be ordered (if it isavailable) to connect a business location in Mexico to one
coun-in the United States, and the differences coun-in capacity andframing would have to be handled appropriately by thenetwork equipment at each end of the link
In some countries, telecommunication is a regulatedindustry subject to many government-imposed rules, and
Trang 32DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
Table 3 TMN Architecture Logical layer Functional responsibilities
Business management Provides an enterprise view that incorporates high-level, business planning
and supports setting goals, establishing budgets, tracking financialmetrics, and managing resources such as products and people
Service management Provides the basic contact point for customers (provisioning, billing and
accounting, troubleshooting, quality monitoring, etc.) as well as forservice providers and other administrative domains
Network management Provides an overall view of the network resources, end to end, based on
the information from below about network elements and links
Coordinates activities at the network level and supports the functionalrequirements of service management
Element management Provides a view of individual network elements or groupings into
subnetworks Element managers (OSs) are responsible for subsets of allnetwork elements, from the perspective of TMN-manageable informationsuch as element data, event logs, and activity Mediation devices belong
in this layer, communicating with OSs via the Q3 interface
Network elements Presents the TMN-manageable information of individual network resources
(e.g., switches, routers, Q-adapters)
there may be no or a limited choice of carriers Other
countries have begun to deregulate, so that multiple
car-riers compete for subscriber business, often creating more
choices in technology and services, as well as better
pric-ing In either case, service availability may differ from one
location to another: DSL access might be easily obtained
in greater Boston, but not be available in a rural area; T1
service might be acquired readily in Singapore but
per-haps not everywhere in New York City
Do not make the mistake, however, of assuming thatmore highly developed areas or countries always have bet-
ter service options than developing ones An established
metropolitan area experiencing rapid growth in demand
for telecommunications may be less able to adapt or
ex-pand existing cable and switching capacity to meet new
orders than a new suburban business park where there
is plenty of room to install new cables and switches to
provide higher-speed services Similarly, developing
coun-tries that have very little investment in old infrastructure
may be able to skip generations of technology, installing
the latest where there was previously none Economics
tend to dictate that this does not happen uniformly, but
rather emphasizes locations more likely to provide rapid
payback for the particular technology investment (e.g.,
urban rather than rural, business rather than residential,
and high-density population areas) Often it is the access
infrastructure that lags behind, because the upgrade costs
cannot be amortized across multiple subscribers the way
backbone investments can This is especially true where
the end-points are individuals with more limited budgets
than business or organizational enterprises
SWITCHING, ROUTING,
AND SIGNALING
Network Architecture
MANs and WANs are usually divided into three logical
seg-ments (Figure 5) Access typically includes the customer
premises equipment (CPE) located in a subscriber’s ing or office area and the link that physically connectsfrom there to the service provider’s point of presence This
build-link is connected to a device at the edge of the service
provider’s network, and the edge device is connected to
devices that compose the core (also called the backbone)
of the service provider’s network Different technologiesare often used in the access and core portions, with theedge required to translate between the two The ratio ofthe aggregate input capacity from all subscriber connec-tions to an edge device to the output capacity from theedge into the core describes the degree of oversubscrip-tion For example, if the sum of all access links is 200 Mbpsand the core link is 100 Mbps, then the oversubscriptionratio is 2:1 A ratio less than or equal to 1 is called non-blocking; the network performance for values greater than
1 depends on the bursty nature of data traffic to minimizethe probability that traffic will be delayed excessively (bybuffering) or discarded (when buffers become full)
MAN or WAN
B A
Trang 33DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
W IDE A REA AND M ETROPOLITAN A REA N ETWORKS
782
Some form of packet switching is employed in most
core data networks today to move traffic through the
net-work Various techniques are used to meet customer
ex-pectations for reliable, timely, and effective delivery of
traffic to its intended destination For example, a virtual
circuit can be established to approximate the service
char-acteristics available in a circuit-switching environment,
such as guaranteed delivery of packets in the same order
as they were transmitted However, virtual circuits do not
dedicate resources along the path from source to
desti-nation, so the network must have sufficient intelligence
to keep traffic moving well enough to meet subscriber
expectations
Choosing the best place to put network intelligence
(at the edge or in the core) has been a subject of
ongo-ing discussion among service providers for many years
For example, packets could be examined and labeled at
the edge in a way that forwarding decisions in the core
are made by simple, high-speed switches This approach
would provide very fast core transit, but the cost of many
intelligent edge devices could be high and core switches
must still be smart enough to accommodate and adapt to
changes in network topology or conditions An alternative
approach makes the edge devices quite simple and
inex-pensive, while requiring the core to have the intelligence
and take the time to understand the characteristics and
accommodate the transport needs of the traffic
Switching Technologies
In the OSI Reference Model, switching takes place at
Layer 2, the Data Link Layer However, much of the WAN
switching technology for data networking was developed
from experience with X.25, an ITU-T packet-switching
protocol standard developed in the 1970s to support
pub-lic data networking, and still in use today X.25 creates
a connection-oriented network out of packet-switching
resources by employing virtual circuits to handle packet
flow, keeping the data link layer simpler but requiring
cir-cuits to be established before packets can be sent Circir-cuits
that are prebuilt from a source to a particular
destina-tion and then left in place are permanent virtual circuits
(PVCs), while switched virtual circuits (SVCs) are
estab-lished only on demand SVCs are like dial-up connections,
requiring circuit establishment to the specified
destina-tion for each call before traffic can flow
X.25
X.25 is a three-layer protocol suite (Figure 6) The OSI
net-work layer equivalent is the packet-layer protocol (PLP),
which has operational modes for call establishment, data
transfer, and call termination, plus idle and restarting
op-erations These functions are implemented through the
Flag(framedelimiter)
Address(command
or responseindicator)
Control(frame type,sequence #,function)
DATA
FCS(frame checksequence)
Flag(framedelimiter)
Figure 7: LAPB frame format.
PLP
LAPB
X.21bis,EIA/TIA-232,EIA/TIA-449,EIA-530,G.703X.25
Figure 6: X.25 protocol suite.
services of a data link protocol called the Link Access cedure, Balanced (LAPB), which is responsible for fram-ing data and control commands and for basic error check-ing through use of a frame-check sequence (Figure 7).During call establishment, the PLP sets up SVCs usingX.121 standard addresses These include the internationaldata number (IDN), made up of a four-digit data networkidentification code (DNIC, to specify the packet-switchingnetwork containing the destination device) and a nationalterminal number (NTN) consisting of as many as 10 digits.The NTN specifies the exact destination device to whichpackets will be forwarded
Pro-Frame Relay
Frame relay is the most widely used packet-switchingWAN technology going into the 21st century As WAN fa-cilities became more reliable during the 1980s, interestrose in streamlining X.25 to improve performance and ef-ficiency Frame relay (FR) was thus designed as a Layer-2protocol suite, with work begun by CCITT in 1984 How-ever, it was not until 1991, when several major telecommu-nication equipment manufacturers formed a consortiumcalled the Frame Relay Forum (FRF) to work out interop-erability issues and foster acceptance, that frame relay be-gan to be more widely deployed In particular, FRF definedextensions to the CCITT work called the local manage-ment interface (LMI) to improve service providers’ abili-ties to provision and manage frame relay services.Frame relay networks (Figure 8) are based on the con-cepts of data-terminal equipment (DTE) and data circuit-terminating equipment (DCE) first defined by X.25 Sub-scriber hosts, servers, workstations, personal computers,
Trang 34DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
FR network
DTEDTE
DCEPSE
Figure 8: Frame relay network elements.
and terminals connected to a frame relay network are all
considered to be DTE The DCE is usually built as an
in-terface into the service provider’s packet-switching
equip-ment (PSE) rather than just being a modem at the edge
of an X.25 network Frame relay also uses virtual circuits
to create a bidirectional communication path between a
pair of DTE devices FR virtual circuits are distinguished
by data link connection identifiers (DLCIs), which may
have local significance only, meaning that each end of a
single virtual circuit could have a different DLCI assigned
by the FR service provider
The format for frame relay data combines LAPB’s dress and control fields into one 16-bit address field that
ad-contains the 10-bit DLCI, an extended addressing
indica-tor bit (for future use), a command/response bit that is
not used, and congestion control information To
mini-mize network overhead, the congestion control
mecha-nisms are quite simple:
one forward-explicit congestion notification (FECN) bit
that tells a DTE that congestion occurred along the
path in the direction from the source to the
destina-tion;
one backward-explicit congestion notification (BECN) bit
that tells a DTE that congestion occurred along the
path in the direction opposite to the transmission from
the source to the destination; andone discard-eligibility (DE) bit to indicate whether this is
a lower priority frame that may be discarded beforeothers in a congested situation
As a packet-switching technology, frame relay also pends on the bursty nature of data traffic to make effi-
de-cient use of its transmission facilities for larger numbers
of subscribers than could be served with physically
ded-icated connections The ability to overbook resources is
fundamental to the service provider’s business model, as
well as being a benefit to subscribers, who may be able to
insert traffic occasionally at a higher rate than nominal
for their access link (called bursting)
Integrated Services Digital Network (ISDN)
Integrated services digital network (ISDN) is a set oftelecommunication standards first developed from theperspective of telephony networks to accommodate multi-ple types of traffic such as voice, fax, data, alarm systems,and video, all in digital format, over a single network Thegoal was to develop standard interfaces, both for accessand within the network, that would allow all types of dig-ital traffic to be transported end to end, reliably, and in
a timely fashion according to the needs of its tion The best-known elements of ISDN are the user in-terface definitions for connecting subscriber equipment
applica-to the network: the primary rate interface (PRI), intended
to replace T1 and E1 services, and the basic rate interface(BRI), designed with multiple channels for voice or datatraffic from an individual subscriber
Asynchronous Transfer Mode (ATM)
Asynchronous transfer mode (ATM) was selected as theOSI Layer-2 transport technology for broadband ISDN(B-ISDN) in 1988 It was designed to be useful acrossWAN, MAN, and LAN communications, as well as to ac-commodate multiple types of traffic in a single network(voice, data, video, etc.) and scale for very large networks.Other design goals included the abilities to support avariety of media types (e.g., fiber and copper), leveragesignaling standards already developed for other tech-nologies, promote low-cost switching implementations(potentially one-tenth the cost of routing), adapt readily tofuture network requirements, and enable new, large-scaleapplications The challenges inherent in such a diverse set
of goals brought together designers from many differentbackgrounds, and resulted in a rather complex architec-ture (Figure 9)
Basically, ATM is a connection-oriented, switching technology that uses fixed-length packets called
packet-cells The 53-byte cell size (5 bytes of header information
and 48 bytes for the payload) was chosen as a compromisebetween the optimal size for voice traffic and the largersize preferred for data applications The fixed size andformat mean that very fast switches can be built across
a broad range of transmission rates, from megabits togigabits per second and beyond ATM interfaces are of-ten characterized by their equivalent optical-carrier lev-els whether they employ fiber or copper media The most
Application Presentation Session Transport Network Data Link Physical
OSI Reference Model
ATM Adaptation Layer ATM Layer Physical Layer
Higher Layers Higher Layers
Management Plane Control Plane Control Plane
Figure 9: ATM reference model.
Trang 35DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
W IDE A REA AND M ETROPOLITAN A REA N ETWORKS
784
popular interfaces tend to be OC-3, OC-12, and OC-48
(Table 2), according to their application in WANs, MANs,
or LANs
An important feature of ATM is the definition of service
categories for traffic management:
Constant Bit Rate (CBR) was designed to emulate
tradi-tional circuit-switched connections It is characterized
by minimum and maximum cell rates specified at the
same, constant value Typical CBR applications include
uncompressed voice and video, or television, all
sensi-tive to both delay and delay variation
Variable Bit Rate real-time (VBR-rt) and non-real-time
(VBR-nrt) are characterized by specified minimum and
maximum cell rates, much like frame relay Typical
ap-plications include compressed voice or video, and
mul-timedia e-mail VBR-rt handles applications sensitive
to delay variation, while VBR-nrt is suitable for bursty
traffic
Unspecified Bit Rate (UBR) handles traffic on a best-effort
basis, without guaranteeing delivery or any
particu-lar rate This is used to carry data (such as
store-and-forward e-mail) not sensitive to delay In a highly
con-gested network situation, UBR cells may be discarded
so that the network can meet its traffic contracts for
the other types
Available Bit Rate (ABR) is characterized by a guaranteed
minimum cell rate, but may offer additional bandwidth
when network resources are available Rate-based flow
control provides the adjustment mechanism When it
is offered, ABR is often preferred for data traffic
ATM’s service categories are crucial to meeting user
de-mands for quality of service (QoS), which generally means
guaranteed, timely delivery of traffic to match the needs
of particular applications An ATM end system will
re-quest a particular level of service for traffic entering the
network, forming a traffic contract with the network The
ATM switches throughout the network are responsible for
meeting the terms of the contract by traffic shaping (using
queues to smooth out traffic flow) and by traffic policing
to enforce the limits of the contract The capabilities of
ATM to provide QoS end to end across a network for
mul-tiple types of traffic simultaneously are the most
sophis-ticated to date, and distinguish ATM from other
packet-switching technologies Its suitability for LAN, MAN, and
WAN applications makes ATM especially popular with
service providers, because they can use one technology
throughout to manage their own infrastructure and to
support a large variety of service offerings to their
cus-tomers
Fiber Distributed Data Interface (FDDI)
Fiber distributed data interface (FDDI) was developed
by the American National Standards Institute (ANSI) in
the mid-1980s as a 100-Mbps standard for ring-based
networks that had outgrown their capacity to handle
high-speed workstations or provide nonblocking
back-bone connections It was designed originally to expand
the typical LAN environment, using a timed token access
method for sharing bandwidth at OSI Layer 2 and
read-start of data flow end of data flow
(head of Bus A)
(head of Bus B) Bus B
Phys-a highly reliPhys-able network with deterministic, predictPhys-ableperformance FDDI was the first LAN technology suitablefor distances beyond a building or small campus, and wasused by some to cover the geographic scope of a MAN
Distributed Queue Dual Bus (DQDB)
Distributed queue dual bus (DQDB) was also developedduring the 1980s, specifically to address the needs ofmetropolitan area networking for integrated services such
as voice, data, and video The IEEE 802.6 working groupfinally ratified it as a Layer-2 standard in 1990 As itsname suggests, DQDB specifies a network topology of twounidirectional buses that are able to interconnect multi-ple nodes (Figure 10) The supporting physical layer forDQDB initially offered various transmission interfacesand speeds from DS3 (45 Mbps) to STM-1 (155 Mbps).The idea of DQDB was that multiple subnetworks could beinterconnected to form a MAN, with the goal of support-ing connectionless and connection-oriented data trans-fers, along with isochronous traffic, sharing the total com-munication capacity available
DQDB may be most familiar as the basis for
defini-tion of switched multimegabit data service (SMDS)
packet-switched public data networks SMDS was designed byBell Communications Research (Bellcore) for high-speed,connectionless delivery of data beyond the LAN Its vari-able frame size up to 9188 octets is large enough to en-compass as payload any of the popular LAN technologyframes (i.e., Ethernet, token ring, and FDDI) The SMDSinterface protocol was defined as a three-level protocolthat specifies how subscribers access the network As aservice, SMDS was intended to be independent from anyunderlying transport technology Thus it was first offered
at DS1 to DS3 access speeds, with a goal of increasinglater to OC-3
Ethernet
Ethernet became the dominant LAN technology in the ter 1990s, as extensions from the original 10 Mbps weredefined for 100 Mbps, then 1,000 Mbps ( = 1 Gbps), and be-came widely deployed In the same time period, new com-munication companies with no telephony heritage beganlaying optical fiber, and leasing capacity for short-haul(i.e., MAN) or long-haul (i.e., WAN) connections rather
Trang 36lat-DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
router
Figure 11: Routing within and between autonomous domains.
than selling services, as was typical in public networks
This meant that customers could specify the technology
used to put bits on the medium rather than subscribing
only to specific services offered by providers As advances
in optics and use of switching allowed Ethernet to cover
even greater distances, the geographic limits that
distin-guished LAN from MAN technologies began to disappear
In fact, new providers sprang up offering Ethernet
con-nectivity from the business doorstep to other locations
across town or beyond The great competitive question
was whether Ethernet MANs could be made as reliable
and fault-tolerant as more traditional MAN/WAN
tech-nologies built over SONET
Resilient Packet Ring (RPR)
Resilient packet ring (RPR) is an effort begun by the IEEE
802.17 working group in late 2000 to design a high-speed
access protocol combining familiar Ethernet interfaces
with the fault-tolerance and rapid restoration capability
of ring-based MAN technologies like SONET RPR defines
a new medium access control (MAC sublayer of OSI Layer
2) protocol that extends Ethernet framing from the LAN
into the MAN/WAN environment As seen by the RPR
Alliance (an industry consortium designed to promote
adoption of RPR), this approach combines the
cost-effective scalability of Ethernet access interfaces with a
MAN that can be optimized for rapidly increasing
vol-umes of data traffic Because it focuses on the MAC
sublayer, RPR is independent of the underlying Layer-1
technology, making it suitable to run over much of the
MAN infrastructure already in place
Routing Technologies
In the OSI Reference Model, routing takes place at Layer
3, the Network Layer Essentially routing consists of three
major functions: maintaining information about the
net-work environment, finding a path through the netnet-work
from particular sources to destinations, and forwarding
packets at each relay point The Internet protocol (IP) is
the dominant method of interconnecting packet-switched
networks (i.e., for internetworking) at Layer 3 It provides
connectionless network services (CLNS), with no antee of delivery or packet ordering, and is widely usedtoday for private and public LANs, MANs, and WANs, in-cluding the Internet IP is primarily concerned with theformat for packets (also called datagrams), the defini-tion and structure of addresses, a packet-forwarding al-gorithm, and the mechanisms for exchanging informationabout conditions in and control of the network
guar-Routing responsibility in an internetwork is divided tween intradomain or interior routing protocols (IRPs)and interdomain or exterior routing protocols (ERPs) asshown in Figure 11 IRPs are used for internetworks thatbelong to a single administrative authority, such as an en-terprise LAN, a single service provider’s MAN, or a privateWAN ERPs are used when routers tie together networksbelonging to multiple independent authorities, as in theInternet These protocols differ in how much informa-tion is kept about the state of the network and how rout-ing updates are performed using the mechanisms defined
be-by IP
IP Version 4 (IPv4)
IP version 4 (IPv4) was defined by the Internet neering Task Force (IETF) for the original ARPAnet andpublished as (Request for Comments) RFC 791 in 1981
Engi-It specifies that each interface capable of originating orreceiving internetwork traffic be identified by a unique32-bit address consisting of an ordered pair containing anetwork identifier (net ID) and a host/interface identifier(host ID) Three primary classes of network addresses (A,
B, and C) were designed to promote efficient routing, withadditional classes defined for special or future uses (Fig-ure 12) Although the Internet is not centrally managed, itwas necessary to establish a single authority to assign ad-dresses so that there would be no duplicates or conflicts
As the Internet grew through the 1980s, a number oflimitations in the design of IPv4 became apparent The al-location of addresses, especially classes A and B, tended to
be wasteful For example, a single class B address assigned
to one organization accommodates one network with over64,000 IP interfaces—much larger than is practical or
Trang 37DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
W IDE A REA AND M ETROPOLITAN A REA N ETWORKS
786
001011011111
Host_IDNet_ID
Multicast addressNet_ID
24
31bit position
A B C D E
Figure 12: IPv4 addressing format.
needed for most, meaning that a lot of address space can
be wasted On the other hand, a single class C address
ac-commodates only 255 interfaces, which is too small for
most organizations, requiring them to have more than
1 From a routing perspective, the two-level hierarchical
address structure means that routers need to keep track
of over 16 million net IDs just for class C networks, as
well as calculate paths through the Internet to each one
A number of schemes were developed to solve some of the
addressing and router problems (subnet masking,
class-less interdomain routing or CIDR), but those were not
the only issues Rising interest in using the Internet to
carry voice, video, multimedia application, and
commer-cial transaction traffic increased the demand for security
and quality of service support, neither of which were built
into IPv4 Consequently, the IETF began work on a new
version, IP-ng, to handle the next generation
IP Version 6 (IPv6)
IP version 6 (IPv6) represents that next generation of
Net-work Layer services It extends the addressing space from
32 to 128 bits, simplifies the packet header and allows for
future expansion, and adds new capabilities to label flows
of packets (same source to a single destination), to assign
packets priority in support of QoS handling, and to
pro-vide authentication and security Several of these features
(CIDR, DiffServ, and IPsec) were designed so they could
be added onto IPv4 In fact, such retrofitting solved IPv4
problems well enough in the late 1990s that people
be-gan to question whether a move to IPv6 was necessary
Upgrading the large numbers of routers involved with
In-ternet traffic would be expensive, time-consuming, and
require careful coordination Transition strategies and
mechanisms would likely be needed over a considerable
period of time Unfortunately, retrofits cannot do much
about the size of IPv4 addresses Sufficient growth in the
numbers and types of devices people want to connect
to or through the Internet (handheld devices, household
appliances, automobile systems, etc.) and international
pressure from countries without enough addresses will
eventually make IPv4 addressing inadequate The only
question seems to be when
Border Gateway Protocol (BGP)
Border gateway protocol (BGP) is the exterior routing
pro-tocol used by independent or autonomous systems (ASs)
to exchange routing information throughout the Internet
Published in 1995 as RFC 1771, it defines procedures to
establish neighbor relationships, and to test the
reachabil-ity of neighbors and other networks A router at the edge
of an AS uses BGP to work with adjacent (i.e., directlyconnected) routers in other ASs Only after two routers(one in each AS) have agreed to become neighbors canthey exchange routing information or relay traffic for eachother’s AS Unlike IRPs, which use the services of IP toaccomplish their communication, BGP uses the reliabletransport services of TCP (transmission control protocol,running over IP) In this way, BGP can be simpler because
it depends on the error control functions of TCP, and itsmessages are not limited in size by the constraints of an
IP datagram
BGP is purposefully designed to allow an AS to trol what detail of internal information is made visibleoutside the AS (aggregating routes using CIDR, for ex-ample) Typically each BGP router screens potential rout-ing updates or reachability advertisements against a con-figuration file that specifies what type of information it
con-is allowed to send to each particular neighbor Thcon-is proach promotes policy-based routing, but at the expense
ap-of needing to calculate paths from incomplete detail aboutthe network topology Thus BGP will not always choosethe optimal path across an internetwork to reach a par-ticular destination It does, however, allow a country orcompany constituting an AS to make appropriate political
or business decisions about when and where to route itstraffic
Questions about the scalability of BGP have beenraised in light of predictions for continued substantialgrowth in Internet traffic, and particularly as more orga-nizations consider deploying delay-sensitive applicationsover the Internet (e.g., voice, video, conferencing) Intelli-gent route control, virtual routing, and new approaches totraffic engineering are among the options being explored
to solve performance problems before they become ous impediments to effective use of the Internet
seri-Multiprotocol Label Switching (MPLS)
Multiprotocol label switching (MPLS) has been designed
by the IETF to improve the performance of routed works by layering a connection-oriented framework over
net-an IP-based internetwork MPLS requires edge routers toassign labels to traffic entering the network so that in-termediate routers (called label-switching routers, LSRs)can make forwarding decisions quickly, choosing the ap-propriate output port according to the packet’s label andrewriting that label (which is intended to have local
Trang 38DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
significance only) as necessary MPLS represents a
sig-nificant shortcut from the usual IP approach, where every
relay node must look deeply into the packet header, search
a routing table for the best match, and then select the best
next hop toward the packet’s destination All packets with
the same MPLS label will follow the same route through
the network In fact, MPLS is designed so that it can
ex-plicitly and flexibly allocate network resources to meet
particular objectives such as assigning the fastest routes
for delay-sensitive packet flows, underutilized routes to
balance traffic better, or multiple routes between the same
end-points for flows with different requirements This is
called traffic engineering and serves as the foundation for
both optimizing performance and supporting QoS
guar-antees
Nothing about the MPLS design limits its use to the
IP environment; it can work with suitably equipped ATM
and frame relay routers as well In fact, it can coexist with
legacy routers not yet updated with MPLS capabilities,
and it can be used in an internetwork that contains a mix
of IP, ATM, and frame relay Another powerful feature is
the ability to stack labels on a last-in-first-out basis, with
labels added or removed from the stack by each LSR as
appropriate This allows multiple label-switched paths to
be aggregated into a tunnel over the common portion of
their route for optimal switching and transport MPLS
is also a convenient mechanism to support virtual
pri-vate networks, especially when multiple Internet service
providers are involved along the path from one end to the
other
Signaling and Interworking
Connection-oriented networks require specific
mecha-nisms for establishing a circuit (physical or virtual) prior
to traffic flow, and for terminating the circuit
after-ward In the circuit-switched telephony environment, call
setup and termination are part of a well-developed set of
telecommunication system control functions referred to
as signaling MANs and WANs that were built for voice
included signaling as an integral part of their designs,
because resources were dedicated to each call as it was
established and needed to be released after call
comple-tion
The ITU-T began developing standards for digitaltelecommunication signaling in the mid-1960s; these have
evolved into common channel interoffice signaling system
7 (CCIS7, known in the United States as Signaling System
7, or just SS7 for short), currently in use around the world.
SS7 is an out-of-band mechanism, meaning that its
mes-sages do not travel across the same network resources as
the conversations it was designed to establish and control
In fact, SS7 uses packet switching to deliver control
mes-sages and exchange data, not just for call setup, but also
for special features such as looking up a toll-free number
in a database to find out its real destination address, call
tracing, and credit card approvals Out-of-band delivery
of the messages allows SS7 to be very fast in setting up
calls, to avoid any congestion in the transport network,
and also to provide signaling any time during a call
The SS7 network has a number of elements that worktogether to accomplish its functions (Figure 13):
STP
STP
SCPSSP
SSP
datatransportnetwork
SS7network
DB1
1
Figure 13: SS7 network elements.
Signal switching points (SSPs) are the network edge vices responsible for setting up, switching, and termi-nating calls on behalf of connected subscriber devices,and thus insert user traffic into, and remove it from,the service provider’s backbone network
de-Signal transfer points (STPs) are packet switches sible for getting SS7 messages routed through thecontrol network
respon-Signal control points (SCPs) house the databases thatsupport advanced call processing
In packet-switched MANs and WANs, signaling hadbeen associated primarily with establishing and tearingdown SVCs that required no further control during thedata transfer phase With a rising interest in multime-dia communications (e.g., video, and especially voice overIP) however, the ITU-T quickly recognized a need for ad-
ditional capabilities Their H 323 recommendations
en-compass an entire suite of protocols that cover all aspects
of getting real-time audio and video signals into packetform, signaling for call control, and negotiation to ensurecompatibility among sources, destinations, and the net-work H.323 takes advantage of prior ITU work (such asISDN’s Q.931 signaling protocol) and defines four majorelements (Figure 14):
Terminals are the end-user devices that originate and
re-ceive multimedia traffic
Gateways primarily handle protocol conversions for
par-ticipating non-H.323 terminals, as would be found inthe public switched telephone network (PSTN)
Gatekeepers are responsible for address translation, call
control services, and bandwidth management
Multipoint Control Units (MCUs) provide
multiconferenc-ing among three or more terminals and gateways
The IETF took a simpler approach to signaling with the
session initiation protocol (SIP), which was designed as a
lightweight protocol simply to initiate sessions betweenusers SIP borrows a great deal from the hypertext trans-fer protocol (HTTP), using many of the same header fields,
Trang 39DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
W IDE A REA AND M ETROPOLITAN A REA N ETWORKS
Figure 14: H.323 network elements.
encoding rules, error codes, and authentication methods
to exchange text messages Like H.323, SIP assumes that
the end-point devices (i.e., terminals) are intelligent,
run-ning software known as the user agent The agent has two
components: User Agent Client, which is responsible for
initiating all outgoing calls, and the User Agent Server,
which answers incoming calls In the network itself, SIP
provides support with three types of server:
Registration servers keep track of where all users are
lo-cated
Proxy servers receive requests and forward them along to
the next appropriate hop in the network
Redirect servers also receive requests and determine the
next hop, but rather than forwarding the request, they
return the next-hop server address to the requester
An alternative approach to multimedia
communica-tion control developed by the IETF is called the media
gateway control protocol (MGCP) It is quite different
from H.323 and SIP because it assumes that the
end-user devices are not very intelligent Consequently MGCP
takes a central server approach to communication
coor-dination and control Two elements are defined: the
Me-dia Gateway Controller (also known as the call agent),
which provides the central intelligence and controls all
of the Media Gateways, which perform a variety of
in-terface functions such as with the PSTN, residential
de-vices, and business private branch exchanges (PBXs)
MGCP defines the communication that takes place
be-tween the call agent and the Gateways that execute its
commands
In practice H.323, SIP, and MGCP will likely coexist to
support multimedia communication in the Internet
en-vironment because each has advantages for specific
ap-plications or coverage MGCP is particularly useful to
MAN/WAN service providers with large installed bases of
unintelligent end-point devices, and its gateway approach
allows for tailored interfaces to each different
underly-ing technology The simplicity of SIP is more attractive
to enterprise networks designed primarily for data traffic
with smaller requirements for supporting voice and video
Finally, H.323 is the most mature and most
comprehen-sive As usual in the telecommunication industry, vendorsupport and suitability to customer business models arelikely to determine which, if any, one approach becomesdominant
PROVIDERS AND SERVICES
Carriers and Service Providers
The public provision of telecommunication services tosubscribers for a fee has a history of being government-regulated in most parts of the world (the term “commoncarrier,” for example, dates back to public transportationfor people, first by stagecoach, then by trains, buses, etc.).Regulation was required because access to telecommuni-cation services depended on cabling that was run fromsubscriber premises (residential or business) across pub-lic property (e.g., along roads) to a provider’s central of-fice as a service point Governments could also imposestandards to ensure that services offered by providers indifferent locations would be compatible enough to inter-operate In some countries, infrastructure was built andservices operated by the government itself (e.g., PTTs thatprovided postal, telegraph, and telephone services nation-wide) In the United States, telephone industry regulationwas divided between LECs whose cabling and local ser-vices go to individual premises, and IXCs who providedthe interconnection (i.e., long-distance services) betweenLECs
The Internet as a means of public data tion has grown up rather differently, driven largely by theU.S regulatory environment, where telecommunicationcompanies were prohibited from providing data services.Consequently, a new type of company called an Internetservice provider (ISP) was born Data would move from asubscriber’s premises, across cables belonging to an LEC,
communica-to ISP equipment in a point of presence, where it wastransferred onto Internet resources The subscriber thushad to be a customer of both the LEC and the ISP unless
a private link could be installed directly to the ISP’s POP.The Internet connections from one ISP location to an-other are most often lines leased from an IXC As telecom-munication services have been increasingly deregulatedworld-wide, the distinctions among voice and data serviceproviders have become blurred
Trang 40DeNoia WL040/Bidgolio-Vol I WL040-Sample.cls June 20, 2003 17:57 Char Count= 0
It is important to remember that “the Internet” is notreally a single entity, but rather an interconnected set of
autonomous networks whose owners have agreed to
co-operate and use a common set of standards to ensure
in-teroperability Peering is a form of interconnection where
ISPs agree to exchange traffic for their respective
cus-tomers, based on a specific set of business terms Peering
points are where the networks actually connect to effect
this exchange The number and location of peering points
and partners is decided by each ISP according to customer
demand and its own business criteria Subscribers may
need to be aware of these agreements in order to
under-stand fully the performance they can expect end to end
across the Internet
Just as the background and emphasis of traditionalvoice and traditional data service providers differ, so do
their business models and their choices of technology
Some offer only transport for traffic, either between
sub-scriber sites or to the Internet Others offer access to
ap-plications or management services Local
telecommuni-cation carriers tend to offer MAN services over an ATM
and SONET infrastructure, while data providers would
be more likely to offer IP services or simply Ethernet
ac-cess and transport Cable television and wireless service
providers also offer access services according to the
char-acteristics of their infrastructure technologies The
op-tions available will likely continue to grow as technology
progresses
Class of Service, Quality of Service
As interest in carrying multimedia or multiple-service
traffic (i.e., voice, data, video) over MANs and WANs
has grown, managing the traffic to provide
perfor-mance appropriate to each application has become more
important Quality of service techniques are expected to
guarantee performance and delivery, usually in terms of
bandwidth allocation, timeliness of delivery, and minimal
variation in delay (e.g., ATM service categories) Class of
service (CoS) techniques do not make such guarantees,
but rather attempt to meet user requests on a best-effort
basis Typically CoS works by grouping together traffic
with similar requirements (e.g., voice or streaming video)
and using a priority queuing system so that switches and
routers forward the traffic accordingly Connectionless
network services such as IP offer CoS traffic management,
while connection-oriented services such as ATM provide
QoS
QoS cannot really be guaranteed unless it is availableall the way from end to end of the connection This creates
a challenge for MAN and WAN environments where
mul-tiple technologies from one or more service providers may
be involved in delivering user traffic, and especially when
the traffic originates or terminates in a LAN of yet another
different technology Several groups are involved in
devel-oping standard techniques for CoS and QoS The problem
is making sure that appropriate translation mechanisms
can carry user application requirements across network
and SP boundaries:
IEEE 802.1p is a Layer-2 tagging mechanism to specify
priority using 3 bits in the Layer-2 frame header
IETF’s differentiated services (DiffServ) indicates howpackets are to be forwarded using per-hop behavior(PHB) queuing, or discarded if there is not sufficientbandwidth to meet performance requirements.ATM traffic management defines service categories andtraffic classes
Virtual Private Networks
A virtual private network (VPN) is a special service thatamounts to establishing a closed user group capabil-ity over a shared or public network infrastructure Thismeans that access is restricted to authorized users only,privacy of data content is assured, traffic belonging withinthe VPN does not get out or become visible to unautho-rized users, and outside traffic does not get in VPNs arebecoming a very attractive way for organizations to re-duce the cost of private WANs while improving the secu-rity for traffic that travels over public networks Wherehigh-speed MAN and WAN services are available, long-distance performance can even be kept reasonably close
to what the remote users would experience if they weredirectly connected to the LAN VPNs may also be built tosend traffic across the Internet, with one or more SPs pro-viding the access links between the Internet and variousgeographically dispersed customer sites Internet VPNscan be significantly less expensive than the private lines
or networks they replace
Management
The OSI model for network management encompassesfive functional areas: configuration management, perfor-mance management, fault management, accounting man-agement, and security management A MAN or WAN ser-vice provider must cover these from the perspective ofboth operating the entire network effectively and balanc-ing the needs and expectations of paying customers whocould always choose to take their business elsewhere Op-eration must be reliable, there must be sufficient capacity
to meet traffic needs and performance expectations, andprivacy must be maintained not only for the content ofthe traffic carried but also for data about the customers
At the same time, subscribers typically want the ability
to manage the performance and flow of their own trafficthrough their allotment of SP resources SP operation sys-tems must be capable and sophisticated to meet all theserequirements
A primary mechanism used to establish and manage
expectations between customers and providers is the
ser-vice level agreement (SLA) SLAs are the defining
docu-ments (contracts) that spell out what services and levels
of support will be provided to the customer at a specifiedprice Successful SLAs are built on a solid, shared un-derstanding of business priorities and service impact, forboth the service user and the service provider Detail aboutroles and responsibilities, metrics and reporting, addedcost for incremental services or enhancements, escala-tion procedures, and change management are just some
of what should be covered in an SLA Many customersalso build in penalties in case the provider fails to de-liver services at the level specified in the SLA This may be