2006, Identifying significant predictors of injury severity in traffic accidents using a series of artificial neural networks Accident Analysis and Prevention ; 38:434-444.. 2006, Time seri
Trang 1Dai Y., Nakano Y (1998), Recognition of facial images with low resolution using a Hopfield memory model Pattern Recognition ; 31:159-167
Dasu T., Johnson T (2003), Exploratory Data Mining and Data Cleaning New Jersey: Wiley
De Groot D., Wurtz D (1991), Analysis of univariate time series with connectionist nets: A case study of two classical examples Neurocomputing ;3:177-192
Deboeck G., Kohonen T (1998), Visual Explorations in Finance with Self-organizing Maps London: Springer-Verlag
Delen D., Sharda R., Bessonov M (2006), Identifying significant predictors of injury severity
in traffic accidents using a series of artificial neural networks Accident Analysis and Prevention ; 38:434-444
Dhar V., Chou D (2001), A comparison of nonlinear methods for predicting earnings sur-prises and returns IEEE Transactions on Neural Networks ; 12:907-921
Dia H (2001), An object-oriented neural network approach to short-term traffic forecasting European Journal of Operation Research ; 131:253-261
Dittenbach M., Rauber A., Merkl, D (2002), Uncovering hierarchical structure in data using the growing hierarchical self-organizing map Neurocompuing ; 48:199-216
Doganis P., Alexandridis A., Patrinos P., Sarimveis H (2006), Time series sales forecasting for short shelf-life food products based on artificial neural networks and evolutionary computing Journal of Food Engineering ; 75:196-204
Dutot A.L., Rynkiewicz J., Steiner F.E., Rude J (2007), A 24-h forecast of ozone peaks and exceedance levels using neural classifiers and weather predictions Modelling and Software; 22:1261-1269
Dutta S., Shenkar S (1993), “Bond rating: a non-conservative application of neural net-works.” In Neural Networks in Finance and Investing, Trippi, R., and Turban, E., eds Chicago: Probus Publishing Company
Enke D., Thawornwong S (2005), The use of data mining and neural networks for forecast-ing stock market returns Expert Systems with Applications ; 29:927-940
Evans O.V.D (1997), Discovering associations in retail transactions using neural networks ICL Systems Journal ; 12:73-88
Fahlman S., Lebiere C (1990), “The cascade-correlation learning architecture.” In Advances
in Neural Information Processing Systems, Touretzky, D., ed
Fletcher R (1987), Practical Methods of Optimization 2nd Chichester: John Wiley & Sons Frean M (1990), The Upstart algorithm: a method for constructing and training feed-forward networks Neural Computations ; 2:198-209
Funahashi K (1998), Multilayer neural networks and Bayes decision theory Neural Net-works ; 11:209-213
Gallinari P., Thiria S., Badran R., Fogelman-Soulie, F (1991), On the relationships between discriminant analysis and multilayer perceptrons Neural Networks ; 4:349-360 Geman S., Bienenstock E., Doursat T (1992), Neural networks and the bias/variance dilemma Neural Computation ; 5:1-58
Gorr L (1994), Research prospective on neural network forecasting International Journal of Forecasting ; 10:1-4
He H., Wang J., Graco W., Hawkins S (1997), Application of neural networks to detection
of medical fraud Expert Systems with Applications ; 13:329-336
Hebb D.O (1949), The Organization of Behavior New York: Wiley
Hinton G.E (1992), How neural networks learn from experience Scientific American; 9:145-151
Trang 2Hornik K., Stinchcombe M., White H (1989), Multilayer feedforward networks are universal approximators Neural Networks ; 2:359–366
Hopfield J.J (2558), (1982), Neural networks and physical systems with emergent collective Hopfield J.J., Tank D.W (1985), Neural computation of decisions in optimization problems Biological Cybernetics ; 52:141-152
Hu J.Q., Rose, E (1995), On-line fuzzy modeling by data clustering using a neural network Advances in Process Control , 4, 187-194
Huang J.S., Liu H.C (2004), Object recognition using genetic algorithms with a Huang Z Chen, H., Hsu, C.J Chen, W.H and Wu, S., Credit rating analysis with support vector machines and neural networks: a market comparative study Decision Support Systems ; 37:543-558
Hopfield’s neural model (1997) Expert Systems with Applications 1997; 13:191-199 Jain L.C., Vemuri V.R (1999), Industrial Applications of Neural Networks Boca Raton: CRC Press
Kiang M.Y., Hu, M.Y., Fisher D.M (2006), An extended self-organizing map network for market segmentation—a telecommunication example Decision Support Systems ; 42:36-47
Kiang M.Y., Kulkarni U.R., Tam K.Y (1995), Self-organizing map network as an interactive clustering tool-An application to group technology Decision Support Systems ; 15:351-374
Kim T., Kumara S.R.T., (1997), Boundary defect recognition using neural networks Inter-national Journal of Production Research; 35:2397-2412
Kim T.Y., Oh K.J., Sohn K., Hwang C (2004), Usefulness of artificial neural networks for early warning system of economic crisis Expert Systems with Applications ; 26:583-590
Kirkos E., Spathis C., Manolopoulos Y., (2007), Data Mining techniques for the detection of fraudulent financial statements Expert Systems with Applications ; 32: 995-1003 Kiviluoto K (1998), Predicting bankruptcy with the self-organizing map Neurocomputing ; 21:203-224
Klein B.D., Rossin D F (1999), Data quality in neural network models: effect of error rate and magnitude of error on predictive accuracy Omega ; 27:569-582
Kohonen T (1982), Self-organized formation of topologically correct feature maps Biolog-ical Cybernetics ; 43:59-69
Kolehmainen M., Martikainen H., Ruuskanen J (2001), Neural networks and periodic com-ponents used in air quality forecasting Atmospheric Environment ; 35:815-825 Law R (2000), Back-propagation learning in improving the accuracy of neural network-based tourism demand forecasting Tourism Management ; 21:331-340
Lee D.L (2002), Pattern sequence recognition using a time-varying Hopfield network IEEE Transactions on Neural Networks ; 13:330-343
Lewis O.M., Ware J.A., Jenkins D (1997), A novel neural network technique for the valua-tion of residential property Neural Computing and Applicavalua-tions ; 5:224-229
Li W.J., Lee T., (2002), Object recognition and articulated object learning by accumulative Hopfield matching Pattern Recognition; 35:1933-1948
Lim G.S., Alder M., Hadingham P (1992), Adaptive quadratic neural nets Pattern Recogni-tion Letters ; 13: 325-329
Lisboa P.J.G., Edisbury B., Vellido A (2000), Business Applications of Neural Networks : The State-of-the-art of Real-world Applications River Edge: World Scientific
computational abilities Proceedings of National Academy of Sciences; 79(8):2554-2558
Trang 3McCulloch W., Pitts W (1943), A logical calculus of the ideas immanent in nervous activity Bulletin of Mathematical Biophysics; 5:115-133
Min S.H., Lee J., Han I (2006), Hybrid genetic algorithms and support vector machines for bankruptcy prediction Expert Systems with Applications ; 31: 652-660
Minsky M L., Papert S A (1969), Perceptrons MA: MIT press
Miyake S., Kanaya F (1991), A neural network approach to a Bayesian statistical decision problem IEEE Transactions on Neural Networks ; 2:538-540
Mozer M.C., Wolniewics R (2000), Predicting subscriber dissatisfaction and improving re-tention in the wireless telecommunication IEEE Transactions on Neural Networks; 11:690-696
Nag A.K., Mitra A (2002), Forecasting daily foreign exchange rates using genetically opti-mized neural networks Journal of Forecasting ; 21:501-512
Nelson M., Hill T., Remus T., O’Connor, M (1999), Time series forecasting using neural networks: Should the data be deseasonalized first? Journal of Forecasting ; 18:359-367 O’Connor N., Madden M.G (2006), A neural network approach to predicting stock exchange movements using external factors Knowledge-Based Systems ; 19:371-378
Paik J.K., Katsaggelos, A.K (1992), Image restoration using a modified Hopfield neural network IEEE Transactions on Image Processing ; 1:49-63
Pajares G., Cruz J.M., Aranda, J (1998), Relaxation by Hopfield network in stereo image matching Pattern Recognition ; 31:561-574
Panda C., Narasimhan V (2007), Forecasting exchange rate better with artificial neural net-work Journal of Policy Modeling ; 29:227-236
Parker D.B (1985), Learning-logic: Casting the cortex of the human brain in silicon, Tech-nical Report TR-47, Center for Computational Research in Economics and Management Science, MIT
Palmer A., Monta˜no J.J., Ses´e, A (2006), Designing an artificial neural network for forecast-ing tourism time series Tourism Management ; 27: 781-790
Partovi F.Y., Anandarajan M (2002), Classifying inventory using an artificial neural network approach Computers and Industrial Engineering ; 41:389-404
Petersohn H (1998), Assessment of cluster analysis and self-organizing maps International Journal of Uncertainty Fuzziness and Knowledge-Based Systems ; 6:139-149 Prybutok V.R., Yi J., Mitchell D (2000), Comparison of neural network models with ARIMA and regression models for prediction of Houston’s daily maximum ozone concentrations European Journal of Operational Research ; 122:31-40
Qi M (2001), Predicting US recessions with leading indicators via neural network models International Journal of Forecasting ; 17:383-401
Qi M., Zhang G.P (2001), An investigation of model selection criteria for neural network time series forecasting European Journal of Operational Research ; 132:666-680 Qiao F., Yang H., Lam, W.H.K (2001), Intelligent simulation and prediction of traffic flow dispersion Transportation Research, Part B ; 35:843-863
Raudys S (1998), Evolution and generalization of a single neuron: I., Single-layer perceptron
as seven statistical classifiers Neural Networks ; 11:283-296
Raudys S (1998), Evolution and generalization of a single neuron: II., Complexity of statis-tical classifiers and sample size considerations Neural Networks ; 11:297-313 Raviwongse R Allada V., Sandidge T (2000), Plastic manufacturing process selection methodology using self-organizing map (SOM)/fuzzy analysis International Journal of Advanced Manufacturing Technology; 16:155-161
Reed R (1993), Pruning algorithms-a survey IEEE Transactions on Neural Networks ; 4:740-747
Trang 4Remus W., O’Connor M (2001), “Neural networks for time series forecasting.” In Princi-ples of Forecasting: A Handbook for Researchers and Practitioners, Armstrong, J S ed Norwell:Kluwer Academic Publishers, 245-256
Reutterer T., Natter M (2000), Segmentation based competitive analysis with MULTICLUS and topology representing networks Computers and Operations Research; 27:1227-1247
Richard, M (1991), D., Lippmann, R., Neural network classifiers estimate Bayesian
aposteriori probabilities Neural Computation ; 3:461-483.
Ripley A (1993), “Statistical aspects of neural networks.” In Networks and Chaos - Statisti-cal and Probabilistic Aspects, Barndorff-Nielsen, O E., Jensen J L and Kendall, W S eds London: Chapman and Hall, 40-123
Ripley A (1994), Neural networks and related methods for classification Journal of Royal Statistical Society, Series B ; 56:409-456
Roh T H (2007), Forecasting the volatility of stock price index Expert Systems with Ap-plications ; 33:916-922
Rosenblatt F (1958), The perceptron: A probabilistic model for information storage and organization in the brain Psychological Review ; 65:386-408
Rout S., Srivastava, S.P., Majumdar, J (1998), Multi-modal image segmentation using a modified Hopfield neural network Pattern Recognition ; 31:743-750
Rumelhart D.E., Hinton G.E., Williams R.J (1986), “Learning internal representation by back-propagating errors.” In Parallel Distributed Processing: Explorations in the Mi-crostructure of Cognition Press, Rumelhart, D.E., McCleland, J.L and the PDP Research Group, eds MA: MIT
Saad E.W., Prokhorov D.V., Wunsch, D.C II (1998), Comparative study of stock trend pre-diction using time delay, recurrent and probabilistic neural networks IEEE Transactions
on Neural Networks; 9:456-1470
Salcedo-Sanz S., Santiago-Mozos R.,Bousono-Calzon, C (2004), A hybrid Hopfield network-simulated annealing approach for frequency assignment in satellite communi-cations systems IEEE Transactions on System, Man and Cybernetics, Part B:108-116 Sarle W.S (1994), Neural networks and statistical models Poceedings of the Nineteenth Annual SAS Users Group International Conference, Cary, NC: SAS Institute
Schumacher M., Robner R., Vach W (1996), Neural networks and logistic regression: Part I., Computational Statistics and Data Analysis ; 21:661-682
Smith K.A., Ng, A (2003), Web page clustering using a self-organizing map of user naviga-tion patterns Decision Support Systems ; 35:245-256
Smith K.A., Willis R.J., Brooks M (2000), An analysis of customer retention and insur-ance claim patterns using data mining: a case study Journal of the Operational Research Society; 51:532-541
Soulie F.F., Gallinari P (1998), Industrial Applications of Neural Networks River Edge, NJ: World Scientific
Suganthan P.N., Teoh E.K., Mital D.P (1995), Self-organizing Hopfield network for at-tributed relational graph matching Image and Vision Computing; 13:61-71
Sun Z.Z., Yu S (1995), Improvement on performance of modified Hopfield neural network for image restoration IEEE Transactions on Image processing; 4:683-692
Suykens J.A.K., Vandewalle J.P.L., De Moor B.L.R (1996), Artificial Neural Networks for Modeling and Control of Nonlinear Systems Boston: Kluwer
Swanson N.R., White H (1995), A model-selection approach to assessing the information in the term structure using linear models and artificial neural networks Journal of Business and Economic Statistics; 13;265-275
Trang 5Tatem A.J., Lewis H.G., Atkinson P.M., Nixon M.S (2002), Supre-resolution land cover pattern prediction using a Hopfield neural network Remote Sensing of Environment; 79:1-14
Temponi C., Kuo Y.F., Corley H.W (1999), A fuzzy neural architecture for customer satis-faction assessment Journal of Intelligent & Fuzzy Systems; 7:173-183
Thieme R.J., Song M., Calantone R.J (2000), Artificial neural network decision support systems for new product developement project selection Journal of Marketing Research; 37:543-558
Vach W., Robner R., Schumacher M (1996), Neural networks and logistic regression: Part I Computational Statistics and Data Analysis; 21:683-701
Wang T., Zhuang X., Xing X (1992), Robust segmentation of noisy images using a neural network model Image Vision Computing; 10:233-240
Webb A.R., Lowe D., (1990), The optimized internal representation of multilayer classifier networks performs nonlinear discriminant analysis Neural Networks; 3:367-375 Werbos P.J., (1974), Beyond regression: New tools for prediction and analysis in the behav-ioral sciences Ph.D thesis, Harvard University, 1974
West D., (2000), Neural network credit scoring models Computers and Operations Research; 27:1131-1152
West P.M., Brockett P.L., Golden L.L., (1997), A comparative analysis of neural networks and statistical methods for predicting consumer choice Marketing Science; 16:370-391 Widrow B., Hoff M.E., (1960), Adaptive switching circuits, 1960 IRE WESCON Convention Record, New York: IRE Part 4 1960:96-104
Widrow B., Rumelhart D.E., Lehr M.A., (1994), Neural networks: applications in industry, business and science, Communications of the ACM; 37:93-105
Wong B.K., Bodnovich T.A., Selvi Y., (1997), Neural network applications in business: A review and analysis of the literature (1988-1995) Decision Support Systems; 19:301-320
Young S.S., Scott P.D., Nasrabadi, N.M., (1997), Object recognition using multilayer Hop-field neural network IEEE Transactions on Image Processing; 6:357-372
Zhang G.P., (2007), Avoiding Pitfalls in Neural Network Research IEEE Transactions on Systems, Man, and Cybernetics; 37:3-16
Zhang G.P., Hu M.Y., Patuwo B.E., Indro D.C., (1999), Artificial neural networks in bankruptcy prediction: general framework and cross-validation analysis European Jour-nal of OperatioJour-nal Research; 116:16-32
Zhang G.P., Keil M., Rai A., Mann J., (2003), Predicting information technology project escalation: a neural network approach European Journal of Operational Research 2003; 146:115–129
Zhang G.P., Qi M (2002), “Predicting consumer retail sales using neural networks.” In Neu-ral Networks in Business: Techniques and Applications, Smith, K and Gupta, J.eds Hershey: Idea Group Publishing, 26-40
Zhang G.P., Patuwo E.P., Hu M.Y., (1998), Forecasting with artificial neural networks: the state of the art International Journal of Forecasting; 14:35-62
Zhang W., Cao Q., Schniederjans M.J., (2004), Neural Network Earnings per Share Fore-casting Models: A Comparative Analysis of Alternative Methods Decision Sciences; 35: 205–237
Zhu Z., He H., Starzyk J.A., Tseng, C., (2007), Self-organizing learning array and its appli-cation to economic and financial problems Information Sciences; 177:1180-1192
Trang 6Granular Computing and Rough Sets - An
Incremental Development
Tsau Young (’T Y.’) Lin1and Churn-Jung Liau2
1 Department of Computer Science
San Jose State University
San Jose, CA 95192
tylin@cs.sjsu.edu
2 Institute of Information Science
Academia Sinica, Taipei 115, Taiwan
liaucj@iis.sinica.edu.tw
Summary This chapter gives an overview and refinement of recent works on binary granular computing For comparison and contrasting, granulation and partition are examined in parallel from the prospect of rough Set theory (RST).The key strength of RST is its capability in representing and processing knowledge in table formats Even though such capabilities, for general granulation, are not available, this chapter illustrates and refines some such capability for binary granulation In rough set theory, quotient sets, table representations, and concept hierarchy trees are all set theoretical, while in binary granulation, they are special kind of pretopological spaces, which is equivalent to a binary relation Here a pretopological space means a space that is equipped with a neighborhood system (NS) A NS is similar to the classical NS of a topological space, but without any axioms attached to it3
Key words: Granular computing, rough set, binary relation, equivalence relation
22.1 Introduction
Though the label, granular computing is relatively recent, the notion of granulation has in fact been appeared, under different names, in many related fields, such as pro-gramming, divide and conquer, fuzzy and rough set theories, pretopological spaces, interval computing, quantization, data compression, chunking, cluster analysis, be-lief functions, machine learning, databases, and many others In the past few years,
we have seen a renewed and fast growing interest in Granular Computing (GrC) Many applications of granular computing have appeared in fields, such as medicine, economics, finance, business, environment, electrical and computer engineering, a number of sciences, software engineering, and information science
3This is an expansion of the article (Lin, 2005) in IEEE connections, the news letter of the IEEE Computational Intelligence Society
O Maimon, L Rokach (eds.), Data Mining and Knowledge Discovery Handbook, 2nd ed.,
DOI 10.1007/978-0-387-09823-4_22, © Springer Science+Business Media, LLC 2010
Trang 7Granulation seems to be a natural problem-solving methodology deeply rooted
in human thinking Many daily ”things” have been routinely granulated into sub”things;” human body has been granulated into head, neck, and so forth; geo-graphic features into mountains, planes, and others The notion is intrinsically fuzzy, vague and imprecise Mathematicians idealized it into the notion of partitions, and developed it into a fundamental problem-solving methodology; it has played major roles throughout the entire history of mathematics
Nevertheless, the notion of partitions, which absolutely does not permit any over-lapping among its granules, seems to be too restrictive for real world problems Even
in natural science, classification does permit small degree of overlapping; there are
beings that are both appropriate subjects of zoology and botany A more general theory is needed
Based on Zadeh’s grand project on granular mathematics, during his sabbati-cal leave (l996/l997) at Berkeley, Lin focused on a subset of granular mathematics, which he called granular computing (Zadeh, 1998) To stimulate research on granu-lar computing, a special interest group, with T Y Lin as its Chair, was formed within BISC (Berkeley Initiative in Soft Computing) Since then, granular computing has evolved into an active research area, generating many articles, books and presen-tations at conferences, workshops and special sessions This chapter is devoted to present some of such development over the past few years
There are two possible approaches: (1) One is starting from fuzzy side and mov-ing down, and (2) the other one is from extreme crisp side and movmov-ing up In this chapter, we take the second approach incrementally Recall that algebraically a parti-tion is an equivalence relaparti-tion, so a natural next step is the binary granulaparti-tion defined
by a binary relation For contrasting, we may call a partition A-granulation and the more general granulation B-granulation
22.2 Naive Model for Problem Solving
An obvious approach to a large-scaled computing problem is: (1) To divide the prob-lem into subtasks, might be point by point and level by level (2) To elevate or abstract the problem into concept/knowledge spaces, could be in multilevels (3) To integrate the solutions of subtasks and quotient tasks (knowledge spaces) of several levels 22.2.1 Information Granulations/Partitions
In the first step, we select an appropriate system of granulation/partition so that only the summaries of granules/equivalence classes may enter into the higher level computing The information in data space is transformed to a concept space, pos-sibly in levels, which may be locally at each point or globally at eh whole uni-verse (Lin, 2003b) Classically, we granulate by partitioning (no overlapping on gran-ules) Such examples are plentiful: in mathematics (quotient groups, quotient rings and etc (Birkhoff and MacLane, 1977)), in theoretical computer science
(divide-and-conquer (Aho et al., 1974)), in software engineering (the structural, object oriented,
Trang 8and component based design and programming (Szyperski, 2002)), in artificial intel-ligence (Hobbs, 1985, Zhang and Zhang, 1992), in rough set theory (Pawlak, 1991) among others However, these are all partition based, where no overlapping of gran-ules is permitted As we have observed, even in biology, classification does allow
some overlapping The focus of this presentation will be on non-partition theory, but
only in an epsilon step away from partitioning method
22.2.2 Knowledge Level Processing and Computing with Words
The information in each granule is summarized and the original problem is expressed in terms of symbols, words, predicates or linguistic variables Such re-expressing is often referred to as knowledge representations Its processing has been termed computing with symbols (table processing, computing with words, knowl-edge level processing, even precisiated natural language, depending on the complex-ity of the representations
In this chapter, we are computing on the space of granules or ”quotient space.” in which each granule is represented by a word that carries different degree of seman-tics For partition theory, the knowledge representation is in table format (Pawlak, 1991) and its computation is syntactic in nature For binary granulation, that we have focused here, is semantic oriented We expand and streamline the previous works (Lin, 1998a, Lin, 1998b, Lin, 2000); the main idea is to transfer the computing with words into computing with symbols
Loosely speaking computing with symbols or symbolic computing is an “ax-iomatic” Computing: all rules of computing symbols are determined by the axioms The computation follows the formal specifications Such computing occurs only in
an ideal situation In many real world applications, unfortunately, such as non-linear computing, the formal specifications are often unavailable So computing with words are needed; it can be processed informally Semantics of words often may not be completely or precisely formalized Their semantic computing is often carried out
in the systems with human helps (the semantics of symbols are not implemented) Human enforced semantic computing are common in data processing environment 22.2.3 Information Integration and Approximation Theory
Most applications require the solutions be presented in the same level as input data
So the solutions often need to be integrated from subtasks (solutions in granules) and quotient tasks (solutions in the spaces of granules) For some applications, such
as Data Mining and some rough set theory, are aimed at high level information;
in such cases this step can be skipped In general, the integration is not easy In partition world, many theories have been developed in mathematics; e g., extension functors The approximation theory of pretopological spaces and rough set theory can be regarded as in this step
Trang 922.3 A Geometric Models of Information Granulations
For understanding the general idea, in this section, we recall and refine a previous formalization in (Lin, 1998a) The goal is to formalize Zadeh’s informal notion of granulation mathematically
As original thesis is informal, the best we could do is to present, hopefully, con-vincing arguments We believe our formal theory is very close to the informal one According to Zadeh (1996):
Information granulation involves partitioning a class of objects(points) into granules, with a granule being a clump of objects (points) which are drawn together by indistinguishability, similarity or functionality
We will literally take Zadeh’s informal words as a formal definition of granula-tion We observe that:
1 A granule is a group of objects that are draw together (by indistinguishability, similarity or functionality)
The phrase ”drawn together” implicitly implies certain level of symmetry among
the objects in a granule Namely, if p is drawn towards q, then q is also drawn towards p.
Such symmetry, we believe, is imposed by imprecise-ness of natural language
To avoid such an implications, we will rephrase it to ”drawn towards an object
p,” so that it is clear the reverse may or may not be true So we have first revision:
2 A granule is a group B(p) of objects that are draw toward an object p Here p
varies through every object in the universe
3 Such an association between object p and a granule B(p) induces a map from
the object space to power set of object space This map has been called a binary granulation (BG)
4 Geometric View:
We may use geometric terminology and refer to the granule as a neighborhood of
p, and the collection {B(p)} a binary neighborhood system (BNS) It is possible that B (p) is an empty set In this case we will simply say p has no neighborhood (abuse of language; to be very correct, we should say p has an empty
neighbor-hood) Also it is possible that different points may have the same neighborhood
(granule) B (p) = B(q) The set of all q, where B(q) is equal to B(p), is called the centers C(p) of B(p).
5 Algebraic View:
Consider the set R = {(p,u)}, where u in B(p) and p in U It is clear that R is a subset of U ×U, hence defines a binary relation (BR), and vice versa.
Proposition 1 A binary neighborhood system (BNS), A binary granulation (BG), and a binary relation (BR) are equivalent.
From the analysis given above, we propose the following mathematical model for information granulation
Trang 10Definition 1 By a (single level) information granulation defined on a set U we mean
a binary granulation (binary neighborhood system, binary relation) defined on U.
Let us goes a little bit further Note that the binary relation is a mathematical expression of Zadeh’s ”indistinguishability, similarity or functionality.” We abstract the three properties into a list of abstract binary relations{B j | j run through some
index set}, where each B jis a binary relation
Note that at each point p, each B j induces a neighborhood B j (p) Some may
be empty, or identical By removing empty set and duplications, the family have
been we re-indexed N i (p) As in the single level case, we will define directly the
granulation
N : U → 22U
; p "→ {B i (p) | i run through some index set }.
The collection{B i (p)} is called a neighborhood system(NS)or (LNS); the latter
one is used to distinguish itself from the neighborhood system (TNS) of a topological space (Lin, 1989a, Lin, 1992)
Definition 2 By a local multi-level information granulation defined on U, we mean
a neighborhood system (NS) is defined on U By a global multi-level information granulation defined on U, we mean a set of BG is defined on U.
All notions can be fuzzified The right way to look at this section is to assume implicitly there is a modifier ”crisp/fuzzy” to all notions presented above
22.4 Information Granulations/Partitions
Technically, granular computing is actually computing with constraints Especially
in “infinite world”, granulation is often given in terms of constraints In this chapter,
we concerns primarily with constraints that are mathematically represented as binary relations
22.4.1 Equivalence Relations(Partitions)
Partition is a decomposition of the universe into a family of disjoint subsets They are called equivalence classes, because a partition induces an equivalence relation and vice versa In this chapter, we will view the equivalence class in a special way Let
A ⊆ U ×U be an equivalence relation (a reflexive, symmetric and transitive binary relation) For each p, let
A p is the equivalence class containing p, and will be called A-granule for the purpose
of contrasting with general cases Elements in A pare equivalent to each other Let us summarize the discussions in: