This project report was drafted in the first year of the SPIN!-project to brief partners and prepare for a report deliverable due as part of Work Package 5. The report was placed online in August 2003 along with a follow up report concerning the State of the Art Geographical Data Mining that was drafted in January 2003. This report is also available as a word document.
Fundamentally this report details what exploratory spatial data analysis (ESDA) involves, outlining the special roles that inductive analytical techniques and visual environments play in supporting it. The report is based on a search for and the use of a variety of computer software that supports ESDA, discussions with users and developers of these tools, an examination of relevant literature and a careful consideration of all the available expert opinion. The report aims to encourage the development and integration of appropriate ESDA functionality in the spatial data mining system for data of public interest we are building (SPIN). It does not review the technology currently being integrated in SPIN, rather it provides an overview of the functionality of other available ESDA tools and suggests what of this and what additional functionality could further enhance the ESDA capabilities of SPIN.
Data mining systems (DMS), geographical information systems (GIS), mathematical software packages (MATHS), statistical software packages (STATS), and bespoke software technology are complementary tools with wide ranging functionality that can be applied to analyse data that relate to real world systems. The spatial data mining system we are building (SPIN) is an attempt to integrate much of the useful functionality of these tools in a highly extensible, open, internet-enabled plug in architecture that enhances ESDA capabilities for data of public interest. This report describes a number of these ESDA tools and suggests what functionality to integrate in SPIN in order to enhance its ESDA capabilities.
The main aims of the SPIN!-project are:
This report does not review the technology currently being integrated in SPIN. Instead it aims to increase awareness of the special nature of spatial data, space-time data and geographical information, outline the problems concerning geographical data mining (GDM), describe the utility of available ESDA tools, and suggest some ideas for developing the analytical capabilities of SPIN. Table 1 lists some of the special features of geographical information.
More basic than geographic information is spatial data, which is data attributed to a point line area or region defined by some geometrical system. Geographical data is spatial data that also relates to some period or periods in time. The spatial reference relates to some three-dimensional region on or near the surface of the earth that is abstracted in some simplified form. Often a projection is used to cope with the curvature of the earth surface so that data can be viewed as flat surface. Different projections distort distance area direction and orientation in different amounts using different surface constraint. Some projections are designed to minimise a specific type of distortion all over the map, but this is very difficult and no matter what is done there is distortion which tends to vary over the map and be greatest at its extremes. There are many different types of projections and various ways of transforming geographical data from one to another. Projection functionality is available in most standard GIS. Many geographical data sets relate to small spatial regions for which the effects of the curvature of the earth are negligible compared to the effects of study area definition. For small area data the problems of map projection do not take effect until data sets are merged to form significantly larger regions.
Geographical data can be spatially referenced in a number of ways. Commonly it is either referenced irregularly to points lines surfaces or regions (using a co-ordinate system with a specified origin and level of precision), or it is represented by a regular regionalisation or fixed set of evenly spaced measurements with an implicit topology (for a specified origin and spatial resolution). These are often referred to as vector data and raster data respectively. Functionality for converting data from one form into another is available in most standard GIS. Most standard GIS can manipulate display and analyse both types of data but they have tended to specialise in methods relevant to one or the other. Much geographical data is two-dimensional and relates to small areas on or near the surface of the earth. However, there is an increasing amount of data collected and referenced spatially with three-dimensional co-ordinate systems (e.g. soil data, seismic data, hydrological data and climatological data) nd increasingly two-dimensional data is being made three-dimensional by adding a height variable. For example a two-dimensional set of lines representing rivers can be draped over coincident digital terrain models to produce more useful three-dimensional sets of lines. Many GIS now include three-dimensional data handling display and analysis functionality.
As mentioned above, geographic data is often also temporally referenced, that is, it is collected and related to a specific period or to various periods in time. Often the temporal reference is fixed and implicit for a set of spatial data, but increasingly it is an explicitly recorded variable. Space-time referenced data are sometimes referred to as tri-space data, that is, they may have a single or multiple set of attributes and they relate specifically to region(s) in space and a period(s) of time. Really, GDM goes beyond ESDA and spatial data mining where time is either fixed or regarded no differently to any other aspatial variable. GDM involves searching for and generalising patterns and relationships in tri-space data, which is notoriously difficult, see Openshaw (1998).
ESDA and spatial data mining are terms that appear frequently in geographical contexts, but it is worth while remembering that it is a generic term that might be used in other areas of spatial science. Fundamentally, ESDA demands an awareness of the special nature of spatial data and the exploratory approach to data analysis. The exploratory approach involves visualising data and generalisations (derivatives) of them in tables, graphs, maps and various other types of display. Human analysts interpret these displays to gain an appreciation of the nature of the data, which often helps them to define interesting subsets of data and guide the search for interesting and potentially useful patterns and interrelationships in it. ESDA and data mining in general are interactive and iterative search processes where patterns and relationships in data are used to refine the search for more interesting patterns and relationships. With a very large amount of data, ESDA is equivalent to spatial data mining. To clarfy the terminology, geographical data mining (GDM) is spatial data mining of data relating to geographical processes, and spatial data mining is ESDA involving very large quantities of spatial data. GDM, spatial data mining and ESDA are essentially very similar activities that can begin without any hypothesis to test regarding the nature of patterns in the data and how they relate to real world processes. The basic idea is to encourage the data to express themselves in order to identify interesting patterns that help generate hypotheses about the processes underlying them in the real world.
Data generated theories or models of a process can only be partly corroborated by empirical measurement, there is always some interpretation going on as to the processes at work. In an exploratory data led approach to theory generation it is very important to try and keep as separate as possible data for building and data for validating models. As well as this it is also advisable when developing tools to find specific patterns that their capabilities are tested using synthetic data where the patterns are known to some extent. Further to this there is a need to interpret the quality of data going in and results coming out. It is often a good idea to test how likely any patterns observed in the data are caused by error, and how sensitive observed patterns are when random noise is added to the data. In a way the null hypothesis in the exploratory approach is that all data patterns are a consequence of randomness and data error.
DMS, GIS, MATHS, STATS, bespoke programs and computer operating systems are used in conjunction for ESDA and spatial data mining. Although each different type of software is specialised for a particular use there is an overlap in functionality between them albeit that some are faster and more efficient than others and can cope better with large volumes of data. Individual software tools are becoming increasingly flexible and there is increasingly less need to have many different tools to analyse data in general. The remainder of this section introduces GIS, DMS, MATHS and STATS in turn, then outlines the remainder of this report.
Standard GIS packages offer only very basic sorts of data mining functionality and nothing in the way of inductive analysis techniques that are the core functionality of DMS. Standard GIS also offer only very limited statistical functionality and support only fairly basic mathematical operations on tables of data. Perhaps the most important thing to appreciate about standard GIS is that they tend to be limited to fairly basic forms of spatial analysis, (perhaps with the exception of network analysis at which some are quite sophisticated). A further criticism of most standard GIS is that they do not offer linked dynamic display environments that can be used to search for interesting data patterns and relationships.
However, in the last few years a new generation of GIS has been emerging which do provide linked dynamic display environments. This new generation have mainly been developed separately from the more standard proprietary GIS by academic researchers who already have these tools available. Consequently the new generation do not have much of the more standard functionality, but they do enhance ESDA capabilities by facilitating an interactive exploratory and graphical approach to analysing spatial data. Like standard GIS - new generation GIS still lack data mining tools and useful geometry related spatial analysis functionality, but this should not detract from what they do provide. LiveMap, CDV, SAGE, and GeoTools are examples of new generation GIS that are described in Section 3.
DMS have developed largely unconnected to GIS research and offer a range of useful data analysis tools. DMS aim to provide user-friendly interfaces and visual programming environments that promote an interactive search for patterns in data. The majority of DMS are equipped to analyse multivariate time series data taking account of the linear or cyclical transgression of such data. Unfortunately, most fail to recognise the need to treat spatial data and spatial variables any differently from their aspatial relatives. Contemporary DMS do not offer an easy to use highly automated push button technology and instead rely greatly on the users understanding of how the data represents reality and how the available tools can be sensibly used to analyse it. Due to the nature of the data mining industry there are a vast number of commercial systems that exist and a vast amount of hype surrounding them. This means that in descriptions of the systems there are many exaggerated and misleading claims about their caabilities and a lack of information about their weaknesses and indeed the functionality of the tools they contain and how they work. The systems themselves are also very expensive and some are tuned to work on very specific hardware making them hard to evaluate and compare. Many DMS developed primarily as interfaces to neural networks with some kind of data input and display functionality. The majority of packages also contain other methods for classifying data, tools for linear and logistic regression, for generating decision trees, for association rules detection and for sequence discovery. Section 2 first outlines the functionality and capability of contemporary DMS in more detail, then outlines GeoMiner, which is the only spatial data mining system that was found during the search for them. It falls some way short of reasonable expectations of such a system.
In MATHS the most useful additional spatial analysis functionality relates to geometry and fuzzy logic. Geometry forms the foundations of spatial analysis and is a vast branch of mathematics and thus cannot be detailed here. Capability to construct and use geometrical shapes for generalising spatial data is undoubtedly a requirement for SPIN. Fuzzy logic based methods of classification, prediction and inference are more relevant and are outlined in Section 4.
STATS support statistical inference methods that are commonly used for analysing the distribution of aspatial variable values. There is a problem applying many of these methods sensibly to spatial data due to the inherent assumptions about the nature of the data on which they are based. This does not mean that they cannot be useful for some forms of spatial analysis, only that the results of applying these methods should be interpreted cautiously. Other statistical methods, which do not infer the nature of variable distributions or data dependencies are more relevant to ESDA and GDM. There are some spatial statistical methods can be applied to analyse spatially irregular or sampled data, but most only work well with spatially regular data. One recipe for spatial statistical analysis is outlined here: The first step involves plotting all the data for each variable and assessing whether it comes from a stationary random field by checking for stationarity (a constant mean within specified regions across he study region). If a spatial variable is found not to exhibit stationarity then various trend surfaces that relate nearby values of the variable are fitted and assessed by testing the residual differences between observed and expected values for stationarity. Having taken away the univariate trend the analysis proceeds by mapping and attempting to correlate variables in the data set to form more complex models of spatial association. In fact, few STATS facilitate this kind of spatial statistical analysis, an exception is SpaceStat that is described in Section 5. Much of the hypothesis testing functionality concerning probability distributions commonly found in STATS is difficult to apply sensibly in a spatial data mining context due to fundamentally flawed assumptions. Local indicators of spatial association (such as Geary's C and Moran's I) and local statistics (like the mean, standard deviation and correlation coefficients) are useful and so SPIN should provide ways of computing these for various spaial subsets. Testing patterns against randomness and using bootstrap methods is of course also very relevant to SPIN.
Altogether there are a vast number of tools that can be used for ESDA and spatial data mining. Perhaps one of the most challenging problems in developing SPIN concerns what functionality to include, what order to develop it in and what levels of automation to provide. This review is a reference point in addressing these issues aiming to provide a synoptic overview rather than a comprehensive and detailed evaluation of what is currently available. The main assumption of this report is that many end users will not have expertise in the statistical or spatial sciences but will hope to perform some sort of geographical data mining. The reader of this report is assumed to be aware of the functionality available in proprietary GIS and the proposed SPIN. Section 2 describes the exploratory or data mining approach to the analysis of large spatial databases again and outlines the core functionality of DMS briefly reviewing Clementine, Intelligent Miner for Data and GeoMiner. Section 3 outlines some of the new eneration GIS that offer linked visual display environments including CDV, SAGE, LiveMap, GeoTools and GeoVista. Section 4 is concerned with fuzzy logic. Section 5 outlines two other spatial STATS packages SpaceStat and S+ and a bespoke ESDA tool called GWR. Finally, Section 6 presents a summary.
To recap, spatial data mining systems are essentially ESDA tools. ESDA and spatial data mining involve an interactive search for patterns or regularities in spatial data. The processes could begin by visualising data in maps, graphs, charts, plots and tables. Following this some form of data classification and some form of variable prediction could be performed, the results visualised and the predictions and models tested in some way. This process could continue until something interesting or useful emerged or until the analyst became satisfied that the data contained no detectable patterns of interest.
Vastly increasing quantities of spatially referenced data of all kinds are being collected all the time, much of it relates to geographical and geophysical processes and much of it is of public interest. This has led to data rich but theory poor environments that fundamentally require an exploratory rather than a confirmatory hypothesis driven approach. Tools needed to help generalise and visualise patterns in these data have begun to emerge, but tools that use all the available data at the highest level of resolution with a minimum of preselection as yet have not.
ESDA tools, in particular SPIN, should be designed to be end-user friendly in that the results should be understandable and capable of being safely used by people who do not have higher degrees in the statistical or spatial sciences. Outputs from the tools should aim to stimulate the imagination. If there are patterns hidden in spatial data or indeed relationships encrypted in spatial information then the ESDA tools in SPIN should help users to find them by at least pointing them in the right direction. An important goal is to provide an end-user friendly ESDA environment containing highly automated analysis tools that are easy to apply, whose visualised outputs are easy to interpret and whose techniques are explained in simple terms so that non-experts can understand how they work.
It is important to appreciate the vast difference between exploring a manageable small data set with few variables and the need to perform the same process on massive databases (with many more cases) and possibly high levels of multivariate complexity. Manual graphical exploration of multi dimensional spatio-temporal data does not scale well, that is, the larger the volume of data gets - the greater is the need for generalisation and automated analysis methods.
There are a number of data mining tools designed to assist the process of exploring large amounts of data in search of recurrent patterns and relationships. Many developers and users of these tools appear to believe that their tools will work with any and all data regardless of its subject origins. This overlooks the features of geographical information that make it special that were summarised above in Table 1. GDM is a special type of data mining that has to take into account the special features of geographical information, the rather different styles and needs of analysis and modelling relevant to the world of GIS, and the peculiar nature of geographical explanation.
It can be argued that conventional data mining tools are still useful, in the same way that conventional statistical methods can be applied to spatial data. However, the view expressed here is that conventional statistical techniques and conventional data mining tools are not sufficiently focused on, and tailored to, the special needs of geographical analysis today. Most data mining packages offer: exploratory data analysis tools to display data in various ways; most multivariate statistical methods including linear and logistic regression; various classification and modelling tools, such as decision trees, rule induction methods, neural networks, memory or case based reasoning; and, sequence discovery (time series analysis). Many of these methods can be usefully applied to spatial data. For example, data reduction tools, such as multivariate classification, can be useful as a means of summarising the essential features of large spatial data sets. Modelling tools such as neural networks and decision tees can be readily applied to some geographic problems. Although it can be argued that these methods ignore some of the special features of geographical information outlined in Table 1 - they can still be usefully applied to spatial data.
Conventional DMS have no mechanism for handling location or spatial aggregation or for coping with spatial entities or spatial concepts such as a region or circle or distance or map topology. What are therefor needed are new types of data mining tools that can handle the special nature of spatial information.
Most analytical functionality available in DMS are simple extensions and generalisations of techniques used for decades. The core functionality in most DMS is based on artificial intelligence and other techniques that can be set to search for patterns and trends in large volumes of data. Neural networks (NN) are universal approximators that mimic the workings of animal nervous systems that learn to classify and generalise patterns in their local environments. They have powerful pattern recognition and generalisation capabilities derived from their ability to learn to represent complex multi-variate data patterns. There are three main types of neural network commonly used in research; the multilayer perceptron, the learning vector quantisation network, and the self organising map. Multilayer perceptrons are perhaps the most common, but the majority of DMS also incorporate some form of self organising map. Genetic algorithms (GA), also called evolutionary algorithms, constitut a set of powerful global search heuristics that are also biologically inspired. They mimic the basic idea of the evolution of life forms as they adapt to their local environments over many generations. Both NN and GA provide a means of approaching highly complex search problems by simulating in the computer the biological process of natural selection and we should aim to incorporate them in SPIN.
Clementine Data Mining System (Clementine) was originally developed by Integral Solutions Ltd who were recently bought by SPSS. SPSS offer various training courses, provide a support service for users of Clementine and undertake consultancy work. Clementine has a user friendly visual programming GUI for developing data mining applications, see Figure 1 below. There is a menu across the top which can be used to load and save features specific to the data mining application, set parameters, run modelling scripts and load a basic help system.
Clementine stores input data as a table and most operations are performed sequentially record by record. The skeleton of a data mining application is represented by a set of stream diagrams which are usually constructed in the GUI model building pallet but can also be defined using CLEM, the Clementine macro modelling and scripting language. Streams can be saved and loaded from file and can be executed more efficiently in batch mode using modelling scripts. All Clementine streams are comprised of a set of nodes which are linked together to form a sequence of operations on a table of data. A variety of nodes can be selected from the node pallet at the bottom of the GUI and connected into a stream, the following types of node are available:
A useful feature, particularly for streams that split into multiple channels, is that as the data table passes through a node it can be cached (stored at that node). Nodes, streams, caches, generated models and modelling scripts can be individually saved and loaded from file as can the state of any Clementine session which stores the stream, any generated models, current parameter values and the modelling script associated with the stream. Clementine has several types of parameter including slot parameters, supernode parameters, stream parameters and global parameters. Slot parameters are used to control some of the editable characteristics of nodes, whereas, supernode parameters, stream parameters and global parameters are more like declared variables used in conventional programming. A major limitation of the system is that data or parameter values cannot be passed back to a modelling script from the stream it is executing. This means that Clementine cannot automatically perform many potentially ver useful recursive operations. First impressions of Clementine are that the GUI is tidy, elegant and intuitively laid out, but its data manipulation functionality and visualisation capability are basic. For more information on Clementine refer to SPSS.
The IBM intelligent miner for data (IMiner) is described by IBM as: “... a suite of statistical, processing, and mining functions that you can use to analyse large databases. It also provides visualisation tools for viewing and interpreting mining results…” IMiner provides access to on-line analytical processing functions for cleaning and re-arranging data in a similar way to the data manipulation nodes provided in Clementine. These processing functions enable missing values to be defined, rows and columns of data to be filtered, data to be discretised into specific ranges, randomly sampled, aggregated, joined, and converted to lower or uppercase. IMiner enables you to specify cyclic fields where values have a cyclical relationship to each other such as days of the week. IMiner can also create taxonomy objects to define a hierarchy of relations between data categories. Each item of a data table can be categorised and corresponding relations between these item categories can be used to define associations and sequential patterns.
The standard statistical functions available include bivariate statistics, factor analysis, linear regression, principle component analysis, and univariate curve fitting. The following IMiner data mining functions are available:
These functions have been developed to exploit RS/6000 SP hardware (configured by IBM) and use parallel DB2 software (licensed by IBM). For more information on IMiner refer to IBM.
The functionality of Clementine and IMiner are almost identical. Clementine has a nicer interface and is better in terms of visualisation, but is slower.
GeoMiner claims to be a knowledge discovery system for spatial databases and GIS. It is software implemented on a MapInfo proprietary GIS platform with access to on-line analytical processing functionality. The available information regarding GeoMiner is hyped up containing exaggerated claims about its capabilities. The functionality of the system is described using simple examples of the sorts of analysis that can be done with it. The main spatial data mining operations are referred to as drill down and roll up. Drill down is the process of examining data and mining results at higher general levels of spatial resolution, and roll up is the reverse process, which involves aggregating and generalising data and mining results at coarser levels of spatial resolution.
The GeoMiner functions include a geo-characterizer, a geo-comparator, a geo-classifier, a geo-associator, and a geo-cluster analyzer. The geo-characterizer module is for finding ‘a set of characteristic rules at multiple levels of abstraction’. Applying this module involves selecting and retrieving a subset of data from the database using some form of query. Then some form of spatial aggregation is performed to group data into larger user defined regions. Selected aspatial data variables are then analysed for each region using what is called the ‘attribute-oriented induction technique’. The geo-comparator module is for defining a set of comparison rules, which contrast the general features of different classes of the relevant sets of data in a database. This compares one set of data, known as the target class, to other sets of data, known as contrasting classes. The geo-associator module is said to find a ‘set of strong spatial-related association rules’. A typical spatial association rule is in th form of IF A THEN B (s%, c%) where A and B are sets of spatial or nonspatial predicates, s is the support of the rule (the probability that A and B hold together among all the possible cases), and c is the confidence of the rule (the conditional probability that B is true under the condition of A). This functionality is already proposed for SPIN. Mining spatial association rules is similar to the process of analysing aspatial association rules. The geo-cluster analyzer module is for finding clusters of points with relevant aspatial attributes. It uses an algorithm called CLARANS to perform spatial clustering. Basically, the spatial clustering module involves drawing a user defined number of convex polygons to represent spatial clusters of point data. Analysis can then be done by generalising and comparing the point attributes of the polygons. Additional analysis might involve repeating this for different expected numbers of clusters and examining differences between the polygons. The geo-classifier moule uses a decision-tree induction method to classify a set of relevant data according to one of the aspatial attributes. The classification tree is then displayed and by clicking on any of the nodes of it a user can highlight corresponding regions on a linked map.
The GeoMiner system although limited in some ways does enhance ESDA capablities. For a more comprehensive review of the functionality in GeoMiner refer to SPIN!-report D3 from Work Package 3.
SAGE is a spatial statistical analysis environment for interactively studying area-based data. It is based on the ArcInfo GIS developed by ESRI and provides some spatial statistical functionality and a linked display environment. The GUI contains a single map view, a table and one or more graphical representations of the data including histograms, scatterplots, rankit plots, boxplots, lagged plots and matrix graphs. Rankit plots are effectively scatterplots of a variable against its Z-scores. Lagged plots are a series of boxplots for a variable where (for a selected area) the ith-boxplot corresponds to the values of that variable in the ith-order adjacent areas. SAGE contains tools for creating these displays, for calculating basic statistics, for fitting regression models and for regionalising data. Data can be selected and queried from the maps and tables and new variables can be defined using basic arithmetic expressions. SAGE creates various n by n weights matrices when area data is loaded, (where n is the number of areas in the data set). One contains adjacency information, so if ni is adjacent to nj then nij=1 else nij=0; another stores the distance between the area centroids; and the other stores the length of the shared boundary between adjacent areas. The matrices are used to create regionalisations and ith-order adjacency information. One of the most useful features in SAGE is the ability to create statistics on the ith-order adjacencies and it is worth considering the incorporation of such functionality in SPIN. For more information see Jingsheng et.al. (1997) available at the following URL:
LiveMap is a spatial statistical tool conceived for ESDA. It is Lisp programming code that adds some map drawing functions to Xlisp-Stat which is a statistical programming environment that enables new statistical techniques to be developed easily and analysis to be performed ad hoc. It does not provide a GUI to a toolkit of commonly used techniques, rather it provides a command line from which interactive graphics and a vast range of statistical techniques can be called with Lisp protocols. The linked display environment and the ease of producing maps and other graphical output aims to encourage users to adopt an exploratory and graphical approach. To really take advantage of LiveMap and extend and customise the system for a particular application a user must become very familiar with the Lisp language and what Xlisp-Stat has to offer in terms of available statistical functionality, (detail of this functionality is not covered here although some of it is outlined in the Section 5). The available XLisp-Stat functionality is extremely diverse and the command line environment makes them all equally accessible and applicable to spatial data once the syntacs of the command is understood. The most notable feature of LiveMap is its ability to link multiple maps and graphical displays. Extensive documentation including examples of the use of the software and a thorough help system are also very attractive features of LiveMap that enable it to be picked up and applied fairly easily by users with intermediate level statistical skills and a basic programming ability, see Brunsdon (1998). LiveMap is a good example of a state of the art exploratory spatial statistical environment providing a linked display environment, which has become a common feature of ESDA environments and is rightly a core component of SPIN. More detailed information about Livemap is available at the following URL:
CDV is a cartographic data visualiser that demonstrates the utility of linked display ESDA environments. It is based on a GUI that is visually impressive and highly responsive in its display of GIS data. It is a demonstration tool that promotes dynamic linked display environments as ESDA tools. It permits numerous linked views containing a number of plots including; box-plots, scatter-plots, polygon maps, graduated circle maps, population cartograms and parallel coordinates plots. Symbols in each view are dynamically linked to corresponding cases in all other views and new views can be created from subsets selected from any view. There is a calculation function that provides an equation parser so as to derive variables using basic mathematical functions. CDV as a demonstration provides the look and feel that SPIN is aiming for. The cartogram display functionality is a particularly novel feature that SPIN should consider including. For more information on CDV refer to Dykes (1997) and the following RL:
GeoTools is an open source Java based mapping toolkit that is designed for viewing maps interactively on web browsers without the need for dedicated server side support. Geotools has provided a stable framework for producing the types of facilities available in CDV in a portable scalable and distributed manner. The functionality of GeoTools and the fact it is written in Java make it a useful reference for what we are trying to achieve. The latest version of GeoTools and further information is always available from the following URL:
GeoVista studio is another open source Java application. It visualises parallel coordinates plots and Self Organising Maps and has a bean architecture which offers a coordinated way of designing linked displays and GUIs. There is a plan to make GeoTools into a bean so it can be directly plugged into GeoVista studio to add mapping functionality. This development is new and little information is available but it will be posted either on the GeoTools web site or at the following URL:
Fuzzy logic and fuzzy set theory resemble human reasoning in that it deals with approximate information and uncertainty to generate decisions. Fuzzy logic is a superset of Boolean logic where everything is either true or false. In fuzzy logic things are both true to a degree and false to another degree. In the same way fuzzy set theory is a superset of conventional set theory, the difference is that fuzzy set theory does not conform to the law of the excluded middle. In other words, fuzzy set theory allows each element of a set to also belong to its compliment to a relative degree.
Most spatial data is fuzzy, imprecise and vague. Any methodology or theory implementing precise definitions such as classical set theory, arithmetic, and programming, may be fuzzified by generalising the concept of a crisp set to a fuzzy set. Fuzzy logic, fuzzy set theory and fuzzy inference are very useful in attempts to solve data modelling and control problems involving imprecision and noise in the data and information used.
Fuzzy logic allows for constructing linguistic variables that enable computation to be done with linguistic IF THEN ELSE statements, such as, if road density is high and population density is low then non-road land use nearby is more likely to be business or commercial than residential else there are a fairly large number of empty or partially populated dwellings. Terms such a 'large' 'medium' and 'small' are defined using membership functions where values can have varying degrees of belonging or membership to each group. Fuzzy set theory encompasses fuzzy logic, fuzzy arithmetic, fuzzy mathematical programming, fuzzy topology, fuzzy graph theory, and fuzzy data analysis, though the term fuzzy logic is often used to describe them all.
Because fuzzy logic can handle approximate information in a systematic way, it is ideal for controlling nonlinear systems and for modelling complex systems. A typical fuzzy system consists of a rule base, membership functions, and an inference procedure. Developing a fuzzy logic component in SPIN could be extremely useful. Details on the vast range of fuzzy software available can be found on the Web at the following URLs:
Statistics is the theory of accumulating information that attempts to answer three basic questions:
The third question concerns part of the process known as statistical inference (SI). Regression, correlation, analysis of variance, discriminant analysis, standard error, significance levels, hypothesis testing, probablility distributions, expectations, confidence intervals, random sampling and the bootstrap are all part of SI. Probability theory provides the mathematical framework for SI. SI concerns learning from experience based on samples and often involves estimating some aspect of a probability distribution such as a maximum likelihood.
SpaceStat provides standard statistical functionality for calculating the mean, range, variance, standard deviation, skewness and kurtosis of a variable. It provides functionality for measuring how close the distribution of variable values is to a normal distribution and provides a way of calculating Pearson's correlation coefficient and principle components. Seemingly more relevant for spatial data is its functionality for calculating multivariate spatial correlation and spatial principle components. Multivariate spatial correlation first involves standardising all values of every variable by taking away the mean value of the variable and dividing by its standard deviation. These standardised values are then stored in a matrix S (where the columns are the variables and the observations are the rows) which is transformed by a stochastic (where all elements sum to 1) spatial weights matrix W by using the following matrix equation:
The values of the elements in the matrix T and its eigenvalues and eigenvectors can be analysed to gain some idea of spatial associations between variable values. This interpretation is far from easy, but a basic comparison with aspatial principal components can provide a useful means of identifying spatial effects although it is difficult to infer much about them. The simplest measures of spatial autocorrelation SpaceStat provides are join count statistics. These simply count the number of joins between areas of one class and areas of another class. SpaceStat is restricted to using binary classes, however the method is useful and can be extended to handle fuzziness by using continuous values, multiple sets, and ith-order contiguity data and as such should be considered for SPIN. SpaceStat can also be used to calculate Moran's I and Geary's C statistics which provide some indication of the nature (positive or negative) of spatial autocorrelation. The magnitude of these statistics is har to interpret. Further inference can be attempted by assuming normality or randomness in the distribution of permutations (created by omitting observations) of these statistics then testing the assumption by detecting significant differences of the observed values from theoretically expected ones. Some way of omitting observations based on spatial criteria may be useful in detecting spatial effects in such inference, however, this is all very statistically intricate and the non-normality or non-random assumptions are inherently flawed. Like most statistics or mathematical formulae applied to spatial data, both Moran's I and Geary's C statistics are best applied to different and possibly overlapping sets of localities and then analysed in terms of the differences or variations between the spatial subsets. Unless spatial subsets or localities are used the statistics are global providing little useful spatial information that is not highly sensitive to the problem of modifiable areal units. When applied in localities, such statistics are termed local indicators of spatial association (LISA). SpaceStat does not provide functionality that makes this easy, but it does have the building blocks required for it. Providing the functionality for applying a range of statistics for different types of strictly tessellating and overlapping spatial subsets is something worth considering for SPIN. For more information on Spacestat refer to Anselin (1992), Anselin (1995) and the following URL:
S+ is a widely used flexible package and programming language for statistical research. Two official S+ extensions that have appeared in recent years are S+SpatialStats and S+ArcView. These are specifically geared for use by geographers and the like who need to analyse spatial data. S+SpatialStats offers a variety of graphical display functionality of interest for ESDA. The graphical functionality includes ways to display scatterplots and 3D point clouds, hexagon plots (where each hexagon is shaded in terms of the density of points), boxplots, contour plots, variogram plots and correlograms. In terms of analysis it offers: Variograms, which are effectively a means of displaying the variance between points at various distances or in various directions; Kriging, which is a linear interpolation method for predicting cell values for a regular grid based on a set of observations; methods for assessing spatial autocorrelation and spatial randomness; methods for building spatial regression models; methods for stimating the intensity function of spatial point patterns; and methods for investigating spatial interaction and spatial dependence. S+ArcView integrates the statistics, graphical and GIS functionality of S+SpatialStats and ArcView. Further details of S+ can be found on the Web at:
Details of S+SpatialStats are on the Web at:
Geographically Weighted Regression (GWR) is based on multiple regression. Multiple regression is a commonly used statistical technique that involves estimating the relationship between one dependent variable and a set of predictor variables that are assumed to be independent of one another and identically distributed about the mean. Thus an assumption that is inherent when applying multiple regression to spatial data is that there is no spatial variation in these attribute distributions within spatial subsets of these data. These assumptions are fundamentally flawed when applied to geographical data, however there are many examples of the use of this statistical technique in geographical enquiry. The difference between standard multiple regression and GWR is that the later enables parameter estimates which control the importance of specific dependent variables to vary locally according to a defined kernel that weights the importance of observations from a given point by some kind of distance function. he resulting parameter estimates from GWR vary locally and can be mapped in order to examine local variations in the relationship between the dependent variable and its predictor variables. This examination of model deviations or errors is sometimes referred to as econometrics and this specific type is sometimes referred to as the analysis of heteroskedasticity. Basically it involves testing the inequality of variance in the regression errors, which can inform as to where the assumption that the variables used are independent and identically distributed normal variables is most wrong. In other words it can be used examine spatial autocorrelation effects. Monte Carlo simulation provides a means of testing whether the variations in the parameter estimates are due to chance, which is a way of informing about how non-random the fitted model parameters are. Applying this econometric analysis in conjunction with GWR can provide both evidence that the underlying assumptions of the model fitted are wrong and an intricate way of examining spatial autocorrelation effects in spatial data. Mapping parameter estimates or their standard deviations is perhaps the simplest and most easily followed method of providing information about the nature of the autocorrelation. All things considered, the GWR technique can be usefully applied to spatial data and it is well worth considering incorporating it or something similar in SPIN. An important thing to keep in mind though is that the complexities of the econometrics and the intricacies of analysing the statistical models probably make this method inappropriate for use by anyone who does not have a degree in statistics or a good appreciation of the correct use of such methods for spatial analysis. For more information on GWR refer to the following URL:
The ESDA tools outlined in this report along with the stated expert opinion aimed to provide a useful and representative account of the current state of the art providing a guide to the functionality that we should be aiming to incorporate and develop in SPIN. Not everything was covered in this review explicitly. There are some useful techniques that have not being discussed, such as, spatial interaction models. There are also some useful integrated environments that have not even received a mention.
ESDA involves the concept of spatial patterns or locations of values and their relative magnitude over time. Various aspects of these patterns are measurable and of interest. This includes things like; the extent of the pattern in terms of non-randomness, the extent of the pattern in terms of clustering, the evidence of spatial correlation, and the implied spatial association between variables over various (possibly overlapping) time ranges. Sometimes there is a temporal lag indicative of cause and effect relationships in spatial data values that can be measured.
SPIN should not be aiming to incorporate absolutely everything available that can be useful in an ESDA context. One reason is that a bulky system is hard to maintain, however, the main reason is that this would be very hard to make user friendly. Even experts would find it hard to fathom what tools are best to apply to their data and in what order if the system contained a plethora of tools that essentially did the same thing. In other words: Why have a hundred different tests for clustering when the best few will do? This should not distract from the need to have available example analyses and an extensive tutorial and help system.
SPIN should be capable of: producing summary statistics (mean, median, mode, standard deviation, maximum, minimum, range, etc.); producing exploratory data analysis displays (boxplots, histograms, graphs, etc.); and dynamically linking maps and graphical displays. It should contain standard data mining functionality (NN, GA, rule induction, decision trees) and some multivariate statistical methods (linear and logistic regression). Perhaps most importantly SPIN should contain new types of spatio-temporal data mining tools that can handle the special nature of geographical information.
Multivariate data with both spatial and temporal reference is complex and the challenge is in designing analysis tools able to cope with this complexity. One approach is to create intelligent multidimensional entities that search for interesting patterns and relationships and visualise them as animated colour maps. The search task can be viewed as identifying localised chunks of multidimensional space-time data that contain unusual amounts of unexpected patterns.
In summary, the most important tools to have available are; classification tools that enable data to be evaluated in terms of their membership to an aggregate set, prediction tools which model the data in some way to relate one set of variables to another, and visualisation tools that display the data and can be used interactively as part of the process. There also needs to be a high level of automation, and an extensive help system.