POST
|
Hi Paula, It's great to hear that you are interested in learning more about some of the spatial statistics tools in ArcGIS! We have a ton of resources that I think will be very helpful to you as you start learning about the tools and how they can help you analyze your data. To start with, it sounds like Hot Spot Analysis should work really well for the questions that you are trying to answer, as well as some other spatial statistics tools. We have all of our resources listed at http://bit.ly/spatialstats , including tutorials, short videos, hour-long free training seminars, and a lot more. The Hot Spot Analysis Tutorial and the Spatial Pattern Analysis Tutorial will be great places for you to start. They come with data, and walk you step-by-step through the analysis process. As far as Hierarchical Nearest Neighbor clustering, ArcGIS doesn�??t have that tool (we found results are very dependent on the first cluster found). I think Hot Spot Analysis will provide you with the results that you are looking for. However, if you feel that Hierarchical Clustering is an important solution to your problem, you may want to check out CrimeStat, which does include it in their package. Hope this helps! Lauren Rosenshein Geoprocessing Product Engineer
... View more
12-15-2010
08:52 AM
|
0
|
0
|
13
|
POST
|
Hi Paula, Even though your point data does not have an attribute associated with it that you would like to analyze, there are some techniques that you can use that will allow you to use the spatial statistics tools to run a hot spot analysis. We have a great tutorial that you can download that will walk you through a hot spot analysis of 911 calls. It should be very helpful because it walks through the workflow of aggregating the point data for use in tools that require a value field. You might also find this video useful as it describes the importance of thinking through parameters like the search radius. There are also a ton of additional spatial statistics resources available at http://bit.ly/spatialstats . Hope this helps. Lauren Rosenshein Geoprocessing Product Engineer
... View more
12-15-2010
08:39 AM
|
0
|
0
|
1
|
POST
|
Hi Steve, I am very sorry to hear that you are having trouble! We do everything that we can to ensure that users never get 999999 errors, so that sounds like a bug, and I am hoping that we will be able to figure out what is going wrong and fix it as soon as possible. There are a few things that I'll need to know in order to reproduce the issue, and ultimately fix the problem. What version of ArcGIS are you working on? Is it possible for you to share your data so that we can reproduce your exact workflow, and what are the precise steps that you are taking in your analysis (i.e. all of the parameter choices, inputs, etc.)? If you can share your data, you can send it to me directly at lrosenshein@esri.com . Again, I am very sorry for the trouble this has caused you! Lauren Rosenshein Geoprocessing Product Engineer
... View more
11-30-2010
01:46 PM
|
0
|
0
|
60
|
POST
|
That is a great question, Bill! So great, in fact, that I had to go ask a colleague down the hall. 🙂 The response that I got is as follows: A given point can move up to 1.4 X the cluster tolerance, but it�??s an iterative process, so the movement of a point, could move it within the cluster tolerance of other points, and shift it further. I think in this case, yes, all the points will shift, and will shift to the center of the grouping (2.5, 0). My colleague also pointed me to the help topic, Topology in ArcGIS , which I think you will find provides a lot of additional information on cluster tolerances. Hope this helps! Lauren
... View more
11-23-2010
09:13 AM
|
0
|
0
|
6
|
POST
|
Hi Bin, Another option for this type of grouping is to start with the Integrate tool (Data Management), which allows you to snap features together that are within a specified distance of each other. The integrate tool will change the coordinates of all input features so that those features within a specified distance of each other will be coincident. Please note that the Integrate tool will change your input feature class, so it is always a good idea to run Copy Features beforehand to ensure that you do not lose your original data. Once you have run Integrate, you can use the Collect Events tool from the Spatial Statistics toolbox to create a count of the number of coincident points at each location. This workflow is detailed in the Hot Spot Analysis Tutorial as a way to aggregate incident data for a hot spot analysis. The tutorial provides step-by-step instructions on creating a model to Integrate and then Collect Events using point data. Hope this helps! Lauren Rosenshein Geoprocessing Product Engineer Check out the latest Spatial Statistics resources at http://bit.ly/spatialstats !
... View more
11-22-2010
07:02 AM
|
0
|
0
|
6
|
POST
|
Hi Tim, I cannot be sure about the problem with your data, but I do know that Joseph Kerski's Population Drift tutorial recently helped us find a bug. The problem with that data is that all of the counties for one of the states (Alaska I believe) has -999 for population for at least one year. When all of the features in one of the cases have all negative values (in this case the State Name was the case field), the tool fails to execute. The fix for this bug should be in 10 SP2. To get around the issue, you can remove the states where all counties have -999 values, or change the values to 0. We are very sorry for any inconvenience! That being said, it doesn't sound like the "all negatives in a case" issue is what was giving you trouble with your original data. If you are still having trouble with this, and you don't mind sharing your data, I'd be happy to try to troubleshoot the issue. You can send me the data at lrosenshein@esri.com . Lauren Rosenshein Geoprocessing Product Engineer
... View more
10-28-2010
08:27 AM
|
0
|
0
|
0
|
POST
|
Hi Amanda, There is a tool in the Data Management Toolbox, under the Graphs toolset, called Save Graph . You can use the Save Graph tool to save the graph to a number of different image formats, and you can then insert that image into your layout. Hope this helps! Lauren Rosenshein Geoprocessing Product Engineer
... View more
10-21-2010
11:31 AM
|
0
|
0
|
0
|
POST
|
Hi Arnau, My first guess as far as your NULL output is the fact that when the result of a computation is infinity or undefined, the result for nonshapefiles will be Null; for shapefiles the result will be -DBL_MAX = -1.7976931348623158e+308. If your output is a featureclass, and the result is infinity, then the NULL would make sense. The real question then becomes "Why are you getting results of Infinity". One thing to keep in mind is that the data you are working with (i.e. continuous surfaces created from rasters, where presumably you have many neighboring points with identical or extremely similar values), you are very likely dealing with issues of local multicollinearity . The type of data that you are working with is one clue that you have multicollinearity, and another clue is the Condition Numbers that are part of the GWR output. In general, a condition number over 30 indicates issues with multicollinearity...and from looking at your images it would appear that many of your condition numbers are in the hundreds, some closer to the thousands. This type of local multicollinearity indicates that your results are unstable. To learn more about these issues, see the conceptual documentation on GWR, found here: How GWR Works . One potential way to deal with this type of multicollinearity is to increase the cell size of your raster, which might have the effect of increasing the variation in both your dependent and independent variables. The help states some additional methods: Try creating a thematic map for each explanatory variable. If the map reveals spatial clustering of identical values, consider removing those variables from the model or combining those variables with other explanatory variables to increase value variation. If, for example, you are modeling home values and have variables for both bedrooms and bathrooms, you may want to combine these to increase value variation or represent them as bathroom/bedroom square footage. Avoid using spatial regime dummy/binary variables, spatially clustering categorical/nominal variables, or variables with very few possible values when constructing GWR models. Hopefully this will help point you in the right direction. The online documentation for GWR is a great resource, and may help you solve some of your problems! Lauren Rosenshein Geoprocessing Product Engineer Spatial Statistics
... View more
09-08-2010
03:26 PM
|
0
|
0
|
42
|
POST
|
Hi Uday, This is a great question. The simple answer is that the Kernel density function and the Getis Ord Gi* statistic are completely different analyzes. The Kernel density function does a simple density calculation based on the user-specified radius and raster cell size. For more information on the mathematics involved in Kernel Density calculations, you can check out the conceptual documentation here . The Getis-Ord Gi* statistic, also known as Hot Spot Analysis, works by looking at the value of each feature in your dataset in the context of its neighbors' values. The Hot Spot Analysis tool reports a z-score and a p-value, both of which represent statistical significance of clustering, which can help make the resulting maps less subjective (i.e. red represents statistically significant hot spots and blue represents statistically significant cold spots, as opposed to a density surface where the color choices are largely subjective cartographic decisions). To learn more about how Hot Spot Analysis (Getis-Ord Gi*) works, you can take a look at the conceptual documentation here . There is also a video that looks at the differences between Hot Spot Analysis and Density surfaces, which you can see here: Performing Proper Density Analysis . And for more information about the Spatial Statistics tools, you can check out our blog post which will point you to tons of videos, tutorials, seminars, and more. The post can be found at http://bit.ly/spatialstats , and we update it whenever new materials are available! Hope this helps! Lauren Rosenshein Geoprocessing Product Engineer Spatial Statistics
... View more
09-08-2010
01:48 PM
|
1
|
0
|
43
|
POST
|
And don�??t forget�?� The most important thing to remember when you are going through these steps is that the goal of your analysis is to understand your data and use that understanding to contribute to solving problems and answering questions. I am not going to lie�?�you may try a number of models, with and without transformed variables, explore several small study areas, analyze your coefficient surfaces, and still not find a good OLS model (again, I speak from experience!). But, and this is a big but, you will still be contributing to the body of knowledge on the phenomenon that you are modeling! If the model that you hypothesized would be a great predictor is not significant at all, that is incredibly important information. If one of the variables that you thought would be important has a positive relationship in some areas and a negative relationship in other areas, that is important information. It is true that you may ultimately want to move on to additional tools that will do other types of regression analysis (perhaps one that does not involve linear models, perhaps a spatial autoregressive model, the possibilities are endless). The work that you do here, however, trying to find a good model using OLS, and using GWR to understand your data and improve your model, is valuable. And speaking from experience�?�just writing up all of the work that you do and the valuable information that you have uncovered along the way will be more than enough to fill a thesis with meaningful content! All the best, Lauren Rosenshein Geoprocessing Product Engineer ESRI | Redlands, CA
... View more
05-14-2010
02:48 PM
|
0
|
0
|
101
|
Online Status |
Offline
|
Date Last Visited |
11-11-2020
02:23 AM
|