Dr. Stuart Wright, Senior Scientist EBSD, EDAX
For better or worse, I’ve been long involved in trying to set forth some guidelines for the measurement of grain size using EBSD. This involvement has included serving on some of the standard committees, advising customers and reading, reviewing and publishing papers [1]. However, I’ve always felt unsettled about the outcome of those efforts. Part of that discomfort comes from the fear that an inexperienced EBSD user could be misled to an incorrect conclusion based on using a canned procedure. However, some recent experiments have given me more confidence in obtaining reliable grain size statistics using EBSD.
There are a couple of challenges associated with the measurement of grain size using EBSD. First, there are the usual factors associated with collecting good EBSD data: determining a polishing procedure that will produce good patterns, finding good SEM and EBSD camera settings, ensuring the sample is in the expected geometry [2]. However, I will assume these factors are well under control so that good quality EBSD data is obtained, that is data with a high indexing rate so that very little cleanup is required. Now, the challenge is making sure the choices for the parameters associated with grain size are appropriate. The first of these is the grain tolerance angle. I will focus on recrystallized materials at this point so the choice of grain tolerance angle is not as critical as for a deformed material. In the case of deformed materials, I’ve found you need to experiment with different values to get a feel for the best choice. However, for recrystallized material using the default value of 5° generally works well. The next critical parameter is the choice of the minimum grain size in terms of the number of grid points. Once this value is selected the analysis software will exclude grains with less than this value from the grain size distribution.
There are several approaches to selecting a good value for the minimum grain size – I will call it N_{min} for brevity. It is entirely possible that a single point in an OIM scan could be a grain. Imagine the grain structure in three dimensions; it is easy to imagine a single scan point being associated with the top or bottom tip of a grain just intersecting the sampling plane. Thus a single point in an OIM scan could represent a grain especially if that point has a high confidence index. Of course, we have also seen many cases where an individual point is located at a grain boundary with a low confidence index. This arises because the resulting pattern is a combination of two competing patterns from the two grains on either side of the grain boundary (or three at a triple point) and thus the Hough transform will find bands from all the competing patterns resulting in an incorrect indexing [3]. Another consideration that has gone in to the choice of N_{min} is is how well a shape can be reconstructed from a given number of grid points. For example, consider a circle. The following figure shows how well a circle can be approximated by a given number of grid points. To approximate the area of circle with less than 5% error requires about 30 points. However, if very many grains are measured the errors in approximating a given grain will be averaged out so I personally don’t think this argument should carry much weight in the choice of N_{min}. In fact, I hope by the end of these ramblings that I will have convinced you that if using the right approach, the choice of N_{min} is not as important as one might presume.
Figure 1: Approximation of the area of a circle using a square grid and a hexagonal grid.
I’ve done some experiments with a very nice set of data provided by my colleague Matt Nowell. These measurements were performed as part of a RoundRobin test conducted for the development of ISO Guidelines for the measurement of grain size using EBSD [4]. The figure below shows five data sets measured on one sample merged together. I have excluded the grains touching the edges from my analysis. On average, each field contains 750 grains. The step size is 2 microns and each scan contains almost 250,000 data points. The total number of grains analyzed was 3748 grains. Prior to any analysis a grain dilation cleanup was applied to each individual dataset using a 5° tolerance angle, two pixels for the minimum grain size and the grains were required to span multiple grid rows. The percentage of points changed in the cleanup process was only 0.7% confirming the high fidelity of the data.
Figure 2: Grain map of the merged data.
The grain size distributions for the individual data sets and the merged data set are shown in Figure 3 using both a linear axis and a log axis for the grain diameter (the horizontal axis). The vertical axis is the usual area fraction.


Figure 3: Log and linear grain size distributions of the individual data sets (colors) and the merged data set (black) 
The next step in such analyses is to calculate the average grain size (I will use the diameter). This is simply done by adding up the diameters of all the grains and then dividing by the total number of grains. Figure 4 shows the average grain diameter overlaid on the distribution for the merged data.
Figure 4: Area fraction grain diameter distribution for the merged data overlaid with the average diameter.
As usual, the location of the average diameter does not correlate well with the center of the distribution. I always find this disconcerting. A common response is to simply pass this off to more data being needed especially since the distribution curve is still a bit jagged even with 3748 grains. You can experiment with the number of bins used to create the distribution as well as the grain definition parameters but the average grain size is always a bit left of where you think it should be. The reason for this mismatch is that the distribution is given in area fraction but the average is calculated as a number average. If you plot the distribution as a number fraction then the number average appears to fit the data better. However, what originally approximated a Gaussian curve in the area fraction plot now becomes skewed to the left in the number fraction plot.
Figure 5: Number fraction grain diameter distribution for the merged data overlaid with the average diameter.
Another solution to overcoming this mismatch between the average and the peak in the distribution is, instead of using a number averaging approach, using an area weighted averaging approach. The area weighted average is a relatively simple calculation given by:
Where n is the number of grains, A_{i} is the area of grain i and d_{i} is the diameter.
The area weighted averaging leads to an average value that matches that approximated “byeye”.
Figure 6: Area fraction grain diameter distribution for the merged data overlaid with the average diameter calculated using number averaging and area weighted averaging.
In fact, I’ve found that the weighted area average provides a very good seed to any automatic fitting of the distribution data. The advantage of the area average relative to a curvefit determination, is that the area average is calculated from the raw grain size data independent of the binning used to create the distribution plot.
Figure 7: Gaussian distributions for the merged data grain diameter distributions.
So why do I bring this whole area weighted averaging approach up when the accepted approach is the number averaging approach? What does it have to do with selecting an appropriate N_{min} ? The reason lies in the following plot of the area and number averages versus N_{min}. In this plot the area average appears less sensitive to the choice of N_{min} than the number average. This observation has held up on across many different samples on which I’ve performed grains size analysis.
Figure 8: Plot of average grain diameter as a function of the choice of N_{min}. .
The sensitivity of the area average to N_{min} seems to increase in the plot at about a value of 50. If we look at this same data plotted on a log scale the increase in sensitivity at higher N_{min} become more apparent. These curves start at a value of N_{min} = 5 due to the fact that this was the minimum grain size I selected for the grain dilation cleanup.
Figure 9: Log plot of average grain diameter as a function of the choice of N_{min}
The argument among the grain size community is to select a much larger N_{min} value than that typically used in the EBSD community (2 to 5 points). The ASTM standard states that the minimum grain that should be included should contain 100 points and that the average grain should contain at least 500 points [4]. This standard evolved to this methodology primarily for historical consistency with optical measurements. However, these results show that the choice of N_{min} is not so critical in the determination of the average value provided it is not too large. Another way to look at this is to plot the data differently. If we change the horizontal axis to N_{min} divided by the number of points corresponding to the average grain size and change the vertical axis to be the average diameter divided by the average diameter at N_{min} equal to five then we get the following plot.
Figure 10: The normalized average grain size as a function of the ration of the number of points in the minimum sized grain to the number of points in the average sized grain.
From this plot it appears that using a small number for N_{min} is fine. The key is not to pick too large of a number – it should not exceed 10% of the number of points in the averaged sized grain. However, this is with the proviso that area weighted averaging is used in the analysis. The contribution of the small grains to the average is very much less for area weighted averaging as opposed to number averaging. Thus using a small number, even just 1, will lead to nearly the same result as using a larger value such as the value of 5 I used in this example. Because we (the EBSD community) can very much more definitively calculate grains of any size relative to traditional optical measurements, I believe this approach is the correct one despite its inconsistency with the historical approach.
References
[1] Wright, S. I. (2010). “A Parametric Study of Electron Backscatter Diffraction based Grain Size Measurements.” Practical Metallography 47: 1633.
[2] Nolze, G. (2007). “Image distortions in SEM and their influences on EBSD measurements.” Ultramicroscopy 107: 172183.
[3] Wright, S. I., M. M. Nowell, R. de Kloe and L. Chan (2014). “Orientation Precision of Electron Backscatter Diffraction Measurements Near Grain Boundaries.” Microscopy and Microanalysis: (in press but available on First View).
[4] ISO 13067: Microbeam analysis – Electron backscatter diffraction – Measurement of average grain size.
[5] ASTM E262710: Standard practice for determining average grain size using electron backscatter diffraction (EBSD in) fully recrystallized polycrystalline materials.
Like this:
Like Loading...