Dr. Stuart Wright, Senior Scientist EBSD, EDAX
For better or worse, I’ve been long involved in trying to set forth some guidelines for the measurement of grain size using EBSD. This involvement has included serving on some of the standard committees, advising customers and reading, reviewing and publishing papers . However, I’ve always felt unsettled about the outcome of those efforts. Part of that discomfort comes from the fear that an inexperienced EBSD user could be misled to an incorrect conclusion based on using a canned procedure. However, some recent experiments have given me more confidence in obtaining reliable grain size statistics using EBSD.
There are a couple of challenges associated with the measurement of grain size using EBSD. First, there are the usual factors associated with collecting good EBSD data: determining a polishing procedure that will produce good patterns, finding good SEM and EBSD camera settings, ensuring the sample is in the expected geometry . However, I will assume these factors are well under control so that good quality EBSD data is obtained, that is data with a high indexing rate so that very little clean-up is required. Now, the challenge is making sure the choices for the parameters associated with grain size are appropriate. The first of these is the grain tolerance angle. I will focus on recrystallized materials at this point so the choice of grain tolerance angle is not as critical as for a deformed material. In the case of deformed materials, I’ve found you need to experiment with different values to get a feel for the best choice. However, for recrystallized material using the default value of 5° generally works well. The next critical parameter is the choice of the minimum grain size in terms of the number of grid points. Once this value is selected the analysis software will exclude grains with less than this value from the grain size distribution.
There are several approaches to selecting a good value for the minimum grain size – I will call it Nmin for brevity. It is entirely possible that a single point in an OIM scan could be a grain. Imagine the grain structure in three dimensions; it is easy to imagine a single scan point being associated with the top or bottom tip of a grain just intersecting the sampling plane. Thus a single point in an OIM scan could represent a grain especially if that point has a high confidence index. Of course, we have also seen many cases where an individual point is located at a grain boundary with a low confidence index. This arises because the resulting pattern is a combination of two competing patterns from the two grains on either side of the grain boundary (or three at a triple point) and thus the Hough transform will find bands from all the competing patterns resulting in an incorrect indexing . Another consideration that has gone in to the choice of Nmin is is how well a shape can be reconstructed from a given number of grid points. For example, consider a circle. The following figure shows how well a circle can be approximated by a given number of grid points. To approximate the area of circle with less than 5% error requires about 30 points. However, if very many grains are measured the errors in approximating a given grain will be averaged out so I personally don’t think this argument should carry much weight in the choice of Nmin. In fact, I hope by the end of these ramblings that I will have convinced you that if using the right approach, the choice of Nmin is not as important as one might presume.
I’ve done some experiments with a very nice set of data provided by my colleague Matt Nowell. These measurements were performed as part of a Round-Robin test conducted for the development of ISO Guidelines for the measurement of grain size using EBSD . The figure below shows five data sets measured on one sample merged together. I have excluded the grains touching the edges from my analysis. On average, each field contains 750 grains. The step size is 2 microns and each scan contains almost 250,000 data points. The total number of grains analyzed was 3748 grains. Prior to any analysis a grain dilation cleanup was applied to each individual dataset using a 5° tolerance angle, two pixels for the minimum grain size and the grains were required to span multiple grid rows. The percentage of points changed in the clean-up process was only 0.7% confirming the high fidelity of the data.
The grain size distributions for the individual data sets and the merged data set are shown in Figure 3 using both a linear axis and a log axis for the grain diameter (the horizontal axis). The vertical axis is the usual area fraction.
|Figure 3: Log and linear grain size distributions of the individual data sets (colors) and the merged data set (black)|
The next step in such analyses is to calculate the average grain size (I will use the diameter). This is simply done by adding up the diameters of all the grains and then dividing by the total number of grains. Figure 4 shows the average grain diameter overlaid on the distribution for the merged data.
As usual, the location of the average diameter does not correlate well with the center of the distribution. I always find this disconcerting. A common response is to simply pass this off to more data being needed especially since the distribution curve is still a bit jagged even with 3748 grains. You can experiment with the number of bins used to create the distribution as well as the grain definition parameters but the average grain size is always a bit left of where you think it should be. The reason for this mismatch is that the distribution is given in area fraction but the average is calculated as a number average. If you plot the distribution as a number fraction then the number average appears to fit the data better. However, what originally approximated a Gaussian curve in the area fraction plot now becomes skewed to the left in the number fraction plot.
Another solution to overcoming this mismatch between the average and the peak in the distribution is, instead of using a number averaging approach, using an area weighted averaging approach. The area weighted average is a relatively simple calculation given by:
Where n is the number of grains, Ai is the area of grain i and di is the diameter.
The area weighted averaging leads to an average value that matches that approximated “by-eye”.
In fact, I’ve found that the weighted area average provides a very good seed to any automatic fitting of the distribution data. The advantage of the area average relative to a curve-fit determination, is that the area average is calculated from the raw grain size data independent of the binning used to create the distribution plot.
So why do I bring this whole area weighted averaging approach up when the accepted approach is the number averaging approach? What does it have to do with selecting an appropriate Nmin ? The reason lies in the following plot of the area and number averages versus Nmin. In this plot the area average appears less sensitive to the choice of Nmin than the number average. This observation has held up on across many different samples on which I’ve performed grains size analysis.
The sensitivity of the area average to Nmin seems to increase in the plot at about a value of 50. If we look at this same data plotted on a log scale the increase in sensitivity at higher Nmin become more apparent. These curves start at a value of Nmin = 5 due to the fact that this was the minimum grain size I selected for the grain dilation clean-up.
The argument among the grain size community is to select a much larger Nmin value than that typically used in the EBSD community (2 to 5 points). The ASTM standard states that the minimum grain that should be included should contain 100 points and that the average grain should contain at least 500 points . This standard evolved to this methodology primarily for historical consistency with optical measurements. However, these results show that the choice of Nmin is not so critical in the determination of the average value provided it is not too large. Another way to look at this is to plot the data differently. If we change the horizontal axis to Nmin divided by the number of points corresponding to the average grain size and change the vertical axis to be the average diameter divided by the average diameter at Nmin equal to five then we get the following plot.
From this plot it appears that using a small number for Nmin is fine. The key is not to pick too large of a number – it should not exceed 10% of the number of points in the averaged sized grain. However, this is with the proviso that area weighted averaging is used in the analysis. The contribution of the small grains to the average is very much less for area weighted averaging as opposed to number averaging. Thus using a small number, even just 1, will lead to nearly the same result as using a larger value such as the value of 5 I used in this example. Because we (the EBSD community) can very much more definitively calculate grains of any size relative to traditional optical measurements, I believe this approach is the correct one despite its inconsistency with the historical approach.
 Wright, S. I. (2010). “A Parametric Study of Electron Backscatter Diffraction based Grain Size Measurements.” Practical Metallography 47: 16-33.
 Nolze, G. (2007). “Image distortions in SEM and their influences on EBSD measurements.” Ultramicroscopy 107: 172-183.
 Wright, S. I., M. M. Nowell, R. de Kloe and L. Chan (2014). “Orientation Precision of Electron Backscatter Diffraction Measurements Near Grain Boundaries.” Microscopy and Microanalysis: (in press but available on First View).
 ISO 13067: Microbeam analysis – Electron backscatter diffraction – Measurement of average grain size.
 ASTM E2627-10: Standard practice for determining average grain size using electron backscatter diffraction (EBSD in) fully recrystallized polycrystalline materials.