Distance sampling


Distance sampling is a widely used group of closely related methods for estimating the density and/or abundance of populations. The main methods are based on line transects or point transects. In this method of sampling, the data collected are the distances of the objects being surveyed from these randomly placed lines or points, and the objective is to estimate the average density of the objects within a region.

Basic line transect methodology

A common approach to distance sampling is the use of line transects. The observer traverses a straight line. Whenever they observe an object of interest, they record the distance from their current position to the object, as well as the angle of the detection to the transect line. The distance of the object to the transect can then be calculated as x = r * sin. These distances x are the detection distances that will be analyzed in further modeling.
Objects are detected out to a pre-determined maximum detection distance w. Not all objects within w will be detected, but a fundamental assumption is that all objects at zero distance are detected. Overall detection probability is thus expected to be 1 on the line, and to decrease with increasing distance from the line. The distribution of the observed distances is used to estimate a "detection function" that describes the probability of detecting an object at a given distance. Given that various basic assumptions hold, this function allows the estimation of the average probability P of detecting an object given that is within width w of the line. Object density can then be estimated as, where n is the number of objects detected and a is the size of the region covered.
In summary, modeling how detectability drops off with increasing distance from the transect allows estimating how many objects there are in total in the area of interest, based on the number that were actually observed.
The survey methodology for point transects is slightly different. In this case, the observer remains stationary, the survey ends not when the end of the transect is reached but after a pre-determined time, and measured distances to the observer are used directly without conversion to transverse distances. Detection function types and fitting are also different to some degree.

Detection function

The drop-off of detectability with increasing distance from the transect line is modeled using a detection function g. This function is fitted to the distribution of detection ranges represented as a probability density function. The PDF is a histogram of collected distances and describes the probability that an object at distance y will be detected by an observer on the center line, with detections on the line itself assumed to be certain.
By preference, g is a robust function that can represent data with unclear or weakly defined distribution characteristics, as is frequently the case in field data. Several types of functions are commonly used, depending on the general shape of the detection data's PDF:
Detection functionForm
Uniform
Half-normal
Hazard-rate
Negative exponential

Here w is the overall detection truncation distance and a, b and σ are function-specific parameters. The half-normal and hazard-rate functions are generally considered to be most likely to represent field data that was collected under well-controlled conditions. Detection probability appearing to increase or remain constant with distance from the transect line may indicate problems with data collection or survey design.

Covariates

Series expansions

A frequently used method to improve the fit of the detection function to the data is the use of series expansions. Here, the function is split into a "key" part and a "series" part; i.e., g = key. The series generally takes the form of a polynomial and is intended to add flexibility to the form of the key function, allowing it to fit more closely to the data PDF. While this can improve the precision of density/abundance estimates, its use is only defensible if the data set is of sufficient size and quality to represent a reliable estimate of detection distance distribution. Otherwise there is a risk of overfitting the data and allowing non-representative characteristics of the data set to bias the fitting process.

Assumptions and sources of bias

Since distance sampling is a comparatively complex survey method, the reliability of model results depends on meeting a number of basic assumptions. The most fundamental ones are listed below. Data derived from surveys that violate one or more of these assumptions can frequently, but not always, be corrected to some extent before or during analysis.
AssumptionViolationPrevention/post-hoc correctionData example
All animals on the transect line itself are detected This can often be assumed in terrestrial surveys, but may be problematic in shipboard surveys. Violation may result in strong bias of model estimatesIn dual observer surveys, one observer may be tasked to "guard the centerline".
Post-hoc fixes are sometimes possible but can be complex. It is thus worth avoiding any violations of this assumption
Animals are randomly and evenly distributed throughout the surveyed areaThe main sources of bias are
a) clustered populations but individual detections are treated as independent
b) transects are not placed independently of gradients of density
c) transects are too close together
a) record not individuals but clusters + cluster size, then incorporate estimation of cluster size into the detection function
b) place transects either randomly, or across known gradients of density
c) make sure that maximum detection range does not overlap between transects
Animals do not move before detectionResulting bias is negligible if movement is random. Movement in response to the observer will incur a negative/positive bias in detectabilityAvoidance behaviour is common and may be difficult to prevent in the field. An effective post-hoc remedy is the averaging-out of data by partitioning detections into intervals, and by using detection functions with a shoulder
Measurements are exactRandom errors are negligible, but systematic errors may introduce bias. This often happens with rounding of angles or distances to preferred values, resulting in heaping at particular values. Rounding of angles to zero is particularly commonAvoid dead reckoning in the field by using range finders and angle boards. Post-hoc smoothing of data by partitioning into detection intervals is effective in addressing minor biases

Software implementations