Articles | Volume 16, issue 4/5
https://doi.org/10.5194/se-16-367-2025
https://doi.org/10.5194/se-16-367-2025
Research article
 | 
15 May 2025
Research article |  | 15 May 2025

Unbiased statistical length analysis of linear features: adapting survival analysis to geological applications

Gabriele Benedetti, Stefano Casiraghi, Daniela Bertacchi, and Andrea Bistacchi
Abstract

A proper quantitative statistical characterization of fracture length or height is of paramount importance when analysing outcrops of fractured rocks. Past literature suggests adopting a non-parametric approach, such as circular scanlines, for the unbiased estimation of the fracture-length mean value. This is due to the fact that, in the past, estimating any type of statistical distribution was difficult and there was no real interest in defining precise parametrical models. However, due to the recent rise in popularity of digital outcrop models (DOMs) and of stochastic discrete fracture networks (DFNs), there is an increasing demand for distribution-based solutions that output a correct estimation of parameters for a given proposed model (e.g. mean and standard deviation). This change in demand highlights in the geological literature the absence of properly structured theoretical works on this topic. Our methodology, presented for the first time in this contribution, represents a powerful alternative to non-parametrical methods, aimed at specifically treating censoring bias and obtaining an unbiased trace-length statistical model. As our first objective, we propose to tackle the censoring bias by applying survival analysis techniques: a branch of statistics focused on modelling time-to-event data and correctly estimating model parameters with data affected by censoring. As a second objective, we propose a novel approach for selecting the most representative parametric model. We combine a direct visual approach and the calculation of four statistics to quantify how well proposed models reflect the data. We apply survival analysis to correctly estimate statistical parameters of a censored length dataset in three different case studies and show the effects of the censoring percentage on parametrical estimations that do not use this paradigm. The presented analyses are carried out using an open-source Python package called FracAbility, which we purposefully created to carry out the described workflow (https://github.com/gecos-lab/FracAbility, last access: 8 September 2024).

Share
1 Introduction

Fractured rock masses are complex systems composed of intact rock and discontinuities (Hoek1983). Characterizing the statistical distribution of three-dimensional (3D) geometrical properties of such discontinuities (e.g. aperture, roughness, area, orientation, height / length ratio) is fundamental for understanding and modelling mechanical and hydraulic properties of rock masses and fluid–rock interaction. The recent increase in computing power and the emergence of new approaches based on digital outcrop models (DOMs) allow the extraction of large datasets and facilitate the sampling of properties instead of just their estimation (Bistacchi et al.2015; Tavani et al.2016; Healy et al.2017; Thiele et al.2017; Marrett et al.2018; Nyberg et al.2018; Bistacchi et al.2020; Martinelli et al.2020; Bistacchi et al.2022; Mittempergher and Bistacchi2022; Storti et al.2022). The need for a more rigorous statistical approach to structural data analysis is also motivated by the popularization of stochastic discrete fracture networks (DFNs) as a modelling approach for rock masses. In DFNs, discontinuities are represented, in a simplified way, as finite planar surfaces, generally rectangular, polygonal, or elliptical (Cacas et al.1990; Dershowitz et al.1992; Tavakkoli et al.2009; Hyman et al.2015). In fact, stochastic algorithms used to generate these surfaces are guided by statistical distributions obtained from field or well data or are assumed based on some prior knowledge (Andersson et al.1984; Cacas et al.1990; Davy et al.2018). For example, DFNs where fractures are represented as rectangular surfaces (the most common implementation) require the definition of fracture size via parametric distributions of length (measured along strike) and height (measured along dip) or, alternatively, length and length / height ratio (Hyman et al.2015). However, it is worth noting that DFNs are not the only viable approach to modelling fractured rock volumes. Other methods, such as tensorial approximations (Brown and Bruhn1998; Suzuki et al.1998) based on the crack tensor measure (Oda1983), are also used (Healy et al.2017) and, with details depending on the implementation, require similar parameters in input.

The main obstacle that geologists encounter while trying to characterize a fractured rock volume is the impossibility of directly measuring the 3D properties of discontinuities, since only indirect geophysical methods may provide complete 3D datasets. However, imaging discontinuities with geophysical methods are strongly limited by spatial resolution and/or by the absence of contrast in the physical properties investigated by a particular technique (Martinelli et al.2020). Therefore, a vast body of research focuses on the characterization of properties of discontinuity traces or lineaments, i.e. the two-dimensional (2D) intersections of 3D discontinuity surfaces with the outcrop surface or with topography (Dershowitz and Herda1992; Bonnet et al.2001; Manzocchi2002; Bistacchi et al.2011; Sanderson and Nixon2015; Bistacchi et al.2020; Martinelli et al.2020; Storti et al.2022). In this contribution, we focus on the problem of defining accurate and unbiased length (or height) distributions based on data collected on DOMs. In any given outcrop, lineament length measurements will always be affected by four main biases: length (i.e. size), censoring, truncation, and orientation (Baecher1983). The correction of these biases has been thoroughly researched, and the standard solution currently adopted by many authors is based on circular scanlines (Mauldon1998; Zhang and Einstein1998; Mauldon et al.2001; Rohrbaugh et al.2002). The method consists of drawing a circle directly on the outcrop or on an image and counting the intersections of the circle with lineaments. With a chain of assumptions, defining an indirect relationship between mean length and the intersections of lineaments with the scanline, an unbiased non-parametric estimation of mean length, fracture density, and intensity can be obtained (Mauldon et al.2001). Thanks to its simple implementation in the field, this technique is widely used; however, it has an important limitation: lineament lengths are never directly measured. For this reason, analysis carried out with the circular scanline method yields estimates of mean length values but is unable to provide a complete characterization of the lineament length distribution and therefore lacks any real statistical significance. Despite this problem, the popularity of this method was motivated by the fact that, in the past, without modern digital imaging techniques, length data acquisition was slow and tedious; thus datasets were usually small. Moreover, calculating the length and estimating any distribution other than the exponential was difficult and computationally intensive (Baecher and Lanney1978; Baecher1980), and, due to limitations in early algorithms used to generate stochastic fracture networks, there was no real interest in using precise distribution parameters in input. These limitations have mostly been overcome by modern characterization methods; thus new tools and techniques based on the maximum likelihood estimator (MLE), such as FracPaQ (Healy et al.2017; Rizzo et al.2017), have been developed, enabling researchers to apply quantitative statistical inference on dense digital datasets.

Our methodology, presented for the first time in this contribution, represents a powerful alternative to circular scanline methods to treat specifically the censoring bias and to obtain unbiased trace length statistical models. This specific bias is defined when, for some traces, one or both ends cannot be seen due to the limited size of the outcrop (Baecher and Lanney1978). This effect is present from thin section to satellite image scale, and it is caused by the inability to see beyond the study area (i.e. thin-section limits, outcrop extension; Mauldon et al.2001). From a statistical point of view, this problem is analogous to the censoring bias affecting some medical, biological, and engineering datasets, and the techniques used in these disciplines to solve or limit the effects of this bias go under the names of survival analysis, life testing, or reliability analysis (Kaplan and Meier1958; Leung et al.1997; Lawless2003; Cox2017; Karim and Islam2019). Even though, in these disciplines, the recorded random variables are time spans (e.g. lifetime of a patient, time to failure of a mechanical part), we demonstrate that this statistical technique can also be adapted to length measurements. Measuring length is straightforward with dedicated code or with a simple GIS software; however, applying survival analysis and fitting robust unbiased parametric statistical distributions needs a more detailed treatment. As the main topic of this contribution, we propose to adapt survival/reliability analysis techniques to correctly account for censored lengths and estimate robust trace length distributions derived from DOMs. Furthermore, we define a quantitative methodology to select the most representative estimated statistical model (i.e. parametric distribution) from a list of proposed models. The theory and techniques presented in this paper are available as an open-source Python package called FracAbility that accepts shapefiles as input and allows one to carry out a complete and unbiased statistical analysis workflow for fracture-length data (https://github.com/gecos-lab/FracAbility. last access: 8 September 2024).

2 Fracture surveys and terminology

A discontinuity in a rock mass can be defined as a surface across which a material has lost its cohesion or was originally discontinuous. This definition thus includes faults, fractures, foliations, stylolites, compaction or deformation bands, and bedding interfaces. Fractures are classified as shear or tensional fractures. Tensional fractures, when possible, can be further divided into joints when empty or veins when filled by minerals (Twiss and Moores2007; Davis et al.2012). This distinction, however, is often difficult to make without an in-depth field and sample validation, since some fractures may have hybrid fill attributes and may be only partly filled with inconspicuous mineral deposits and thus resemble joints. Moreover, the degree of fill may depend on fracture width so that small fractures resemble veins (Laubach et al.2019). Although three-dimensional by nature, most of the times discontinuities are mapped as 2D lineaments or traces over a surface. These lineaments are the intersections of such discontinuities with an exposed surface, such as the topography, an outcrop, a borehole, or a sample. In this work, following a common usage in outcrop studies, the term fracture is used as a generic term to indicate any type of discontinuity trace. Fractures with the same formation age, kinematics, and orientation can be grouped into families or sets. Multiple fracture sets present within an area form a fracture network or system (Davis et al.2012).

Additionally, fracture networks present two other fundamental components: boundaries and nodes (Fig. 1a). Boundaries are the limits of the sampling area, at scales from thin sections to satellite images, within which the sampled fracture traces are assumed to be complete. Ideally, boundaries are strictly convex; however, this is often not the case. Boundaries often show tight bends, coves, and holes, and the final shapes are mostly controlled by localized alteration, anthropogenic activity, vegetation, etc. (Fig. 1b).

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f01

Figure 1Panel (a) shows an example of a simple fracture network and its components. Panel (b) shows a modified version of the boundary with a “hole” (in red) in which no interpretation can be carried out. The presence of these holes can increase the number of censored fractures and introduce uncertainty into the interpretation, splitting fractures into multiple pieces (i.e. fracture 10 is split in two).

Download

Nodes are points in the network that define how fractures interact or do not interact with other elements of the network (other fractures, boundary, holes). Nodes can be classified by the number and type of segments (branches) that insist on the given point (Sanderson and Nixon2015) (Fig. 2):

  • Isolated I nodes are connected to only one branch belonging to a fracture.

  • Y/T and X nodes are respectively connected to three and four branches belonging to two fractures.

  • “Boundary” B nodes are connected to one branch belonging to a fracture and two branches belonging to the boundary.

Node elements describe useful hydraulic properties of a fractured rock mass (Gueguen et al.1991; Manzocchi2002; Sanderson and Nixon2015).
In a non-porous rock with all open fractures, a network with a prevalence of I nodes is less connected; thus fluid flow is usually more restrained. Conversely, both Y/T and X nodes increase the connectivity of the network and thus increase both fluid flow and the permeability. However, sometimes this is indeed not the case, e.g. sealed faults and opening-mode fractures (Forstner and Laubach2022). Moreover, if a rock is porous, then length becomes the key parameter for controlling fluid flow; thus, even in a network with only I nodes, length can markedly augment fluid flow (Philip et al.2005). Additionally, nodes give important chronological information about the network. I nodes indicate that fractures terminate within the observational boundary and that they do not end or cross with other fractures. Y/T nodes can be evidence of either abutting or cross-cutting relationships of non-coeval discontinuities; the former is relevant for discontinuities with no displacement, while the latter applies to discontinuities with visible displacement. X nodes represent a mutual cross-cutting relationship, and the chronological interpretation depends also on other observations. Finally, B nodes identify whether a fracture is complete or censored and, as such, define if measured properties along a trace are also complete or incomplete. These nodes could be regarded as a sort of equivalent of the U (undefined) nodes in Nyberg et al. (2018), however, with a slightly different meaning and more precise definition.

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f02

Figure 2Topological abstraction of the simple fracture network in Fig. 1 following the described node classification.

Download

3 Statistical modelling of censored length data

Having a length dataset, there is usually the necessity of estimating the parameters of one or multiple statistical distributions. When doing so, censoring is inevitable, as the area within which measures are carried out will always be limited. Then, how can we carry out an unbiased estimation of such parameters? On the one hand, common and simplistic approaches, such as ignoring censoring (i.e. regarding censored lengths as if they were complete) or excluding censored measurements (i.e. cherry-picking only non-censored data), should be avoided, as they will always lead to an underestimation of the model parameters (see Discussion for a more in-depth analysis). On the other hand, circular scanline methods offer an unbiased estimate of the mean length; however, being non-parametric, they yield neither the distribution type (e.g. normal, exponential) nor distribution shape parameters (e.g. standard deviation, variance). This in turn makes the estimate's use quite limited and not apt for downstream statistical modelling applications (such as DFNs). To solve these problems, we propose using survival analysis, a specialized field of statistics, developed to deal with censored data. Survival analysis focuses on the analysis of time of occurrence until an event of interest (Kalbfleisch and Prentice2002). Although, in literature, the terms survival times, time to failure, and (more generally) lifetimes (Lawless2003) seem to imply that time is the only valid variable, any non-negative continuous variable, such as length, can be valid (Kalbfleisch and Prentice2002; Lawless2003). The advantage of survival analysis over the other methods discussed above is that it regards censored data as a valid data point and as a carrier of the information that the event did not occur up to the censoring time. This is a necessary shift in perspective that allows an unbiased estimation of statistical models that will be described in this paper. We will start by describing the theory of survival analysis in function of time, then we will show how the same theory can be applied in space to sets of length or distance measurements.

3.1 Survival analysis theory and standard terminology

Since this technique is rooted in medical and biological applications, the nomenclature from this type of literature is carried along. The event of interest (for which we measure the time to event) is often defined as death, while a loss indicates that the observation has been lost because it was hindered by a secondary event, called a censoring event (Kaplan and Meier1958). Censoring events can be classified as three main types depending on when censoring happens in respect of the observation period (Karim and Islam2019) (Fig. 3):

  1. Right censoring. The event happens after the end of the study period; thus the partial lifetime of the event is observed.

  2. Left censoring. The event happens before the start of the study; thus it is not observed.

  3. Interval censoring. The event happens somewhere between observation intervals within the study period because it was not possible to continuously monitor the occurrence of the event.

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f03

Figure 3Different censoring types in a fixed observation period. Right censoring is defined when the event happens after the end of the study. Left censoring occurs when the event happens before the start of the study. Interval censoring happens somewhere between observation intervals within the study period.

Download

The most treated and common type is right censoring, which can be further divided into the following (Fig. 4):

  1. Single censoring. When there is an imposed end time equal for all events (i.e. controlled loss).

  2. Random censoring. When each measured event is characterized by a random censoring event (i.e. accidental loss).

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f04

Figure 4Different right-censoring types for the same seven subject experiments. Panel (a) represents single censoring, where events 1, 3, and 6 are complete measurements, while the remaining events are censored at time C. In panel (b), only event 7 is complete, while the others are all censored at different times.

Download

3.2 The survival curve and the Kaplan–Meier estimator

A fundamental assumption in survival analysis is the independence of the right-censoring mechanisms, i.e. the assumption that the probability of occurrence of the event does not depend on the censoring mechanism (Kalbfleisch and Prentice2002; Lawless2003; Kleinbaum and Klein2012). Independent censoring is also called non-informative censoring, as it does not affect inference and only indicates that the lifetime exceeds the censoring time (Kalbfleisch and Prentice2002). This assumption is the basis of the non-parametric Kaplan–Meier estimator of the survival function (SF), i.e. the population probability that Xi exceeds a given value x (Kaplan and Meier1958; Kalbfleisch and Prentice2002):

(1) SF ( x ) = P ( X i > x ) .

The survival function is often called the complementary cumulative function, since it adds to 1 with the cumulative distribution function (CDF):

(2) CDF ( x ) = P ( X i x ) = 1 - P ( X i > x ) = 1 - SF ( x ) .


The non-parametric Kaplan–Meier estimator P^(x) is defined by ordering N values of complete xco and censored xce data by increasing magnitude, with x1x2xN, such that (Kaplan and Meier1958; Kalbfleisch and Prentice2002)

(3) P ^ ( x ) = r N - r N - r + 1 ,

where r is the indexes in which the complete values are smaller than a value x (i.e. xcox).

Therefore, P^(x) is a step function that (Fig. 5)

  1. remains constant in any given time interval where no new events are recorded or where censoring occurs;

  2. decreases by a degree (step) depending on the number of events that occur in a time interval.

Empirical survival curves are fundamental for representing, comparing, and understanding the survival rates of different censored datasets. The curve's steepness is directly proportional to the survival rate, and the median of the (random) survival time (i.e. time value corresponding to a 50 % chance of survival) can be used as a simple indicator to compare different survival curves.

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f05

Figure 5An example of the Kaplan–Meier estimated survival curve. In panel (a), the estimated curve seems continuous; however, when zooming in (red rectangle in panel a), one can see it is a step function (b). The red crosses indicate the censoring events and show how the function never jumps at these values.

Download

3.3 The time–length dimensional shift

Since length is, as time, a non-negative continuous variable, it is theoretically possible to apply survival analysis techniques to length datasets by considering (Fig. 6)

  1. the complete fracture trace length analogous to the lifetime;

  2. the event as the end (death) of a trace, marked by a node (I and/or Y/T nodes in the case of complete traces);

  3. the study area, defined by its boundary, analogous to the study period;

  4. the censored event to be the intersection between the fracture trace and the boundary (marked by a B node).

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f06

Figure 6Censoring effect on an example of a simple fracture network and corresponding survival diagram. The survival diagram is a 1D representation of the fracture length. On the y axis the fracture number is indicated, and on the x axis the length is represented. Solid lines indicate the actual measured length, while dashed lines indicate the possible continuation of the fracture. Yellow pentagons represent the censoring event given by the intersection of the fracture with the boundary.

Download

By applying the definitions of the different types of censoring (described in Sect. 3.1) to our specific application, it is reasonable to assume that only random censoring occurs in trace length analysis. Moreover, censoring is non-informative, since the boundary is the product of secondary events (i.e. alteration, erosion, debris covering parts of the outcrop, vegetation, human activity) that occur after fracture genesis and thus do not inform the occurrence of the event (see Discussion for a more in-depth analysis).

3.4 Unbiased MLE model estimation for censored data

Considering that the censored event does not modify the probability of occurrence of the measured event, survival analysts try to solve the problem of estimating a parametric statistical model (i.e. distribution) by using optimization algorithms such as linear regression or the maximum likelihood estimator (MLE). Both methods are valid; however, in these types of applications, MLE is most used; thus we will discuss this technique in more detail.

Given a statistical parametric model with density g(x,θ) (i.e. an assumed theoretical distribution) and a sample x of size n, the main objective of MLE is to estimate the parameters θ^ such that the observations x are the most likely under g(x,θ^) (Burnham and Anderson2002, 2004; Karim and Islam2019).

In the simplest one-dimensional (1D) case (estimation of the parameter of a one-parameter distribution), the likelihood can be described as a chain product of probabilities carried out on all n individuals in the following sample:

(4) L ( θ | x , g ) = i = 1 n g ( x i , θ ^ ) .

Given the fixed (assumed) distribution type g with parameter θ and a sample x, the likelihood will be the product of all the probabilities g(xi,θ^), i.e. the distribution's probability density function (PDF) calculated at value xi (P(θ|xi)). In practice, since g and x are fixed, the likelihood is a function only of the distribution parameter θ. The maximum of this function will be θ^, i.e. the parameter that maximizes the likelihood for that model type and data combination. This process can be extended to estimate multiple parameters, increasing the complexity, but the core concept remains unaltered. However, Eq. (4) is valid only for complete datasets because the probabilities associated with right-censored data cannot be calculated using the PDF. To solve this problem, it is possible to calculate the MLE for censored datasets, under the assumption of random censoring, using the model's survival function (SF) S(θ|xi) instead of the PDF for the censored data (Karim and Islam2019):

(5) L ( θ | x , g ) = i = 1 n P ( θ | x i ) δ i × S ( θ | x i ) 1 - δ i ,

where δ is an on–off switch for complete (δ=1) or censored (δ=0) values. In other words, when data are censored, we use the probability information that the individual survived up to the censoring event, given by the survival function.

MLE is a consistent (i.e. estimation approaches the population parameter as sample size increases) and efficient (i.e. lowest variance) estimator (Enders2005); however, it has its limitations. For example, if the model has more than one parameter, the weight of influence of each parameter is not known; thus it is difficult to know which parameter controls the fit. Moreover, simplistic MLE algorithms output a single likelihood value, with a possibility that the optimization gets trapped in a local minimum. This ultimately leads to questioning whether the estimated parameter or combination of parameters is the absolute best or if there are other more optimal solutions. These (and other) problems culminate to an even greater uncertainty related to the choice of the distribution, as an infinite variety of functions can be used for the same dataset. These uncertainties tie nicely to the following subsection discussing model selection comparison criteria and how we propose to solve this problem.

3.5 Model selection

With a series of fitted models, it is important to understand which model better represents the data. In the geological literature, this is usually achieved using a specific type of null hypothesis test defined as a goodness-of-fit (GOF) test (i.e. Storti et al.2022). These types of tests, in general, do not identify the best-fitting distribution among a set of possible distributions. A GOF test takes a distribution L and the sample data and tells if the model L is not plausible; that is, if the probability of drawing such a set of data, from a population of distribution L, is too small. This means that more than one distribution may be deemed plausible and that the conclusion is strong only when a distribution is an unlikely model for the data. Moreover, GOF tests usually have underlying assumptions that, if ignored, undermine the test's accuracy (Storti et al.2022). For example, the Kolmogorov–Smirnov test (Kolmogorov1933) is biased if the parameters of the tested distribution are estimated from the data (Bistacchi et al.2020). As an alternative to GOF tests, we propose a combined visual and quantitative approach that guides the researcher in an informed choice of the most representative model out of a list of sensible candidates. By sensible candidates, we mean those models that make physical sense for the case study. For example, for fracture lengths, a power law can be considered a sensible candidate, since it describes an observed pattern of fracture self-similarity. A lognormal is also an equally sensible candidate because of the effect of truncation. Conversely, a normal distribution is not very sensible, since values in the left tail can be negative. Thus, for a normal model, either the average length is very high and the standard deviation is very low or a truncated normal should be tested. In the Case studies section, we will cover different models, purposefully adding non-sensible models (such as a standard normal) to showcase the model selection workflow.

3.5.1 A visual approach using the probability integral transform (PIT)

The probability integral transform (PIT) is a well-known transformation of continuous distributions which converts random variables with continuous distribution to standard uniform distributed random variables (Fisher1990), given

  1. a random variable Y,

  2. FY as the CDF of Y,

  3. Z=FY(Y) as the PIT of Y and a standard uniform random variable.

This means that, given a random sample X, sampled from a population with distribution Y, the transformed sample Z=FY(X) has, with large probability, an empirical distribution which is close to the standard uniform distribution. Clearly, if we transform X with another distribution W, which is not the true distribution of the population, then FW(X) is not uniformly distributed. This observation provides a visual tool to compare different fits: indeed, suppose that sample X might have been sampled from a population of distribution Y or W. We may compute the two transformed samples Z1=FY(X) and Z2=FW(X): if X was drawn from distribution Y, then the empirical distribution of Z1 is likely to be close to the standard uniform, while, if the true distribution was W, then we expect that Z2, instead of Z1, would be close to the standard uniform. A synthetic example of PIT is presented in Fig. 7. A random sample X of 10 000 (non-censored) measures is extracted from a standard normal distribution Y (Fig. 7a). The exponential YE model (hypothesis) has been fit with an MLE to the random sample. Figure 7b and c represent the empirical frequency histogram and the empirical CDF (ECDF) of Z for both models (ZN=Fn(X) and ZE=Fe(X)). Both figures show that the normal estimated model follows a standard uniform, while the exponential model does not. The ECDF visualization of Fig. 7c is preferred over the simple frequency binning of Fig. 7b, since multiple models can be tidily represented with the reference uniform distribution appearing as the diagonal of the plot. The closer the ECDF of Z is to the diagonal of the plot, the closer the fit is to the true underlying model. Values in the PIT plot that are below the diagonal indicate an overestimation of the model's CDF in respect to the empirical cumulative. Conversely, values that are above the diagonal will result in an underestimation of the model's CDF. Hence, the PIT provides a simple visual, yet powerful, method to estimate which parametric model better fits the empirical data.

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f07

Figure 7Synthetic example of the probability integral transform. Panel (a) shows 10 000 random samples X drawn from a standard normal distribution Y. Panel (b) shows the frequency distribution of Z from two different models (hypotheses): normal (YN) and exponential (YE). Panel (c) shows the empirical cumulative of Z for both models. In panels (b) and (c), it is possible to observe the effect of the probability integral transform. In panel (b), the empirical frequency histogram of the normal model is remarkably close to a uniform distribution, while the exponential model is not. This is visualized much more clearly in panel (c), where the normal model is the closest to the diagonal line y=x (i.e. the standard uniform).

Download

3.5.2 A quantitative approach using distances

While the PIT visual approach is very intuitive, we propose also using a quantitative approach, calculating four different statistics. The first is the Akaike information criterion (AIC) (Akaike1974), which, by using the result of the MLE, quantifies the distance between the true natural phenomena and the estimated model as

(6) AIC = 2 k - 2 ln L ( θ ^ ) ,

where k is the number of parameters used by the model (i.e. the dimension of the parameter vector) and θ^ is the k-dimensional parameter vector that maximizes the likelihood of the model ().
This formulation outputs a real number (either positive or negative) which should be small as the distance between the true population and the model decreases. If multiple models are tested, then it is possible and advised to calculate the Δi parameter, i.e. the distances between the different models to the best-scoring one (AICmin):

(7) Δ i = AIC i - AIC min ,

where i is the index of the proposed model.

From Δi it is possible to obtain the weight of evidence (wi) of the given model with the following formula:

(8) w i = e - Δ i / 2 r = 1 R e - Δ R / 2 ,

where R is the total number of tested models (i.e. models deemed reasonable by the researcher).

The weight of evidence outputs a value between [0;1] that represents, in a set of proposed hypotheses, how likely it is that the model comes close to the true underlying process. The closer to 1, the more likely it is that, in the pool of candidates, the model represents the true underlying process (Burnham and Anderson2004). Since AIC and the derived formulas are directly based on the MLE, they are affected by the same limitations discussed in Sect. 3.4 (Akaike1974). We thus propose calculating three different distances between the model and the data (Kim2019). Usually, distances are calculated by comparing the empirical cumulative frequency with the cumulative frequency of the model; however, the data are censored; thus we use the empirical cumulative frequency estimated from Kaplan–Meier. Moreover, Kim (2019) proposes to calculate the distances using the data transformed with PIT (i.e. Z=Fy(X)). We would like to highlight that, purely from the point of view of the calculation, the distances are the same with or without using the transform; however, with PIT, the data are “normalized”; thus the different models are compared over the same scale (0, 1). Equation (3) then can be rewritten as

(9) G ^ n ( z ) = 0 , z < Z 1 1 - Z r z ( N - r N - r + 1 ) δ r , z Z N 1 , z > Z N ,

where r is the indexes of each data point and δr is the same on–off switch of Eq. (5).

Kim (2019) then proposes to calculate

  • the Kolmogorov–Smirnov statistic (DCn), representing the maximum distance between the two cumulative curves (Kolmogorov1933; Smirnov1939);

  • The Koziol and Green statistic (Ψ2), representing the sum of squared distances between the two cumulative curves (Koziol and Green1976);

  • the Anderson–Darling statistic (ACn2), representing the weighted sum of squared distances between the two cumulative curves, imparting more weight than the Kolmogorov–Smirnov on tail observation (the closer to 0 or 1, the higher the weight; Anderson and Darling1954).

The three sets of statistics can be modified to accommodate the presence of censored data by using the Kaplan–Meier estimator.

The Kolmogorov–Smirnov statistic is generally calculated as

(10) DC n = max ( DC n + , DC n - ) ,

where, with censored data,

(10a)DCn+=max(G^n(Zr)-Zr),(10b)DCn-=max(Zr+1-G^n(Zr)).

The Koziol–Green statistic is a generalization of the Cramér–von Mises statistic (Koziol and Green1976):

(11) Ψ 2 = N 0 1 ( G ^ n ( z ) - z ) 2 d z ,

which, considering censored data, can be written as (Appendix 2 in Koziol and Green1976)

(11a) Ψ 2 = 1 3 N + N r = 1 N G ^ n ( Z r ) × ( Z r + 1 ) - Z r × [ G ^ n ( Z r ) - ( Z r + 1 - Z r ) ] .

Finally, the Anderson–Darling distance (Anderson and Darling1954), generally defined as

(12) AC n 2 = N 0 1 ( G ^ n ( z ) - z ) 2 z ( 1 - z ) d z ,

where 1z(1-z) is the weight, can be calculated accounting for censoring as a finite sum (Kim2019):

(12a) AC n 2 = N r = 1 N - 1 ( G ^ n ( Z r ) ) 2 × ( - ln ( 1 - Z r + 1 ) + ln ( Z r + 1 ) + N × ( 1 - Z r ) - ln ( Z r ) ) + - 2 N r = 1 N - 1 ( G ^ n ( Z r ) ) × ( - ln ( 1 - Z r + 1 ) + ln ( 1 - Z r ) ) + - N × ln ( 1 - Z N ) + N × ln ( Z N ) + N .

Once these distances are calculated for each model, we propose to rank the models independently for each distance (from minimum to maximum). If all distances converge (i.e. for the same model, the same rank is assigned), then we regard this as sufficient proof for the overall ranking of the model in the list. By comparing multiple rankings, even if the calculated distances are ranked differently, it is still possible to make a sensible guided choice, for example, by using the PIT representation together with the mean ranking position or by following a specific type of distance.

4 Case studies

We demonstrate the applicability of the discussed theory by using the FracAbility package in three different case studies. The first case study focuses on the characterization of a very regular and densely fractured fracture network with a simple boundary geometry. The second focuses on analysing a typical average-quality outcrop, with multiple fracture sets and complex boundary geometry. Finally, the third case study presents an application with spacing analysis to demonstrate that any type of length can (and must) be corrected from right-censoring bias. We propose to test six statistical models of length, frequently adopted in the literature (Bonnet et al.2001) or used in stochastic DFNs:

  • Lognormal

  • Truncated power law (power law in the following)

  • Normal

  • Weibull

  • Exponential

  • Two-parameter gamma (gamma in the following).

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f08

Figure 8Overview maps of the first case study area. Pictured on the left is a general overview of the area with the boundary geometry overlaid on the orthophoto and a small sample of the digitized fracture set. Pictured on the right is a subarea of the quarry with the boundary, fracture traces, and boundary nodes (yellow pentagons).

4.1 First case study: Cava Pontrelli, Puglia (IT)

This first case study focuses on the analysis of a single NW–SE-striking fracture set present in an abandoned quarry near the town of Altamura in Puglia (Italy). The study area is located on the Apulian Platform, representing the forebulge of the southern part of the fold and thrust belt of the Apennines (Vai and Martini2001; Patacca and Scandone2007). The outcrop is characterized by an extensive horizontal pavement of about 18.000 m2, showing densely fractured Cretaceous platform limestone of the Altamura Limestone Formation (Ricchetti and Luperto Sinni1979). The continuous maintenance that followed the discovery of thousands of dinosaur footprints on the quarry pavement (Nicosia et al.1999) made it possible to obtain an exceptionally clean outcrop surface. This in turn resulted in the definition of a simple boundary geometry (with a couple of interpretational holes) and the digitalization of a very regular and dense fracture network (1941 fractures). The combination of these factors led to a very low percentage of censoring, with only 8.9 % of the total fractures being censored (Fig. 8).

We applied survival analysis as implemented in FracAbility, and, in Fig. 9, the resulting PIT plot is shown. The most representative model appears to be the lognormal, confirmed by all the distances calculated in Table 1 (ordered by ascending Akaike scores; see Fig. 10 for a summary of the fit). Figure 9 shows that the PIT of the lognormal model is quite linear, with gentle undulations indicating a slight underestimation for lengths between 1 and 2 m (accounting for 20 % of the measures) and a slight overestimation for lengths between 3.5 and 6.5 m (accounting for another 20 % of the measures). For the calculated distance scores, the lognormal is followed by the gamma and Weibull models; however, the rank scores of these last two models are not uniform. While the Koziol and Green distance favours the gamma model, the Anderson–Darling distance favours the Weibull model. Finally, the power-law and normal models rank lowest in comparison to the other models, indicating that they are less able to represent the dataset.

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f09

Figure 9PIT visualization for the proposed length models in the Pontrelli quarry. The red line represents the reference U(0,1): the closer the model line is to this reference line, the more representative the model. For this dataset, the lognormal model is the most representative, following the reference line almost perfectly, with some minor underestimation between 1.16 and 2.11 m and overestimation between 3.58 and 6.47 m.

Download

Table 1Models' distance tables and ranking score tables for the Pontrelli quarry. The closer to 0, the better. For this dataset, the lognormal is the most representative of the data in all the different distances, while the power-law and normal models are the least representative. The positions of the other models in between are less certain (especially the gamma and exponential).

Download Print Version | Download XLSX

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f10

Figure 10Summary of the best-fitting model (lognormal) for the Pontrelli quarry dataset. (a) The probability density function against the histogram of the dataset. (b) The cumulative distribution function and (c) the survival function against the empirical counterparts calculated with the Kaplan–Meier estimator. (d) The summary table of the main statistics for the data (e.g. sample mean, sample variance) compared with the estimated model.

Download

4.2 Second case study: Colle Salza, Valle d'Aosta (IT)

The second case study focuses on the analysis of a less ideal, albeit more realistic, outcrop. The study area is located in the basement of the Western Alps (Dal Piaz et al.2003), on paragneiss of the Monte Rosa Nappe (Dal Piaz and Lombardo1986). The outcrop is cross-cut by several brittle fractures, Tertiary in age (Bistacchi and Massironi2000), and is characterized by a main central area of 1234 m2 and two secondary smaller satellite areas of 9 and 3 m2. The boundary geometry is highly concave, thus leading to a high censoring fraction of fracture traces. Moreover, the outcrop is exposed on top of a small topographical height with a slightly ellipsoidal shape due to glacial erosion (i.e. a roche moutonnée), with the main axis directed NW–SE. This leads to an inevitable deformation of the orthophoto (and thus of the measured lengths) along the extremities of the analysed area. The analysed fractures are subdivided in two main sets, striking NE–SW (Set 1) and NW–SE (Set 2), conforming to the general trend of the area for brittle deformation (Bistacchi et al.2000; Bistacchi and Massironi2000; Bistacchi et al.2001). The total number of fractures is 1718 (24.7 % censored), while, set-wise, the number of fractures is 692 (24.3 % censored) for Set 1 and 1026 (24.9 % censored) for Set 2 (Fig. 11).

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f11

Figure 11Overview maps of the second case study area. Pictured on the left is a general overview of the area with the boundary geometry overlaid on the orthophoto and a small sample of the digitized fracture sets. On the right is a subarea of the quarry with the boundary, fracture traces, and intersection nodes between the two (yellow pentagons).

The different estimated models are represented for both sets in Fig. 12. In both cases, the lognormal model better fits the data overall. Nonetheless, the fit of estimated models is visibly worse than in the first case study, particularly for Set 1 (Fig. 12a). The lognormal fit of Set 1 underestimates lengths between 0.34 and 1.5 m (about 50 % of the measures), while lengths above this value are generally overestimated. For the lognormal fit of Set 2 (Fig. 12b), there is a less relevant underestimation of the length values between 0.44 and 2.57 m (about 60 % of the measures) and a slight overestimation for longer length values. The distance values in Table 2 (ordered by ascending values of Akaike) confirm the lognormal as the most representative for both sets (see Fig. 13 for the summary plots of the best fit). For the other models, the mean rank value can be used to assign a final ranking. The mean ranks show that the gamma distribution is less representative than both the exponential and the Weibull (respectively the second- and third-most representative models).

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f12

Figure 12PIT visualization for the proposed length models is shown for Set 1 (a) and Set 2 (b) of the Colle Salza dataset. The red line represents the reference U(0,1); the closer a model's line is to this reference, the more representative the model. Among the models, the lognormal distribution demonstrates the closest fit to the reference line in both sets, although showing a worse fit to that observed in the first case study. Across both sets, all estimated models exhibit less linearity, with notable underestimation between 0.34 and 1.5 m in Set 1 (a) and between 0.44 and 2.57 m in Set 2 (b).

Download

Table 2Models' distance and rank tables for both Colle Salza sets. The closer to 0, the better. For this length dataset, the lognormal is the most representative of the data for all the different distances. Using the mean rank, the lognormal is followed by the exponential, Weibull, and gamma distributions. As with the first case study, the power-law and normal distributions are the least representative.

Download Print Version | Download XLSX

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f13

Figure 13Summary of the best-fitting model for the Colle Salza dataset (Set 1 and 2). (a) The probability density function against the histogram of the dataset. (b) The cumulative distribution function and (c) the survival function against the empirical counterparts calculated with the Kaplan–Meier estimator. (d) The summary table of the main statistics for the data (e.g. sample mean, sample variance) and estimated model.

Download

4.3 Third case study: spacing analysis

Survival analysis can also be used to analyse the spacing length distribution for each fracture set. Thus, the same workflow was applied to spacing measurements of both Set 1 and Set 2 of the Colle Salza dataset. It is worth noting that the censoring analysis is a secondary part of the analysis for spacing. Analysing the spatial arrangement of the fractures in the network (such as Marrett et al.2018; Bistacchi et al.2020) is of fundamental importance; however, we decided not to include this analysis and focus mainly on censoring to avoid increasing the length of an already dense text. Spacing is defined as the distance between two fractures of the same set, measured perpendicular to the average fracture plane attitude (Bistacchi et al.2020, and references therein). Traditionally, this type of statistic is obtained in the field with scanline surveys, with properly oriented scanlines, or by applying the Terzaghi correction (Terzaghi1965). With DOMs, this procedure can easily be sped up with custom scripts, programmatically tracing a large number of parallel scanlines and thus increasing the number of spacing measurements (De Toffoli et al.2021; Storti et al.2022). However, this leads to a higher number of censored spacing measurements occurring at the outcrop boundary and where “holes” are present. As with standard lengths, the introduction of a boundary affects the actual estimation of the spacing by a degree depending on the outcrop's convexity (Fig. 14). Perfectly convex boundaries (Fig. 14b) will lead to minor censoring, affecting only the ends of the scanlines, while more realistic non-convex boundaries will increase the censoring percentage because the boundary will cross the same scanline at multiple points (Fig. 14c).

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f14

Figure 14The effect of boundary geometry on censoring for spacing distribution modelling. (a) The ideal situation where no censoring occurs. The visible intersections between the scanline and the fracture are green, while the measured spacing segments are blue. (b) The effect of a perfectly convex boundary (i.e. a circle) on the estimation of censoring. For the “visible” scanlines, only the final segments are censored. (c) The effect of a complex boundary geometry. Here, the number of censoring measurements increases drastically, showing that censoring is no longer limited to the ends of the scanlines.

Download

For the Colle Salza dataset, 1282 and 1119 spacing measurements were obtained for fracture sets 1 and 2 respectively (with 52.5 % and 36.5 % censoring). Figure 15 shows the distances of the different models for the obtained spacing values. In this case, for Set 1, an extreme underestimation of the lognormal model is present for values longer than 4 m (Fig. 15a), while, for the other proposed models, particularly the Weibull, the distances at the same values are smaller. This is also visible in the distance tables (Table 3). Although the Akaike distance for the lognormal model is the smallest, the difference from the other models (Δi) is not as big as in the other examples. Thus, the weight of evidence (wi) is also not completely in favour of the lognormal model. The poor fit of the lognormal model is also confirmed by the other distances that rank the lognormal lower than the Weibull and gamma models (ranking first and second respectively). All these factors point to the fact that, in this case, the Weibull is the most representative model (see Fig. 16 for a summary representation of the best fit). Conversely, the lognormal model is clearly the most representative for Set 2 (Fig. 15b). The PIT plot shows a quite linear behaviour and all the calculated distances ranking first.

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f15

Figure 15PIT visualization for the proposed length models is shown for Set 1 (a) and Set 2 (b) of the Colle Salza dataset. The red line represents the reference U(0,1); the closer a model's line is to this reference, the more representative the model. In contrast with the other case studies, the lognormal distribution demonstrates a marked underestimation for values longer than 4.4 m for the spacing of Set 1 (a). In this case, the closest fit is the Weibull. Conversely, for the spacing of Set 2 (b), the lognormal model is again the model closest to the reference line, performing similarly to the first case study.

Download

Table 3Models' distance and rank tables for the spacing calculated on both Colle Salza sets. The closer to 0, the better. For Set 1, the low performance of the lognormal in the estimation of longer values (Fig. 15a) is visible at all distances. Even if the Akaike distance calculated for the lognormal is the smallest, the difference from the other calculated values is not enough to completely justify the selection of this model. This is also shown in the other distances where the lognormal clearly underperforms. On the other hand, for Set 2, the lognormal model is the most representative for the spacing values for all the different distances.

Download Print Version | Download XLSX

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f16

Figure 16Summary of the best-fitting model for the spacing dataset (Set 1 and 2). (a) The probability density function against the histogram of the dataset. (b) The cumulative distribution function and (c) the survival function against the empirical counterparts calculated with the Kaplan–Meier estimator. (d) A summary table of the main statistics for the data (e.g. sample mean, sample variance) and estimated model. The spacing of Set 1 is the only case where the data are not lognormally distributed, and the frequency histogram reflects such a conclusion.

Download

5 Discussion

With this work, we delineate a robust statistical framework to quantitatively analyse and statistically model fracture trace length and spacing. The statistical distributions are estimated using the entire length of fracture traces, following the original works of Baecher and Lanney (1978) and Baecher (1980) and not the length of branches as defined in Sanderson and Nixon (2015). Branches offer a useful topological abstraction of the network (making it possible to classify node intersections); however, they do not carry a real geological or physical meaning; thus a distribution obtained by fitting branch length will have a different meaning compared to one obtained from complete lengths. Furthermore, length can sometimes be affected by a subjective bias leading to an uncertainty as to which structure should be measured. For example, with segmented or en echelon fractures, do we measure the single segments or consider their cumulative sum? By approaching this problem with branches, a clear objective answer cannot be reached, while, by regarding a 2D lineament as a representation of the 3D geological structure, contextualized in the geological history of the outcrop, a more rational and informed decision can be made (i.e. Forstner and Laubach2022).

The first and crucial point of discussion is the objective of having a representative unbiased statistical model that is sufficiently likely to reproduce the observed length data. In particular, we want to stress that there is no real interest or practical reason and no theoretical possibility to find the real underlying distribution, while there is a strong necessity to fit specific simplified parametric models, useful for stochastic modelling applications (such as DFNs). Due to this, non-parametrical methods, such as those proposed by Mauldon et al. (2001), are unfit, since they are not linked to any model. Furthermore, having an unbiased statical model can also be extremely important for engineering applications. For example, at the end of the paper by Pahl (1981) it is stated that, in geomechanical application, it could be useful to know the longest trace likely to occur in a group of traces. With a parametric model, this can easily be estimated by checking the length values associated to a probability chosen, depending on the safety margin that is needed for the use case. Modern alternatives that use a simple implementation of MLE, such as FracPaQ (Healy et al.2017), are a good step forward; however, the censoring bias is still present; thus the results are biased, always underestimating length. The approach proposed here provides for the middle ground that was missing. We have shown that it is possible to use survival analysis to model censored length datasets, and we have showcased a new approach in model selection via quantitative distances to rank and choose the most representative model from a pool of simple candidates that can be used in common modelling applications.

As stated, to use MLE for a given model, we assumed that, in geological applications, the only censoring mechanism present is random censoring. Both left and interval censoring, by definition, cannot happen because the event of interest coincides with the end of the fracture. Thus, for left censoring, fractures terminate outside the boundary and thus cannot be measured (this is analogous to a patient never entering the study because they are already dead). The same applies to interval censoring that could be erroneously identified when holes are present in the sampling area or when boundaries are concave. However, since the measured event is the end of the fracture, on the other side of the interruption no fractures should be present. In terms of classic survival analysis theory, this would compare to a patient's loss of follow-up in the study period, caused by voluntary or involuntary exit from the study group, thus classifying it again as right censoring (Leung et al.1997).

To correctly classify censoring as random, we must assume independence between the censoring and length distribution. By “independence”, it is intended that the mechanisms behind the generation of a fracture-length distribution are different from the mechanisms that censor such lengths. The boundary, which represents censoring, is usually the product of secondary events that occur after fracture genesis (i.e. alteration, debris hiding part of the outcrop, vegetation, human activity). Thus, while it is often the case that such events are controlled by pre-existing structures, the physical processes that caused censoring are not the same ones that generated the fracture set and thus the original length distribution. This leads to an important implicit caveat where the measured lengths must be related only to the mechanism that we are interested in modelling, for example, lengths that are surely linked to tectonics and no other secondary events. Such discussion highlights that the assumption of independence is difficult to rigorously prove, since the true distribution of the length of fractures is not known (we only observe a set of complete and censored data). In some applications (Eppes et al.2024), this assumption may not hold, and a more in-depth study may be required to prove the independence hypothesis before proceeding. Nonetheless, we believe that it can be safely assumed in geological applications when the appropriate field work and a posteriori analysis are carried out.

As a second point, we would like to discuss the relationship between the fraction of censored data and the uncertainty in the estimated statistical model. There is not much literature on this relationship, probably because, since traditional survival analysis is applied in the context of time, the time boundary can be expanded if necessary. For example, if a study shows a censoring rate that is deemed too high, the researcher could repeat the experiment and simply extend the study period to observe the event of interest. In our case, the spatial boundary usually cannot be expanded; thus knowing the effects of censoring on estimation is an important aspect. For example, it would be useful to know how censoring influences the estimations based on simplistic approaches described in Sect. 3, the censoring percentage value above which survival analysis must be used, or the value above which even survival analysis fails. Moreover, it can also be useful to estimate the precision of estimation with survival analysis, knowing the censoring percentage. The case studies discussed in this work show a censoring percentage ranging from 8.96 % (first case) to 52.50 % (third case) and effectively showcase the influence of censoring on the estimation. For each dataset, it is possible to use the best-fitting statistical model defined in the results and to use PIT plots to visualize the differences. The comparison with the circular scanline approaches was intentionally omitted because it cannot be compared in any way. Being a non-parametrical approach, cumulative frequencies cannot be compared, and comparing mean values without a variance is not informative. Figure 17 shows that, in all cases, the estimated model using survival analysis closely follows the reference uniform, while the other two approaches (i.e. ignoring censoring and removing censored data) always overestimate the cumulative frequency, diverging with a rate positively correlated with the censored fraction. This translates to an underestimation of a model's parameters proportional to the censoring percentage. This effect is due to the fact that measured lengths of censored fractures will always be shorter than their true lengths. Thus, by using the first simple approach, the dataset is essentially “polluted” by shorter fractures, thus always decreasing the measured mean. The second simple method will also lead to an underestimation of the mean because of the size bias. However, this second method can be less impacted by censoring. For example, if a fracture population has a very small standard deviation (i.e. almost all fractures have the same length) and/or fractures are occurring in an outcrop that is much bigger than the characteristic fracture length, then removing censored values would not have a great impact on the estimation. But, even if small, the underestimation will always be present. Overestimation of the mean length would be possible in these scenarios when we do not regard censoring as independent from the length distribution (for example, if only fractures shorter than a certain value are systematically censored). However, this would violate both the core underlying hypothesis of random censoring and standard geological experience; thus we do not deem it possible under these imposed limits. Furthermore, the model obtained by removing censored values is always the worst. This is because MLE is a consistent estimator that converges as the number of data increase; thus, by removing data, we are effectively hindering the estimator's precision (in the last case study, we removed 52 % of the data, i.e. 667 fractures). This effect is also visible when comparing the estimation of important distribution parameters such as mean (μ) and standard deviation (σ2). Table 4 compares the values of these parameters calculated with the different methods, showing a consistent underestimation of both the mean and the standard deviation, with a greater underestimation coming from the method that removes censored data.

https://se.copernicus.org/articles/16/367/2025/se-16-367-2025-f17

Figure 17PIT visualization for the lognormal model estimated with three different approaches: ignoring censoring (full line), removing censored data (dashed line), and using a survival analysis approach (dashed–dotted line). The survival model follows the true model almost constantly (red line), while the other two approaches increasingly diverge on longer fractures with a rate depending on the censored fraction.

Download

Table 4Summary table comparing mean (μ) and standard deviation (σ2) values for increasing censoring fractions. The method ignoring censoring and the method removing censored data both underestimate the values, with the latter always showing a greater underestimation.

Download Print Version | Download XLSX

The outlined effect could be useful to predict what to expect from any survival analysis and shed some light on the exploration of the censoring mechanics that govern these types of datasets. Still, the comparison briefly discussed in this chapter is not rigorous enough to fully constrain this relationship. Even if we have a clear increase in the censoring percentage in the different datasets, the genetic factors, lithology, and censoring mechanics are completely different. Moreover, the number of measures in each case study are different and thus not statistically comparable. Because of these reasons, this behaviour should be isolated and modelled in controlled synthetic experiments to be carried out in the future. Nevertheless, we found that our preliminary results and the possible implications were too interesting not to be discussed.

As a third point, the readers that support and firmly believe in power-law distribution of length may be shocked by the results of this paper. In all the different case studies, the power-law distribution always ranked last or second to last together with the normal distribution. Power-law distributions theoretically describe many natural events, and, in our case, if fractures, for example, are self-similar, then the lengths could show a power-law distribution (Bonnet et al.2001). However, it is shown that, in general, truncated power-law models better fit natural and realistic data; thus usually only the tails of the distribution follow a power law (Clauset et al.2009). This means that, to properly fit a power law, it is necessary to estimate both the scaling parameter (α) and the minimum truncation value (xmin). Truncation is a constantly present bias caused by the resolution limits of the data source above or under which no data can be acquired (right and left truncation respectively). In our case, left truncation (i.e. xmin) is the most common type, and it is caused by the pixel resolution of the images under which fracture traces cannot be seen. Due to this limitation, for a constant-resolution scale, the modelled length distribution will always have an underestimation of the frequencies for the shorter lengths; thus usually only the tail of the distribution follows a power law. Adding to this bias is also the human error caused by countless different factors (one of which is the inevitable boredom of digitizing an outcrop). If the length distribution truly follows truncated power laws, then shorter fractures near the truncation limit would be much more frequent but harder to spot. Moreover, for some fracture systems, the smaller fractures are more prone to be mineral-filled and potentially less obvious features on images. This usually results in fractures that are left uninterpreted even if visible, thus leading to an increase in the real xmin value. Truncation bias and fitting power laws are a very active field of research still with no definitive solution (Clauset et al.2009; Deluca and Corral2013); thus, the implementation, testing, and discussion of such solutions would lie outside this work. For the same reason, a proper estimation procedure could not be included, and the results obtained in this work relative to the power-law model must be taken with a grain of salt. Nonetheless, we still wanted to leave the estimation results and discussion of power laws to again show, as Healy et al. (2017) and Rizzo et al. (2017) did, how this model cannot be blindly applied to geological data without carrying out necessary important considerations.

Finally, censoring is one of many biases that influence the length measurements, and the presented work only treats this specific bias. Because of this, if other biases are present, they will also be carried out after the correction. Moreover, the underlying statistical model between different sets can be different; each fracture set has its own set of biases that influence measurements; and variation in the statistics can also occur within the same set, for example, in proximity of a fault or local changes in rheology. Because of all these reasons, grouping all entities in the same statistical model without any kind of consideration leads to inevitable misinterpretation of the statistics and provides a wrong statistical parametrization of the whole network; thus the analysis of a fractured rock system must be carried out in a homogeneous domain or different stationary domains must be defined before any further analysis is carried out (Bistacchi et al.2020). By applying this divide-and-conquer approach, a more precise and robust characterization is assured because statistical models and inevitable biases related to different families will not mix. These observations further highlight the crucial importance of field data acquisition and geological reasoning to avoid blindly applying these methods to outcrops and making severe inferential errors.

6 Conclusions

In this paper, the effects of censoring bias on the estimation of statistical length models are delineated and discussed. In particular, we demonstrated that censoring bias leads to a general underestimation of any distribution. The typical approaches found in the literature are not suitable, and we have shown that using survival analysis is also the better way to treat censoring in length datasets. In particular, survival analysis offers a valid alternative to the popular circular scanline method. Firstly, survival analysis methods are based on the relationship between the event of interest and the censoring mechanism. Without any underlying geometrical assumptions (basis of the circular scanline method), this methodology is quite flexible and can easily be applied to any geological case study. Secondly, the parameters that are estimated with survival analysis are directly based on the measured length (i.e. the quantity that we want to model) and are always associated with a model; thus they have a higher statistical significance of the scanline method. Regarding the other parametrical methods (i.e. ignoring censoring and removing censored data), we have shown that the censoring percentage heavily influences the estimation quality. In particular, with only 8.96 % censoring, the two classical methods underestimate the mean and variance, and the increase in censoring percentage positively correlates with such underestimation. Also, survival analysis is visibly impacted by the increase in censoring; however, its output always remains stable around the natural model (i.e. the reference diagonal). Nonetheless, the influence of censoring percentage is not yet fully constrained, since, in this study, the visible variation is probably also given by different datasets, with different genetic factors. Thus, a more robust and in-depth analysis must be carried out to fully understand this relationship. The proposed combination of PIT plots and distance calculations demonstrates an effective approach to quantitatively rank a list of length distribution models. The workflow has the objective of comparing sensible (simple) models useful to practical applications (such as DFNs) and finding which one best represents the data. The theory and necessary tools to carry out a complete statistical characterization of a fracture network are included in the specifically developed FracAbility Python library (https://github.com/gecos-lab/FracAbility, last access: 8 September 2024), which encapsulates the necessary steps to carry out as easily and quickly as possible the workflow starting from input shapefiles. The methodology and library were successfully tested in three different use cases, the first having almost no censoring bias and the last with more than 50 % of the total measurements being censored. In all cases, except for the last, in which one of the two spacing distributions follows a Weibull distribution, the lognormal model was the most accurate, followed by either gamma, exponential, or Weibull.

Code and data availability

The code and data are available in a GitHub repository (https://github.com/gecos-lab/FracAbility, last access: 8 September 2024; https://doi.org/10.5281/zenodo.14893964, Benedetti2025).

Author contributions

AB first tested and proved the possibility of analysing and modelling censored fracture-length data. GB and SC formulated the initial key ideas and goals of the article. GB, SC, DB, and AB designed the model-testing procedure. GB wrote all the code, tested it, and improved it in collaboration with SC. SC provided the datasets. AB and DB both greatly contributed to defining the final structure and played a major role in the revision process. AB provided the funding used to gather the data, develop the code, and present and publish the results of this work.

Competing interests

The contact author has declared that none of the authors has any competing interests.

Disclaimer

Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors.

Acknowledgements

We thank Christian Albertini, Francesco Bigoni, Fabio La Valle, Sylvain Mayolle, Mattia Martinelli, and Silvia Mittempergher for the stimulating discussions carried out during numerous project meetings. We thank Stephen Laubach, Sarah Weihmann, and David Healy for the stimulating discussions and for providing important suggestions to improve the value of the text. We also thank Eni S.p.A. for the collaboration in our Joint Research Agreement and for funding Gabriele Benedetti and Stefano Casiraghi's scholarships. We warmly thank the Soprintendenza Archeologia, belle arti e paesaggio per la città metropolitana di Bari; the Parco Nazionale dell'Alta Murgia; and the Ufficio cultura e turismo del comune di Altamura for their cooperation and for authorizing access to the Cava Pontrelli site. We finally acknowledge support of an early stage of our project by the Italian–Swiss Interreg RESERVAQUA project (grant no. 551749).

Review statement

This paper was edited by Stefano Tavani and reviewed by Sarah Weihmann and David Healy.

References

Akaike, H.: A New Look at the Statistical Model Identification, IEEE T. Automat. Contr., 19, 716–723, https://doi.org/10.1109/TAC.1974.1100705, 1974. a, b

Anderson, T. W. and Darling, D. A.: A Test of Goodness of Fit, J. Am. Stat. A., 49, 765–769, https://doi.org/10.1080/01621459.1954.10501232, 1954. a, b

Andersson, J., Shapiro, A. M., and Bear, J.: A Stochastic Model of a Fractured Rock Conditioned by Measured Information, Water Resour. Res., 20, 79–88, https://doi.org/10.1029/WR020i001p00079, 1984. a

Baecher, G. B.: Progressively Censored Sampling of Rock Joint Traces, J. Int. Ass. Math. Geol., 12, 33–40, https://doi.org/10.1007/BF01039902, 1980. a, b

Baecher, G. B.: Statistical Analysis of Rock Mass Fracturing, J. Int. Ass. Math. Geol., 15, 329–348, https://doi.org/10.1007/BF01036074, 1983. a

Baecher, G. B. and Lanney, N. A.: Trace Length Biases In Joint Surveys, in: 19th U.S. Symposium on Rock Mechanics (USRMS), 1–3 May 1978, Reno, Nevada, OnePetro, 1978. a, b, c

Gabriele Benedetti: gecos-lab/FracAbility: FracAbility: Python toolbox to analyse fracture networks for digitalized rock outcrops (v1.5.3), Zenodo [code and data set], https://doi.org/10.5281/zenodo.14893964, 2025. a

Bistacchi, A. and Massironi, M.: Post-Nappe Brittle Tectonics and Kinematic Evolution of the North-Western Alps: An Integrated Approach, Tectonophysics, 327, 267–292, https://doi.org/10.1016/S0040-1951(00)00206-7, 2000. a, b

Bistacchi, A., Eva, E., Massironi, M., and Solarino, S.: Miocene to Present Kinematics of the NW-Alps: Evidences from Remote Sensing, Structural Analysis, Seismotectonics and Thermochronology, J. Geodynam., 30, 205–228, https://doi.org/10.1016/S0264-3707(99)00034-4, 2000. a

Bistacchi, A., Dal Piaz, G., Massironi, M., Zattin, M., and Balestrieri, M.: The Aosta–Ranzola Extensional Fault System and Oligocene–Present Evolution of the Austroalpine–Penninic Wedge in the Northwestern Alps, Int. J. Earth Sci., 90, 654–667, https://doi.org/10.1007/s005310000178, 2001. a

Bistacchi, A., Griffith, W. A., Smith, S. A. F., Di Toro, G., Jones, R., and Nielsen, S.: Fault Roughness at Seismogenic Depths from LIDAR and Photogrammetric Analysis, Pure Appl. Geophys., 168, 2345–2363, https://doi.org/10.1007/s00024-011-0301-7, 2011. a

Bistacchi, A., Balsamo, F., Storti, F., Mozafari, M., Swennen, R., Solum, J., Tueckmantel, C., and Taberner, C.: Photogrammetric Digital Outcrop Reconstruction, Visualization with Textured Surfaces, and Three-Dimensional Structural Analysis and Modeling: Innovative Methodologies Applied to Fault-Related Dolomitization (Vajont Limestone, Southern Alps, Italy), Geosphere, 11, 2031–2048, https://doi.org/10.1130/GES01005.1, 2015. a

Bistacchi, A., Mittempergher, S., Martinelli, M., and Storti, F.: On a new robust workflow for the statistical and spatial analysis of fracture data collected with scanlines (or the importance of stationarity), Solid Earth, 11, 2535–2547, https://doi.org/10.5194/se-11-2535-2020, 2020. a, b, c, d, e, f

Bistacchi, A., Massironi, M., and Viseur, S.: 3D Digital Geological Models: From Terrestrial Outcrops to Planetary Surfaces, in: 3D Digital Geological Models, Chap. 1, 1–9, John Wiley & Sons, Ltd, ISBN 978-1-119-31392-2, https://doi.org/10.1002/9781119313922.ch1, 2022. a

Bonnet, E., Bour, O., Odling, N. E., Davy, P., Main, I., Cowie, P., and Berkowitz, B.: Scaling of Fracture Systems in Geological Media, Rev. Geophys., 39, 347–383, https://doi.org/10.1029/1999RG000074, 2001. a, b, c

Brown, S. R. and Bruhn, R. L.: Fluid Permeability of Deformable Fracture Networks, J. Geophys. Res.-Sol. Ea., 103, 2489–2500, https://doi.org/10.1029/97JB03113, 1998. a

Burnham, K. P. and Anderson, D. R. (Eds.): Model Selection and Multimodel Inference, Springer, New York, NY, ISBN 978-0-387-95364-9, https://doi.org/10.1007/b97636, 2002. a

Burnham, K. P. and Anderson, D. R.: Multimodel Inference: Understanding AIC and BIC in Model Selection, Sociol. Method. Res., 33, 261–304, https://doi.org/10.1177/0049124104268644, 2004. a, b

Cacas, M. C., Ledoux, E., de Marsily, G., Tillie, B., Barbreau, A., Durand, E., Feuga, B., and Peaudecerf, P.: Modeling Fracture Flow with a Stochastic Discrete Fracture Network: Calibration and Validation: 1. The Flow Model, Water Resour. Res., 26, 479–489, https://doi.org/10.1029/WR026i003p00479, 1990. a, b

Clauset, A., Shalizi, C. R., and Newman, M. E. J.: Power-Law Distributions in Empirical Data, SIAM Review, 51, 661–703, https://doi.org/10.1137/070710111, 2009. a, b

Cox, D. R.: Analysis of Survival Data, Chapman and Hall/CRC, New York, ISBN 978-1-315-13743-8, https://doi.org/10.1201/9781315137438, 2017. a

Dal Piaz, G. V. and Lombardo, B.: Early Alpine Eclogite Metamorphism in the Penninic Monte Rosa-Gran Paradiso Basement Nappes of the Northwestern Alps, in: Geological Society of America Memoirs, 164, 249–266, Geological Society of America, ISBN 978-0-8137-1164-5, https://doi.org/10.1130/MEM164-p249, 1986. a

Dal Piaz, G. V., Bistacchi, A., and Massironi, M.: Geological Outline of the Alps, Episodes, 26, 175–180, https://doi.org/10.18814/epiiugs/2003/v26i3/004, 2003. a

Davis, G. H., Reynolds, S. J., and Kluth, C. F.: Structural Geology of Rocks and Regions, Wiley, Hoboken, NJ, 3rd edn., ISBN 978-0-471-15231-6, 2012. a, b

Davy, P., Darcel, C., Le Goc, R., Munier, R., Selroos, J.-O., and Mas Ivars, D.: DFN, Why, How and What For, Concepts, Theories and Issues, in: 2nd International Discrete Fracture Network Engineering Conference, 20–22 June 2018, Seattle, Washington, USA, OnePetro, 2018. a

Deluca, A. and Corral, Á.: Fitting and Goodness-of-Fit Test of Non-Truncated and Truncated Power-Law Distributions, Acta Geophys., 61, 1351–1394, https://doi.org/10.2478/s11600-013-0154-9, 2013. a

Dershowitz, W. and Herda, H.: Interpretation of Fracture Spacing and Intensity, in: The 33rd US Symposium on Rock Mechanics (USRMS), 3–5 June 1992, Santa Fe, New Mexico, OnePetro, 1992. a

Dershowitz, W., Hurley, N., and Been, K.: Stochastic Discrete Fracture Modelling of Heterogeneous and Fractured Reservoirs, in: ECMOR III – 3rd European Conference on the Mathematics of Oil Recovery, cp, European Association of Geoscientists & Engineers, ISBN 978-90-6275-785-5, https://doi.org/10.3997/2214-4609.201411069, 1992. a

De Toffoli, B., Massironi, M., Mazzarini, F., and Bistacchi, A.: Rheological and Mechanical Layering of the Crust Underneath Thumbprint Terrains in Arcadia Planitia, Mars, J. Geophys. Res.-Planets, 126, e2021JE007007, https://doi.org/10.1029/2021JE007007, 2021. a

Enders, C. K.: Direct Maximum Likelihood Estimation, in: Encyclopedia of Statistics in Behavioral Science, edited by: Everitt, B. S. and Howell, D. C., Wiley, 1st edn., ISBN 978-0-470-86080-9 978-0-470-01319-9, https://doi.org/10.1002/0470013192.bsa174, 2005. a

Eppes, M. C., Rinehart, A., Aldred, J., Berberich, S., Dahlquist, M. P., Evans, S. G., Keanini, R., Laubach, S. E., Moser, F., Morovati, M., Porson, S., Rasmussen, M., and Shaanan, U.: Introducing standardized field methods for fracture-focused surface process research, Earth Surf. Dynam., 12, 35–66, https://doi.org/10.5194/esurf-12-35-2024, 2024. a

Fisher, R. A.: Statistical Methods, Experimental Design, and Scientific Inference, Oxford University Press, Oxford, England, New York, ISBN 978-0-19-852229-4, 1990. a

Forstner, S. R. and Laubach, S. E.: Scale-Dependent Fracture Networks, J. Struct. Geol., 165, 104748, https://doi.org/10.1016/j.jsg.2022.104748, 2022. a, b

Gueguen, Y., David, C., and Gavrilenko, P.: Percolation Networks and Fluid Transport in the Crust, Geophys. Res. Lett., 18, 931–934, https://doi.org/10.1029/91GL00951, 1991. a

Healy, D., Rizzo, R. E., Cornwell, D. G., Farrell, N. J., Watkins, H., Timms, N. E., Gomez-Rivas, E., and Smith, M.: FracPaQ: A MATLAB™ Toolbox for the Quantification of Fracture Patterns, J. Struct. Geol., 95, 1–16, https://doi.org/10.1016/j.jsg.2016.12.003, 2017. a, b, c, d, e

Hoek, E.: Strength of Jointed Rock Masses, Géotechnique, 33, 187–223, https://doi.org/10.1680/geot.1983.33.3.187, 1983. a

Hyman, J. D., Karra, S., Makedonska, N., Gable, C. W., Painter, S. L., and Viswanathan, H. S.: dfnWorks: A Discrete Fracture Network Framework for Modeling Subsurface Flow and Transport, Comput. Geosci., 84, 10–19, https://doi.org/10.1016/j.cageo.2015.08.001, 2015. a, b

Kalbfleisch, J. D. and Prentice, R. L.: The Statistical Analysis of Failure Time Data, Wiley Series in Probability and Statistics, J. Wiley, Hoboken, N.J, 2nd edn., ISBN 978-0-471-36357-6, 2002. a, b, c, d, e, f

Kaplan, E. L. and Meier, P.: Nonparametric Estimation from Incomplete Observations, J. Am. Stat. A., 53, 457–481, https://doi.org/10.2307/2281868, 1958. a, b, c, d

Karim, M. R. and Islam, M. A.: Reliability and Survival Analysis, Springer, Singapore, ISBN 978-981-13-9775-2 978-981-13-9776-9, https://doi.org/10.1007/978-981-13-9776-9, 2019. a, b, c, d

Kim, N.: Tests Based on EDF Statistics for Randomly Censored Normal Distributions When Parameters Are Unknown, Communications for Statistical Applications and Methods, 26, 431–443, https://doi.org/10.29220/CSAM.2019.26.5.431, 2019. a, b, c, d

Kleinbaum, D. G. and Klein, M.: Survival Analysis: A Self-Learning Text, Statistics for Biology and Health, Springer New York, New York, NY, ISBN 978-1-4419-6645-2 978-1-4419-6646-9, https://doi.org/10.1007/978-1-4419-6646-9, 2012. a

Kolmogorov, A. N.: Sulla Determinazione Empirica Di Una Legge Di Distribuzione, Giornale dell'istituto italiano degli attuari, 4, 89–91, 1933. a, b

Koziol, J. A. and Green, S. B.: A Cramér-von Mises Statistic for Randomly Censored Data, Biometrika, 63, 465–474, https://doi.org/10.1093/biomet/63.3.465, 1976. a, b, c

Laubach, S. E., Lander, R. H., Criscenti, L. J., Anovitz, L. M., Urai, J. L., Pollyea, R. M., Hooker, J. N., Narr, W., Evans, M. A., Kerisit, S. N., Olson, J. E., Dewers, T., Fisher, D., Bodnar, R., Evans, B., Dove, P., Bonnell, L. M., Marder, M. P., and Pyrak-Nolte, L.: The Role of Chemistry in Fracture Pattern Development and Opportunities to Advance Interpretations of Geological Materials, Rev. Geophys., 57, 1065–1111, https://doi.org/10.1029/2019RG000671, 2019. a

Lawless, J. F.: Statistical Models and Methods for Lifetime Data, Wiley Series in Probability and Statistics, Wiley-Interscience, Hoboken, N.J, 2nd edn., ISBN 978-0-471-37215-8, 2003. a, b, c, d

Leung, K.-M., Elashoff, R. M., and Afifi, A. A.: Censoring Issues in Survival Analysis, Annu. Rev. Publ. Health, 18, 83–104, 1997. a, b

Manzocchi, T.: The Connectivity of Two-Dimensional Networks of Spatially Correlated Fractures, Water Resour. Res., 38, 1–1–1–20, https://doi.org/10.1029/2000WR000180, 2002. a, b

Marrett, R., Gale, J. F. W., Gómez, L. A., and Laubach, S. E.: Correlation Analysis of Fracture Arrangement in Space, J. Struct. Geol., 108, 16–33, https://doi.org/10.1016/j.jsg.2017.06.012, 2018. a, b

Martinelli, M., Bistacchi, A., Mittempergher, S., Bonneau, F., Balsamo, F., Caumon, G., and Meda, M.: Damage Zone Characterization Combining Scan-Line and Scan-Area Analysis on a Km-Scale Digital Outcrop Model: The Qala Fault (Gozo), J. Struct. Geol., 140, 104144, https://doi.org/10.1016/j.jsg.2020.104144, 2020. a, b, c

Mauldon, M.: Estimating Mean Fracture Trace Length and Density from Observations in Convex Windows, Rock Mech. Rock Eng., 31, 201–216, https://doi.org/10.1007/s006030050021, 1998. a

Mauldon, M., Dunne, W., and Rohrbaugh, M.: Circular Scanlines and Circular Windows: New Tools for Characterizing the Geometry of Fracture Traces, J. Struct. Geol., 23, 247–258, https://doi.org/10.1016/S0191-8141(00)00094-8, 2001. a, b, c, d

Mittempergher, S. and Bistacchi, A.: Image Analysis Algorithms for Semiautomatic Lineament Detection in Geological Outcrops, in: 3D Digital Geological Models, Chap. 6, 93–107, John Wiley & Sons, Ltd, ISBN 978-1-119-31392-2, https://doi.org/10.1002/9781119313922.ch6, 2022. a

Nicosia, U., Marinotti, N., and Sacchi, E.: The Late Cretaceous Dinosaur Tracksite near Altamura (Bari, Southern Italy), Geologica Romana, 35, 237–247, 1999. a

Nyberg, B., Nixon, C. W., and Sanderson, D. J.: NetworkGT: A GIS Tool for Geometric and Topological Analysis of Two-Dimensional Fracture Networks, Geosphere, 14, 1618–1634, https://doi.org/10.1130/GES01595.1, 2018. a, b

Oda, M.: A Method for Evaluating the Effect of Crack Geometry on the Mechanical Behavior of Cracked Rock Masses, Mech. Mater., 2, 163–171, https://doi.org/10.1016/0167-6636(83)90035-2, 1983. a

Pahl, P. J.: Estimating the Mean Length of Discontinuity Traces, Int. J. Rock Mech. Min., 18, 221–228, https://doi.org/10.1016/0148-9062(81)90976-1, 1981. a

Patacca, E. and Scandone, P.: Geology of the Southern Apennines, Bollettino della Società Geologica Italiana, 7, 75–119, 2007. a

Philip, Z. G., Jennings, J. W., Olson, J. E., Laubach, S. E., and Holder, J.: Modeling Coupled Fracture-Matrix Fluid Flow in Geomechanically Simulated Fracture Networks, SPE Reservoir Evaluation & Engineering, 8, 300–309, https://doi.org/10.2118/77340-PA, 2005. a

Ricchetti, G. and Luperto Sinni, E.: Osservazioni Stratigrafiche e Paleontologiche Preliminari Sugli Strati Con Raadshoovenia Salentina e Murciella Cuvillieri Delle Murge e Della Penisola Salentina, Studi geologici e morfologici sulla regione pugliese, 6, 26, 1979. a

Rizzo, R. E., Healy, D., and De Siena, L.: Benefits of Maximum Likelihood Estimators for Fracture Attribute Analysis: Implications for Permeability and up-Scaling, J. Struct. Geol., 95, 17–31, https://doi.org/10.1016/j.jsg.2016.12.005, 2017. a, b

Rohrbaugh, M., Dunne, W., and Mauldon, M.: Estimating Fracture Trace Intensity, Density, and Mean Length Using Circular Scan Lines and Windows, AAPG Bulletin, 86, 2089–2104, https://doi.org/10.1306/61EEDE0E-173E-11D7-8645000102C1865D, 2002. a

Sanderson, D. J. and Nixon, C. W.: The Use of Topology in Fracture Network Characterization, J. Struct. Geol., 72, 55–66, https://doi.org/10.1016/j.jsg.2015.01.005, 2015. a, b, c, d

Smirnov, N. V.: On the Estimation of the Discrepancy between Empirical Curves of Distribution for Two Independent Samples, Moscow University Mathematics Bulletin, 2, 3–14, 1939. a

Storti, F., Bistacchi, A., Borsani, A., Balsamo, F., Fetter, M., and Ogata, K.: Spatial and Spacing Distribution of Joints at (over-) Saturation in the Turbidite Sandstones of the Marnoso-Arenacea Fm. (Northern Apennines, Italy), J. Struct. Geol., 156, 104551, https://doi.org/10.1016/j.jsg.2022.104551, 2022. a, b, c, d, e

Suzuki, K., Oda, M., Yamazaki, M., and Kuwahara, T.: Permeability Changes in Granite with Crack Growth during Immersion in Hot Water, Int. J. Rock Mech. Min., 35, 907–921, https://doi.org/10.1016/S0148-9062(98)00016-3, 1998. a

Tavakkoli, M., Shaheabadi, A., Khajoee, S., Malakooti, R., and Beidokhti, M. S.: Deterministic versus Stochastic Discrete Fracture Network (DFN) Modeling, Application in a Heterogeneous Naturally Fractured Reservoir, in: SPE Kuwait International Petroleum Conference and Exhibition, vol. All Days of SPE Kuwait International Petroleum Conference and Exhibition, SPE–127 086–MS, https://doi.org/10.2118/127086-MS, 2009.  a

Tavani, S., Corradetti, A., and Billi, A.: High Precision Analysis of an Embryonic Extensional Fault-Related Fold Using 3D Orthorectified Virtual Outcrops: The Viewpoint Importance in Structural Geology, J. Struct. Geol., 86, 200–210, https://doi.org/10.1016/j.jsg.2016.03.009, 2016. a

Terzaghi, R. D.: Sources of Error in Joint Surveys, Géotechnique, 15, 287–304, https://doi.org/10.1680/geot.1965.15.3.287, 1965. a

Thiele, S. T., Grose, L., Samsu, A., Micklethwaite, S., Vollgger, S. A., and Cruden, A. R.: Rapid, semi-automatic fracture and contact mapping for point clouds, images and geophysical data, Solid Earth, 8, 1241–1253, https://doi.org/10.5194/se-8-1241-2017, 2017. a

Twiss, R. J. and Moores, E. M.: Structural Geology, Freeman, New York, 2nd edn., 3. print edn., ISBN 978-0-7167-4951-6, 2007. a

Vai, G. B. and Martini, I. P. (Eds.): Anatomy of an Orogen: The Apennines and Adjacent Mediterranean Basins, Springer Netherlands, Dordrecht, ISBN 978-90-481-4020-6 978-94-015-9829-3, https://doi.org/10.1007/978-94-015-9829-3, 2001. a

Zhang, L. and Einstein, H. H.: Estimating the Mean Trace Length of Rock Discontinuities, Rock Mech. Rock Eng., 31, 217–235, https://doi.org/10.1007/s006030050022, 1998. a

Download
Short summary
At any scale, the limited size of a study area introduces a bias in the interpretation of linear features, defined as right-censoring bias. We show the effects of not considering such bias and apply survival analysis techniques to obtain unbiased estimates of multiple parametrical distributions in three censored length datasets. Finally, we propose a novel approach to select the most representative model from a sensible candidate pool using the probability integral transform technique.
Share