816
Views
2
CrossRef citations to date
0
Altmetric
Articles

Aggregation Methods Using Bathymetry Sources of Differing Subjective Reliabilities for Navigation Mapping

ORCID Icon, ORCID Icon, , ORCID Icon &
Pages 99-128 | Received 14 Sep 2022, Accepted 02 Jan 2023, Published online: 06 Feb 2023

Abstract

When planning for ship navigation or compiling data for a bathymetry map, the navigator or mapper uses many different sources of bathymetry information and navigation hazards. The quality of these sources is inconsistent in general, however, making it challenging to provide a coherent picture for planning. Here, we describe an approach for consistent planning/mapping that uses a combination of soft computing and Bayesian estimation. The case study used to exercise this system involves NOAA Electronic Nautical Charts for an area in the Chesapeake Bay. We first interpolate each set of irregularly spaced soundings to gridded versions of each point-cloud set. Each of these intermediate grids is then aggregated into a fused bathymetric realization using order weighted averaging (OWA) to provide the weights for each source based on their subjective reliabilities. The OWA allows for fusion informed by the user’s subjective risk allowed in the reconstruction of the seafloor surface and provides quantitative methods to generate, use, and record subjective reliability weights. Each sounding point that went into the bathymetry estimate is then categorized as “no-go,” “caution,” or “go” status. Reliability estimates are reused for weighted Bayesian categorization of each output grid cell to compute the navigable surface.

Introduction

Effective and safe ship transit is of great concern to the development of products for navigational decision making. It is well-known that mistakes or imprecision in navigation charts have led to numerous shipwrecks. A review of Canadian, United States, and European shipwreck data (Bakdi et al. Citation2019) reports a total of over 100,000 maritime accidents during various periods over the 2000 to 2019 era. This includes cargo, military, and passenger transport ships. The grounding of the USS Guardian () is one recent example (Walhey Citation2013).

Figure 1. Grounding of the USS Guardian on the Tubbataha Reef (130 km southeast of Palawan, Philippines) in January 2013 (US Navy photo, public domain).

Figure 1. Grounding of the USS Guardian on the Tubbataha Reef (130 km southeast of Palawan, Philippines) in January 2013 (US Navy photo, public domain).

Over 90% of worldwide commerce uses shipping (Naevestad et al. Citation2019), often from ports located on the major rivers of the world. Where major rivers flow into the ocean there are many navigation problems. For example, the Columbia River of the western United States has the Columbia Bar, a shifting sandbar that makes the river’s mouth one of the most hazardous stretches of water to navigate in the world (Saddler Citation2006). Since 1792, approximately 2,000 large ships have sunk in and around the Columbia Bar, and because of the danger as well as the numerous shipwrecks the mouth of the Columbia River, this location has acquired a reputation worldwide as the graveyard of the Pacific (Wilma Citation2006).

The most important task of a ship navigator is to decide how to safely traverse the water surface between start and ending points of a ship’s course. While there are multiple factors related to safe navigation (e.g. avoidance of hazardous weather, areas of civil unrest, etc.), in this article, we use the term “safe navigation” to mean that the vessel does not make contact with the seafloor or underwater obstructions, where the depth of the keel dictates the shallowest depth that it can navigate (Calder Citation2015; Gilardoni and Presedo Citation2017).

An experienced navigator uses multiple sources of information for multicriteria decision making and forms a mental-model of safe navigation areas based on the location of soundings (geospatial points on a chart attributed with water depth) as well as obstruction or “hazard” locations to avoid (Cutler Citation2003). Part of the rationale for this mental processing is that point representations of seafloor depth and hazards on Electronic Navigation Charts (ENC; International Hydrographic Office Citation2000) are sparse geospatially. Soundings are quantitative but often lack information regarding the uncertainty of the depth, such as one standard deviation, an interval, a percentage of water depth or, as defined by the International Hydrographic Organization (IHO), a Category Zone of Confidence (CATZOC) (International Hydrographic Organization Citation2020). Hazards, however, have a qualitative aspect on charts. Cartographers generating the chart provide qualitative identification of the hazard using symbols that map the location on the chart and linguistic labels for each symbol (e.g. “shipwreck,” “sounding doubtful,” “rocks,” “value of sounding” (VALSOU), etc.) to identify the hazard type.

A navigator can judge that information sources have differing degrees of reliability based on the nature of the charts being used, the configuration of the soundings and contours, the known or estimated character of the seafloor, the age of the data, and so forth (Calder Citation2015). Such human-in-the-loop judgment is becoming even more important as community-based bathymetry collection becomes further developed (Calder et al. Citation2020), machine learning-based assessments from imagery become implemented (Chénier et al. Citation2020), source type and technology used to collect the data become attributed to digital bathymetry data as well as subsequent charts (Calder Citation2015, Weatherall et al. Citation2015; International Hydrographic Office Citation2019; EMODnet Bathymetry Consortium Citation2020; Jakobsson et al. Citation2020), and CATZOC attributions are assessed more fully in ENCs (International Hydrographic Organization Citation2020). Thus, the critical aspects of this human-based production of a navigation plan are (a) forming a believable forecast of seafloor depth contours in the areas between the soundings and (b) doing so using optimistic and pessimistic outlooks of what the information means for keel clearance as influenced by information sparsity and reliability. Although automated and objective means for multicriteria route planning and navigation risk-evaluation are now being proposed in the literature (Bakdi et al. Citation2019; Jeong et al. Citation2019; Kastrisios et al. 2020), a human still makes the final judgment for execution of the route and will assess the reliability of all information presented through human-in-the-loop processes.

In this article, we describe a hybrid-framework using both soft computing and quantitative methodologies that provide an approach for navigation planning. Our objective is to develop this framework through application of ordered weighted averaging (OWA) (Yager Citation1988; Beliakov, Pradera, and Calvo Citation2007) and Dirichlet categorical estimation (Gelman et al. Citation2013) subprocesses, respectively. The use of the OWA provides a means for allowing the human-in-the-loop judgment of reliability of the information sources and attitudinal risk tolerance than can be used in the quantitative Bayesian estimate of navigation risk category. This process provides a chart of clearly defined areas for safe or unsafe navigation as well as fuzzy areas in between where it may be safe or unsafe for navigation by mapping areas that are “go,” “no-go,” and “caution” areas, respectively. Another reason for the use of the OWA, a fuzzy logic stage, is significantly reduced computation costs versus a strict Bayesian approach for the steps of the process involving injection of linguistic/subjective/nonnumerical knowledge into the analysis.

Section “Approach” of this article presents the overall approach. The OWA is formally introduced. The section provides simple examples of how to use the OWA with a simple bathymetric test example. The OWA is further developed to show how human assessed reliability can be used for the fusion process and how the reliabilities then are reused in the Bayesian process, respectively. Section “Case study setup” introduces a case study of using this framework with ENC data in the Chesapeake Bay area. An overview of the data and computational implementations are discussed. Section “Results” provides results of the test case for two example keel depths, showing how the reliabilities assessments, attitudinal risk tolerance, and number of points used affect the results. Sections “Discussion” and “Summary” provide discussion and summary, respectively.

Approach

The approach here is to allow for a construction of the seafloor surface from the various sources using the OWA and Bayesian techniques. Specifically, the problem being solved is the inconsistency in charted soundings and information due to differences in compilation and generalization. The OWA is needed to generate “safe” realisations, and then, the Bayesian categorical estimation is used to fuse these into an estimated navigation category. Deeper or shallower estimates of depths can be based on parameters that could be adjusted to reflect the opinion of expert observers, environmental events within the area, the current tactical situation, and other factors. The use of this combined approach enables the use of these human-based and other qualitative factors within a recordable and quantitative framework. This framework also allows for record keeping of these qualitative factors so these judgments are inspectable and computations are reproducible.

Overall, irregularly spaced soundings are used as input, which then undergo aggregation and categorical processing steps; the output is gridded map displaying differing navigation categories (). We use charted soundings from ENC data for this input data. We assume that the typical end-user would obtain such information from ENC products rather than accessing databases with multibeam or single beam soundings for input (although use of such database information is permissible in this framework) due to volume of such data and level of expertise required to process it correctly. Here, only the soundings as they appear in the “SOUNDG” layer in each ENC data product are used (source survey soundings, “hydrographic” scale sounding sets for chart production, or other “submerged” chart features having associated a bathymetric value (e.g. depth contours, underwater rocks) are not used here for simplicity in developing the overall approach). For each set of soundings used in the aggregation process, an intermediate gridded bathymetry surface is generated using an inverse distance weighting interpolation scheme. These intermediate surfaces are then used to construct two reference bathymetry grids used for categorical estimation. The first bathymetry grid is the shallowest (i.e. shoalest) surface that can be generated from the intermediate grids (i.e. at each xy-grid point, the shallowest depth from the intermediate surfaces at this xy-position is assigned to this shallow bathymetry grid). The second surface is a deeper bathymetry that is calculated using the OWA process, controlled by a parameter θ[0,1 that adjusts this bathymetry surface in depth to reflect expert opinion, source metadata, present tactical situation, tidal level, vertical datum reference, and so forth. Once these two surfaces are constructed, the point cloud soundings that went into computing the grid point are categorized as “go,” “caution,” or “no-go” based on keel depth, soundings depths, and the two reference bathymetry grids as follows.

Figure 2. Human-in-the-loop framework presented in this article. Source: Author.

Figure 2. Human-in-the-loop framework presented in this article. Source: Author.
  • “No-go”: the keel depth is either deeper than the (a) sounding depth OR (b) deeper reference depth

  • “Go” if the keel depth is shallower than the (a) sounding depth AND (b) shallow reference depth

  • “Caution” if (a) the keel depth is in between the two reference depths AND (b) the sounding is in between the two surfaces.

The output grid-point for the navigable surface is the maximum a posteriori category from the Bayesian estimates of the three categories (ties go to most pessimistic category). The overall map then displays the “go,” “no-go,” and “caution” categories. The “caution” category forms a safety contour that is a “fuzzy” contour between the “go” and “no-go” regions.

Order weighted averaging

The OWA operation (Yager and Kacprzyk Citation1997) provides quantitative methods to use risk tolerance and subjective weighting of the navigation product by reliability to construct a weighted average of a target variable of interest. Since linguistic quantifiers, CATZOCs, tidal datums, and identification of source types are becoming part of the community of practice (e.g. International Hydrographic Organization Citation2019, ch. 17; EMODnet Bathymetry Consortium Citation2020; International Hydrographic Organization Citation2020; Jakobsson et al. Citation2020; Kastrisios and Ware Citation2022), this information can form subjective judgment on the reliability/trust of the information. These attributions can be mapped by an end-user into reliability metrics used as generating functions for OWA weights.

OWA Definition.

An OWA aggregator operator (Yager Citation1988) of dimension n is the mathematical mapping operation RnR (R  is the set of real numbers) that has an associated n dimensional vector of weights (1) W= [w1wn](1) operating on set A = {a1,,an}  such that: (1) wj[0,1]; and (2) jwj=1. Let λ be an index function on A so that a new reordered version of A = {aλ(1),,aλ(n)} is created. Each element is ordered such that aλ(1)aλ(2)aλ(n) according to some ordering operator that is problem specific (e.g. the “greater than” operator for simple cases). Then, we can define (2) OWA(a1,,an)= j=1nwjaλ(j).(2)

Section “Inclusion of reliability in OWA for safety of navigation” below proposes an ordering scheme based on reliability of each ai.

Common cases

The OWA operator allows for a wide variety of different types of aggregation depending on the choice of the weighting vector W and the ordering operator. Some notable examples of weighting vector and their resulting aggregation are as follows.

  1. W*: w1 = 1 and wj = 0 for j ≠ 1. Thus, OWA(a1, …, an) = aλ(1) = Maxi(ai)

  2. W*: wn = 1 and wj = 0 for jn; OWA(a1, …, an) = aλ(n) = Mini(ai)

  3. Wn: wj = 1/n for all j = 1 to n; OWA(a1, …, an) = 1ni=1nai=mean(a1,,an)

  4. W[K]: wK = 1 and wj = 0 for jK; OWA(a1, …, an) = aλ(K)

We observe that as more weight is assigned to the wj with smaller indices (terms near the top of W), the larger the aggregated value. On the other hand, if more of the weight is assigned to the wj with the larger index (terms near the bottom of W), the smaller the aggregated value. The choice of weighting vector W determines the form of OWA aggregation.

OWA weight generation with dependence on attitudinal character, θ

A class of approaches for obtaining wj’s that can be used for safety of navigation is based on regular increasing monotone quantifiers (Yager, Elmore, and Petry Citation2017). Let f: [0, 1] → [0, 1] be a monotonic function (i.e. f(x)f(y) if xy with f(0) = 0 and f(1) = 1). Using this function, we can obtain the wj for j = 1 to n such that (3) wj=f(jn)f(j1n)(3)

With EquationEq. (3), wj ∈ [0, 1] and j=1nwj=1 as required for the OWA weights.

Following from Yager, Elmore, and Petry (Citation2017), let f(x) be specified by f(x) = xm for m ≥ 0 in EquationEq. (3), with x being abstract for the moment. Then, the “attitudinal character” (to be applied to the bathymetry context below) is defined in Yager, Elmore, and Petry (Citation2017) to be the integral of f(x) from 0 to 1. (4) θ=01xmdx=(m+1)1.(4)

Given EquationEq. (4), define mθ1θ (which is different from Yager, Elmore, and Petry (Citation2017), as our order of the data will be reversed from the application in that paper). With this functional form, the OWA weights for j = 1 to n are (5) wj=f(jn)f(j1n)=(jn)m(j1n)m.(5)

Defining mθ1θ in EquationEq. (5), (6) wj=(jn)θ1θ(j1n)θ1θ.(6)

Hence, for θ=0, m = 0, and the weights are heavier on w1. As θ1, m  , and the weights are heavier on wn. The value θ=0.5 implies m = 1, so wj=1nj, equally weighting all inputs. Consequently, user choice of θ can adjust the reconstruction using the W=[wj] from largest to smallest value, or in bathymetric terms, deepest to shoalest plausible reconstruction.

The parameter θ[0,1] is the subjective attitudinal character the navigator may have in the ambiguity of the “caution” category of safety when processing the ordered set of bathymetry values, ordered from deepest to shallowest. The values θ=0, 0.5, and approximately 1 linguistically mean “completely pessimistic” (i.e. the navigator wants to consider only the deepest values of bathymetry for extra keel clearance and have more “go” categories changed to “caution”), “neutral” (i.e. use a simple weighted average), and “completely optimistic” (i.e. the navigator is willing to route the ship as shallow as possible or may be operating an amphibious craft). Note that complete pessimism forces the OWA to provide the deepest bathymetry (i.e. maximum depth) being considered in the OWA process for additional keel clearance.

Example 1 - OWA by attitudinal character alone

This section and subsequent sections for Examples 2 and 3 provide simple examples for using the OWA and, later, Bayesian categorical estimation. The intent is to provide simple examples that can be worked by hand, spreadsheet program or software for reader intuition. in Appendix A provides graphs of the depth estimates and OWA weights as a function of attitudinal character for the cases discussed.

Consider the following three sets of depth information, z1*=18, z2*=16, z3*=20. One first reorders by depth to get z1=20, z2=18, z3=16. With this ordering, OWA weights for this information for the following three different degrees of attitudinal character.

Case (a) θ = 0.5: Here, mθ/(1θ) = 1, and hence, wj=(jn)1(j1n)1 for each j. Thus, for neutral optimism, the OWA reduces back to the standard arithmetic mean. OWA(z,w;θ=0.5)=13(16+18+20)=18

Case (b) θ = 0.2: Generation of “deeper” bathymetry for θ = 0.2 (recall that θ = 0 yields the maximum value of the set) has m = θ1θ=0.20.8=0.25, so that wj=(jn)0.25(j1n)0.25. For this case,   (jn)0.25(j1n)0.25wjwjzj j=10.7600.7615.2 j=20.900.760.142.6j=310.900.101.5   wj=1¯wjzj=19.3¯

Case (c) θ = 0.8: The “shallower” value of θ = 0.8, (recall that θ -> 1 yields the minimum value of the set) m = θ1θ=0.80.2=4, so that wj=(jn)4(j1n)4. For this case,   (jn)4(j1n)4 wj wjzj j=1 1/810 1/81 0.25j=2 16/811/8115/81 3.33j=31 16/81 65/8112.84   wj=1¯wjzj=16.42¯

Inclusion of reliability in OWA for safety of navigation

Consider the situation in which each jth set of soundings, Zj, has an integer measure of relative reliability; Rj[Rmin, Rmax], Rmin1, Rj Z+; which can be objective, e.g. from a statistical process, or subjective, e.g. from linguistic quantifiers mapped to an interval of integers (use of an Rj=0 would amount to discarding the dataset). A value of Rj = Rmin, could mean “least reliable of sources” (e.g. due to source provider, age of the data, storm, or tsunami events making the information unreliable, etc.), Rmax means “most reliable of sources,” and intermediate values provide degrees of “intermediate reliability.” Importantly, Rj depends on the number and scale of the other reliabilities, so that reliability is in relative terms (i.e. “twice as good as that one”). Setting Rmin=Rmax means that all sources have equal weight.

Following Yager, Elmore, and Petry (Citation2017), the reliability index for an OWA, rj, is then generated from Rj as follows. First, for each Zj, the normalized reliability is computed as rj=Rj/j=1n(Rj). For the application purpose of navigation safety, there will always be an Rj>0 such that jrj=1. Next, define (7) Sji=1jri(7) where Sj is the partial cumulative summation of the normalized reliabilities associated from 1 to j, with SN =1. Since Sj is monotone increasing (and therefore meets the requirements used to create EquationEq. (3)), we can substitute Sj for j/n  in EquationEq. (6) to get (8) wj=(Sj)θ1θ(Sj1)θ1θ(8) where the dagger superscript indicates weighting by reliability. Note that in the special case of all reliabilities being equal, EquationEq. (8) reverts to EquationEq. (6). Example 2 illustrates use of EquationEq. (8) for OWA computations for nonequal degrees of reliability.

Example 2 - OWA by reliabilities

Example 2 repeats Example 1 except now let there be two sets of subcases that have differing reliabilities: (i)  r1=0.3, r2=0.6, r3=0.1; and (ii)  r1=0.1, r2=0.1, r3=0.8. For the special case of r1=r2=r3=1/3, the cases below revert back to their counterparts in Example 1. correspond to these examples

  • Case (a) θ = 0.5: As shown in Example 1(a), m = 1. Thus, in this special case, the OWA revert to standard weighted average, with the weights given by the normalized reliabilities,  rj. (i) OWA(z={20,18,16}|θ=0.5, r={0.3, 0.6, 0.1})=0.3(20)+0.6(18)+0.1(16)=18.4 (ii)  OWA(z={20,18,16}|θ=0.5, r={0.1, 0.1, 0.8})=0.1(20)+0.1(18)+0.8(16)=16.6

  • Case (b): θ = 0.2 (i) θ=0.2,m=0.25, r={0.3, 0.6, 0.1} j=1j=2j=3  zj201816  rj0.30.60.1  Sj0.30.91  Sj100.30.9  wj=Sj0.25Sj10.250.740.230.03wj=1¯wjzj14.84.20.4wjzj=19.4¯ (ii) θ=0.2,m=0.25, r={0.1, 0.1, 0.8} j=1j=2j=3  zj201816  rj0.10.10.8  Sj0.10.21  Sj100.10.2  wj=Sj0.25Sj10.250.560.110.33wj=1¯wjzj11.21.95.3wjzj=18.4¯

  • Case (c): θ=0.8 (i)  θ=0.8,m=4, r={0.3, 0.6, 0.1} j=1j=2j=3  zj201816  rj0.30.60.1  Sj0.30.91  Sj100.30.9  wj=Sj4Sj140.010.650.34wj=1¯wjzj0.211.75.5wjzj=17.4¯ (ii) θ=0.8,m=4, r={0.1, 0.1, 0.8} j=1j=2j=3  zj201816  rj0.10.10.8  Sj0.10.21  Sj100.10.2  wj=Sj4Sj140.00010.00150.9984wj=1¯wjzj0.0020.02715.974wjzj=16.003¯ 

Bayesian component

Expectation values

The OWA analysis allows construction of attitudinally-adjusted surfaces for each source of data, and therefore, classification of source estimates of depth. Bayesian categorical estimation is used to compute the posterior probability of each category of navigational safety label, and hence, fuse all sources of information. Appendix B provides mathematical details for deriving mean estimates discussed in this section.

Categorical estimation problems for multinomial cases are usually based on the use of the Dirichlet distribution as the natural conjugate prior (Gelman et al. Citation2013, ch. 23). Maximum a posteriori probability reconstruction then results in a ternary classification that best represents the available estimate from the irregularly scattered data while still preserving the “caution” area that provides information on the uncertainty of the safety contour’s location. The posterior mean probability for each category; μi, i={1=Go, 2=Caution, 3=NoGo}; is (9) μi=ni+αi,0N+ iαi,0(9) where ni is the number of counts for the ith category, N is the total number of counts, and αi,0 is the parameter that specifies the initial Dirichlet distribution used. For this application, αi,0=1 for every i, which just sets the initial Dirichlet prior to be a uniform distribution across all three categories.

Expectation values with reliability weighting

For the case where each chart is given a different set of reliabilities, Rj,  this weighting amounts to recounting the category Rj times. In other words, the chart with Rj>1 gets Rj “more votes” versus the chart with Rj=1. Note that Rj is being used twice in the entire process for consistent use of reliability for (1) computation of the deep bathymetry surface via OWA and (2) weighted Bayesian category estimation.

With this need to recount, Eq. (9) becomes (10) μi=j=1JRjni,j+j=1JRjαi,ji=1Ij=1JRjni,j+ ij=1JRjαi,j=j=1JRjni,j+αi,0i=1Ij=1JRjni,j+ iαi,0(10) with αi,jαi,0/JRJ. This choice of αi,j enables setting the initial probability density function (pdf) based on knowledge of the reliabilities. In the special case of Rj=1  j, Eq. (10) reduces to Eq. (9). Equation (10) provides the reuse of these reliabilities that are used for the OWA weight generation for the bathymetry surfaces for weighted categorical estimation. Example 3 provides examples for using Eq. (10).

Example 3 - Bayesian categorical estimation

Let 8 nearest neighbor points to an output grid point be classified by the categories {“go,” “caution,” “no-go”} for each chart. Let the data of the category counts for the charts be as follows.

  • Case (a): R = {1, 1, 1}. For this case, N=24. μgo=1·3+1·6+1·5+124+ 3=0.56 μcaution=1·4+1·1+1·1+124+ 3=0.26 μnogo=1·1+1·1+1·2+124+ 3=0.18

  • Case (b): R = {1, 5, 10}. For this case, N=16*24=384. μgo=1·3+5·6+10·5+1384+ 3=0.64 μcaution=1·4+5·1+10·1+1384+ 3=0.15 μnogo=1·1+5·1+10·2+1384+ 3=0.21

In both cases, the output point would be classified as “go”; however, in the second case, more weight is given to “no-go” versus “caution.”

Overall process flow

The overall process is as follows.

  1. First, each irregular point cloud of soundings is interpolated to an output grid. For every output grid point, a polar-coordinate system with the output grid point being at the origin is first set-up. The polar coordinates of each irregularly spaced sounding relative to this origin are then computed.

  2. Soundings are then radially binned into 45-degree sectors. Soundings in each sector are then sorted by radial distances to find the nearest neighbors in each sector to the output grid point; radial coordinates of these points were stored for reuse during the categorization process.

  3. The bathymetry at the output grid point is then computed using an inverse distance weighted average of these 8*K nearest neighbor points, with K = 1 (higher values of K will be shown later along with discussion of why we choose K = 1). To avoid division by zero or bias by a heavily weighted point, radial distances have a nominal uncertainty value added of 5 m.

  4. Once the gridded bathymetries for each chart type are created, the bathymetry information is interpolated to a set of two grids for the OWA parameters θ={θmin, 1}, 0θmin<1, using reliability of each irregularly spaced input dataset for weight generation as shown in EquationEq. (8).

  5. For each output grid point, the irregularly spaced data point that went into estimation of bathymetry at that grid point is then assigned a category of 1, 2, or 3 for “Go”/”Caution”/”No-Go,” respectively. Counts are replicated for datasets that have higher reliabilities as discussed in Appendix B (Derivation of Eq. (10)).

  6. EquationEquation (10) is then used to compute the posterior probability density function (with αi=1 for i={1, 2, 3}) at each output grid point.

  7. The category for the maximum from the a posteriori pdf is assigned to the output point for the categorical surface, with any ties resolved in favor of the most pessimistic category.

Case study setup

Case study area

A test area was identified based on NOAA electronic navigation charts (ENCs) of the Norfolk, Virginia area (National Oceanic and Atmospheric Administration – Office of Coastal Survey Citation2022) using QGIS Version 3.12 (QGIS Citation2020). The ENC used to obtain the point cloud data were US3EC08M, US4VA12M and US5VA13M. In addition, charts US2EC03M, US3DE01M, US4NC32M, and US5VA19M were used to complete the scene shown in . The S-57 format chart data’s SOUNDG field were exported as ASCII files of position and depth for subsequent processing.

Figure 3. Case study data and area used as displayed in QGIS software. Black dots show soundings from the SOUNDG field in the ENC files, dark yellow signifies land, and light blue signifies potentially shallow water. Source: Author.

Figure 3. Case study data and area used as displayed in QGIS software. Black dots show soundings from the SOUNDG field in the ENC files, dark yellow signifies land, and light blue signifies potentially shallow water. Source: Author.

ENC data are provided at a scale appropriate for use, which affects the sounding density. Bands 3, 4, and 5 (“US3x,” “US4x,” and “US5x” here) are, respectively, “Coastal” (lowest density), “Approach,” and “Harbor” (highest density) charts. To simulate the effects of having data in just one of these categories, data from all source charts were exported for “Coastal” (Band 3), “Approach” (Bands 3 and 4), and “Harbor” (Bands 3, 4, and 5) use in the same area (). Candidate test areas along both the eastern portion of the dense polygon and the northern portion were considered; the northern portion within the Chesapeake Bay was chosen as it contains a dredged shipping channel that provides diversity in the sounding depths for numerical experiments.

Figure 4. Sounding for the (a) Coastal chart, (b) Approach chart, and (c) Harbor Chart. Panels (d–f) are the gridded bathymetry for these cases, respectively.

Figure 4. Sounding for the (a) Coastal chart, (b) Approach chart, and (c) Harbor Chart. Panels (d–f) are the gridded bathymetry for these cases, respectively.

Numerical implementations

The point cloud data require processing to the common grid for bathymetry and (subsequent) navigable surfaces. The gridding is a two-step process, as outlined in Appendix C. Implementation was done in MATLAB (Mathworks Citation2021). Algorithm 1 first augments the xyz point-clouds with additional nodes for stability in subsequent gridding processes in Algorithm 2. Specifically, Algorithm 1 augments the source point cloud data by using (a) Delaunay triangulation routines (Preparata and Shamos Citation1988) as implemented by MATLAB software and (b) linear interpolation within each triangle to increase the number of points in each irregularly spaced dataset.

Algorithm 2 then (a) grids bathymetry for each of the three datasets, (b) fuses these charts into the shallow and deeper fused surfaces using the OWA for the specific value of θ0, 1 of interest for the later surface, and (c) computes the navigation category in accord with the process flow given in Section “Overall” process flow. Note from that for edge and corner points, data still exists outside of the boundary. This process is repeated for each chart so that each xyz point-cloud set has a corresponding gridded bathymetry surface. f show the gridded surfaces of the –c, respectively, for K = 1.

The OWA is then used to obtain the shallow and deep surfaces from these gridded bathymetry surfaces. The shallow reference surface is created by computing the grid for θ=0.99. Deep bathymetry surfaces are then computed for each of differing values of θ in the numerical experiments. Soundings used to grid each grid point of the charts are retrieved to compute the navigation category at each grid point. To account for the subjective reliabilities in the Bayesian classifier, the set of reliabilities are divided by the smallest nonzero reliability and then rounded to the nearest integer. This set of integers provides the number of repetitions each irregularly spaced point used in EquationEq. (10) to provide the final posterior distribution for that point. Grids for both bathymetry and navigation classes are output from the algorithm.

Results

–c shows the irregularly spaced point cloud data from the ENCs as discussed in Section “Case study area”. For the gridded surfaces, the nearest neighbor in each 45-degree bin was used (i.e. K = 1), so that eight points were used in total for computing each grid point using inverse distance weighting. The corresponding gridded bathymetry surfaces are mapped below each point cloud set in . The unnormalized reliabilities for grids corresponding to the Coastal, Approach, and Harbor were subjectively set to R={1, 5, 10}, respectively, based on the qualitative density of points between the datasets. Hence, the normalized reliabilities were r={1/16, 5/16, 10/16}. and show the shallow and deep surfaces bathymetry surfaces for the case of θ=0.1. shows the gridded difference between these two surfaces.

Figure 5. Panels (a, b) are aggregated bathymetry grids for the (a) shallow bathymetry surface formed by setting θ=0.99 in the OWA process and (b) deeper bathymetry surface corresponding to θ=0.1 Panel (c) maps the difference between these bathymetry grids.

Figure 5. Panels (a, b) are aggregated bathymetry grids for the (a) shallow bathymetry surface formed by setting θ=0.99 in the OWA process and (b) deeper bathymetry surface corresponding to θ=0.1 Panel (c) maps the difference between these bathymetry grids.

In through , two keel depth cases will be discussed: 2.3 m (7.5 feet) and 3.05 m (10 feet). Due to the depth ranges that are present, these cases enable visualization of different effects on the navigation surface with respect to reliability weightings, number of points used for the inverse distance weighted averaging process, keel depth, and attitudinal character.

and shows the navigable surfaces for θ=0.1 resulting for (a) R={1, 1, 1}, i.e. each surface is equally weighted (so that r={1/3, 1/3, 1/3}), and (b) for R={1, 5, 10} (so that  r={1/16, 5/16, 10/16}). The yellow circle indicates a shoal feature that is lost by the equally weighted case. This shoal surface is retained in by using higher reliabilities for denser datasets.

Figure 6. Navigation surface using the OWA for unnormalized reliability weightings of (a) R={1, 1, 1} and (b) R={1, 5, 10} as discussed in the text. Source: Author.

Figure 6. Navigation surface using the OWA for unnormalized reliability weightings of (a) R={1, 1, 1} and (b) R={1, 5, 10} as discussed in the text. Source: Author.

Continuing to use θ=0.1 and R={1, 5, 10}, the results of using K = {1, 3, 5, 10} are illustrated in for a 2.3 m (7.5 foot) keel depth. The shoal area in the southeastern portion of the channel no longer maintains the “no-go” category as more nearest neighbors are counted for the navigation category. Using more points leads to smoothing away of the “caution” category for one of the cases. show these same results for the 3.05 m (10 foot) keel depth. The shoal area exhibits the same patterns as the 2.3 m keel case. In the northwest and southeast corners, shallower points that are retained in the irregularly scattered data have an effect on changing the “go” category to “caution.”

Figure 7. Navigation surface as nearest neighbors used vary from (a) K = 1, (b) K = 3, (c) K = 5, and (d) K = 10 for the 7.5-foot keel clearance case, θ=0.1 and R={1, 5, 10}. Source: Author.

Figure 7. Navigation surface as nearest neighbors used vary from (a) K = 1, (b) K = 3, (c) K = 5, and (d) K = 10 for the 7.5-foot keel clearance case, θ=0.1 and R={1, 5, 10}. Source: Author.

Figure 8. Same as , but for the 10-foot keel clearance case. Source: Author.

Figure 8. Same as Figure 7, but for the 10-foot keel clearance case. Source: Author.

Retaining K = 1, shows the navigation categories for θ={0.3, 0.5, 0.7} for the two keel depths of interest. As is observable, the categories change from “caution” to “no-go” with increasing θ for the keel depth of 2.3 m. For the 3.05 m depth, these effects are not as drastic. The shoal at (37.04 N, −76.09E) becomes a “no-go” area, along with more of areas on either side of the channel.

Figure 9. Navigation surface for θ={0.3, 0.5, 0.7} for the 7.5-foot keel clearance case in subfigures a–c, respectively, and for the 10-foot keel clearance case in subfigures d–f. Source: Author.

Figure 9. Navigation surface for θ={0.3, 0.5, 0.7} for the 7.5-foot keel clearance case in subfigures a–c, respectively, and for the 10-foot keel clearance case in subfigures d–f. Source: Author.

shows the bathymetry along the transect line along −76.14E illustrated in the gridded surfaces for K = 1 and R={1, 5, 10}. also shows the keel depths of 2.3 m and 3.05 m and the corresponding navigation categories for the 2.3 m keel depth (the navigation category changes for the 3.05 m keel depth simply change from “no-go” to “go” after crossing 37.09 N from the south to the north). Consistent with the graphs in Appendix A and in and , plots the 2D effect of adjusting the deep surface for θ={0.1, 0.3, 0.5, 0.7,0.9}. This transect shows that higher values of the attitudinal character leads to a shallowing for the deeper surface. We also inspected the difference surfaces as displayed in for higher values of the attitudinal character; the maximum of the difference between shallow and deep surfaces (given by the upper range of the colorbar) decreased with increasing θ.

Figure 10. Transect lines for the vertical line shown in the previous figures for θ={0.1, 0.3, 0.5, 0.7, 0.9}.

Figure 10. Transect lines for the vertical line shown in the previous figures for θ={0.1, 0.3, 0.5, 0.7, 0.9}.

Discussion

The key point in creating the navigable surface is that using lower values of θ for the deep bathymetry surface results in more “go” categories becoming “caution” categories along the edges of the channel. The deeper that surface is to the reference surface, the more the navigator is hedging for higher safety by putting more of the “go” categories into “caution.” The trade off, it seems, may be false alarms of “caution.” This classification trend of the shoal area in the southern portion of the scene does not occur here until the attitudinal character gets closer to one. This trend indicates that human inspection still is required at this location; such false alarms of “caution” could make the transit longer or impossible (e.g. closed channels) if accepted without inspection.

The concept of optimistic and pessimistic reconstructions in each area directly leads to a model of uncertainty for safety. It is clear that in the region shoalward of the deeper reconstruction, it will be unsafe to travel. In the region deepward of the reconstruction, it will be safe to travel. In between these regions there is a “caution” area. The size of the “caution” area reflects the attitudinal character of risk acceptance in the construction of the navigation surface.

Different categories could be developed that involve the use of linguistic quantifiers for a judgment on the reliability/trust of the information used for assigning R values. These quantifiers could be standardized into a defined set of linguistic terms or identifiers (i.e. terms placed in a look-up table, e.g.). Examples are type identifications used by GEBCO for the GEBCO grid (International Hydrographic Organization Citation2019) as mentioned in Section “Order weighted averaging” or CATZOCs. The terms then would be translated into R values that are then used by the weight generating function and Bayesian categorical estimator.

The derivation of relative reliability estimates and attitudinal character are subjective, and may develop over time, or through the current tactical situation. By their nature, it is reasonable to assume that reliability estimates will be more stable between observers, however. For example, it seems that a most observers would agree that more dense sources and more recent surveys are likely to be more reliable. Thus within the algorithm, R values would reflect the change of horizontal precision due to cartographic generalization, with higher R values assigned to ENCs with large scale (e.g. Harbor charts should have higher R than Coastal charts) or with more recent survey information.

It also is likely that reliabilities will need to vary spatially, rather than have a single fixed set of values as used here for simplicity, primarily to take into account the large variability in quality in some of the existing ENCs (as depicted by the presence of ENC areas with different CATZOC). Such values might therefore be recommended by a hydrographic agency, or through international agreement and change rarely; recommended values associated with charted data by the hydrographic office might eventually be possible as an adjunct, or alternative, to CATZOC or other data quality measures (Calder et al. Citation2020).

The attitudinal character, however, is likely to be much more variable since it depends strongly on the navigator’s attitude to risk, which is likely to be affected by a large number of own-ship factors. For example, higher risk might be acceptable for a bulk carrier than a liquified natural gas (LNG) tanker (Calder Citation2015), although the total cost of an incident might be higher for a large crude oil tanker than for an LNG ship (i.e. the clean-up for an oil incident is significantly higher than for natural gas). Methods developed to elicit level of risk acceptance in operational research (Raiffa and Schlaifer Citation2000) might be used to assess appropriate values based on ideas of risk, total cost, and utility.

The Concept of Operations (CONOP) for the proposed model is primarily in the planning stages of a ship transit where the goal is to have a more reproduceable (and less mentally taxing) method to fuse a common navigational picture from all of the available data sources. The expectation is that this would most likely take place through a specialized tool attached to the Electronic Chart System (ECS) or Electronic Chart Display and Information System (ECDIS), since they provide the only generally available “point of use” source data for the mariner. (The ECS route is more likely due to International Maritime Organization carriage requirements for ECDIS systems.) The goal would be to allow the navigator to specify reliabilities and attitudinal character, and to vary the assumptions for the planned transit to assess the most plausible route.

The output would most likely be an overlay of the fused categorical classes to be viewed in conjunction with other ENC information; an IHO S-100 series (International Hydrographic Organization Citation2022) special-purpose cartographic product would be the most likely current implementation model. Use of the S-100 framework for the output has the added benefit of using such standards on the input side as well as it would facilitate the standardized ingestion of other types of information (e.g. S-102 - high resolution bathymetry, S-104 - tidal information). Implementations will likely require software engineering, starting with a relatively simple implementation that relates the algorithm parameters to existing ENC quality indicators, such as CATZOC and date of the latest applied updates. Such implementations could then be made available for end-user software, such as a plugin for open-source OpenCPN (Citation2022) in the Safety plugins category.

Some research questions involving human factors remain open and include the question of what visualization model would work best for this data. For planning, a simple “traffic light” scheme of red (no-go), yellow (caution), and green (go) transparent coloring might be sufficient, but overlays of this kind are known to obscure detail, making them hard to use in practice. Current research (Kastrisios and Ware Citation2022) on display of uncertainty data for bathymetry in a navigational context might be used to provide better approaches. Another open question with regard to CONOPs is whether the proposed method is better for planning, or if they can be used during operations (e.g. to provide decision support if the ship has to deviate from the planned route). The requirements for real-time support are quite different from planning, and further research would likely be required to determine the best method to provide this information, and effective visualization techniques that provide support without clutter on the display.

Further work also could include the use of available types of S57 hazards that make the reconstruction more adherent to the human interpretation of a real ENC; the paper by Masetti, Faulkes, and Kastrisios (Citation2018) can serve as a starting point. Some of these objects, for instance wrecks (WRECKS) and obstructions (OBSTRN), will require a slightly more sophisticated treatment that the simple extraction of the value of sounding (VALSOU, if available) as they could necessitate the creation of a buffer area around the feature position to reflect horizontal uncertainty of the report. One could translate linguistic terms for hazards into automatic “no-go” regions that would then override the Bayesian output for that position through possibilistic conditioning (Yager Citation2012; Elmore, Petry, and Yager Citation2014) or possibility-probability transformations (Petry, Elmore, and Yager Citation2015).

Additional work also could include temporal dependency built into the algorithm that makes grid nodes become more pessimistic based on age of the information. In addition, analysis of the shape of the estimated categorical distribution (e.g. through the kurtosis, measuring the “peaked-ness” of the distribution) at each location can be used to determine the degree of consistency of the various sources, which could also be displayed to the user as qualifying information for the algorithm and data. Further, other algorithms for performing the irregular interpolation of the bathymetry soundings other than the average of enclosing TIN points could be implemented, such as those based on B-Splines (Steinbeck and Koschke Citation2021). In addition, the algorithms and framework developed in this article could further be used in a secondary fusion process of navigation risk surfaces and safety contours as emerging navigation risk tools (Bakdi et al. Citation2019; Jeong et al. Citation2019) become available.

Summary

This article provides a hybrid-framework for human-in-the-loop information fusion of navigation planning. Because the human-in-the-loop is core to this problem, a soft computing methodology (OWA) designed for such information fusion is used along with Bayesian estimates of risk categories. The framework shows how to compute OWA weights when human judgment of reliability of the information sources is involved, along with attitudinal character of risk tolerance. The estimates of reliability are reused in the Bayesian process of computing the categorical risk by providing the number of extra counts used in the Dirichlet-based maximum likelihood estimation of the probability for each category. The final category assigned is the category with the highest posterior probability, with ties broken toward the most pessimistic category.

Simple examples of using this method to compute an average of bathymetry were presented. This process forms the basis of the overall computational flow used in the case study. In the case study, maps that show areas for safe, cautionary, or unsafe navigation (“go,” “caution,” and “no-go” areas) are presented for an area of the Chesapeake Bay. Differing maps are constructed with differing sounding densities from ENC data. The framework takes irregularly spaced sounding data and computes a common gridded bathymetry surface for each chart set. Demonstrations of navigation risk surfaces computed through the OWA and Bayesian fusion are shown for differing levels of attitudinal character of risk tolerance. As the attitudinal character is lowered (tending to a more conservative approach), more area of navigation surface that were “go” change to “caution” areas along classification boundaries. This change reflects the navigator’s desire for extra keel clearance. The case study also shows how keel depth, risk tolerance, judgment of data reliability, and spatial extent of information used (by number of nearest neighbors) can affect the risk surface calculation by averaging away hazardous areas.

This framework provides a means of being able to record judgments made of data reliability. This recoding capability enables reproducibility of the resultant human-based map of navigation risk. This framework can be further be used for subsequent fusion of differing assessments of navigation risk as future capabilities and software of estimating navigation risk and safety are developed and emerge for use.

Acknowledgments

The main authors wish to thank R. Wade Ladner, Raymond Sawyer, and Dr. David Fabre of the Naval Oceanographic Office and Brett Hode of the Naval Research Laboratory for their helpful discussions in the initial phase of research. The authors wish to thank all of the funding sources for their sponsorship of this research and the anonymous peer-reviewers for their guidance in revising this manuscript.

Disclosure statement

No potential conflict of interest was reported by the authors.

Data availability statement

The data that support the findings of this study are openly available at https://nauticalcharts.noaa.gov/charts/noaa-enc.html.

Additional information

Funding

Initial funding for the conceptual development of this work was provided by the Office of Naval Research10.13039/100000006. Additional funding for generation of this paper was given to Elmore through a Janney grant of the Johns Hopkins University/Applied Physics Laboratory10.13039/100012314 and to Petry through the Naval Research Laboratory’s Base Program. Funding for Calder and Massetti was also provided by NOAA10.13039/100000192 grant NA20NOS4000196.

References

  • Bakdi, A., I. K. Glad, E. Vanem, and O. Engelhardtsen. 2019. AIS-based multiple vessel collision and grounding risk identification based on adaptive safety domain. Journal of Marine Science and Engineering 8 (1):Article 5.
  • Beliakov, G., A. Pradera, and T. Calvo. 2007. Aggregation functions: A guide for practitioners. Heidelberg: Springer. ISBN: 978–3540737209.
  • Calder, B. 2015. On risk-based expression of hydrographic uncertainty. Marine Geodesy 38 (2):99–127.
  • Calder, B. R., S. J. Dijkstra, S. Hoy, K. Himschoot, and A. Schofield. 2020. A design for a trusted community bathymetry system. Marine Geodesy 43 (4):327–58.
  • Chénier, R., M. Sagram, K. Omari, and A. Jirovec. 2020. Earth observation and artificial intelligence for improving safety to navigation in Canada low-impact shipping corridors. ISPRS International Journal of Geo-Information 9 (6):383.
  • Cutler, T. 2003. Dutton’s nautical navigation. 15th ed. Annapolis, MD: Naval Institute Press. ISBN 978-1-55750-248-3.
  • Elmore, P., F. Petry, and R. Yager. 2014. Comparative measures of aggregated uncertainty representations. Journal of Ambient Intelligence and Humanized Computing 5 (6):809–19.
  • EMODnet Bathymetry Consortium. 2020. EMODnet digital bathymetry (DTM). Accessed January 13, 2023. https://doi.org/10.12770/bb6a87dd-e579-4036-abe1-e649cea9881a
  • Gelman, A., J. B. Carlin, H. S. Stern, D. B. Dunson, A. Vehtari, and D. B. Rubin. 2013. Bayesian data analysis. 3rd ed. Boca Raton, Florida: Chapman and Hall/CRC. ISBN: 978–1439840955
  • Gilardoni, E., and J. Presedo. 2017. Navigation in shallow waters. Livingston Scotland, UK: Witherbys Publishing. ISBN: 978–1856096676.
  • International Hydrographic Organization. 2000. Intergovernmental Oceanographic Commission, IHO Transfer Standard for Digital Hydrographic Data, IHO Publication S-57, Edition 3.10. Accessed January 27, 2022. https://iho.int/uploads/user/pubs/standards/s-57/31main.pdf.
  • International Hydrographic Organization. 2019. Intergovernmental Oceanographic Commission, The IHO-IOC GEBCO Cook Book, IHO Publication B-11, Monaco, October 2019. Accessed January 27, 2022. https://www.gebco.net/data_and_products/gebco_cook_book/.
  • International Hydrographic Organization. 2020. Intergovernmental Oceanographic Commission, Mariner’s Guide to Accuracy of Depth Information in Electronic Navigational Charts (ENC), IHO Publication S-67, Edition 1.0.0. Accessed November 21, 2022. https://iho.int/en/standards-and-specifications.
  • International Hydrographic Organization. 2022. Intergovernmental Oceanographic Commission, S-100 Universal Hydrographic Data Model. Accessed November 25, 2022. http://iho.int/en/s-100-universal-hydrographic-data-model.
  • Jakobsson, M., L. A. Mayer, C. Bringensparr, C. F. Castro, R. Mohammad, P. Johnson, T. Ketter, D. Accettella, D. Amblas, L. An, et al. 2020. The International Bathymetric Chart of the Arctic Ocean Version 4.0. Scientific Data 7 (1):176.
  • Jeong, M. G., E. B. Lee, M. Lee, and J. Y. Jung. 2019. Multi-criteria route planning with risk contour map for smart navigation. Ocean Engineering 172:72–85.
  • Kastrisios, C., and C. Ware. 2022. Textures for coding bathymetric data quality sectors on electronic navigational chart displays: Design and evaluation. Cartography and Geographic Information Science 49 (6):492–511.
  • Kastrisios, C., C. Ware, B. Calder, T. Butiewicz, L. Alexander, and O. Hauser. 2020. Nautical chart data uncertainty visualization as the means for integrating bathymetric, meteorological, and oceanographic information in support of coastal navigation. Presented at the Proceedings of the American Meteorological Society 100th Meeting. Boston, MA. Accessed January 27, 2022. http://ccom.unh.edu/sites/default/files/publications/Kastrisios_etal_AMS2020_Abstract_ChartDataUncertaintyVisualizationAndIntegration.pdf.
  • Masetti, G., T. Faulkes, and C. Kastrisios. 2018. Automated identification of discrepancies between nautical charts and survey soundings. ISPRS International Journal of Geo-Information 7 (10):392.
  • Mathematicalmonk. 2012. Expectation of a Dirichlet random variable. Palo Alto, CA: YouTube video hosted by Google. Accessed March 25, 2022. https://youtu.be/emnfq4txDuI.
  • Mathworks. 2021. MATLAB (Version 2021b). Natick, MA: The Mathworks Inc. Accessed January 27, 2022. www.mathworks.com.
  • Naevestad, T. O., R. O. Phillips, K. V. Størkersen, A. Laiou, and G. Yannis. 2019. Safety culture in maritime transport in Norway and Greece: Exploring national, sectorial and organizational influences on unsafe behaviours and work accidents. Marine Policy 99:1–13.
  • National Oceanic and Atmospheric Administration – Office of Coastal Survey. 2022. NOAA ENC – Electronic Navigation Charts. Accessed January 27, 2022. https://nauticalcharts.noaa.gov/charts/noaa-enc.html.
  • OpenCPN. 2022. OpenCPN Chart Plotter Navigation, Version 5.6.2. Accessed November 25, 2022. https://www.openCPN.org.
  • Petry, F., P. Elmore, and R. Yager. 2015. Combining uncertain information of differing modalities. Information Sciences 322:237–56.
  • Preparata, F., and M. Shamos. 1988. Computational geometry - An introduction. Springer-Verlag Berlin: GDR. ISBN: 978–0387961316.
  • QGIS.org. 2020. QGIS Geographic Information System, Version 3.12. QGIS Association. Accessed January 13, 2023. http://www.qgis.org.
  • Raiffa, H., and R. Schlaifer. 2000. Applied statistical decision theory. New York: Wiley Classics Library. Wiley.
  • Saddler, R. 2006. Graveyard of the Pacific: Gateway to the Northwest. Accessed November 24, 2021. https://www.blueoregon.com/2006/01/graveyard_of_th/.
  • Steinbeck, M., and R. Koschke. 2021. TinySpline: A small, yet powerful library for interpolating, transforming, and querying NURBS, B-splines, and Bezier curves. Paper presented at the 28th IEEE International Conference on Software Analysis, Evolution and Reengineering (SANER), (Virtual Meeting), Honolulu, HI, March 09–12.
  • Walhey, F. 2013. U.S. navy to scrap vessel stuck on Philippine reef. New York Times, January 31. Accessed January 27, 2022. https://www.nytimes.com/2013/02/01/world/asia/us-navy-to-scrap-vessel-stuck-on-philippine-reef.html.
  • Weatherall, P., K. M. Marks, M. Jakobsson, T. Schmitt, S. Tani, J. E. Arndt, M. Rovere, D. Chayes, V. Ferrini, R. Wigley, et al. 2015. A new digital bathymetric model of the world’s oceans. Earth and Space Science 2 (8):331–45.
  • Wilma, D. 2006. Graveyard of the Pacific: Shipwrecks on the Washington Coast. Accessed November 24, 2021. https://www.historylink.org/File/7936.
  • Yager, R. 1988. On ordered weighted averaging aggregation operators in multicriteria decision making. IEEE Transactions on Systems, Man, and Cybernetics 18 (1):183–90.
  • Yager, R., P. Elmore, and F. Petry. 2017. Soft likelihood in combining evidence. Information Fusion 36:185–90.
  • Yager, R. R. 2012. Conditional approach to possibility-probability fusion. IEEE Transactions on Fuzzy Systems 20 (1):46–56.
  • Yager, R. R., and J. Kacprzyk. 1997. The ordered weighted averaging operators: Theory and applications. Norwell, MA: Kluwer. ISBN: 978–0792399346

Appendix A

Figure associated with examples

below provides graphs of the depth estimates and OWA weights as a function of attitudinal character for the cases discussed in the body of this article.

Figure A1. Graphs corresponding to the examples given in this Appendix showing depth estimates and weightings as a function of attitudinal character. Subfigure (a): depth estimates for equal reliability weighting in blue, r={0.3, 0.6, 0.1} in red, and r={0.1, 0.1, 0.8} in yellow. Subfigures (b–d): OWA weights for the three data points (z = 20 feet in blue, z = 18 feet in red, and z = 16 feet in yellow) with attitudinal character for (b) the equal reliability case, (c) the r={0.3, 0.6, 0.1} case, and (d) the r={0.3, 0.6, 0.1} case.

Figure A1. Graphs corresponding to the examples given in this Appendix showing depth estimates and weightings as a function of attitudinal character. Subfigure (a): depth estimates for equal reliability weighting in blue, r={0.3, 0.6, 0.1} in red, and r={0.1, 0.1, 0.8} in yellow. Subfigures (b–d): OWA weights for the three data points (z = 20 feet in blue, z = 18 feet in red, and z = 16 feet in yellow) with attitudinal character for (b) the equal reliability case, (c) the r={0.3, 0.6, 0.1} case, and (d) the r={0.3, 0.6, 0.1} case.

Appendix B

Derivation of weighted Dirichlet expectation values

This appendix derives Eqs. (9) and (10). Equation (9) is expectation value for the ith category which is derived in Appendix Section B (Derivation of Eq. (9)). EquationSection Derivation of Eq. (10) will make use of the result of Section B (Derivation of Eq. (9)) to derive Eq. (10). The goal of these derivations is to show that weighting the jth dataset by Rj (where RjZ+) is equivalent to recounting the dataset Rj times for our use of Dirichlet categorical estimation.

Derivation of Eq. (9)

This section follows derivation methods from the YouTube series by Mathematicalmonk (Citation2012). The derivation below has three major components: (1) initial setup of the problem in the context of counting categories, (2) providing useful substitutions and relations, and (3) use of substitutions to complete the derivation. The gamma function, Γ(n) is core to the Dirichlet probability density function. In the present problem, n is an integer, which means that (B1) Γ(n)=(n1)!; nZ(B1)

The expectation is derived explicitly for the category with i=1, but extension to categories with i1 is obvious.

Initial setup

Let αΣi=1i=Iαi where αi is the number of counts for the ith category, with an initial value for each being αi,0>0  i. For categories i=1, 2,,I, the Dirichlet probability density function is (B2) p(θ1, ,θI; α1,,αI)=Γ(αΣ)Γ(α1)Γ(α2)××Γ(αI)i=1Iθiαi1(B2) where i=1i=Iθi=1, with each θi representing the probability mass for the ith category. The expectation value for parameter θ1 is (B3) μ1E(θ1)=θ1Γ(αΣ)Γ(α1)Γ(α2)××Γ(αI)i=1Iθiαi1dθ(B3)

Notice that θ1i=1Iθiαi1=θ1θ1α11i=2Iθiαi1=θ1α1i=2Iθiαi1. Thus, EquationEq. (B3) becomes (B4) μ1E(θ1)=Γ(αΣ)Γ(α1)Γ(α2)××Γ(αI)θ1α1i=2Iθiαi1dθ(B4)

Useful substitutions and relations

A useful recursion relation of the gamma function that this derivation will exploit is (B5) Γ(n+1)=nΓ(n)(B5)

Define, β1α1+1, βiαi for i2, and βΣi=1i=Iβi. Note that βΣ=αΣ+1. Thus, with Eq. (B5), the following useful relations hold (B6) Γ(α1+1)=α1Γ(α1)(B6) (B7) Γ(αΣ+1)=αΣΓ(αΣ)(B7) (B8) Γ(β1)=Γ(α1+1)=α1Γ(α1)(B8) (B9) Γ(βΣ)=Γ(αΣ+1)=αΣΓ(αΣ)(B9) (B10) Γ(α1)=Γ(β1)/α1(B10) (B11) Γ(αΣ)=Γ(βΣ)/αΣ(B11)

Completion of the derivation

Using EquationEqs. (B10) and Equation(B11) in EquationEq. (B4) yields (B12) μ1E(θ1)=α1αΣΓ(βΣ)θ1α1Γ(β1)Γ(β2)××Γ(βI)i=2Iθiβi1dθ(B12)

Note that α1=(α1+1)1=β11 so that (B13) μ1E(θ1)=α1αΣΓ(βΣ)θ1(α1+1)1Γ(β1)Γ(β2)××Γ(βI)i=2Iθiβi1dθ(B13) (B14) =α1αΣΓ(βΣ)θ1β11Γ(β1)Γ(β2)××Γ(βI)i=2Iθiβi1dθ(B14) (B15) =α1αΣΓ(βΣ)Γ(β1)Γ(β2)××Γ(βI)i=1Iθiβi1dθ(B15)

The function in the integral is a Dirichlet probability density function, which must integrate to 1. Thus, (B16) μ1E(θ1)=α1αΣ(B16)

Now, set αi,0=1  i for the initial prior pdf, and ni=αi after counting of the ith category is complete and N=ini. Thus α1=ni+αi,0  and αΣ=N+iαi,0.  These substitution in (B16) yields Eq. (9).

Derivation of Eq. (10)

From EquationEq. (B11), let ni=j=1Jni,j where j=1, 2,,J indexes the dataset used. Then, N=i=1Ij=1Jni,j and ai,0=j=1Jαij,0. (For now, the initial setting for αij,0 is abstract, but will set later in this derivation.) With these substitution, EquationEq. (B11) becomes (B17) μi=ni+αi,0N+ iαi,0=j=1Jni,j+j=1Jαij,0i=1Ij=1Jni,k+ i=1Ij=1Jαij,0(B17)

If the data are now weighted by normalized reliability rk (i.e. j=1Jrj=1), then, EquationEq. (B17) becomes (B18) μi=j=1Jrjni,j+j=1Jrjαij,0i=1Ij=1Jrjni,j+ i=1Ij=1Jrjαij,0(B18)

This result follows from the derivation in Section B (Derivation of Eq. (9)) with substitution of αi = j=1Jrjαi,j and softening the restriction that the gamma function has an integer argument.

As discussed in the main text of the article, rjRj/j=1JRj, where RjZ+. Thus, EquationEq. (B18) becomes (B19) μi=j=1J(Rj/j=1JRj)ni,j+j=1J(Rj/j=1JRj)αij,0i=1Ij=1J(Rj/j=1JRj)ni,j+ i=1Ij=1J(Rj/j=1JRj)αij,0(B19) (B20) =j=1JRjni,j+j=1JRjαij,0i=1Ij=1JRjni,j+ i=1Ij=1JRjαij,0(B20) since all terms in (B20) are divided by j=1JRj. Setting αij,0=αi,0/(JRj), EquationEq. (B20) becomes EquationEq. (B12) in the text (i.e. one is informed of how to construct the initial Dirichlet pdf using the reliabilities.) Thus, one can account for reliability weighting of the jth chart by recounting each category Rj times. In the special case of Rj=1  j, Eqs. (10) and (B20) reduces to Eq. (9) as required.

Appendix C

Algorithm sketches

Algorithm 1

  • Input Parameters: (a) depth clearance; (b) degree of optimism (from 0 to 1)

  • Load chart point cloud

  • Perform Delaunay triangulation

  • FOR each triangle set bv

    • Evaluate the depth of the end nodes for each created triangle

    • IF the end nodes are within the depth clearance

      • Compute depth in the triangle center from linear interpolation

      • Append new triplet to the point cloud.

    • ELSE - skip to next triangle

  • Write augmented XYZ point-cloud to output file.

Algorithm 2

  • Input Parameters: (a) XY output grid points; (b) depth clearance; (c) number of points to examine at each node; (d) number used indicate no data

  • Load augmented point cloud files from Algorithm 1; retain file name string in memory

  • FOR each θ{θ*,θ*}

    • FOR each XY grid point

      • Divide the area around each node into octants

      • For each octant, identify the closest xyz-points from the augmented point-cloud

      • IF xyz-data in at least one octant,

        • XY gridded bathymetry value = average of these nearest xyz-points weighted by the inverse of the squared distance

      • ELSE

        • XY gridded bathymetry value = no data flag (manual intervention will be required to augment the data)

    • WRITE gridded bathymetry to output files for each θ{θ*,θ*}

  • Transform bathymetry to trinary output classes is finally created by using pair of grids (with different degree of optimism) as follows:

  • For each cell in the grid pair for θ{θ*,θ*}, set the quaternary value as follows:

    • 3 (no-go): keel depth is either deeper than the (a) sounding depth OR (b) deeper reference depth

    • 1 (go): keel depth is shallower than the (a) sounding depth AND (b) shallow reference depth

    • 2 (maybe): keel depth is in between the two reference depths AND (b) the sounding is in between the two surfaces

  • WRITE trinary output to output file grid