Skip to main content
search

This online seminar includes industry best practices for using UX-Analyze

to process advanced electromagnetic sensor data to classify unexploded ordnance targets. An overview of UX-Analyze, along with practical tips for experienced users to help improve their workflows.

Overview

Speakers

Darren Mortimer
Product Owner – Seequent

Duration

58 min

See more on demand videos

Videos

Find out more about Oasis montaj

Learn more

Video Transcript

[00:00:01.640]
<v Darren>Hi everyone.</v>

[00:00:02.670]
My name is Darren Mortimer

[00:00:04.120]
and I’m a product owner here at Seequent.

[00:00:06.950]
On behalf of Seequent

[00:00:07.860]
I’d like to welcome you to today’s webinar

[00:00:10.150]
on industry Best Practice’s

[00:00:11.590]
for Advanced Geophysical Classification of UXO Survey Data.

[00:00:17.350]
So here’s what we’re going to cover.

[00:00:19.890]
What is classification

[00:00:21.020]
and why consider using it for your UXO projects?

[00:00:24.890]
And an introduction to Advanced Geophysical Classification,

[00:00:28.150]
along with working with dynamic and static survey data.

[00:00:32.720]
I also have some tips

[00:00:33.870]
of features that you may not be aware of.

[00:00:36.120]
Time savers to make your project workflows more efficient.

[00:00:40.740]
So whether you’re a new user to Oasis montaj

[00:00:43.195]
and UX-Analyze, or a seasoned pro,

[00:00:46.630]
I have something for everyone.

[00:00:48.320]
So, let’s get started.

[00:00:51.770]
So what is classification?

[00:00:53.170]
It’s the action or process of classifying something

[00:00:55.590]
according to shared qualities or characteristics.

[00:00:58.470]
Anyone can do classification.

[00:01:00.220]
In fact, we learned to do this a quite at an early age.

[00:01:04.200]
I went and found some experts

[00:01:05.430]
and see how well they would do.

[00:01:09.800]
You can see they did a pretty good job

[00:01:11.610]
of being able to classify or group the items

[00:01:14.040]
based on their property.

[00:01:16.270]
Things like size, shape and color.

[00:01:20.690]
However our classification problems aren’t quite so easy.

[00:01:23.500]
We need to find things like UXOs, unexploded ordinance

[00:01:28.530]
or ERW, explosive remnants of war.

[00:01:32.490]
And we must look in places like fields and forests,

[00:01:35.550]
where they’re not easily visible.

[00:01:39.970]
Now why would we want to do classification?

[00:01:43.760]
Several years ago, the defense science board did a study

[00:01:47.540]
on the typical pro of cost breakdowns

[00:01:51.440]
for munitions projects.

[00:01:54.340]
And the typical munitions clean up,

[00:01:56.590]
an overwhelming fraction of the money

[00:01:58.450]
is spent removing non-hazardous items.

[00:02:01.410]
So if we can save money,

[00:02:02.920]
if we can identify these items beforehand

[00:02:06.400]
and either remove them with fewer safety precautions

[00:02:10.500]
or simply leave them in the ground.

[00:02:14.450]
Another way to think about this

[00:02:16.340]
is if we can reduce the digging scrap or clutter by 90%,

[00:02:21.100]
we can see a reduction in project costs.

[00:02:26.460]
I should note there are sites

[00:02:28.210]
where classification isn’t recommended.

[00:02:30.440]
If you’re working on heavily impacted areas

[00:02:34.160]
and looking for small items,

[00:02:36.080]
or when you know you’re going to need to dig everything up

[00:02:38.810]
because of the nature of the final land use of the site.

[00:02:46.370]
So what is Advanced Geophysical Classification?

[00:02:48.770]
It’s using a principled physics-based approach

[00:02:52.530]
to reliably characterize the source

[00:02:55.050]
of a geophysical anomaly as either a target of interest,

[00:02:59.060]
a UXO or as a non target of interest,

[00:03:03.210]
clutter, debris or scrap.

[00:03:05.950]
And you must recognize that even the current

[00:03:08.710]
survey or field methods

[00:03:10.490]
already involve some kind of implicit discrimination.

[00:03:13.600]
Mag and flag, how sensitive is the instrument that’s using

[00:03:17.660]
and how attentive is that human that’s

[00:03:20.600]
working and listening to the tones

[00:03:22.530]
and reading the dial as they go along.

[00:03:25.160]
Or in digital geophysics when we set our target thresholds.

[00:03:30.240]
Above this, we will pick it and call it an anomaly,

[00:03:32.900]
below that, we don’t.

[00:03:34.800]
Those themselves are some levels of classification.

[00:03:43.996]
We found that electromagnetic geophysical methods

[00:03:47.770]
are the most useful.

[00:03:50.130]
Compared to magnetic methods,

[00:03:51.668]
EM is minimally affected by magnetic soils

[00:03:54.730]
and can detect both ferrous and non-ferrous items,

[00:03:58.410]
and also provides more information

[00:04:01.120]
or properties about the source.

[00:04:04.460]
Things like distance, orientation, it’s size and shape,

[00:04:09.690]
material type and thickness.

[00:04:13.620]
Some of these can be called extrinsic properties.

[00:04:16.780]
They’re external to the item, other one they’re intrinsic.

[00:04:22.450]
These are the properties that are the most important ones,

[00:04:25.710]
because then we can look at these

[00:04:29.060]
and use those for classification.

[00:04:35.360]
The EM response can be decomposed into components

[00:04:40.949]
along three orthogonal principal directions.

[00:04:47.960]
These magnetic polarizabilities are specific responses

[00:04:52.000]
to the EM excitation or the electromagnetic excitation

[00:04:57.020]
along the target’s or source’s principal axes.

[00:05:00.600]
Basically these things called polarizabilities,

[00:05:04.610]
completely describe the EM response of the target

[00:05:07.860]
and the are intrinsic to the target.

[00:05:11.900]
And we’ll see a little bit more about that coming up.

[00:05:18.750]
So thinking about some of the conventional sensors

[00:05:20.750]
which you may be familiar with.

[00:05:22.230]
These are really good for the detection,

[00:05:24.090]
but the not usually good for classification.

[00:05:27.200]
They have a limited number of measurements.

[00:05:30.500]
Often only a couple of time gates or maybe even one.

[00:05:36.000]
And the generally usually a single monostatic transmitter

[00:05:39.420]
and receiver.

[00:05:40.253]
That means the transmitter and receiver

[00:05:43.210]
are pointing in the same direction

[00:05:45.760]
and they’re at the same location.

[00:05:49.610]
To be able to get a full look of the target,

[00:05:52.750]
we need to move the sensor around

[00:05:55.920]
and even small errors and locating the sensor creates noise.

[00:06:01.220]
The end result of all of this,

[00:06:02.720]
that the sensors aren’t good for classification because they

[00:06:08.210]
don’t allow us to generate good, accurate,

[00:06:11.300]
reliable polarizabilities.

[00:06:15.630]
So along comes the advanced electromagnetic sensors.

[00:06:19.390]
These guys are designed for classification.

[00:06:22.090]
They observe the response

[00:06:23.770]
and allow us to calculate reliable polarizabilities.

[00:06:26.370]
And there kind of is two types of flavors.

[00:06:29.260]
There’s either a single axis planar array

[00:06:32.240]
where we just have a array of coils,

[00:06:35.860]
very similar to what you’re working with already.

[00:06:38.920]
Or we can mount these in

[00:06:41.570]
the transmit and receiver coils

[00:06:43.330]
in many different orientations and directions.

[00:06:46.890]
So it’s both allowing us to fully illuminate

[00:06:50.180]
or excite the target and measure it from several directions.

[00:06:57.460]
Here are some examples of current system sensors

[00:07:00.143]
that are available and are in use today.

[00:07:04.330]
Things like the TEM two by two

[00:07:05.880]
and the MetalMapper two by two,

[00:07:08.210]
they have four transmitter coils and a planer array.

[00:07:14.890]
And in the center of each of those coils is a receiver

[00:07:20.150]
which is a multi-axis receiver.

[00:07:21.840]
It is orientated in both X, Y and Z.

[00:07:27.110]
On the other hand, you’ve got things like the MetalMapper

[00:07:29.330]
and the Man-Portable-Vector or NPV.

[00:07:32.290]
These guys have multiple access transmitters,

[00:07:35.980]
and you can see them there sticking up above

[00:07:39.250]
looking kind of like an egg beater along with

[00:07:43.200]
in the case of the MetalMapper seven multi-axis receivers

[00:07:46.380]
and the case of the MPV, five multi-axis receivers

[00:07:51.970]
on that sort of a round circular head.

[00:07:57.420]
We also record over a much larger window.

[00:08:00.340]
Typically in a static survey mode,

[00:08:02.480]
we record 122 gates over 25 milliseconds,

[00:08:06.940]
collecting much more data.

[00:08:11.612]
And we can use this data to help us to determine

[00:08:14.570]
or develop these intrinsic properties of the source.

[00:08:18.210]
We can take our response data here which has shown

[00:08:23.000]
all the responses from a two by two sensor.

[00:08:29.310]
The plots are shown with a log time and along the x-axis,

[00:08:34.350]
and it’s the log voltage along the y-axis.

[00:08:37.700]
We can take all of this response data and invert it

[00:08:40.810]
to give us reliable polarizabilities.

[00:08:43.860]
And I have a little example here for you.

[00:08:47.310]
Here we have a gain that a two by two type system.

[00:08:52.150]
Each of the large squares represents the transmitter

[00:08:56.000]
in a planar array.

[00:08:57.350]
In the center of that there is a receiver

[00:09:00.950]
that has got the three receiver coils on it

[00:09:04.160]
in each of the three orthogonal directions.

[00:09:09.800]
We have the response data

[00:09:12.930]
and then there’s the polarizabilities.

[00:09:15.540]
And if I take this source

[00:09:18.030]
and we’re going to move it around here,

[00:09:19.680]
you’ll be able to see how changing the source locations

[00:09:24.040]
changes the response,

[00:09:25.630]
but the polarizabilities essentially remain the same.

[00:09:30.100]
So we can move it down there to the bottom

[00:09:33.000]
and then move it over to the top.

[00:09:35.200]
I’ll just go back and forth there where you can see

[00:09:37.970]
how the response keeps changing,

[00:09:41.070]
but the polarizabilities essentially don’t.

[00:09:47.950]
So we can use these polarizabilities

[00:09:50.730]
since they completely describe the EM response of the source

[00:09:53.750]
and they’re intrinsic to the source

[00:09:56.220]
and they really don’t change due to the depth

[00:10:00.530]
that we will bury the source or its orientation.

[00:10:05.520]
We can also extract from them

[00:10:07.190]
a number of properties which are directly related

[00:10:09.700]
to the physical properties of the source.

[00:10:12.450]
We can look at the decay rate

[00:10:14.380]
which will give us the wall thickness.

[00:10:17.040]
We can look at the relative magnitude

[00:10:20.280]
of the various polarizability that gives us

[00:10:22.920]
an indication of the shape of the item.

[00:10:26.460]
And we can also look at the total magnitude

[00:10:32.120]
of the polarizability and that will give us an indication

[00:10:35.510]
of the overall volume or size of the object or source.

[00:10:43.870]
These features or properties can be easily shown

[00:10:48.790]
in a feature space plot.

[00:10:52.540]
For example here’s the size and decay.

[00:10:55.560]
Remember size is kind of the overall volume of the object

[00:10:58.330]
and decay is that notion of the wall thickness.

[00:11:00.750]
And when we can use that to classify items.

[00:11:04.950]
Well, we can see here that we’ve got a grouping of

[00:11:08.220]
targets or sources there related to 75 millimeters

[00:11:13.450]
and other ones related to a 37-millimeter,

[00:11:16.750]
but the 57 millimeters, they’re a little spread out.

[00:11:19.010]
It’s not quite as helpful.

[00:11:21.730]
These feature plots or the features alone

[00:11:24.450]
have a limited classification power

[00:11:27.460]
compared to the overall curve.

[00:11:32.220]
These are really what the source looks like in the EM sense.

[00:11:37.560]
So we could compare the polarizabilities,

[00:11:40.090]
the whole entire curve from our unknown item

[00:11:43.740]
to a bank of signatures of items

[00:11:46.490]
that we would expect to find

[00:11:48.550]
or we have for expected munitions and other items.

[00:11:53.560]
Here on the left,

[00:11:54.393]
we have something that’s typical of a target of interest

[00:11:56.970]
or TOI.

[00:12:00.124]
It’s a 37-millimeter projectile.

[00:12:02.680]
And you can see there,

[00:12:04.170]
it’s got one strong primary polarizability

[00:12:07.750]
and two weaker and equal secondary

[00:12:11.180]
and tertiary polarizabilities.

[00:12:14.170]
This is typical of what we expect to see for munitions

[00:12:19.220]
because of their actual symmetry.

[00:12:22.760]
They’re mostly pipe type shapes.

[00:12:28.020]
Non targets of interest or none TOI, things like horseshoes,

[00:12:32.830]
scrap metal, the debris.

[00:12:35.270]
These typically have different polarizabilities,

[00:12:39.990]
they tend not to be equal.

[00:12:44.010]
They tend to sort of just be very irregular

[00:12:46.800]
because that’s what the shape of

[00:12:49.080]
most scrap pieces of metal are.

[00:12:53.113]
They are regular in shape.

[00:12:59.420]
And to give you an idea of,

[00:13:00.437]
you know, how well these kind of things work,

[00:13:03.010]
we can look here at a couple of different items.

[00:13:05.700]
Here we have a 37 and 75-millimeter.

[00:13:08.290]
They kind of have a different shape

[00:13:09.750]
but you can see clearly they have a different size,

[00:13:13.290]
see where they sort of would be coming in

[00:13:14.970]
and that Y axis intercept is located.

[00:13:25.810]
And this one always amazes me.

[00:13:28.900]
We can even detect and tell the presence of something

[00:13:33.031]
such as the driving band.

[00:13:34.710]
The driving band is usually a thin band of soft metal,

[00:13:38.400]
often copper, that is around the shell

[00:13:42.840]
that cause it to rifle or spin

[00:13:44.920]
as it travels through the barrel.

[00:13:47.030]
And whether that is located at the end of the round,

[00:13:50.770]
whether it’s lost during firing altogether

[00:13:53.820]
or it’s located in the middle round

[00:13:56.070]
causes slight changes in our polarizabilities.

[00:13:58.847]
And the fact that we can see that

[00:14:00.560]
I think is it’s pretty amazing and pretty cool stuff.

[00:14:05.610]
It does point out that we need to make sure that

[00:14:08.200]
our classification system and our methodology

[00:14:12.090]
can deal with subtypes, maybe damage to the item.

[00:14:18.360]
Also possibly just noise and inversion errors

[00:14:22.010]
as slight errors as we go along through our process.

[00:14:32.580]
So doing an Advanced Geophysical Classification survey

[00:14:36.770]
or AGC kind of comes down into two parts these days.

[00:14:42.070]
There is work being carried out

[00:14:43.410]
by the hardware manufacturer and others

[00:14:45.510]
for us to be able to detect and classify in a single pass.

[00:14:48.800]
But right now, we currently need to do things in two passes.

[00:14:52.330]
We do it as in a dynamic survey, kind of a mapping mode.

[00:14:56.970]
It’s kind of like mowing the grass

[00:14:58.530]
where we find all the possible locations

[00:15:00.570]
that we may have a source.

[00:15:03.630]
This can be done with conventional sensors,

[00:15:07.300]
things like the EM sensors that you’re familiar with,

[00:15:12.350]
magnetometers, you can use those,

[00:15:14.840]
but where appropriate the advanced EM sensors

[00:15:17.490]
do give you more accurate locations

[00:15:19.930]
and make the second part of the survey more efficient

[00:15:24.140]
because of these improved locations.

[00:15:27.380]
The second half of the survey is the

[00:15:29.980]
sort of the static survey or classification survey,

[00:15:33.400]
where we go and park our sensor at a flag location

[00:15:37.090]
and collect data to classify that source.

[00:15:41.150]
To give you some idea of production rates,

[00:15:43.020]
it’s often depending on the nature of the site,

[00:15:45.190]
how far the locations are apart.

[00:15:48.590]
We’ve found that people can collect roughly

[00:15:51.570]
three to 400 locations per day.

[00:15:59.550]
So looking at the dynamic survey, that mapping mode,

[00:16:03.950]
kind of breaks down into three easy steps.

[00:16:07.920]
We’re going to prepare the data, identify the sources

[00:16:12.460]
and they review those sources and create our list,

[00:16:14.830]
our flag list that we will use again in the static survey.

[00:16:20.150]
Some people like to call

[00:16:21.770]
the last two parts of this workflow,

[00:16:24.040]
the informed source selection or ISS,

[00:16:27.640]
because we’re using some idea

[00:16:30.010]
or knowledge of the sources that we’re looking for

[00:16:32.710]
to help us pick those targets.

[00:16:38.090]
In your UX-Analyze,

[00:16:39.140]
we have an easy workflow that will step you through this.

[00:16:43.380]
Here I’ve highlighted some of the key items

[00:16:46.000]
with the same cause as we just saw in the general workflow.

[00:16:50.210]
The items with the arrows on are the ones that

[00:16:54.560]
I would call them the must do’s.

[00:16:56.140]
That if you’re going to process data,

[00:16:57.630]
these are the things that we’re going to need to

[00:17:00.170]
step through.

[00:17:02.000]
So why don’t we take a moment here

[00:17:03.400]
and I’ll flip to Oasis Montaj

[00:17:05.905]
and we can take a look at

[00:17:07.430]
some of these parts of the workflow.

[00:17:11.770]
So here I have some data. I’ve imported it already.

[00:17:15.060]
And to save time,

[00:17:16.390]
I’ve gone through some of the processing steps,

[00:17:18.850]
because with the dynamic data

[00:17:20.330]
you do collect very large volumes of data

[00:17:23.630]
and it just takes sometimes,

[00:17:25.670]
a few moments for us to go through and do that processing.

[00:17:31.270]
The first step that you would do is do the data processing.

[00:17:36.100]
This is where we will filter and make sure

[00:17:39.070]
that any data that is outside

[00:17:41.400]
of our quality control specifications

[00:17:44.400]
is dummy down or removed from subsequent processing.

[00:17:49.750]
Like if the sample stations are too far apart because

[00:17:53.257]
the guys in the field went too fast

[00:17:55.630]
or there was some other problem with the sensor.

[00:17:59.580]
When this runs it will give you

[00:18:01.429]
sort of a plot similar to that.

[00:18:05.880]
And as we wrote down through our processing workflow,

[00:18:09.470]
the next thing you’ll need to do

[00:18:10.810]
is do some latency correction

[00:18:13.920]
for just timing aspects of how fast the system fires,

[00:18:19.200]
the GPS, all that kind of coming together.

[00:18:21.970]
Create the located database and then beta grid that up.

[00:18:27.320]
And I’ve got one of these here where I’ve prepared that

[00:18:31.570]
and shown it on here with our survey tracks.

[00:18:38.860]
One of the tips that I’d like to share with you,

[00:18:41.150]
often I like to see where was the data collected?

[00:18:44.960]
What you’re seeing there on that path there

[00:18:46.970]
is the original path of the cart of the sensor itself.

[00:18:52.100]
But remember it’s got in the case of this,

[00:18:54.320]
this is a two by two system, it’s got, as I was showing you,

[00:18:57.107]
those are the diagrams, those four receivers on those.

[00:19:00.320]
Well, where were those traveling?

[00:19:01.960]
And we can load and display those,

[00:19:04.210]
but whoa, that is just crazy.

[00:19:07.100]
I can’t see anything there.

[00:19:09.630]
One of my tips is set the transparency,

[00:19:14.020]
lower the transparency down of something of that

[00:19:16.560]
sort of other paths, where we can still see them.

[00:19:19.450]
We can still see where the original, the cart went

[00:19:23.070]
and then where those air receiver coils

[00:19:26.270]
traveled across our dataset.

[00:19:36.260]
And one of the things that

[00:19:37.240]
once we’ve created this amplitude grid,

[00:19:41.530]
one of the other steps that we like to look at

[00:19:44.460]
is something we call the coherence anomaly.

[00:19:48.810]
And this is where we look at a sample of the data

[00:19:52.910]
and see how well an item will model

[00:19:58.180]
under a window of that data.

[00:20:00.170]
And I’ll show you some examples of the threshold plots

[00:20:04.630]
in a moment.

[00:20:08.240]
The coherence anomaly map, let’s just move this over here.

[00:20:13.480]
I have gone and created one.

[00:20:16.410]
It makes it very easy to detect targets which you may miss

[00:20:23.819]
in the amplitude plot.

[00:20:29.300]
Now maybe we’d like to see our line paths on here as well.

[00:20:32.540]
And since I’ve already created them

[00:20:34.820]
over on my amplitude map,

[00:20:37.120]
it’s as easy as just I can drag and drop them onto this map

[00:20:42.020]
and we can see them there.

[00:20:44.530]
And they’ll come across

[00:20:45.770]
with their same transparencies and everything that we have

[00:20:51.230]
on the previous map.

[00:20:53.510]
And don’t forget, if you’re looking to look at multiple maps

[00:20:58.850]
and gee, wouldn’t it be nice if they were at the same place.

[00:21:02.110]
At the top of the map there is a series of buttons.

[00:21:05.840]
If I click on the second button from the left,

[00:21:09.250]
that will make all my maps zoom to the same area

[00:21:13.040]
that I see on the map that I am controlling it from.

[00:21:20.760]
For those, if you have run that UX-Analyze before,

[00:21:25.050]
any of you have noticed areas where you might see this

[00:21:28.410]
or some of these like broad high features

[00:21:31.330]
in your coherence anomaly data?

[00:21:35.840]
This is generally caused by over-correcting your data

[00:21:40.550]
or over leveling your data.

[00:21:42.470]
When you’re doing the leveling,

[00:21:46.100]
look at going in and adjusting parameters,

[00:21:51.620]
the leveling of filtering the data

[00:21:53.710]
to changing perhaps your amplitude threshold

[00:21:57.520]
or the width of the signal

[00:22:00.100]
that you’re using to do the filtering.

[00:22:04.417]
It says those types of areas where you see the broad,

[00:22:09.196]
in the broad high in the coherence anomaly,

[00:22:13.070]
or perhaps a broad low in the amplitude

[00:22:16.410]
is an indication that you’ve over leveled

[00:22:18.870]
or over filtered your data.

[00:22:21.260]
And after I went through and adjusted that,

[00:22:25.790]
we can see how I can make those a little bit more clearer.

[00:22:35.840]
Now I mentioned earlier that the coherence anomaly

[00:22:38.420]
allows us to see anomalies which we might not

[00:22:43.250]
be readily see in the amplitude.

[00:22:46.470]
Here I’ve just got to made a match snapshot,

[00:22:48.057]
you know, if you send maps to your friends

[00:22:51.890]
or to your manager or senior scientists to have a look at

[00:22:57.020]
and review?

[00:22:58.270]
So here, if I want to send this map on,

[00:23:00.090]
I’ll say, Dave, look, there’s a spot here

[00:23:02.460]
where I’ve got these low amplitude anomalies

[00:23:04.280]
and they’re only coherence elements.

[00:23:05.640]
What do you think?

[00:23:07.330]
He’s like, you know, rather and he’s got to look at this data

[00:23:09.690]
and go, well, where does he Darren want me to look?

[00:23:13.160]
He can just come unload that snapshot

[00:23:16.590]
and it will go to that area

[00:23:18.190]
and if he turns on the changing stance on all maps,

[00:23:23.010]
you can now quite easily,

[00:23:25.560]
we can go on and turn on our shadow cursor

[00:23:27.600]
and see which of those anomalies

[00:23:29.800]
we can see quite clearly on the coherence map,

[00:23:34.440]
but not so much in just the amplitude or response alone.

[00:23:46.370]
We would go on after we’ve, can pick our anomalies

[00:23:50.370]
from both the coherence and the amplitude picks

[00:23:59.840]
and using the thresholding tools,

[00:24:02.300]
we can sort of decide which threshold we used.

[00:24:04.660]
In this data set I used 0.5 and three

[00:24:07.495]
and I’ll show you in a moment how I picked those

[00:24:11.200]
when I flipped back to the slides.

[00:24:13.730]
And finally, you will invert the data

[00:24:19.610]
and generate the sources and then you need to be to filter

[00:24:23.180]
and look at those sources

[00:24:27.520]
and determine which ones are something

[00:24:30.160]
that you would like to go on and go on to

[00:24:33.770]
and collect a static survey data.

[00:24:38.720]
And I had a, so I just took slow open up my source database

[00:24:43.410]
where I’ve gone and done this.

[00:24:50.980]
In the source database.

[00:24:56.320]
So we go from survey data, we pick targets,

[00:25:00.140]
we then invert those targets to generate sources.

[00:25:04.890]
And then we might learn to look at,

[00:25:07.150]
and I’ll just overwrite the one I made earlier

[00:25:09.540]
of being able to filter out some of the sources

[00:25:13.240]
because some of them may be things which just,

[00:25:17.330]
there’s no way that they can be the

[00:25:20.590]
type of target of interest or UXO that we’re looking for.

[00:25:24.650]
And to make this easier or to help you with that,

[00:25:27.100]
when you do the filtering step,

[00:25:29.260]
we create a bunch of channels and we look for

[00:25:33.940]
how big was the source?

[00:25:37.535]
Did we get a good inversion result?

[00:25:38.368]
If we didn’t get a good inversion result

[00:25:40.520]
because of noise or something in the data,

[00:25:42.630]
then we can’t trust that result.

[00:25:44.630]
And we want to keep that on our list

[00:25:47.158]
for further investigation.

[00:25:49.740]
Some things we might say,

[00:25:51.740]
well, look, there’s no way it could have a size and decay

[00:25:57.580]
that that could be something that we could

[00:26:00.050]
possibly classify.

[00:26:01.760]
And these then are indicated in one of the channels here,

[00:26:07.150]
they’re sort of flagged off of that.

[00:26:09.440]
And there’s a nice, clear channel that tells you

[00:26:13.510]
or a column in the data as to why they were filtered out.

[00:26:17.290]
And you can see those various symbols represented on this

[00:26:21.420]
feature space of scatter plot.

[00:26:24.570]
The red lines represent my thresholds for my size and decay.

[00:26:31.490]
So you can see symbols out there turned off or gray,

[00:26:36.560]
whether it be green and turned on in here,

[00:26:40.450]
some of them were found just to simply have no anomaly

[00:26:43.910]
like we had no signal

[00:26:45.940]
in that when we do some of the inversions,

[00:26:48.820]
we will look for what we call three dipoles, three objects.

[00:26:53.920]
And one of the dipoles will have an object.

[00:26:58.490]
One or two of the others may not

[00:27:00.860]
if there’s only one physical object there.

[00:27:04.720]
And those will be filtered out and removed.

[00:27:09.340]
Some of the ones you see here with a little brown Xs

[00:27:12.230]
or orange Xs all over the site and some just on the edge,

[00:27:17.620]
these are the ones that due to some noise in the data

[00:27:20.510]
that we had a poor inversion result.

[00:27:22.430]
And we want to keep a new ongoing revisit in our static survey

[00:27:26.980]
to make sure that there is no targets of interest there

[00:27:30.860]
’cause remember, we’re dealing with UXOs here,

[00:27:32.920]
we want to be conservative.

[00:27:36.360]
So there’s a little bit of a high-level overview

[00:27:38.830]
and a few tips on the dynamic processing.

[00:27:45.250]
And just flipping back to our slides here,

[00:27:47.950]
we kind of walked through some of that workflow,

[00:27:54.880]
and I promise you to looking at the coherence threshold

[00:28:00.790]
or how we pick those thresholds.

[00:28:03.000]
So there’s a tool called

[00:28:04.140]
determined coherence anomaly threshold

[00:28:06.450]
that creates this plot for you.

[00:28:10.480]
And one of the questions I often get asked is,

[00:28:14.070]
well, how many samples should we use when we do this tool?

[00:28:19.210]
And you can see there on the right

[00:28:20.420]
I’ve just kind of highlighted where that sample box is.

[00:28:24.880]
We generally recommend that people use around

[00:28:27.400]
about 2,000 samples.

[00:28:29.530]
And the question is, well, why do I need to use 2,000?

[00:28:33.060]
I want to get it done faster. I want to use less.

[00:28:37.030]
This example where I use just 20 points,

[00:28:44.190]
where we go and find a background area.

[00:28:47.340]
So an area that we believe

[00:28:49.240]
because of its signal is free of metallic debris.

[00:28:52.820]
We synthetically insert a signal response into that

[00:28:59.360]
and invert it and see how well that inversion goes.

[00:29:04.465]
If it’s a very good match to something,

[00:29:06.100]
that gives us a strong coherence.

[00:29:09.280]
And we can look at how noisy that is

[00:29:11.720]
compared to when we just try to invert it

[00:29:14.760]
when nothing’s there.

[00:29:17.020]
With the object, gives us the black dots.

[00:29:19.860]
Without the synthetic object there,

[00:29:21.990]
it gives us the red dots.

[00:29:24.680]
And there’s the thresholds that I picked before

[00:29:28.950]
in our example.

[00:29:32.200]
And you can see there’s with 20 points.

[00:29:33.610]
If I ran it another time with this 20 points,

[00:29:36.250]
because we randomly picked the locations,

[00:29:40.860]
you get a slightly different result.

[00:29:43.330]
So maybe 20 is not good enough.

[00:29:46.260]
So use 200.

[00:29:49.050]
It’s better,

[00:29:50.950]
but I run it with another 200 points.

[00:29:53.900]
Things will slow the shift again.

[00:29:57.450]
Yes, I did run a couple of examples

[00:29:59.550]
to run several runs to cherry pick,

[00:30:02.760]
to give you some examples where you could clearly see

[00:30:05.570]
the shifts occurring.

[00:30:09.120]
But if I ran it with 2,000 points,

[00:30:14.060]
all of those variations that you see do get covered in,

[00:30:17.390]
it gives you a much more reliable set of curves

[00:30:21.730]
to bid a pick from, making an awesome,

[00:30:24.640]
makes them very easy to interpret and see.

[00:30:31.610]
If you’re wondering why the two levels

[00:30:34.020]
on the coherence plot,

[00:30:37.299]
depending on the nature of your munition,

[00:30:42.760]
we can get this sort of down curve into the y-axis.

[00:30:46.830]
And I went and picked a sort of more conservative point.

[00:30:51.700]
The purple line is the depth of investigation

[00:30:55.660]
that we’re being asked for on this project.

[00:30:58.890]
Is we’ve been asked to find things like

[00:31:03.590]
a 37-millimeter or something that can also be represented

[00:31:07.000]
by a medium ISO down to 30 centimeters.

[00:31:10.850]
ISO stands for Industry Standard Object.

[00:31:14.620]
And so that’s the level I want to pick for.

[00:31:18.600]
I want to be maybe a little conservative

[00:31:20.510]
and that’s why I’ve gone and lowered my threshold

[00:31:23.490]
down to 0.5.

[00:31:25.770]
Yes, could I go lower?

[00:31:27.780]
But that would be going sort of above and beyond

[00:31:30.470]
what we were asked for in our project scope.

[00:31:39.890]
So now we’ve chosen our thresholds.

[00:31:42.270]
There’s two places that we use the thresholds.

[00:31:44.470]
And my other little tip is when you take your values

[00:31:48.530]
that you’ve used in your target picking,

[00:31:51.700]
I ended up typing in three I guess or could have used 3.7.

[00:31:55.320]
I think I had 3.6 sort of in around there.

[00:32:02.570]
Those are the values that we picked targets with.

[00:32:05.140]
When we do the inversion because we have this full coverage

[00:32:08.240]
that we get with dynamic data.

[00:32:10.630]
If a source is found to be on the edge of the data chip,

[00:32:15.340]
that piece of data that we use to do the inversion with,

[00:32:19.690]
our results aren’t going to be as good or as reliable.

[00:32:23.130]
But because we have that full coverage,

[00:32:25.000]
we can reposition that chip.

[00:32:29.960]
During that repositioning

[00:32:31.300]
we want to look to see if we’ve got data there

[00:32:34.080]
and it’s got some signal to it.

[00:32:36.230]
Like if it’s not gone into total background.

[00:32:41.100]
And we recommend there for your thresholds

[00:32:44.020]
that you use roughly about 50%

[00:32:46.130]
of the original target picking threshold.

[00:32:49.100]
But be careful, don’t go down into the noise.

[00:32:51.730]
If your site tends to be a little bit noisier

[00:32:53.670]
in your original target picking threshold

[00:32:56.040]
is close to a noise threshold,

[00:32:58.760]
you might not be able to actually go that full 50%

[00:33:02.460]
and we’ll need to modify that value.

[00:33:05.210]
But on hand, generally,

[00:33:06.720]
you can use 50% of your target picking threshold

[00:33:09.930]
for your repositioning threshold

[00:33:12.290]
when we’re doing the inversions.

[00:33:18.640]
So that’s a bit of a walkthrough

[00:33:19.990]
and some tips around the dynamic workflow.

[00:33:22.640]
Remember if you have any questions,

[00:33:24.170]
do enter them into the questions box

[00:33:25.860]
and we’ll respond to you after the webinar.

[00:33:32.200]
Next here, I’ll take a look at the static survey

[00:33:35.450]
or classification workflow.

[00:33:38.660]
For this workflow there’s really just two steps.

[00:33:41.910]
Prepare the data and classifying rank.

[00:33:45.260]
There is this step in the middle there,

[00:33:47.780]
construct and validate a site library.

[00:33:50.480]
This is something you generally just do at the beginning

[00:33:52.840]
or end of your project to make sure

[00:33:55.680]
that the library that you’re using

[00:33:57.410]
is a complete library for your site.

[00:34:01.670]
And we’ll come and look at that

[00:34:03.150]
notion of a complete library for your site

[00:34:05.090]
a couple points through this part of the presentation.

[00:34:10.310]
So here’s our a static workflow.

[00:34:12.950]
Much like my title that seemed to be,

[00:34:14.810]
today seem to be very long

[00:34:15.890]
and might look to be very complicated.

[00:34:18.090]
But again, I can sort of highlight for you

[00:34:20.810]
with the same closes on a general workflow.

[00:34:22.690]
Those must do points are shown there with those arrows.

[00:34:28.100]
And it’s really, you know, import some data, level it.

[00:34:33.180]
If it’s the first time through

[00:34:34.650]
and you haven’t got a library,

[00:34:39.241]
you should validate your library

[00:34:41.600]
and then it’s just classifying rank.

[00:34:43.340]
But for most of your project import level classify.

[00:34:50.480]
And it’s pretty much that simple.

[00:34:52.690]
So let’s go over to Oasis and I’ll show you just

[00:34:57.760]
exactly how simple that is.

[00:34:59.470]
I’ve created a project,

[00:35:01.290]
I’ve already imported my background data

[00:35:04.660]
and then now we will go in and import my survey data.

[00:35:14.470]
I’ll just turn off the filter here.

[00:35:16.910]
In this folder

[00:35:17.860]
I’ve got all of my different types of survey files.

[00:35:20.750]
We have SAM or static anomaly measurements.

[00:35:23.940]
We have other QC ones which I’ll touch on

[00:35:26.620]
a little bit later as QC sense of function tests.

[00:35:30.850]
And then my background measurements as it says,

[00:35:32.800]
I’ve already done those.

[00:35:34.650]
The import with the HDF import,

[00:35:38.400]
you can just give it all of your data

[00:35:41.270]
and we’ll figure it and put the data

[00:35:43.500]
into the right databases for you based on its data type.

[00:35:47.400]
I’ve just gone and selected for this demonstration here,

[00:35:50.700]
just for files, static anomaly measurements over some items.

[00:35:56.990]
So we can import those in.

[00:35:59.110]
You’ll get two databases.

[00:36:01.930]
A data database and a target database

[00:36:05.020]
and there’s a similar ones in the dynamic workflow.

[00:36:07.840]
The data database, tongue twister it is,

[00:36:10.660]
contains the actual transient or response data.

[00:36:13.940]
The target database contains this a list,

[00:36:17.470]
in this case of all the measurements that you made

[00:36:22.410]
and various sort of other parameters

[00:36:25.870]
of the sensor that was UBIT size windows,

[00:36:29.880]
other premises that we use that describe the sensor

[00:36:32.670]
that we then use in the inversion.

[00:36:36.400]
So once I brought it in, I need to level it.

[00:36:39.270]
Leveling it is removing the geology or background

[00:36:43.220]
or drift or the sensor out of the readings.

[00:36:49.950]
So we’re here.

[00:36:50.860]
I just need to pick my survey database

[00:36:54.780]
and my background database.

[00:36:57.090]
We find those based on the codes

[00:36:59.260]
that are there in those names.

[00:37:01.290]
And then we will just a couple options that we need to pick.

[00:37:05.550]
Most people will pick time and location.

[00:37:08.280]
You going to pick the nearest background reading

[00:37:12.110]
that was taken in an area that was free of metal objects

[00:37:17.220]
and subtract that from our measurement.

[00:37:21.030]
And we want to use the one that’s the closest in space,

[00:37:23.830]
most likely the same geology and the closest in time

[00:37:28.450]
to remove any drift.

[00:37:32.810]
And that will give us a new level channel.

[00:37:36.640]
And now we are ready to do our classifying rank.

[00:37:41.200]
I’ve already got a library. It’s a good library.

[00:37:44.520]
I’ve done my validate library work.

[00:37:46.200]
So I can just hit classifying rank.

[00:37:55.320]
We give it our database to use.

[00:37:59.920]
Is asking us what channels you may be using

[00:38:02.120]
as a mass channel.

[00:38:03.820]
A mass channel is a way that you can turn off

[00:38:06.700]
individual targets or flag or measurements and not use them.

[00:38:13.670]
So if you just wanted to do a sample or redo something,

[00:38:17.220]
this is a way that you could toggle that.

[00:38:20.520]
Some parameters about the sensor,

[00:38:24.930]
but the library database that we’re going to match to,

[00:38:27.690]
I can go onto the tools tab and these are all the individual

[00:38:31.100]
sort of steps that this tool

[00:38:35.020]
of classifying rank will go through.

[00:38:37.010]
We’ll invert your sources, do the library match skill,

[00:38:40.070]
look to see if there’s any self matches or clusters,

[00:38:43.900]
identify those and ultimately do that set the thresholds

[00:38:48.350]
and do our classification and prioritization.

[00:38:51.280]
We can also create several plots.

[00:38:54.540]
I’ll just do the one plot at the end of this to speed it up

[00:39:00.810]
for our demonstration today.

[00:39:05.630]
So there we go.

[00:39:06.463]
That’s going to take about two minutes to run

[00:39:11.820]
and just kind of go through all of those steps for us

[00:39:15.330]
and generate plots for each of those targets that

[00:39:21.215]
we read in.

[00:39:26.530]
But while that’s running,

[00:39:28.180]
we can just take a look at

[00:39:33.800]
a data set that I have already brought in,

[00:39:36.720]
much more targets and have ran through,

[00:39:44.910]
sorry, this one.

[00:39:48.750]
And this is after I’d ran my validate library workflow.

[00:39:54.110]
Is where this one is at.

[00:39:56.220]
And some tips of things that are good to look out on here.

[00:40:05.780]
We will load a series of channels, there’s lots of data

[00:40:09.330]
in the database table that you can look at.

[00:40:13.500]
That’s just my progress bar.

[00:40:15.350]
But two that are good to add and show here

[00:40:18.600]
are the ones that look for clusters.

[00:40:22.390]
So I can hit down our list,

[00:40:24.850]
drop down and there’s are two clusters, channels.

[00:40:30.911]
And you can see over in my plot

[00:40:32.840]
where you see a whole bunch of gray lines

[00:40:34.700]
on this one that I happened to be sitting on in behind.

[00:40:37.330]
And that’s the notion of a cluster that they are the same,

[00:40:41.480]
has the same set of polarizabilities.

[00:40:43.390]
In the EM sense, they look alike.

[00:40:48.420]
And during the validation stage,

[00:40:50.490]
you want to look for these unknown items and find those.

[00:40:53.540]
And one of the easy ways to do that

[00:40:56.480]
is to come to our tool here.

[00:41:01.950]
That just keeps popping up

[00:41:03.150]
and just do show a symbol profile.

[00:41:06.240]
And in your profile window, we’ll give you,

[00:41:11.410]
we create an ID for each of the unknown clusters

[00:41:15.430]
and in this window now we can sort of, you know, see those.

[00:41:20.000]
And so we can see this like three lines here if you would,

[00:41:24.700]
that these guys are all the same as this

[00:41:27.620]
’cause this is the ID for this cluster.

[00:41:30.110]
So is a cluster called number two.

[00:41:33.590]
And there’s a bunch of things that look like each other

[00:41:35.970]
that are in this cluster.

[00:41:37.880]
And I can flick between those.

[00:41:40.210]
I’m just going to move that to my second screen

[00:41:44.861]
and it will generate and bring up the plots for those.

[00:41:49.847]
And we can see, well, oops, didn’t click on it quite there.

[00:41:55.680]
There we go.

[00:41:56.513]
Click into the database and you can see it of those.

[00:41:59.420]
So as a time saving tip,

[00:42:02.170]
when you’re doing your validation of your library,

[00:42:09.770]
you can bring this up

[00:42:11.770]
and just do a simple plot of the cluster IDs

[00:42:15.990]
and see quite easily then any of the things

[00:42:19.050]
of the eye of the clusters that we can explain

[00:42:22.760]
in that they don’t have a strong match

[00:42:24.410]
to something in our library.

[00:42:27.380]
Maybe you want to look at, well, what did they match to?

[00:42:30.910]
And you could bring up one of the other plots

[00:42:32.700]
which I have shown up here in the top right.

[00:42:38.800]
But I could show those what they match to

[00:42:41.770]
and what the match metric,

[00:42:43.260]
how well those curves match to each other.

[00:42:46.300]
We come up with a quantitative value there.

[00:42:49.890]
I could load those,

[00:42:51.220]
but one of the key powers of using the tables,

[00:42:56.260]
because we have a lot of information

[00:42:57.890]
is to use the database views.

[00:43:01.270]
And I have gone prior to this and saved myself a view

[00:43:04.410]
that I would like to have and I can get that.

[00:43:07.640]
And just by simply coming in and loading my view,

[00:43:15.600]
I can do that same step that I did

[00:43:19.040]
and showed you manually a moment ago,

[00:43:20.940]
loading those cluster channels,

[00:43:22.930]
but I can also then go and load a whole bunch of

[00:43:26.440]
whatever other channels I would like to look at.

[00:43:28.980]
Here look where we’ve matched all three curves,

[00:43:33.530]
one, one, one.

[00:43:36.760]
Where we’ve just matched two of the curves are primary

[00:43:39.130]
and the secondary curve.

[00:43:41.810]
And what did they all match to.

[00:43:46.270]
From a real power user point of view

[00:43:48.170]
for you guys there that are advanced users,

[00:43:53.730]
these database of view files are text files.

[00:43:57.510]
You can go in and edit them.

[00:43:58.880]
There’s the one that I just loaded up.

[00:44:03.370]
When we run the bundles,

[00:44:04.800]
we give you an automatically load up a database view.

[00:44:10.380]
If you don’t like or continually want to add things

[00:44:14.250]
to the database view that we have,

[00:44:16.740]
you can go and edit that file.

[00:44:18.960]
And add your own channels that you would like to see

[00:44:22.370]
loaded in there.

[00:44:23.430]
And then every time you run classifying rank

[00:44:26.060]
or the validate library,

[00:44:29.040]
that series of channels will be loaded.

[00:44:36.310]
So at this point a moment ago, when we saw there was

[00:44:43.800]
progress priors stop are example that we did there.

[00:44:50.490]
We brought in in here has completed.

[00:44:52.890]
It’s found that those items match to

[00:44:56.470]
and we’ve got some pretty good match metrics.

[00:44:58.740]
And if I click on the item,

[00:45:01.910]
we will automatically bring up

[00:45:03.630]
and show us one of the plots.

[00:45:06.400]
And we can see the polarizabilities on here.

[00:45:11.670]
We can see the polarizabilities

[00:45:13.690]
and how well that they’ve matched.

[00:45:15.910]
We can see where the sensor was parked and laid out.

[00:45:19.330]
They parked really good,

[00:45:20.500]
right over top of the potential flag location

[00:45:25.010]
of where the source is

[00:45:26.080]
and it was found to be at that location.

[00:45:28.090]
And will show you some other plots

[00:45:29.480]
which I’ll come to in a moment.

[00:45:30.820]
There’s our size and decay friend again,

[00:45:33.000]
and this other one called the decision plot.

[00:45:37.180]
And I can just kind of go through and look at those.

[00:45:39.900]
This one didn’t match as quite as well.

[00:45:41.520]
You can see the polarizability is from our data

[00:45:43.987]
and this one with a little bit noisier,

[00:45:46.360]
and we didn’t get still matched to

[00:45:49.910]
a one 20-millimeter projectile,

[00:45:55.380]
but just compared to one of those first ones I looked at,

[00:45:58.160]
they were just a little bit more noisier.

[00:46:01.060]
And so you’d see that was just as easy as import data,

[00:46:05.270]
level it and run classifying rank.

[00:46:10.870]
When we want to come and look at our results,

[00:46:15.180]
that was the one we were looking at before.

[00:46:17.030]
Here’s where I have gone and ran a larger sample.

[00:46:21.820]
And you can see there’s our size and decay plot

[00:46:24.490]
that gives us that,

[00:46:25.670]
looking at that feature of those two properties

[00:46:28.420]
and where things might group.

[00:46:30.300]
They’re colored based on red, we think it’s a TOI,

[00:46:34.840]
target of interest.

[00:46:35.710]
Green, it’s a below some threshold that we set

[00:46:40.180]
of how well things match to library items.

[00:46:45.580]
And that’s shown here in a little,

[00:46:48.410]
what we call a decision plot.

[00:46:51.980]
We can bring other plots up

[00:46:56.330]
or any other images that you may have on

[00:46:58.500]
that you want to see in your data.

[00:47:00.600]
We can create what these plots

[00:47:03.100]
that we call interactive image of yours.

[00:47:05.500]
This could be any image that you have.

[00:47:07.601]
Some people like to do this when they do their dig surveys,

[00:47:11.440]
that they will use photographs of the items.

[00:47:15.150]
And creating a view as simply as looking at the images

[00:47:23.080]
that you’ve created.

[00:47:23.913]
So this is my folder of some images that I,

[00:47:28.780]
polarization plots that I created earlier.

[00:47:31.970]
Seeing what the name is there and generally

[00:47:34.370]
we’ve got everything there with a prefix.

[00:47:36.590]
And then the name of the item is the last part of that.

[00:47:43.580]
And in this case that name is referring to

[00:47:49.070]
its initial acquisition ID.

[00:47:52.290]
So I’m going to go find that in the list here

[00:47:57.370]
and then browse into that folder.

[00:48:04.780]
And we’ll look in that folder and try to figure out

[00:48:07.030]
whether you’ve got a prefix as in my case, or,

[00:48:10.087]
you know, maybe you’ve done a suffix on there

[00:48:15.090]
for naming all the images and then whether the PNG, bitmap,

[00:48:22.350]
JPEG, whatever and then you can load it.

[00:48:30.272]
And then when I click on any one of our sources,

[00:48:35.030]
that image will load up along with my other image.

[00:48:38.010]
I can see here and we can look at it

[00:48:41.120]
and help you do the review.

[00:48:43.060]
Now, maybe you’re like me,

[00:48:44.520]
and you’ve got all these over and you’d be like,

[00:48:46.320]
gee, it would be nice if I could get them to arrange

[00:48:50.310]
the way that I would like them to arrange.

[00:48:54.440]
And for that, you could go,

[00:48:56.100]
we have a tool that will let you save a window layout.

[00:48:58.960]
Earlier I saved a window layout

[00:49:01.980]
and I had one from a self to get started,

[00:49:05.200]
but then I also had one for my classifying rank.

[00:49:07.620]
So I can load that one.

[00:49:09.690]
Oh, wait.

[00:49:10.900]
Sorry, I clicked on the wrong button there.

[00:49:27.290]
So is easy as that.

[00:49:29.150]
Now my window’s all arranged

[00:49:31.480]
in a way that I would like them to be

[00:49:34.800]
and I can easily move through

[00:49:37.140]
and they will update as I go forward and look at the plots.

[00:49:43.030]
We also have other tools.

[00:49:44.710]
I have a size and decay plot on that

[00:49:48.326]
main documentation plot that I made,

[00:49:51.560]
but maybe I would like to have an interactive version

[00:49:55.540]
of one of those so that when I click on it,

[00:49:58.840]
things will change and we can open up

[00:50:01.210]
and create one of those.

[00:50:06.860]
You can load overlays onto these plots

[00:50:10.760]
to help you find some of the items

[00:50:13.580]
that you were looking for.

[00:50:14.570]
And I’ve made an overlay earlier.

[00:50:22.100]
You can color them up based on one of the parameters.

[00:50:27.340]
If you do that based on the category,

[00:50:30.120]
we’ve got already a color code pattern in there

[00:50:32.950]
that will load.

[00:50:35.940]
So you can use that and this is always interactive

[00:50:38.770]
so that when I click on one of these items,

[00:50:41.910]
on the scatter plot, my database will go to it.

[00:50:45.920]
And as long with the other plots

[00:50:47.680]
that I am looking at and reviewing.

[00:50:54.620]
So there’s just a little bit of a walkthrough.

[00:50:59.590]
And I’m looking at some features with the static workflow.

[00:51:05.530]
And I’ll just kind of, you know, go back to my slides here.

[00:51:10.220]
And there was a couple of other,

[00:51:11.957]
you could say, frequently asked questions

[00:51:13.880]
I get from people, how to create a library.

[00:51:18.330]
Well, the answer’s real simple.

[00:51:20.940]
Collect some data over a target of interest.

[00:51:25.580]
With UX-Analyze we include an example library,

[00:51:30.270]
and this is based off a assertive ESTC project.

[00:51:34.760]
And I’ve included the link there

[00:51:36.790]
if you’d like to know about it.

[00:51:38.180]
But you can see some examples from that project where

[00:51:40.840]
they just took a sensor.

[00:51:42.210]
In this case, a MetalMapper,

[00:51:43.820]
put it on some plastic book stands

[00:51:48.690]
and then placed items underneath it and collected some data

[00:51:52.520]
that we then use as part of our libraries.

[00:51:57.997]
And as I touched on it in the demonstration.

[00:52:01.070]
Well, once these unknown items that are not on my site,

[00:52:05.953]
sorry, are not in my library, that are on my site.

[00:52:09.290]
Part of the validate library tool

[00:52:11.140]
looks for clusters of similar items

[00:52:14.870]
and then identifies the ones for you

[00:52:16.970]
which we don’t have an explanation for

[00:52:20.360]
that not in our library.

[00:52:22.290]
Then it’s up to you to go out,

[00:52:23.680]
collect some ground truth.

[00:52:25.570]
We identify which one is the most similar.

[00:52:29.735]
If you’re going to dig any of them up,

[00:52:30.800]
at least dig that one up.

[00:52:32.850]
Figure out what it is

[00:52:34.280]
and then there’s a tool where we can

[00:52:35.607]
add the item to our library as either a target of interest

[00:52:39.600]
or not a target of interest.

[00:52:41.830]
And then it will be used in the classifications

[00:52:43.780]
appropriately.

[00:52:48.390]
We have a number of quality control tests

[00:52:51.720]
because we want to make sure that the data

[00:52:53.183]
that we base our classification decisions on

[00:52:56.240]
is of the highest quality possible.

[00:52:58.710]
You don’t have time to go into those today,

[00:53:00.860]
but these have been developed to prevent common issues

[00:53:04.870]
and have all come from user feedback.

[00:53:08.500]
And there’s some,

[00:53:10.020]
it says those are the three main ones there.

[00:53:15.980]
How do we know this stuff works?

[00:53:18.470]
Oh, wait, sorry.

[00:53:21.680]
Reporting.

[00:53:23.690]
As you saw there,

[00:53:25.640]
the database tables can be easily exported out of

[00:53:29.010]
Oasis Montaj and then used in your reports.

[00:53:33.090]
All the plots that we generate are also saved as PNG files

[00:53:37.203]
so that you can then easily print them off,

[00:53:40.170]
include them in an appendix if you want to do that

[00:53:43.845]
and to share them among project stakeholders.

[00:53:51.180]
So how do we know that these things work?

[00:53:53.900]
Well we can look at something called

[00:53:55.330]
a receiver operating characteristic curves.

[00:53:58.870]
For a number of the demonstration projects of being done

[00:54:01.970]
while testing this types of technology.

[00:54:04.340]
They went and dug up all the items on the site

[00:54:07.290]
to see how successful the classification was.

[00:54:11.310]
So just there,

[00:54:12.670]
we have a sort of schematic of a rank dig list.

[00:54:16.570]
Red items are high confidence TOI,

[00:54:21.970]
yellow, most likely TOI and go on and dig them up

[00:54:25.590]
and then green high confidence none TOI.

[00:54:29.070]
Perfect classification, we would see a nice L

[00:54:31.700]
as shown on the left there or in the middle.

[00:54:34.640]
And if we just did a random guess 50, 50 guess,

[00:54:38.350]
you’d get something like a 45-degree line

[00:54:42.700]
on a one of these plots.

[00:54:46.350]
There is a tool in UX-Analyze.

[00:54:48.910]
If you want to do this to help calculate or create these

[00:54:53.050]
receiver operating characteristic or rock curves.

[00:54:57.765]
And from a couple of examples of the demonstration projects.

[00:55:01.560]
You can see, not quite at that sort of,

[00:55:04.330]
you know, perfect L-shape, but we’re pretty close

[00:55:09.400]
on some of them.

[00:55:10.850]
Some of these don’t go through zero by the way,

[00:55:13.150]
is because they asked for examples.

[00:55:15.390]
I hear they went and dug up some items to help them learn

[00:55:22.150]
or train the classification system.

[00:55:26.330]
So that’s how we know that we can reliably go and detect

[00:55:31.870]
and classify munitions or targets of interest

[00:55:36.130]
based on a geophysical response.

[00:55:41.550]
So in summary, today we’ve taken a look at

[00:55:43.042]
what is Advanced Geophysical Classification,

[00:55:46.670]
how we can use the EM response

[00:55:48.490]
to determine intrinsic property of the source,

[00:55:50.963]
the polarizabilities.

[00:55:52.277]
And those polarizabilities can be used for classification,

[00:55:56.480]
both whether it’s just a simple

[00:55:58.110]
matching to a physical property,

[00:56:00.070]
like size, shape, wall thickness.

[00:56:02.330]
Or actually using the signature mapping.

[00:56:06.150]
At the beginning there we saw how using

[00:56:08.730]
a method of classification

[00:56:11.480]
and being able to eliminate the clutter items.

[00:56:13.700]
We could reduce our excavation costs

[00:56:16.460]
and saving time and money on our projects.

[00:56:20.730]
We took a little bit of a look at dynamic

[00:56:24.290]
using dynamic advanced EM data

[00:56:28.544]
that gives you an improved target detection

[00:56:29.790]
and positioning over the conventional sensors

[00:56:31.860]
but you can use conventional sensors for the dynamic phase

[00:56:37.040]
is perfectly okay.

[00:56:39.840]
You’ll find you just have a few more re-shorts

[00:56:42.160]
when you’re doing your static classification survey.

[00:56:46.010]
And it’s possible with the static survey data

[00:56:50.020]
to reliably classify sources of geophysical anomalies

[00:56:53.340]
whether it’s a target of interest or not.

[00:57:01.120]
If you’d like to learn more about

[00:57:02.590]
Advanced Geophysical Classification,

[00:57:04.310]
we’ve got a couple online resources for you.

[00:57:09.308]
Going to their, there’s some papers and other presentations

[00:57:13.150]
that we’ve done and other examples.

[00:57:16.680]
If you’d like training, contact your local office,

[00:57:19.810]
so we can do remote training these days

[00:57:22.290]
or at some point set up a in-person training.

[00:57:26.710]
There is some material online

[00:57:29.430]
where you can go and also register for

[00:57:31.760]
some of our other training sessions.

[00:57:34.720]
Of course there is our support line

[00:57:37.780]
which you can read through a direct email

[00:57:39.600]
or through our website.

[00:57:41.130]
So hopefully you found this interesting and useful,

[00:57:44.510]
and I’d like to thank you for your time today.

[00:57:49.270]
And there’s my email again if you didn’t catch it earlier,

[00:57:53.070]
and thanks for your time

[00:57:55.650]
and hope you have a nice day.

Free training and support

Grow your geoscience knowledge, connect with peers, and refresh your skills

Seequent Community

Discover new opportunities, connect with peers, and share insights and best practices with others in your field.

Visit Seequent Community

Learning Centre

Get started fast or upskill whilst working with our onboarding and advanced free eLearning courses.

Visit Seequent Learning Centre

Seequent Help

Find answers to your technical questions and get the most out of your product, with our in-depth product help information.

Visit Seequent Help