Skip to main content

This is a technical review of the RBF Interpolant tool aimed toward achieving robust and dynamic workflows in your numerical modelling.

We will take a deeper look at how Leapfrog Geo constructs a numerical model in order to build upon and provide a better understanding of the tools available at your disposal.

Duration

49 min

See more on demand videos

Videos

Find out more about Seequent's mining solution

Learn more

Video Transcript

[00:00:11.105]<v Suzzana>All right, hello and welcome everyone.</v>

[00:00:14.420]I make that 10 o’clock here in the UK.

[00:00:16.324]And so, let’s get started.

[00:00:19.231]Thank you for joining us today for this tech talk

[00:00:23.050]on Numeric Modeling elite for GA.

[00:00:26.373]By way of introduction, my name is Suzanna.

[00:00:29.290]I’m a Geologist by background based out of our UK office.

[00:00:33.624]And with me today are my colleagues,

[00:00:36.510]James and Andre both of him will be on hand to help

[00:00:39.930]run this session.

[00:00:41.658]Now, I run for this session is

[00:00:44.680]to refocus in on the interplant settings

[00:00:48.204]that will most effectively improve your numeric models

[00:00:52.240]from the word go.

[00:00:54.320]Let me set the scene for my project today.

[00:00:56.820]This is a copper gold porphyry deposit.

[00:00:59.817]And from the stats,

[00:01:01.325]we know that it is our early diorite unit

[00:01:06.920]that is a highest grade domain.

[00:01:10.560]For the purpose of today’s session,

[00:01:13.220]I’m going to focus on modeling the gold asset information,

[00:01:17.081]Some port from these drill holes, but in reality,

[00:01:20.650]I could choose to model any type of numerical value here,

[00:01:26.320]for example, our QT data.

[00:01:29.010]And if that is the workflow that you’re interested in,

[00:01:31.160]then I’d recommend you head over to the secret website,

[00:01:34.197]where there were a couple of videos

[00:01:36.560]on this topic already available.

[00:01:40.000]So coming down to my numeric models folder,

[00:01:44.520]I’m now going to select a new RBF interplant.

[00:01:51.500]And in the first window,

[00:01:53.840]start to specify my initial inputs and outputs.

[00:01:59.310]First, I’m going to want to specify

[00:02:02.130]some suitable numeric values.

[00:02:04.440]So in this case,

[00:02:05.720]I’m going to use my composited gold ISO table,

[00:02:10.407]but I can also choose to apply any applicable query filter.

[00:02:15.800]Now, in this case, as part of my initial validation steps,

[00:02:20.658]I actually created a new column in my collar table

[00:02:25.847]called valid simply to help me identify which drill holes

[00:02:32.300]I want to use in my modeling.

[00:02:35.017]I did this by first of all,

[00:02:37.380]creating a new category selection

[00:02:40.640]to make those initial selections

[00:02:43.320]then created a very simple query filter

[00:02:46.776]simply to exclude any of those that are invalid.

[00:02:51.900]And actually for the most part,

[00:02:53.830]it’s just this one hole at the bottom

[00:02:57.290]that I chose to exclude.

[00:03:00.960]As I work on my interpretation or indeed bring in more data,

[00:03:04.416]then it’s always useful to have flexibility

[00:03:07.000]like this built into your drill hole management workflow.

[00:03:10.310]So I would encourage you to set up something similar

[00:03:13.580]if not already done so,

[00:03:16.189]but for now though, let me just go back to my,

[00:03:20.660]I’ll be up in turbulent definition again,

[00:03:25.791]and let’s start to talk about the interplant boundary.

[00:03:32.950]Now, the default here is set to be our clipping boundary.

[00:03:40.850]And if you don’t know what that is,

[00:03:42.660]or if you want to set that,

[00:03:43.700]then you can actually do that from either the topography

[00:03:46.130]or indeed the GIS data folder.

[00:03:49.950]So I could choose to manually change this boundary

[00:03:54.360]or to enclose it around any object

[00:03:58.360]that already exists in my project,

[00:03:59.920]or indeed an existing model boundary or volume.

[00:04:05.960]Whenever I set here in my interpret boundary

[00:04:09.950]is going to be linked to the surface filter

[00:04:13.788]or by default is linked to the surface filter,

[00:04:17.350]which in itself controls which values can go into the model.

[00:04:22.920]I’m just going to bring a quick slice view into my scene,

[00:04:27.220]just to sort of help us visualize

[00:04:28.955]what I’m about to talk about.

[00:04:32.020]But I will explain this as this space

[00:04:33.180]is just a slice of my geological model.

[00:04:36.531]Just with those sort of gold values seen in a scene here.

[00:04:43.120]Typically your interplant boundary

[00:04:45.960]would be set to enclose the entire data set

[00:04:51.490]in which case all of your specified input values

[00:04:56.150]will become interpolated.

[00:04:57.570]So that’s kind of what we’re seeing in scene at the moment.

[00:05:00.150]It’s just a simple view of everything

[00:05:02.975]that’s in my data and in my project,

[00:05:08.380]I could choose however,

[00:05:09.746]to set a boundary around an existing volume.

[00:05:14.380]So for example,

[00:05:15.320]if I want to just specifically look at my only diorite

[00:05:18.320]Which is anything here in green,

[00:05:20.830]then I can choose to do that.

[00:05:23.550]And if I just turn that on now, then you can see,

[00:05:28.950]we’re just seeing that limited subset

[00:05:32.090]of input values in that case.

[00:05:35.187]Now, interestingly, if I wanted to add this point,

[00:05:39.230]mimic a soft boundary around the early diorite

[00:05:42.900]So for example, some sort of value buffer

[00:05:47.180]say it’s about 50 meters away,

[00:05:50.800]which is this orange line here.

[00:05:53.700]Then I could incorporate this also into my surface filter.

[00:06:00.300]And again, by doing, say,

[00:06:02.240]so if I now select this distance function here,

[00:06:08.260]then again, we’re going to see this update

[00:06:12.270]on which input values can be used into the model.

[00:06:17.610]But now though let’s not complicate things too much.

[00:06:20.610]I’m just simply going to use the same boundary

[00:06:25.330]as my geological model.

[00:06:28.870]And I’m also going to bring down my surface resolution

[00:06:33.196]to something a bit more reasonable.

[00:06:35.630]Now, a rule of thumb with drill hole data would be

[00:06:39.580]to set this to your composite length.

[00:06:42.690]So as to equal the triangulation

[00:06:45.752]or indeed some multiple of this,

[00:06:48.980]if you find the processing time takes too much of a hit

[00:06:52.752]for this particular data set,

[00:06:55.360]I have six meter lens composites already defined.

[00:06:59.169]So I’m going to bring surface ISO down

[00:07:02.800]to multiple of that 12 in this case.

[00:07:06.490]And of course those composites are there

[00:07:08.660]in order to help normalize

[00:07:10.400]and reduce the variance of my data.

[00:07:13.870]If I didn’t already have numeric complicit set up

[00:07:18.770]in my drill hole data folder,

[00:07:21.186]then I could actually go into find these here

[00:07:25.770]as well directly.

[00:07:28.608]But now though, let me just go back and use those,

[00:07:31.840]these composites that I have.

[00:07:35.180]And for now, let’s just say, okay to this and let it run.

[00:07:44.920]All right, now, there will be some times here

[00:07:47.884]that it’s going to be easier just to jump into models

[00:07:51.179]that have already been set up.

[00:07:52.817]And I think this is one of those cases.

[00:07:54.940]So let me just bring in that is going to generate,

[00:08:02.120]which is essentially our first pass gold numeric model.

[00:08:09.980]And if I stick the legend on,

[00:08:12.330]then you’ll start to get idea

[00:08:14.030]or reference point for that grade.

[00:08:17.390]Now, the power of leap frogs on behalf energy engine

[00:08:21.680]is in its ability to estimate a quantity

[00:08:25.780]into an unknown point by using known drill hole

[00:08:29.760]or point data.

[00:08:31.830]It’s important however,

[00:08:33.540]when we estimate those unknown points

[00:08:36.770]that we’re selecting an interpolation function,

[00:08:40.383]that will make the most geological sense

[00:08:44.730]at this first pass stage.

[00:08:47.220]I hope you’ll agree that we’re seeing something

[00:08:49.732]that’s very unrealistic,

[00:08:52.110]especially in regards to this higher grade sort of blow outs

[00:08:56.520]in my Northwest corner.

[00:09:00.680]Is fairly common for our projects

[00:09:03.970]to have some areas of data scarcity.

[00:09:07.320]If I bring my drill holes in here

[00:09:11.086]and just turn the filter off for a second,

[00:09:17.330]then I’m sure you agree that that sometimes at depth

[00:09:21.639]or indeed on our last full extents,

[00:09:25.280]that we might just not have any drilling information

[00:09:29.060]in that area.

[00:09:31.740]And in this case,

[00:09:32.960]it is the, it’s just three drill holes here.

[00:09:36.852]If I put my filter on,

[00:09:42.370]you can see that it’s just these three drill holes

[00:09:45.241]that is simply causing that interpolation here

[00:09:49.640]to become really quite extrapolated.

[00:09:53.340]And unfortunately there’s many an extreme example in line

[00:09:56.620]of such models.

[00:09:57.660]The C’s finding their way into company reporting.

[00:10:02.220]So what I would encourage you all to do,

[00:10:04.890]is to of course start refining the internal structure

[00:10:07.890]of my model and the majority of the functionality

[00:10:11.290]to do so actually sits under the interplant tab.

[00:10:16.590]So if I go to the model

[00:10:17.640]and I’ll go to the one that’s run individually,

[00:10:22.520]so just double click into it

[00:10:24.260]and I’m going to come to the interplant tab.

[00:10:29.150]For now, I’m simply going to change my interplant type

[00:10:34.640]to a splital and change my drift function to known

[00:10:41.240]I will come back to this in more detail, but for now,

[00:10:43.720]let me just let that run and we’ll have a look

[00:10:47.380]at what that produces instead.

[00:10:50.960]And again, great.

[00:10:52.220]Here’s one I prepared earlier,

[00:10:58.550]which we can now see,

[00:11:00.910]and hopefully you can already see

[00:11:02.800]this big notable difference already,

[00:11:05.630]especially where a high grade

[00:11:10.130]or where that original high-grade was blown out.

[00:11:15.180]This time, if I bring my drill holes back in again,

[00:11:21.870]we can see that high grade interpolation

[00:11:25.677]around those three drill hole still exists,

[00:11:29.310]but just with a much smaller range of influence,

[00:11:34.040]but why is that the case to help answer that question,

[00:11:38.250]I’m going to try and replicate these RBF and parameters

[00:11:42.030]on a simple 2D grid of points.

[00:11:45.290]And this should quite literally help connect the dots

[00:11:49.850]in our 3D picture.

[00:11:52.020]So let me move from this for a second

[00:11:54.410]and just bring in my grid of points

[00:11:59.070]and a couple of arbitrary samples.

[00:12:02.760]Hopefully you can see those values starting to come through.

[00:12:11.340]Here we go.

[00:12:14.520]So what I’ve done so far for this is,

[00:12:17.820]I have created three new RBF models

[00:12:24.040]in order to estimate these six points shown on screen here.

[00:12:29.140]And then I will come back into each of the Interplant tabs

[00:12:34.728]in order to adjust the interplant interests settings.

[00:12:41.600]And that’s just what the naming refers to here.

[00:12:47.380]Now, leap for geezers two main interplant functions,

[00:12:52.670]which in very simple terms will produce different estimates

[00:12:57.530]depending on whether the distance

[00:12:59.190]from our known sample points,

[00:13:01.308]the range is taken into consideration.

[00:13:06.019]And linear interplant will simply assume

[00:13:10.360]that any known values closer to the points

[00:13:13.840]you wish to estimate

[00:13:15.550]we’ll have a proportionally greater influence

[00:13:18.610]than any of those further away.

[00:13:22.040]A splital interplant on the other hand

[00:13:25.480]assumes that there is a finite range or limit

[00:13:30.160]to the influence of our known data

[00:13:32.700]beyond which this should fall to zero.

[00:13:36.141]You may recognize the resemblance here

[00:13:39.130]to a spherical very grim,

[00:13:42.060]and for the vast majority of metallic ore deposits,

[00:13:45.960]this interpolation type has more applicable

[00:13:49.430]exceptions to this, maybe any laterally extensive deposit

[00:13:53.240]like coal or banded iron formations.

[00:13:57.670]So in addition to considering the interpolation method,

[00:14:00.870]we must always also decide how best to control

[00:14:05.510]our estimation in the absence of any data.

[00:14:09.308]In other words,

[00:14:10.680]how should our estimation behave

[00:14:13.510]when we’re past the range of our samples,

[00:14:17.490]say in this scenario we saw just a minute ago

[00:14:20.540]with these three drill holes,

[00:14:23.610]but this, we need to start defining an appropriate drift

[00:14:27.140]from the options available.

[00:14:29.240]And I think that was the point that I sort of got into

[00:14:33.300]looking at on my grid.

[00:14:35.340]So at the moment I have a linear interplant type

[00:14:40.670]with a linear drift shown

[00:14:43.520]and much kind of like the continuous column bludgeoned

[00:14:46.890]than I have up here,

[00:14:48.470]we’re seeing a linear,

[00:14:49.560]a steady linear trend in our estimation.

[00:14:54.130]The issue is, it was the estimation around

[00:14:57.660]our known data points is as expected.

[00:15:00.497]The linear drift will enable values

[00:15:03.690]to both increase past the highest grade.

[00:15:08.420]So in this case,

[00:15:09.253]we’re sort of upwards to about a grade of 13 here,

[00:15:13.937]as well as go into the negatives past the lowest grade.

[00:15:21.700]So for great data of this nature,

[00:15:25.260]we’re going to want to reign that estimation in

[00:15:29.010]and start to factor in a range of influence.

[00:15:34.190]So looking now at our splital interplant

[00:15:37.931]with a drift of known,

[00:15:41.130]we can see how the range starts to have an influence

[00:15:44.430]on our estimation and then when we move away

[00:15:48.820]from our known values.

[00:15:50.780]So for example, if I start to come out

[00:15:52.770]onto the extents here, then the estimation is falling

[00:15:57.910]or decaying back to zero.

[00:16:00.790]And that will be the same

[00:16:02.180]if I sort of go around any of these,

[00:16:05.000]we would expect to be getting down to a value as zero

[00:16:09.338]away from any data points.

[00:16:13.760]And of course, if you are close to the data point,

[00:16:16.170]we would expect it to be an estimation similar.

[00:16:21.700]So where you’re modeling an unconstrained area,

[00:16:26.290]or perhaps don’t have many low grade holes

[00:16:29.770]to constrain your deposit,

[00:16:31.811]using a drift of known will ensure

[00:16:34.890]that you have some control on how far your samples

[00:16:38.630]would have influenced.

[00:16:40.410]That said, if you are trying to model something constraints,

[00:16:44.500]for example, the early diorite domain that we saw earlier,

[00:16:49.450]then using a drift function of constant

[00:16:54.850]could be more applicable.

[00:16:56.588]In this case, our values are reverting

[00:17:00.480]to the approximate mean of the data.

[00:17:03.030]So if I just bring up the statistics our mean here is 4.167,

[00:17:12.010]which means that as I’m getting to the outskirts,

[00:17:14.910]I would expect it to be coming back towards that mean.

[00:17:18.960]So a few different options, but of course,

[00:17:22.440]different scenarios in how we want to apply these.

[00:17:26.240]Now, if I jump back now to my gold model,

[00:17:35.070]then let’s start first of all,

[00:17:36.580]just with a quick reminder of where we started,

[00:17:41.280]which was with this model,

[00:17:44.170]and this is applying that default linear interplant,

[00:17:49.530]and that simply by changing this already

[00:17:54.030]to they splital or interplant type

[00:17:58.360]along with a drift of known,

[00:18:02.010]then we’re starting to see something

[00:18:04.330]that makes much more sense.

[00:18:08.791]So if I go back now and go back to my interplant tab

[00:18:15.870]to sort of look at some of these other settings.

[00:18:20.924]So far we know we want to limit the influence

[00:18:27.030]of our known data to a certain distance,

[00:18:30.820]and that it’s the distance reign

[00:18:33.020]essentially controlling that correlation.

[00:18:37.230]It’s reasonable, therefore that you’re going to want to,

[00:18:40.110]you’re going to want to change the base range

[00:18:42.610]to something more appropriate.

[00:18:44.720]And if you don’t know where to start,

[00:18:46.170]then a rule of thumb could be around twice

[00:18:49.680]the drill hole spacing.

[00:18:51.240]So in this case, let’s, that’s around 700,

[00:18:55.170]which is appropriate for this project.

[00:18:59.350]We also want to consider our nuggets and in leapfrog,

[00:19:04.000]this is expressed as a percentage to our some.

[00:19:09.070]Increasing the value of the nugget will create smoother

[00:19:13.640]results by limiting the effects of extreme outliers.

[00:19:18.660]In other words, we would give more emphasis

[00:19:21.972]to the average grades of our surrounding values

[00:19:26.190]and less on the actual data point.

[00:19:29.490]It can basically help to reduce noise

[00:19:32.330]caused by these outliers

[00:19:34.630]or with inaccurately measured samples.

[00:19:39.090]What value to use here is very much decided on a deposit

[00:19:43.420]by deposit case.

[00:19:44.780]And by all means, if you or someone else

[00:19:48.500]in your organization has already figured this out

[00:19:52.030]for your deposit, then by all means apply it here,

[00:19:55.570]perhaps for a gold deposit like this.

[00:19:58.970]Then a rule of thumb will be,

[00:20:02.280]let’s say, 20 to 30% of the sales.

[00:20:05.410]So let’s just take this down.

[00:20:07.170]For example, it’s 0.2, which is going to be 20% of that.

[00:20:10.793]So I have 10% might be more appropriate

[00:20:14.170]for other metallic deposits or indeed known,

[00:20:17.400]if you have a very consistent data set,

[00:20:21.560]we’ve spoken about drift already,

[00:20:25.220]but now that we are pressed the range of influence,

[00:20:29.430]then it’s really our drift that comes into play.

[00:20:33.520]Using known in this case is going to help

[00:20:37.220]control how far my samples will have an influence

[00:20:41.409]before the estimation to case back to Siri.

[00:20:46.329]I’m not going to delve too much

[00:20:50.140]into the other settings here.

[00:20:51.850]You can always read about these

[00:20:53.410]to your heart’s content online,

[00:20:56.250]but the take home point really is that,

[00:20:58.720]it’s the interplant the base range,

[00:21:03.320]that percentage nuggets to the sill and drift

[00:21:07.840]that will have the most material effects

[00:21:10.480]on your numeric model.

[00:21:13.400]Get these rights and you should be well on your way

[00:21:16.030]to producing a very best model

[00:21:17.970]that makes the best of your geological knowledge.

[00:21:21.060]All right, now we’ve talked a little bit about

[00:21:24.460]these interplant settings.

[00:21:27.310]I thought it would be worth a quick run over

[00:21:30.706]on the ISO surfaces.

[00:21:33.750]So we can see all these,

[00:21:35.860]as we expand out our new model objects,

[00:21:39.530]we’ve got ISO surfaces

[00:21:41.520]and then of course the resulting output volumes.

[00:21:45.420]When we pick cutoffs in our outputs tab,

[00:21:50.020]we are simply asking leap frog to define some boundaries

[00:21:54.047]and space and your contours,

[00:21:56.990]where there is the same consistent grade.

[00:22:00.520]Essentially the iso surfaces perform the exact same task

[00:22:05.910]as our geological surfaces do

[00:22:08.696]when we’re geologically modeling.

[00:22:10.776]But instead of building these surfaces ourselves,

[00:22:14.690]we’re simply picking grade boundaries

[00:22:17.051]that we want to see and leapfrog we’ll go

[00:22:20.330]and build them for us.

[00:22:23.250]It is this contour surface that,

[00:22:26.343]that we’re seeing here in 3D.

[00:22:29.770]And if we want to visualize a specific value,

[00:22:34.020]then we can do so by updating the ISO values here

[00:22:40.760]before I do that, let me just jump back

[00:22:44.850]into the grid of points,

[00:22:46.980]just to highlight this on a sort of simpler dataset.

[00:22:53.947]So back in with my grid of points, let’s say for instance,

[00:22:58.810]I now want to visualize a contour surface

[00:23:03.350]with a greater two, then I can replicate that here

[00:23:07.612]simply by setting some appropriate value filters.

[00:23:12.049]And we can see that now

[00:23:14.520]sort of finding all of these sort of values of two in space.

[00:23:18.746]And these it’s this point that of course

[00:23:22.840]is going to become our contour.

[00:23:27.900]For the hikers amongst us

[00:23:29.401]who are maybe used to seeing elevation contour lines

[00:23:32.550]on a TD map.

[00:23:33.850]This is the same principle,

[00:23:35.492]except in the case of our gold model.

[00:23:38.180]We’re simply looking at a value representation in 3D.

[00:23:43.060]So let me go back to my gold model once again

[00:23:49.690]and actually set some better ISO values for this deposit.

[00:23:57.710]So am just going to go double click back in

[00:23:59.960]and head over to the output tab.

[00:24:02.870]So let’s just set a few, make a little bit more sense

[00:24:06.670]than the ones that are put in as default.

[00:24:10.900]So let’s say 0.5, 0.75.

[00:24:16.350]Let’s do 1.25 and at 1.5, I think we’ll suffice.

[00:24:27.970]The other thing that I’m going to do

[00:24:29.640]is take down the resolution on my higher grades

[00:24:33.990]to try and be as accurate as possible.

[00:24:38.190]And remembering to know you might not have heard it,

[00:24:41.360]but remembering my much earlier points

[00:24:43.650]about matching this where possible

[00:24:46.010]to the drill hole complicit lens.

[00:24:48.500]So that’s why I’m sort of picking the six here.

[00:24:53.500]There’s also the ability in this tab to clump output values,

[00:24:59.740]the default assumption is that

[00:25:02.610]nothing should fall low zero,

[00:25:04.320]which is why we’re seeing that clump,

[00:25:07.151]but you can always change that if need be likewise,

[00:25:11.440]you can set different lower,

[00:25:14.710]and indeed upper bounds in the value transform tab

[00:25:19.280]to cap any values you may deem too low or too high,

[00:25:23.360]but I’m not going to focus on that for now.

[00:25:28.110]So let me just let that run

[00:25:37.030]and again, in the magic of eight forklifts,

[00:25:42.510]so have a look at, so what that has produced,

[00:25:46.940]and again, we’re starting to see something visually

[00:25:50.630]that perhaps makes more sense.

[00:25:52.710]That makes more sense to us and to our gold boundaries.

[00:25:58.990]Now, if I want to see more connectivity

[00:26:03.010]between my drill holes, then aside from the base range,

[00:26:07.550]I could use a trend for this.

[00:26:09.630]Also the central shear zone, for example,

[00:26:15.720]is no doubt playing a dominant structural control

[00:26:21.870]on my mineralization.

[00:26:23.750]Starting to see a lot of our high grades,

[00:26:25.490]sort of following some sort of trend here.

[00:26:29.370]So applying a structural trend should help

[00:26:32.400]to account for any changes in the strength

[00:26:36.409]and direction of continuity along this ridge.

[00:26:41.330]I’ve already defined a structural trends for the shear zone.

[00:26:46.230]I sort of put it as something like this for now.

[00:26:51.130]However we could have caused always update this as need be.

[00:26:55.410]Either way, this is a typical step to take.

[00:26:58.640]Given that nature is very rarely isotropic.

[00:27:03.530]So let me go and apply this structural trends to my model

[00:27:11.960]and which I can do from my trend tab.

[00:27:17.790]And I’ve only got the one in my project at the moment.

[00:27:20.220]So it’s just that stretch or trend,

[00:27:25.240]and let’s just let that run.

[00:27:34.000]And again, in a fantastic here’s one I prepared earlier,

[00:27:43.430]we can see what that has done

[00:27:45.370]this definitely you can see how that trend has changed

[00:27:49.870]there is weighting of points in space

[00:27:52.780]to define that continuity along that ridge.

[00:27:57.550]So though powerful and usually they’ll be applicable

[00:28:01.560]certainly in our sort of metallic gold

[00:28:05.776]or indeed any deposit.

[00:28:09.970]So now that we’ve sort of run to this point,

[00:28:14.870]it’s likely at this stage

[00:28:17.030]that we’re going to want to hone in our mineralize domains.

[00:28:21.590]I spoke earlier about the early diorites unit

[00:28:27.780]having our containing our highest mean gold grade.

[00:28:31.650]So let’s come full circle and create a numeric model

[00:28:36.439]for just this for you also.

[00:28:41.430]So let me clear the scene and let me cheat

[00:28:48.370]by copying my last model

[00:28:51.640]with all of its parameters into the new one.

[00:29:05.230]And now we can go into that copy of our model

[00:29:11.880]and start to apply reasonable premises here as well.

[00:29:16.750]Now, the first thing we’re going to want to do is of course,

[00:29:20.060]set some boundary at the moment

[00:29:21.470]it’s just the exact same as the last one,

[00:29:25.900]which is the extent of my geological model.

[00:29:30.760]Say, let’s go in and set a new lateral extent.

[00:29:34.830]And what I’m going to do is use the volume

[00:29:37.971]for my geological model as that extends.

[00:29:41.290]Say from surface, and then under my geological models

[00:29:48.370]under the output volumes,

[00:29:50.200]I’m going to select that early diorite.

[00:29:57.230]Now, whilst that’s running.

[00:29:59.920]Remember that what that is going to do

[00:30:02.960]is first of all, constrain your model

[00:30:06.710]to that unit of interest.

[00:30:08.340]So this is my early diorite.

[00:30:12.530]And it’s also just only going to use the values

[00:30:22.910]that refer to this unit,

[00:30:26.320]what to do with that initial surface filter set up.

[00:30:29.700]We then of course,

[00:30:30.533]want to go and review our interplant settings here

[00:30:34.120]to check that they’re still applicable.

[00:30:35.693]Now that we’ve constrained our data to one domain.

[00:30:40.740]Let me say, let me go back and double click into this model.

[00:30:46.240]And let’s just double check once again,

[00:30:47.788]that our interplant settings make sense.

[00:30:52.570]It’s probably going to be more applicable in this case

[00:30:58.635]that where we have an absence of data.

[00:31:01.290]So again, on the outskirts, on the extents of our model,

[00:31:08.460]that perhaps we’re going to want to revert

[00:31:10.730]to the mean of the grade in the absence

[00:31:13.610]of any other information.

[00:31:16.110]In which case, the best thing that we can do

[00:31:19.030]is simply to update the drift here to constant.

[00:31:23.890]Hopefully you remember from that grid of points,

[00:31:26.570]how that is always going to revert to the mean of the data.

[00:31:31.200]I think for now,

[00:31:32.740]I’m just going to leave every other setting the same,

[00:31:35.460]but of course we could come in

[00:31:37.630]and start to change any of these,

[00:31:41.270]if we wish to see things a little bit differently,

[00:31:44.340]but for now, it’s that drift that for me

[00:31:47.070]is the most important.

[00:31:48.570]So again, let me rerun and bring in a final output.

[00:32:02.240]And if I just make these a little less transparent,

[00:32:06.520]you can hopefully see now

[00:32:08.230]that whereas before we had sort of that those waste grades

[00:32:14.230]coming in on this sort of bottom corner,

[00:32:17.260]we’re now reverting to what?

[00:32:19.500]To something similar of the mean and I think in this case,

[00:32:23.330]if I remember rightly the mean is yeah, 1.151.

[00:32:29.272]So we would expect to be kind of up in that sort of darker,

[00:32:33.130]darker yellow colors.

[00:32:38.020]At this point,

[00:32:39.620]I’m going to say that I’m reasonably happy

[00:32:42.230]with the models I have.

[00:32:44.720]I would, of course, want to interrogate these lot further.

[00:32:47.900]Maybe make some manual edits

[00:32:50.502]or double check some of the input data,

[00:32:53.578]but for the purpose of this talk,

[00:32:55.490]I hope that this has gone some way

[00:32:57.910]in highlighting which interplant settings

[00:33:01.090]will most effectively improve your numeric models.

[00:33:04.984]I appreciate that this is a very extensive topic

[00:33:08.450]to try and fit into a shorter amount of time.

[00:33:10.550]Please do keep your questions coming in.

[00:33:14.090]In the meantime,

[00:33:15.130]I’m going to hand over to James now,

[00:33:19.030]to run us through the indicator RBF interplant tool.

[00:33:24.550]<v James>Thanks, Suzanna,</v>

[00:33:25.730]bear with me two seconds and I’ll just share my screen.

[00:33:28.850]So a lot of the stuff that Suzanna has run through

[00:33:32.350]in those settings, we’re going to apply now

[00:33:34.830]to the indicator numeric models

[00:33:38.520]indicating numeric models are a tool

[00:33:41.530]that is often underused or overlooked in preference

[00:33:46.220]to the RBF models,

[00:33:47.810]but indicator models can have a really valuable place

[00:33:51.480]in anyone’s workflow.

[00:33:53.870]So I’ve got the same project here,

[00:33:55.910]but this time I’m going to be looking at

[00:33:58.050]some of the copper grades that we have in the project.

[00:34:02.120]If I go into a bit of analysis initially on my data,

[00:34:06.310]I can see that when I look at the statistics for my copper,

[00:34:09.968]there isn’t really a dominant trend to the geology

[00:34:16.350]and where my copper is hosted is more of a disseminated

[00:34:19.879]mineralization that is spread across multiple domains.

[00:34:24.480]If I come in and have a look at the copper itself,

[00:34:27.600]what I want to try and understand

[00:34:30.170]is what it would be a good cutoff to apply

[00:34:34.740]when I’m trying to use my indicator models

[00:34:37.580]and this case here,

[00:34:38.650]I’ve had a look at the histogram of the log,

[00:34:42.470]and there’s many different ways

[00:34:44.380]that you can approach cutoffs.

[00:34:46.954]But in this case, I’m looking for breaks

[00:34:49.440]in the natural distribution of the data.

[00:34:51.640]And for today, I’m going to pick one around this area

[00:34:54.600]where I see this kind of step change

[00:34:56.593]in my grade distribution.

[00:34:59.410]So at this point here,

[00:35:00.900]I can see that I’m somewhere between 0.28 and 0.31% copper.

[00:35:06.770]So for the purpose of the exercise,

[00:35:09.950]we’ll walk through today we’ll use 0.3 as my cutoff.

[00:35:15.051]So once I’ve had a look at my data and I have a better idea

[00:35:18.210]of the kind of cutoffs

[00:35:19.660]I want to use to identify mineralization,

[00:35:23.020]I can come down to my numeric models folder,

[00:35:26.720]right, click to create a new indicator.

[00:35:30.710]I’ll be up in turbulent.

[00:35:34.186]The layout to this is very similar to the numeric modeling.

[00:35:39.010]Again, you can see,

[00:35:39.843]that I can specify the values I want to use.

[00:35:43.050]I can pick the boundaries I want to apply.

[00:35:45.090]So for now, I’m just going to use the title project

[00:35:48.700]in all my data.

[00:35:51.420]And if I want to apply my query filters,

[00:35:53.730]I can do that here as well.

[00:35:56.086]A couple of other steps I need to do,

[00:35:58.670]because this is an indicator interplant.

[00:36:00.820]I need to apply a cutoff.

[00:36:02.730]So what I’m doing here is

[00:36:03.590]I’m trying to create a single surface

[00:36:05.960]in closing grades above my cutoff of no 0.3.

[00:36:11.840]And for the purpose of time,

[00:36:13.240]I’m also going to just composite my data here.

[00:36:16.490]So that it helps to run,

[00:36:18.710]but it also helps to standardize my data.

[00:36:22.770]So I’m going to set my composite length to four

[00:36:26.400]and anywhere where I have residual links, less than one,

[00:36:29.240]I’m going to distribute those equally back through my data

[00:36:33.950]can give it a name to help me identify it.

[00:36:37.350]And we’re going to come back

[00:36:38.500]and talk about the ISO values in a minute.

[00:36:43.520]So for now, I’m going to leave my ISO value

[00:36:45.338]as a default of no 4.5, and then I can let that run.

[00:36:52.460]So what leapfrog does with the indicators

[00:36:55.580]is it will create two volumes.

[00:36:58.090]It creates a volume that is considered inside my cutoff.

[00:37:02.150]So if I expand my model down here,

[00:37:06.440]we’ll see there’s two volumes as the output.

[00:37:08.800]So there’s one that is above my console,

[00:37:11.960]which is this volume.

[00:37:12.960]And one that is below my cutoff,

[00:37:15.890]which is the outside volume.

[00:37:19.775]Now, at the moment,

[00:37:20.690]those shapes are not particularly realistic.

[00:37:23.560]Again, very similar to what we saw

[00:37:25.330]with Suzanna’s explanation initially

[00:37:28.720]because of the settings I’m using,

[00:37:30.100]I’m getting these blow outs to the extent of my models.

[00:37:35.020]The other thing we can have a quick look at

[00:37:36.388]before we go and change any of the settings

[00:37:38.940]is how leapfrog manages the data

[00:37:42.550]that we’ve used in this indicator.

[00:37:45.050]So it’s taken all of my copper values

[00:37:46.850]and I’ve got my data here in my models.

[00:37:49.340]So if I track that on,

[00:37:54.503]just set that up so you can see it.

[00:37:57.690]So initially from my copper values,

[00:38:01.310]leapfrog will go and flag all of the data

[00:38:04.230]as either being above or below my cutoff.

[00:38:07.060]So here you can see my cutoff.

[00:38:09.270]So everything above or below

[00:38:12.420]it’s then going to give me a bit of an analysis

[00:38:15.870]around the grouping of my data.

[00:38:18.130]So here you can see that it looks at the grades

[00:38:22.570]and it looks at the volumes it’s created,

[00:38:25.060]and it will give me a summary of samples

[00:38:28.870]that are above my cutoff and fall inside my volume,

[00:38:32.477]but also samples that are below my cutoff

[00:38:35.080]that are still included inside.

[00:38:37.510]So we can see over here,

[00:38:38.530]these green samples would be an example

[00:38:43.010]where a sample is below my cutoff,

[00:38:45.977]but has been included within that shell.

[00:38:49.360]And essentially this is the equivalent

[00:38:51.200]of things like dilution.

[00:38:52.110]So we’ve got some internal dilution of these waste grades

[00:38:55.130]and equally outside of my indicator volume,

[00:38:59.210]I have some samples here that fall above my cutoff grade,

[00:39:04.290]but because they’re just isolated samples,

[00:39:06.240]they’ve been excluded from my volume.

[00:39:08.610]So with that data and with that known information,

[00:39:13.800]I then need to go back into my indicator

[00:39:16.160]and have a look at the parameters and the settings

[00:39:18.671]that I’ve used to see if they’ve been optimized

[00:39:20.910]for the model I’m trying to build.

[00:39:23.820]What we know with our settings,

[00:39:25.880]and again, when we have a look at the inside volume,

[00:39:29.840]if I come into my indicator here,

[00:39:32.327]and go and have a look at the settings I’m using,

[00:39:35.726]then I come to the interplant tab.

[00:39:39.260]And for the reasons that Suzanna has already run through,

[00:39:42.740]I know when modeling numeric data

[00:39:45.950]and particularly any type of grade data,

[00:39:48.020]I don’t want to be using a linear interplant.

[00:39:50.544]I should be using my splital interplant.

[00:39:54.157]The other thing I want to go and look at change in is that

[00:39:57.380]I don’t have any constraints on my data at the moment.

[00:40:00.410]So I’ve just used the model boundaries.

[00:40:02.950]So I want to set my drift to known.

[00:40:06.410]So again, there’s more of a conservative approach

[00:40:09.060]as I’m moving away from data

[00:40:10.207]and my assumption is my grade is reverting to zero.

[00:40:16.860]We can also come and have a look at the volumes.

[00:40:20.090]So currently our ISO value is five.

[00:40:23.020]So we’ll come and talk about that in a second.

[00:40:25.340]And the other good thing we can do here is

[00:40:26.920]we can discard any small volumes.

[00:40:31.130]So as an example of what that means,

[00:40:33.670]if I take a section through my projects,

[00:40:40.000]we can see that as we move through,

[00:40:41.440]we get quite a few internal volumes

[00:40:44.910]that aren’t really going to be of much use to us.

[00:40:47.920]So I can filter these out based on a set cutoff.

[00:40:52.690]So in this case,

[00:40:53.523]I’m excluding everything less than 100,000 units,

[00:40:57.540]and I can check my other settings to make sure

[00:40:59.450]that everything else I’m doing,

[00:41:00.720]maybe I used the typography is all set up to run how I want.

[00:41:05.480]Again, what Suzanna talked about,

[00:41:07.018]was there any time that you you’re modeling numeric data

[00:41:11.055]best practice would be to typically use

[00:41:14.180]some form of trend to your data

[00:41:16.560]so I can apply the structural trend here

[00:41:18.460]that’s Susanna had in hers, and I can let that one run.

[00:41:24.400]Now, when that’s finished running,

[00:41:27.700]I’ve got my examples here.

[00:41:31.800]So if I load this one on,

[00:41:36.320]I can see the outline here of my updated model

[00:41:39.240]and how that’s changed the shape of my indicator volume.

[00:41:44.810]So if we come back out of my section, put these back on,

[00:41:54.160]I can see by applying those,

[00:41:56.410]simply by applying those additional parameters.

[00:41:58.270]So changing my interplant type from linear

[00:42:01.060]to splital adding a drift of known

[00:42:03.510]and my structural trend,

[00:42:06.830]I’ve got a much more realistic shape now

[00:42:08.620]of what is the potential volume of copper

[00:42:12.460]greater than 0.3%.

[00:42:17.320]Now, the next step into this

[00:42:19.250]is to look at those ISO values

[00:42:20.930]and what they actually do to my models,

[00:42:23.900]the ISO value and if we come into the settings here,

[00:42:28.322]so I’m just going to open up the sentence again.

[00:42:31.030]You can see at the end, I can set an ISO value.

[00:42:34.140]This is a probability of how many samples within my volume

[00:42:39.280]are going to be above my cutoff.

[00:42:41.830]So essentially if I took a sample

[00:42:44.380]anywhere within this volume,

[00:42:46.230]currently there is a 50% chance that,

[00:42:48.690]that sample would be above my no 0.3% copper cutoff.

[00:42:53.871]So by tweaking those indicator ISO values,

[00:42:58.860]you actually can change the way the model is being built

[00:43:01.760]based on a probability factor of how many,

[00:43:07.110]what’s the chance of those samples

[00:43:08.940]inside being above your cutoff.

[00:43:12.020]So I’ve created two more with identical settings,

[00:43:15.190]but just change the ISO value on each one.

[00:43:18.710]If we step back into the model as a on a section,

[00:43:26.640]so here we can see our model

[00:43:28.600]which going to take the triangulations off.

[00:43:30.960]So we’ve got the outline.

[00:43:33.320]So currently with this volume,

[00:43:36.650]what I’m saying is that I have a 50% chance

[00:43:39.497]but if I take a sample, anywhere in here,

[00:43:41.780]it is going to be above my cutoff

[00:43:44.810]can also change my drill holes here to reflect that cutoff.

[00:43:49.810]So you can see here, no 0.3% cutoff.

[00:43:52.460]So you can see on my drilling samples

[00:43:55.470]that are above and below.

[00:43:58.840]If I go into my parameters

[00:44:01.000]and I change my ISO value down to no 0.3.

[00:44:06.060]So we can have a look at the inside value on this one,

[00:44:10.790]what this is saying and I just made some,

[00:44:12.890]so you can see it as well.

[00:44:13.810]It’s basically, this is a volume

[00:44:15.930]that is more of a prioritization around volume.

[00:44:18.830]So this could be an example.

[00:44:20.130]If you needed to produce a mean case and a max case

[00:44:23.677]and a mid case,

[00:44:25.150]then you could do this pretty quickly by using your cutoffs

[00:44:28.620]and using your ISO values.

[00:44:31.620]So drop in an ISO value down to 0.3

[00:44:35.120]is essentially saying that there’s a 30% confidence

[00:44:38.130]that if I take a sample inside this volume,

[00:44:41.240]it will be above my cutoff.

[00:44:43.240]So you can see that changing the ISO value

[00:44:45.960]and keeping everything else the same

[00:44:47.300]has given me a more optimistic volume

[00:44:50.640]for my indicator shell.

[00:44:53.180]Conversely, if I change my indicator ISO value to 0.7,

[00:44:58.780]it’s going to be a more conservative shell.

[00:45:00.660]So here, if I have a look at the inside volume in green,

[00:45:06.150]and again, maybe just to help highlight

[00:45:07.740]the differences with these.

[00:45:11.100]So now this is my, exactly the same settings,

[00:45:15.220]but applying a ISO value.

[00:45:18.260]So a probability or a confidence of 0.7.

[00:45:22.070]And again, just to, for you to review the notes.

[00:45:24.730]So increase in my ISO value

[00:45:28.640]will give me a more conservative case.

[00:45:31.720]This will be prioritizing the metal content,

[00:45:34.500]so less dilution.

[00:45:36.520]And essentially I can look at those numbers and say,

[00:45:38.720]I have a 70% confidence that any sample inside that volume

[00:45:42.700]will be above my cutoff.

[00:45:47.020]So I say it very quickly,

[00:45:48.520]and particularly in the exploration field,

[00:45:50.610]you can have a look at a resource or a volume.

[00:45:53.700]And if you want to have a look at a bit of a range analysis

[00:45:56.300]of the potential of your mineralization,

[00:45:58.500]you can generate your cutoff and then change your ISO values

[00:46:03.340]to give you an idea of how that can work.

[00:46:07.220]Once you’ve built these,

[00:46:08.090]you can also then come in and have a look at the,

[00:46:10.050]it gives you a summary of your statistics.

[00:46:12.830]So if we have a look here at the indicator at 0.3,

[00:46:16.040]I look at the statistics of the indicator at 0.7,

[00:46:22.254]we can go down to the volume.

[00:46:24.840]So you can see the volume here

[00:46:27.610]for my conservative ISO value at 0.7

[00:46:31.900]is 431 million cubic meters,

[00:46:34.640]as opposed to my optimistic shell, 545.

[00:46:40.330]You can also see from the number of parts.

[00:46:42.350]So the number of different volumes

[00:46:43.840]that make up those indicators shells in the optimistic one,

[00:46:48.310]I only have one large volume,

[00:46:50.930]whereas the 0.7 is obviously a bit more complex

[00:46:54.520]and has seven parts to it.

[00:46:57.314]The last thing you can do

[00:46:59.190]is you can have a look at the statistics.

[00:47:02.180]So for example, the insight volume,

[00:47:04.720]I can have a look at how many samples here

[00:47:06.970]for below my cutoff.

[00:47:09.220]So out of the what’s that 3,465 samples,

[00:47:15.260]260 those are below my cutoff.

[00:47:19.200]So you can kind of work out a, again,

[00:47:21.613]a very rough dilution factor by dividing

[00:47:24.100]the number of samples that fall inside your volume

[00:47:28.310]by the total number of samples,

[00:47:29.610]which would bring your answer around

[00:47:31.020]in this case for the 0.3 around seven and a half percent.

[00:47:37.580]So it’s just, this exercise is really just to highlight

[00:47:40.610]some of the, again, the tools aren’t necessarily

[00:47:43.162]used very frequently,

[00:47:45.200]but can give you a really good understanding of your data

[00:47:48.930]and help you to investigate the potential of your deposits

[00:47:53.200]by using some of the settings in the indicator templates,

[00:47:58.669]but ultimately coming back to the fundamentals

[00:48:02.620]of building numeric models.

[00:48:04.870]And that is to understand the type of interplant you use

[00:48:08.530]and how that drift function can affect your models

[00:48:11.750]as you move away from data.

[00:48:15.160]So that’s probably a lot of content for you

[00:48:18.822]to listen to in the space of an hour as always,

[00:48:23.500]we really appreciate your attendance and your time.

[00:48:26.680]We’ve also just popped up on the screen,

[00:48:28.580]a number of different sources of training and information.

[00:48:33.810]So if you want to go

[00:48:35.347]and have a look at some more detailed workflows,

[00:48:38.440]then I strongly recommend you look at the Seequent websites

[00:48:41.980]or our YouTube channels.

[00:48:43.500]And also in the, MySeequent page,

[00:48:45.930]we now have all of the online content

[00:48:47.860]and learning available there.

[00:48:50.110]As always we’re available through the support requests.

[00:48:52.660]So drop us an email

[00:48:54.057]and if you want to get into some more detailed workflows

[00:48:57.820]and how that can benefit your operations and your sites,

[00:49:01.830]then let us know as well.

[00:49:03.570]We’re always happy to support

[00:49:05.420]via project assistance and training.

[00:49:08.900]So that pretty much brings us to the end of the session

[00:49:13.290]to the top of the hour as well.

[00:49:14.480]So again, thanks to you for attending,

[00:49:17.780]thanks to Suzanna and Andre for putting this together

[00:49:21.940]and running the session.

[00:49:23.500]And we’ll look to put another one

[00:49:25.700]of these together again in the new year.

[00:49:27.690]So hopefully we’ll see you all there.

[00:49:30.630]Thanks everybody and have a good day.