Transcript
WEBVTT
00:00:00.180 --> 00:00:02.028
Hello everybody, welcome to the Fire Science Show.
00:00:02.450 --> 00:00:11.320
Today we're going to talk AI in fire safety engineering and I am very excited because we're not going to hypothesize what the use of AI will look like.
00:00:11.320 --> 00:00:16.452
We're going to talk about experiences in doing that and that's something very unique.
00:00:16.452 --> 00:00:28.126
With my guest Professor Sinian Huang from Hong Kong Polytechnic University, we've hypothesized how it could look like and talked about his early experiences three years ago in episode seven.
00:00:28.126 --> 00:00:40.905
Actually, that was one of the first episodes of the Fire Science Show and today, three years fast forward, we can talk about a lot more experiences that Simeon and his group has gained over those years.
00:00:40.905 --> 00:00:46.615
They're on the forefront of implementing AI in various kinds of fire science.
00:00:46.615 --> 00:00:49.546
And we're not talking chatbot AI.
00:00:49.546 --> 00:00:59.591
We're talking neural networks and using it to predict fire behaviors, to predict fire phenomena, to measure fire and to help in engineering design.
00:00:59.591 --> 00:01:21.376
It all started with smart firefighting, so the theme of episode seven and the discussion back then was how we can use AI to assist firefighters and, if you're curious how this ended, it's at the same time interesting, to some extent disappointing and, at the same time, exciting.
00:01:21.376 --> 00:01:37.793
Some pathways did not lead anywhere, but some pathways lead to extremely exciting places and we're also going to talk about that later in the episode, and if you want to learn about that, well, first intro and then let's go with the AI in fire safety engineering.
00:01:43.600 --> 00:01:45.168
Welcome to the Fireiresize Show.
00:01:45.168 --> 00:01:48.670
My name is Wojciech Wigrzyński and I will be your host.
00:01:48.670 --> 00:02:04.510
This podcast is brought to you in collaboration with OFR Consultants.
00:02:04.510 --> 00:02:07.468
Ofr is the UK's leading fire risk consultancy.
00:02:07.468 --> 00:02:18.312
Its globally established team has developed a reputation for preeminent fire engineering expertise, with colleagues working across the world to help protect people, property and environment.
00:02:18.312 --> 00:02:34.116
Established in the UK in 2016 as a startup business of two highly experienced fire engineering consultants, the business has grown phenomenally in just seven years, with offices across the country in seven locations, from Edinburgh to Bath, and now employing more than 100 professionals.
00:02:34.116 --> 00:02:45.760
Colleagues are on a mission to continually explore the challenges that fire creates for clients and society, applying the best research, experience and diligence for effective, tailored fire safety solutions.
00:02:45.760 --> 00:02:56.849
In 2024, ofr will grow its team once more and is always keen to hear from industry professionals who would like to collaborate on fire safety futures this year.
00:02:56.849 --> 00:02:59.743
Get in touch at OFRConsultantscom.
00:03:00.185 --> 00:03:01.990
Hello everybody, welcome to the Fire Science Show.
00:03:01.990 --> 00:03:06.109
I am here today with Professor Sinan Huang from Hong Kong Polytechnic University.
00:03:06.109 --> 00:03:09.647
Hey, sinan Hi Bozsik, good to see you again.
00:03:09.647 --> 00:03:12.288
Very happy to have you back in the podcast.
00:03:12.288 --> 00:03:15.389
You were one of the first 10 episodes.
00:03:15.389 --> 00:03:18.188
I can call you the OG of the Fire Science Show.
00:03:18.188 --> 00:03:21.352
Yeah, I'm F-7.
00:03:21.352 --> 00:03:25.430
Yeah, nice 007, licensed to do AI in fire safety engineering.
00:03:25.430 --> 00:03:30.411
Good, good, and we're going to continue the discussion that stopped in episode seven.
00:03:30.411 --> 00:03:41.068
Gosh, that's a long time ago, but in that episode we've discussed smart firefighting and different ways of using AI to assist firefighters.
00:03:41.068 --> 00:03:48.562
I remember you were very happy back then by a large grant that your unit was given on this topic.
00:03:48.562 --> 00:03:51.330
I know a lot of papers came out of your office.
00:03:51.330 --> 00:03:52.893
So now, fast forward.
00:03:52.893 --> 00:03:54.122
Three years have passed.
00:03:54.122 --> 00:03:57.310
Let's see where we are on smart firefighting today.
00:03:57.310 --> 00:04:06.495
So if you can summarize, what's the biggest change between 21 and 24 in terms of using AI in assisting firefighters?
00:04:07.240 --> 00:04:20.014
Yeah, I think the last time when we first talked about the smart firefighting back 2021, and we just get that grant and I just recruit a few students doing this project.
00:04:20.014 --> 00:04:21.745
Also, I'm new to AI.
00:04:21.745 --> 00:04:24.906
Right now I think I'm still new to AI.
00:04:24.906 --> 00:04:42.310
I haven't got a chance to really run AI as all the students doing the hard works, but I think I know more about the AI application and how it works and what problem AI can solve to help us, no matter it's to support firefighters or support 5G news.
00:04:42.310 --> 00:04:45.769
So I think I know more about the tool.
00:04:46.420 --> 00:04:54.624
The last time we talked, you also mentioned that the use of AI comes from patches of code that are implementing real packages.
00:04:54.624 --> 00:04:57.668
You don't have to be an AI scientist in implementing AI.
00:04:57.668 --> 00:05:05.166
Now, working with this for three years together with your students, was it very hard to enter the world of AI?
00:05:05.166 --> 00:05:08.512
I mean, three years have passed and you've shown some success.
00:05:08.512 --> 00:05:12.343
I wonder if that can inspire others to try out AI.
00:05:12.343 --> 00:05:14.648
Was it very hard to implement a lot of challenges?
00:05:15.209 --> 00:05:18.942
I think it's very simple to use the AI algorithm.
00:05:18.942 --> 00:05:40.975
So most of my students who have most of them have never used AI in their research before, or maybe they haven't done any research before, so when they start to do the AI-related research, I think it takes less than one month to be able to run some simple AI program or reproduce some previous paper.
00:05:40.975 --> 00:05:46.012
So I think running the AI algorithm itself is also a challenge.
00:05:46.012 --> 00:05:59.391
The most challenging thing is to identify the problem that is worth solving by AI, and that requires a lot of knowledge about the file as well as the capacity of AI.
00:06:00.021 --> 00:06:02.290
And how about choosing the correct algorithm?
00:06:02.290 --> 00:06:08.173
Because with my understanding of AI, I understand that there are supervised and unsupervised models.
00:06:08.173 --> 00:06:11.629
You have neural networks, but you have also classification tools.
00:06:11.629 --> 00:06:14.586
There is so much like when you start digging.
00:06:14.586 --> 00:06:16.331
There are so many choices.
00:06:16.331 --> 00:06:22.312
Have you figured out the way how to assign correct algorithm or model to a correct problem?
00:06:22.312 --> 00:06:25.761
Or is it the other way you know an algorithm and you find a problem for it?
00:06:26.062 --> 00:06:37.451
Yes, that's also something we are learning during the process of this project, and overall feeling is I still think the algorithm is not so important.
00:06:37.451 --> 00:06:39.581
First, it can be solved.
00:06:39.581 --> 00:06:48.074
Most of the algorithm problem people are facing can be solved by increasing the size of the database or the number of useful data.
00:06:48.074 --> 00:06:55.687
Sometimes you have a very large database but most of the data could be not valuable, so not really helping training AI.
00:06:55.687 --> 00:07:03.752
But if you have a good database, even if it's small, it can basically solve most of the training problems.
00:07:03.752 --> 00:07:04.694
That's my feeling.
00:07:04.694 --> 00:07:07.559
It can basically solve most of the training problems.
00:07:07.559 --> 00:07:10.882
That's my feeling In some aspect.
00:07:10.942 --> 00:07:12.389
For example, we are doing a lot of file simulation.
00:07:12.408 --> 00:07:14.779
We are predicting the file development showing the smoke movement.
00:07:15.480 --> 00:07:35.833
So if you want to use AI to generate very nice or very real CFD file images, then the algorithm is quite important and from my experience, diffusion model is definitely the best to generate very detailed flow motions, smoke motions.
00:07:36.319 --> 00:07:45.269
But the problem is the training of the diffusion model is very, very long time and even when you use it for prediction, the rendering time of these images the AI prediction images also very long time.
00:07:45.269 --> 00:07:49.045
And even when you use it for prediction, the rendering time of these images the AI prediction images also vary in long time.
00:07:49.045 --> 00:07:57.607
So unless you really want to achieve that detailed structures, usually you don't need a diffusion model.
00:07:57.607 --> 00:08:13.408
In my experience, some models like GAN model or GL model they all solve pretty good problems and, for example, if you just want to predict the ASET, you just want to know when the smoke layer will drop to two meter high.
00:08:13.408 --> 00:08:21.805
Then you don't need to know the detailed flow structure of the smoke, you just need to know when the smoke layer touch the critical line.
00:08:21.805 --> 00:08:26.552
So in that sense there's really no need of using advanced AI model.
00:08:27.074 --> 00:08:34.188
It's really interesting because your group was able to use AI in a way like many of us would use CFD.
00:08:34.188 --> 00:08:41.203
I see so much resemblance, by the way, how you are using AI with how many engineers would be using CFD.
00:08:41.203 --> 00:08:48.347
Also, you know, in the way that we don't really comprehend CFD that well as engineers like, and you don't need to because there are already made packages that you apply.
00:08:48.347 --> 00:08:52.245
You also have to understand the problem definition and set the boundary conditions.
00:08:52.245 --> 00:08:58.563
Like you said here, you choose the appropriate model or appropriate tools for the problem you solve Now in CFD.
00:08:58.563 --> 00:09:06.200
I would say one of the reasons it is such a popular tool in fire safety engineering is because of the realism of output.
00:09:06.200 --> 00:09:10.950
You get those really beautifully looking plots that look like fire.
00:09:11.530 --> 00:09:14.904
People use CFD because they get those lovely images and everything.
00:09:14.904 --> 00:09:21.272
You just said that you can use AI very quickly to just get the ACID value of two meters and yes, I agree you can do that.
00:09:21.272 --> 00:09:24.222
But the question is the perceived value of the tool.
00:09:24.222 --> 00:09:31.549
If, if you lose the beautiful images generated by the diffusion model, will that tool be believed that it truly is two meters?
00:09:31.549 --> 00:09:48.923
You know, without the layer of the graphics, it's hard to convince someone that the result really is the two meters, and it's all a matter of trusting the tool which, let's be honest, the trust to AI generated results is, in general, quite low, I would say.
00:09:49.706 --> 00:09:50.942
Yes, that's the reason.
00:09:50.942 --> 00:10:23.549
A lot of FDS simulations can cheat the public because they look real, and I'm sure the software companies or SmokeView or other rendering tools, they have tried a lot to make these images look real and so far what we can do is first, of course, we can prove that even a rough smoke layer without all these small eddies, they basically have the same smoke height as the one with detailed eddies.
00:10:23.549 --> 00:10:25.030
And that's one way.
00:10:25.030 --> 00:10:33.437
The other way is we can also use AI tools to generate these eddies to make it more real If it's required to convince people.
00:10:33.557 --> 00:10:37.241
Okay, it's just a matter of cost and time, right?
00:10:37.241 --> 00:10:44.327
I am disturbed by the amount of similarities with AI and CFD as a tool, and I wonder how the future will look like.
00:10:44.327 --> 00:10:47.089
Will we be using one or another, or both?
00:10:47.340 --> 00:10:53.307
Yeah, if you print that in paper, probably you cannot tell the difference whether it's AI generated or it's CFD generated.
00:10:53.460 --> 00:10:58.370
It's disturbing because I know at least cell biology and genetics.
00:10:58.370 --> 00:11:16.164
There were hundreds of studies that were retracted with very serious accusations that you know the genome sequencing is like a line with some lines in it, so it's very easy to fire up Adobe Photoshop and just cheat a part of the image.
00:11:16.164 --> 00:11:22.043
Serious scientists would face serious accusations that the images in there these are falsified.
00:11:22.184 --> 00:11:23.808
They cannot blame that for AI.
00:11:23.808 --> 00:11:25.613
That's actually a real human cheating.
00:11:25.960 --> 00:11:34.067
Yeah, I know, I know, I know, but I'm stressed because the ease of this tool and the ease will increase in the future.
00:11:34.067 --> 00:11:36.744
And what if the AI is wrong?
00:11:36.744 --> 00:11:42.164
What if it gives us incorrect output because someone did choose the train set as a too small one?
00:11:42.164 --> 00:11:45.106
How sensitive actually was it to data?
00:11:45.106 --> 00:11:48.267
You said the biggest challenges were with getting data correct.
00:11:48.267 --> 00:11:56.347
Was it truly one of the biggest challenges in your project to get the data and how much of the data you actually needed to get reasonable outcomes?
00:11:56.860 --> 00:11:59.028
Depending on the problems you're trying to solve.
00:11:59.028 --> 00:12:05.971
For example, recently we are trying to use AI to forecast smoke flow in very complex shape atrium.
00:12:05.971 --> 00:12:11.899
In that sense, we need a lot of simulation CFD simulation to form the database.
00:12:11.899 --> 00:12:16.072
But even so, I don't think that database is that big.
00:12:16.072 --> 00:12:19.811
We only have a few hundred case with complex shapes.
00:12:19.811 --> 00:12:24.311
We have another few thousand case with relatively simple shape.
00:12:24.311 --> 00:12:27.769
So I don't think that's large enough.
00:12:27.769 --> 00:12:41.009
Because essentially what we want to do is I mean, nowadays so many consulting companies, they are running CFD simulations for different buildings, different structures, but all these data are not fully used.
00:12:41.399 --> 00:12:43.788
After they finish the project it's in the hard drive.
00:12:43.788 --> 00:12:45.826
Nobody is really using it.
00:12:45.826 --> 00:12:51.068
But in fact, all this data can be trained for improving the AI capacity.
00:12:51.068 --> 00:12:57.788
But if we have this database to train AI, that will be amazing.
00:12:57.788 --> 00:13:04.890
We have a very large database and we are not asking extra input, we're just using what's already there.
00:13:04.890 --> 00:13:15.931
But of course, the pre-processing for these data will take a lot of time because every company, every engineer, they have their own habit of making the model to run the simulation.
00:13:15.931 --> 00:13:25.808
So that's also what we see is that in fact, a lot of time is spent on pre-processing the database rather than creating database itself.
00:13:26.460 --> 00:13:32.065
If you wanted to use this data like the way you said, I think there would be so many human variables with what you just said.
00:13:32.065 --> 00:13:33.446
Different people would do it differently.
00:13:33.446 --> 00:13:38.542
I wonder if it's even possible to quantify all the choices that people do Like.
00:13:38.542 --> 00:13:42.063
There must be doses of choices Like what design fire did you put?
00:13:42.063 --> 00:13:43.065
How do you place it?
00:13:43.065 --> 00:13:44.086
Was the suit healed?
00:13:44.086 --> 00:13:45.289
Was the heat of combustion?
00:13:45.289 --> 00:13:47.413
Was the makeup moisture?
00:13:49.200 --> 00:13:52.571
That's actually the good thing, because everyone chooses different parameters.
00:13:52.571 --> 00:13:55.649
That just makes the database become richer.
00:13:55.649 --> 00:13:57.899
Ah, okay, not just a few settings.
00:13:57.899 --> 00:14:02.369
So everyone has different settings, so the database becomes very good, very large?
00:14:03.019 --> 00:14:06.350
And how about training or integrating experimental data?
00:14:06.350 --> 00:14:11.673
Because I know that you use CFD and you have a good reason for it, which I hope you will reveal.
00:14:11.673 --> 00:14:14.086
But how about using experimental data?
00:14:14.086 --> 00:14:14.909
Let's start with CFD.
00:14:14.909 --> 00:14:19.730
Why do you train more on CFD than on experiments, from what I understand?
00:14:20.231 --> 00:14:20.393
Yes.
00:14:20.393 --> 00:14:24.870
So first of all, we still use some of the experiment data.
00:14:24.870 --> 00:14:34.294
I would say, before we do a large amount of CFD simulation, we always calibrate the model with the experiment data.
00:14:34.294 --> 00:14:46.900
So in that sense I consider we already include some of the experiment into the database, because some parameters used for the boundary condition may come from the experiment.
00:14:46.900 --> 00:14:51.149
I wouldn't say it's 100% numerical input.
00:14:51.149 --> 00:14:53.660
It also has a lot of input from the experiment.
00:14:53.922 --> 00:14:58.211
The problem is, even for the experiment it's very difficult to quantify the result.
00:14:58.211 --> 00:15:01.024
For example, we all have limited sensors.
00:15:01.024 --> 00:15:04.291
For example, even if you have some couple of trees, you have a few points.
00:15:04.291 --> 00:15:09.405
Even you have cameras, you have only a limited view of the smoke motion.
00:15:09.405 --> 00:15:18.982
Compared to the CFD simulations, the data you get from the experiment is extremely limited and have a large uncertainty.
00:15:19.585 --> 00:15:24.740
For example, the fire you used in the experiment may not be so well controlled.
00:15:24.740 --> 00:15:31.394
If you are burning, for example, a wood crib, who knows what kind of large fire, how large is the fire it is, and there's a smoke.
00:15:31.394 --> 00:15:33.727
I mean, every wood burns different smokes.
00:15:33.727 --> 00:15:35.647
It's difficult to quantify that.
00:15:35.647 --> 00:15:41.873
So in that sense I feel it's quite difficult to directly use experimental data.
00:15:41.873 --> 00:15:52.725
And I think, most importantly, for file engineering design is you only consider certain representative scenarios rather than the so-called real scenario.
00:15:52.725 --> 00:15:55.166
There's no such thing as a real scenario.
00:15:55.166 --> 00:16:03.386
Even the same building have the same furniture burning for a hundred times, every time it's completely different.
00:16:03.386 --> 00:16:06.129
So in that sense, we cannot forget.
00:16:06.129 --> 00:16:13.785
Doing design is just following certain framework and test some possible file scenario to give certain confidence.
00:16:13.785 --> 00:16:17.673
We are not trying to simulate a potential real file.
00:16:17.995 --> 00:16:33.067
I would summarize what you've just said is that it's hard to capture all the uncertainties in the learning process, like, if you learn based on CFD, you, if you input it one megawatt, you're certain you've inputted one megawatt.
00:16:33.067 --> 00:16:43.408
And here, if you had a crib that was supposed to give you a megawatt one day it was very hot and dry and the lab was well ventilated, you had 1.1.
00:16:43.408 --> 00:16:45.541
Other day you've done a repeat.
00:16:45.541 --> 00:16:46.841
It was a moist after rain.
00:16:46.841 --> 00:16:50.884
You had 0.95, right, and yet you put something into training.
00:16:50.884 --> 00:16:57.908
You tell the trained model it was one megawatt there's an uncertainty in the input that you've not accounted for.
00:16:57.927 --> 00:16:58.575
So there are a few aspects.
00:16:58.575 --> 00:17:04.428
First, from AI training point of view, the experiment data have their own format.
00:17:04.428 --> 00:17:19.647
So you may have some temperature sensors done one experiment, you have some other group of sensors done in different locations in a different experiment, and these data formats don't match with each other, so it's quite difficult to train them.
00:17:19.647 --> 00:17:24.287
If you run the CFD simulation, you can collect the data in a consistent way.
00:17:24.287 --> 00:17:27.603
Then it's much easier for AI to train them.
00:17:27.603 --> 00:17:38.885
That's one aspect, but I think the second aspect which is most important is when we do the current state-of-the-art practice, we never ask the guys who run CFD.
00:17:38.885 --> 00:17:43.612
We don't question if their model represents the real file or not.
00:17:43.612 --> 00:17:47.665
We just assume okay, your simulation is reasonable and correct.
00:17:47.665 --> 00:17:51.586
So in that sense we only need to compare with the CFD simulation.
00:17:51.586 --> 00:17:55.286
We don't have to compare with the real experiment, because this is a design practice.
00:17:55.807 --> 00:18:04.548
So, basically, assuming that the CFD is the state of the art tool, you basically create a tool that is at the same confidence level as the CFD.
00:18:04.548 --> 00:18:06.317
Yes, okay, that makes sense.
00:18:06.317 --> 00:18:12.017
Out of all the AI implementations you've done, let's pick one and go deeper.
00:18:12.017 --> 00:18:13.722
How about the fire prediction?
00:18:13.722 --> 00:18:15.678
I love the fire prediction, so.
00:18:15.678 --> 00:18:33.969
So your group has built those tools that are able to predict the size of the fire based on images, I believe, and I I found it really interesting, especially that there are videos online that showcase the real, uh, real-time capability of this prediction software, and it's just magical.
00:18:33.969 --> 00:18:40.403
Are the videos fake or it really works like that, like real-time, showing the hit-release rate?
00:18:41.038 --> 00:18:54.224
Yes, we're actually having a latest paper will come out very soon, that we have an online link Everyone can upload their video and we were exporting the real-time hit rate.
00:18:54.546 --> 00:18:57.346
That's good, so I confirm, this is really amazing.
00:18:57.346 --> 00:19:01.740
So tell me, what was the big idea behind starting this and how did it go?
00:19:01.740 --> 00:19:04.040
Actually, what was the point of doing this study?
00:19:04.462 --> 00:19:08.820
So I have to say the idea come from when I was teaching the fire dynamics class.
00:19:09.135 --> 00:19:20.487
So there's one session I have to teach the students what's the definition of the fire heat release rate and then I have to go to the only two methods that we can measure the fire heat release rate.
00:19:20.855 --> 00:19:26.364
One is to measure the mass loss rate of the fuel and the other is oxygen calorimetry.
00:19:26.364 --> 00:19:37.464
You measure the oxygen depletion based on the smoke measurement and eventually the students questioning okay, both methods can only be used in labs.
00:19:37.464 --> 00:19:40.555
Can any method can help us to measure the real fire.
00:19:40.555 --> 00:19:43.441
So I think that's a really a good question.
00:19:43.441 --> 00:20:05.028
Some students I don't really remember the name of the student, but someone asked me about that so I think that's something we have to think about, because if, of course, you can put a big hood above a house or above a burning car, but everything is done in the lab, you cannot put a big hood where you have a fire incident and you put it about there and measure the heat rate.
00:20:05.028 --> 00:20:11.689
So none of the methods that we have so far can actually measure the power of a real fire.
00:20:11.689 --> 00:20:20.561
And that just inspired me to think about AI method and I think during that time we have a lot of advancement.
00:20:20.561 --> 00:20:28.845
For example, using the mobile phone, we can use a facial ID to unlock our phone and we have a lot of facial recognition everywhere.
00:20:28.845 --> 00:20:34.362
In China you can use a face ID to pay, actually Okay, so the image is really powerful.
00:20:34.362 --> 00:20:36.060
That's what I feel.
00:20:36.255 --> 00:20:40.926
And doing experiments and all these fire experiments, we have a very rough view.
00:20:40.926 --> 00:20:46.788
So if the fire size is an area, the volume is larger, of course it's more powerful.
00:20:46.788 --> 00:20:57.079
So I think there is a certain correlation between the size of the fire as well as its power, correlation between the size of the fire as well as its power.
00:20:57.079 --> 00:21:08.547
And if you really look into the details of the fundamental flame sheet and that definitely makes sense because flame is essentially like a coating, it only has a small sheet and all the reactions happen in that sheet.
00:21:08.768 --> 00:21:24.068
So if you can get to the area of that sheet, definitely you are able to quantify the file release, heat release rate, so, and that area of the sheet is proportioned to a certain degree to the image can captured by the camera.
00:21:24.068 --> 00:21:32.404
So that's the original idea, but of course we know it's very challenging to actually train the database Then we are just super lucky.
00:21:32.404 --> 00:21:37.066
I would say super lucky because NIST had such a wonderful database.
00:21:37.066 --> 00:21:54.364
So they are burning all different kinds of things I think they have more than 2,000 different material things burning in the lab and as they record all the heat release rate by oxygen calorimetry, they also have all the images, videos, you can just download from the website.
00:21:54.364 --> 00:21:59.585
So that's just amazing and you can use this database to train.
00:22:00.145 --> 00:22:00.988
I'll quickly plug in.
00:22:00.988 --> 00:22:02.798
I had an interview with Matt Bundy.
00:22:02.798 --> 00:22:10.040
That's episode 110 of the Fire Science Show where we've just discussed what you've just said the NIST fire calorimetry database.
00:22:10.040 --> 00:22:16.782
So Matt told us all the tricks they have for recording cameras, automated storage, processing.
00:22:16.782 --> 00:22:27.244
Like a lot of effort goes into building a database like that and we are very grateful to NIST for developing and maintaining this database.
00:22:27.244 --> 00:22:31.501
So you had images from, or videos from, the database.
00:22:31.501 --> 00:22:39.243
You had the heat release rate plots, what goes from these points to having a trained model that can predict fire size.
00:22:39.243 --> 00:22:45.124
But because I assume it's not a simple like press enter, here's the images and you robot learn.
00:22:45.424 --> 00:22:54.249
But it must be quite a process so basic for all the ai model, you need to identify the input and output and you pair them in the training.
00:22:54.249 --> 00:23:03.931
So for this specific case, the input is the image of that moment and the output is the heat release rate measured by the oxygen calorimetry.
00:23:03.931 --> 00:23:04.934
So you pair them.
00:23:04.934 --> 00:23:07.263
So every second you have a pair.
00:23:07.263 --> 00:23:17.528
If you burn something for 20 minutes let's just take one point per second you have 1,200 pairs of heat release rate as well as images.
00:23:17.528 --> 00:23:28.586
Then you put all of them into the training database and now you have 1,000 different fire burning and each test you have 1,200 images.
00:23:28.586 --> 00:23:32.979
So together you have a million of data pairs for the training.
00:23:32.979 --> 00:23:44.111
So together you have a million of data pairs for the training and that results in a very amazing trained AI model that can basically give you a hit-release rate if you input any image Okay, but the image is a collection of pixels.
00:23:46.414 --> 00:23:47.237
It doesn't reflect the real world.
00:23:47.237 --> 00:23:50.683
If you put a matchstick right next to the camera, it's going to appear huge on the image.
00:23:50.683 --> 00:23:56.115
So how did you solve the dimensionality of the fires at NIST?
00:23:56.115 --> 00:23:58.362
Or were the cameras conveniently positioned?
00:23:58.362 --> 00:23:59.105
Always the same way?
00:23:59.516 --> 00:24:16.704
Yeah, first we removed those cases that clearly the camera is putting a different location and later on we added some additional data from our lab to measure the fire from different distance and use that to calibrate the fire image.
00:24:16.704 --> 00:24:21.346
So we rescale the image to be the same as our database.
00:24:21.346 --> 00:24:29.941
So in the database all the images are rescaled under a certain scale and for any practical applications.
00:24:29.941 --> 00:24:37.545
I think there are three methods you can approach to solve the distance or the scale problem.
00:24:37.545 --> 00:24:39.958
First, you can use the reference lens.
00:24:40.138 --> 00:24:52.019
For example, if you see some fire is burning in a car passenger car and you know roughly how long the passenger car, how tall it is, you can use that as a reference scale to help you to scale that fire image.
00:24:52.019 --> 00:24:54.522
Use that as a reference scale to help you to scale that fire image.
00:24:54.522 --> 00:24:55.364
That's one thing.
00:24:55.364 --> 00:24:59.710
The other is you can use a bimolecular camera.
00:24:59.710 --> 00:25:06.903
You have basically two cameras that can measure the distance between the camera and the fire and that can also give you a reference scale.
00:25:06.903 --> 00:25:18.982
And as a third option we provide is if you put that in UAV and the UAV can measure the height between the ground and a UAV, that actually give you a reference lens as well.