Collecting data is simple; turning it into actionable Information is not! Data management, collation, and contextualisation is the industry’s biggest problem today. Analytics can help your business make informed decisions.
We discuss analytics insights and strategies to;
- Understand the Analytics Continuum in its proper form,
- What data can do for your decision-making process,
- And the latest cool features of Flow.
We also show new Ignition and Canary features and enhancements. And the latest tools to help you with your next project.
Timestamps:
0:00 Welcome
04:54 The Changemakers
09:39 What’s New: Ignition
32:43 What’s New: Canary
40:11 A Better Understanding of the Analytics Continuum
#DLF. LET’S MAKE INDUSTRIAL DATA GREAT AGAIN!
Resources for further learning:
IGNITION
- Download the Inductive Automation Security Hardening Guide.
- Upgrade to the latest Ignition 8.1.17 here.
- What is Digital Transformation?
- Solving Data Problems to Accelerate Digital Transformation.
- Building a Sustainable and Secure SCADA System.
CANARY
- Know Your Industrial DB Options.
- Properly Evaluating Historians.
- Upgrade to the latest Canary 22.1.2 here.
- Native collectors are now available in the updater service; Allen-Bradley, Siemens and Modbus. Learn how to configure them here.
FLOW
- Learn how to pair your Canary Historian with Flow Software and develop even more reporting options.
- Learn how Flow calculates a production run rate using historical data.
- How to become Flow certified.
Follow our Podcast and check out the latest episode with Rob Obree from Next Integration. He explains how organisations can improve operations using data through practical and sustainable methods.
SPEAKERS:
Resources for further learning:
IGNITION
- Download the Inductive Automation Security Hardening Guide.
- Upgrade to the latest Ignition 8.1.17 here.
- What is Digital Transformation?
- Solving Data Problems to Accelerate Digital Transformation.
- Building a Sustainable and Secure SCADA System.
CANARY
- Know Your Industrial DB Options.
- Properly Evaluating Historians.
- Upgrade to the latest Canary 22.1.2 here.
- Native collectors are now available in the updater service; Allen-Bradley, Siemens and Modbus. Learn how to configure them here.
FLOW
- Learn how to pair your Canary Historian with Flow Software and develop even more reporting options.
- Learn how Flow calculates a production run rate using historical data.
- How to become Flow certified.
Follow our Podcast and check out the latest episode with Rob Obree from Next Integration. He explains how organisations can improve operations using data through practical and sustainable methods.
Transcript
00:00
Jaco
Good morning and welcome to the. Do we call it the element eight, the DLF, another three letter acronym for you. Good morning and welcome to the element eight quarterly update. I think it’s been two quarters since our last update. And this is the team of element eight. My name is Jaco. I look after the team. I have with me. I have Laura.
00:22
Laura
Hi.
00:24
Lenny
Yeah, is.
00:25
Laura
I’m Laura. I’m a customer success engineer here at element eight.
00:28
Tebello
And I am Tebello. I am Laura’s customer.
00:36
Lenny
Cool.
00:37
Jaco
And then our guest speaker and also our sponsor for this episode of the quarterly update is Lenny from Flow. Hi, Lenny.
00:44
Lenny
Hi. Good afternoon. Good morning. Sure, almost afternoon. Good morning. Hoping, probably wishing for the weekend to start. Yes. Hi, everybody. Lenny here from Flow software, customer success manager. And, yeah, very proud and very honored to be with elements 18 here today and also to be a sponsor for the session this morning. Cool.
01:02
Jaco
So if you’re not familiar with these sessions, we typically provide things that you may have missed over a couple of months. I suppose if it’s quarterly, over a couple of months, things that you may have missed, that’s important to note. We also try and have conversations with people in our industry that’s a little bit outside of just some of the technical stuff that we’re going to be sharing today. But, yeah. Welcome to this update. If there is an hour that you can spend with us, this is probably the hour that you want to spend with us. And do make sure that you share it with others that are not able to join today in your teams. All right.
01:35
Jaco
I am going to just make sure that I monitor the chat here because I do see, I’ve got a couple of things coming up here, so let me just make sure that I’ve got it. The other chat. And as always, if you have any questions, please make sure that you do let us know in the chat. We’d be happy to answer them. These things very often feels a little bit one directional, a little bit like a monologue. So please, if you have any questions, let us know, and we’d be happy to answer those questions. All right, we are going to cover today, if I look at the agenda, we’re going to talk a little bit about change makers when we talk about the DLF, just a little bit of context. The DLF, if you’re not familiar, it’s something that we recently. It’s a movement.
02:15
Jaco
Is it a movement? It’s not a political party, isn’t it? No, we’re not on the role.
02:23
Tebello
If.
02:23
Jaco
They are declared, but the DLF is the data Liberation Front. So it is a movement. It is something that we in our community and other like minded people are very passionate about, which is the liberation of industrial data everywhere. So that is the purpose of the DLF. It is a community, it’s not a political party. But I think we first introduced the DLF down in Cape Town, in the Western Cape. We spoke about the DLF for the first time. So it is a community of like minded people that is growing. We’d love for you to be part of that community and we’ll share a little bit more about that later. So we’re going to talk about some of our newest partners and customers that we have on board.
03:02
Jaco
We are going to tell you what is new in the world of ignition, canary and flow, and Lenny is going to take us through a new understanding of the analytics. I can’t even say that word. Analytics. Continuum. Lenny is a big words for a Friday.
03:16
Lenny
Very big words for a Friday. Don’t worry, guys, we’ll explain exactly what that means and how flow becomes this analytics hub that will drive your progression through the continuum for analytics. Cool.
03:29
Jaco
Analytics is much more than a buzword. If you followed some of the podcasts and some of the people we’ve been speaking to in the flow of data, from simple on screen KPI to dashboards reporting and then analytics, it’s not always an easy to understand process. And I think Lenny is going to demystify a couple of myths around that process for us today.
03:55
Lenny
Cool.
03:55
Jaco
And then we can end up with some Q A. But again, if you have any questions, don’t have to leave it until the end. You can ask us questions whenever you feel ready or confident enough. We’re all friends here, don’t be shy. But we will save a little bit of time at the end to go through any of those questions. A reminder that we have a flow regalia pack that has been paid for by the party involved. We are doing two live giveaways of some awesome merch or swag. I know that’s one of Laura’s favorite words. It is some amazing sponsored gifts from flow that we’re going to give away live today. And we’ll do that during your session, I think.
04:38
Lenny
Great. Yeah. So you need to listen carefully. Listen properly. Now, the questions before this is going to be based on presentation I’m giving. So, yeah, just make sure that you pay attention.
04:52
Jaco
Okay. Fantastic.
04:54
Lenny
Right.
04:54
Jaco
So looking at our change makers or the people that are leading the charge for data liberation and just doing things a little bit differently in our industry. We’re going to start off with some of our partners. So since our previous update there’s been a couple of new certifications. CSI achieved their gold certification. On ignition we have Donna pcs, global perfide and DRD Gold also doing achieving ignition certified status. It is really fantastic to see customers starting with certification. It is something that is traditionally a little bit unusual. So it’s really nice to see DRD Gold. Mike from DRD Gold getting on the certification board Canary new partners controltech Eritron next integration doing a couple of exciting canary projects on the flow side. New partners Eritron and Bluetech System Solutions quick important update on the flow certification.
05:50
Jaco
You would have received an update our flow partners on the flow certification that there is a new certification for flow that is very relevant to the latest version of flow. So please do make sure that you have a look at that and get that flow certification done at your earliest convenience. A couple of new customers that have come on board. I like the variety game. There’s some food and Bev customers, some water industry customers, manufacturing Grinrod TCM in Mozambique. So joining our community is a good variety of verticals at the moment, which is always encouraging to see that the solutions built isn’t industry specific, it’s cross industry. So well done and welcome to our community, to these new customers and partners. I want to spend 1 minute quickly talking about what is available as far as learning goes.
06:47
Jaco
So we do talk a lot about concepts. A lot of people in our community, in our industry talk a lot about I’m going to call it buzwords. So we call it buzzwords like digital transformation. Iot. There are amazing learning resources available on these topics. I’ve got a couple of them on the screen. You don’t have to worry about asking where you can find them. We will absolutely share in our follow up email. We will share a link to all of these, but there’s a couple of really nice guides or helpful guides that talk about simplistic terms. What is digital transformation? Solving data problems to accelerate digital transformation we’ve often spoken about how data in the context of digital transformation is actually a barrier to digital transformation, building a sustainable and secure scalar system.
07:39
Jaco
So three really nice guides from inductive automation the podcast we haven’t had a steady flow of episodes this year. It’s essentially just a time constraint on our side. But we do have some really valuable podcast episodes that’s available to you right now, where we talk about data and data challenges in the last couple of episodes with some awesome people and men and women in our community. Two really nice guides on industrial database options. The industrial database world is also exploding with different options, some of them open source. So really nice guide to help you understand what those options are and the secondary guide to properly evaluating those options that you have available and understanding what is fitful purpose. Two nice guys there from canary, and then the flow side, another three letter acronym, let’s call it information, decision and action.
08:39
Jaco
A very awesome conversation that you and Vaughan had around the topic of information, decision and action and what that means. Yeah, there’s also next level analytics. The combination of canary and flow. A nice video that was done then. Also very simplistic. Flow KPI projections 100%.
09:01
Lenny
And what I’m going to show today, a little bit as well, is also how to share this out to not just the products in our stack, also to making sure that we liberate the data as we stand for in the DLF. Flow is a great tool to be able to liberate that data and information in that sense, and to make it available to pretty much everybody in the organization. Cool.
09:24
Jaco
Fantastic. So these are a couple of nice guides. It’s easy reading, it is easy to follow, and hopefully you will find some of these valuable just to make sense of some of the noise and some of the confusion that exists. All right, so that was a quick intro in terms of what’s new on that side of the community. We are going to cover what is new in ignition first. Laura, can you.
09:48
Laura
Okay, great. So guys, inductive automation just released their newest version of ignition, which is 8.1.20 on the 23 August. So just a few days ago, and they’ve got a few new features and additions that they have added to the software. So firstly, there’s a new paste and replace operation that has been added in perspective. So what it does is if a component is selected and the user prompts to paste in another component, it will just be replaced by the component copied into the clipboard. So no more worrying about the rest of your components changing sizes, moving around the screen whenever you remove a component. So I can just replace it and keep the same width and height and size of the component that you have replaced. So that’s also just cool. Very nice feature.
10:39
Laura
Something that we have all been waiting and begging for, that I know of, is co branding that they have added to prospective projects. So, guys, this allows the adding of design and implementation strategies against loading and logging screens. Not just that, there’s a few other places where the co branding gets noticed in projects, but those are like the main ones.
11:03
Jaco
That’s very cool.
11:04
Laura
Yeah. And you can also include the project selection on the gateway, which includes the customization of logos, favicons, backgrounds, colors, buttons and so on.
11:15
Jaco
It’s quite important because ignition is a fairly important application within a business. And if you do have, like many companies have, you have a corporate identity with set colors and what’s it called?
11:29
Lenny
Favicon.
11:29
Laura
Favicon.
11:32
Jaco
You could make that application, give it a brief, familiar look and feel to.
11:36
Laura
The people interacting with it, really personalize it a lot more, especially for big companies that have specific brands and color brands. Grant colors. There we go. Okay, so I will demo that to you guys a bit later as well. And then this is just, I know something that a few people have been asking for is they now allow the storing of byte arrays in data sets as well. Cool. And another new feature that has been added to the designer is the new tag report tool. And it actually came out in April 1.19. But I don’t think a lot of people know about this feature. So it actually allows the generation of reports for all of your tags on a system, obviously according to specific criteria that you have set up for it.
12:22
Laura
So not only can you just create reports, you can now also search through the tags, edit them, delete them, copy them, replace them. You can pretty much go crazy. And this is great if you have a lot of tags that you have to work. So I’m thinking of guys that have told me they work with 10,000 no tags in the design. So that’s pretty cool. And a bit of insider information. A tag cross reference tool is coming out in one of the new releases. They don’t want to tell us when, but it’s coming and they say it’s going to be very cool because it can go hand in hand with the tag report.
12:59
Jaco
That’s also been a popular feature request.
13:03
Laura
And as you guys know.
13:06
Jaco
Yes, you wanted to mention diagnostics tags.
13:08
Laura
Yes, please do. Those diagnostic tags that yaka flipped over. Sorry guys, you know that you can view them in the gateway, but they have now also made them available in the design as actual tags. So that’s great. So no more jumping between gateways and designers. You can actually use them and their values in a designer, which is cool. And of course the multiple bug fixes, as always, if you guys are interested in checking them out, a complete list of them are available on their website under the release notes for version 8.1.20. And yeah, guys, you can find the newest version where you would normally find it under the downloads page on inductive automation’s website.
13:52
Jaco
Yeah, and for those of you that are not familiar, the download that is available on the inductive automation page of ignition is actually a full version of ignition. So you are pretty enabled to build at your heart’s content and design and really get quite involved with the project. The only limitation is that gateway will reset every 2 hours.
14:13
Laura
I also have the mining trial mode to show some of the security. Actually moving on to the next slide. Yeah, so for those of you who do not know, I think a lot do, but there might be one or two. Inductive automation took part in the s four x 22 conference this year. So very specifically, guys, it’s a conference where they have a lot of very smart people.
14:43
Jaco
It’s a wide hat hacking conference specifically for industrial automation, I think.
14:47
Laura
Yeah, industrial control system. Yeah. So inductive, they take part yearly, but especially this year, they kind of went in full throttle because it’s very important for them to keep up with security standards and ensure that they provide the best security that they possibly can, especially to ignition. Right. So this year there were a few vulnerabilities that were identified. The dev team jumped on it very quickly and pushed out all of the fixes in 8.1.17. So this is why we would recommend all of our ignition users to please update to at least 8.1.17 or higher to get all of these security features for our 7.9 users. They have backported the security fixes into 7.9.20. So that’s also available for all of you guys that’s still using seven point.
15:41
Jaco
You do still have a number of people that are very happy still on version 7.9.
15:46
Laura
And also if you guys are keen on upgrading your security, as I think a lot of people need to be, actually, you can take a look at the security hardening guide that inductive automation released a while back. It’s such a great tool that provides you recommendations on how to secure your.
16:04
Jaco
Ignition installation and very important for it teams. I think very often the most discussion that we have around security is with the it folks, as you would expect it. And the feedback we’ve had is that this security hardening guide is very valuable to them. It really is an end to end guide around security. Best practice implementation. So yeah, we’ll also include a link to this in the follow up email.
16:26
Laura
Yeah, cool. Okay, so just to dive in a bit deeper about specific things that they have changed. So we’re going to start off with the designer. So some of the biggest changes that they’ve made actually came out in 8.1.18 and obviously then build upon that in 1920 and so forth. But the biggest changes were made to the script editor, the tag browser and the tag editor. Now these can, like I mentioned, all be found in the designer. So for some of you guys that don’t know the designer is the application that pretty much puts you in the driver’s seat, that provides you the robust.
17:04
Jaco
It’s your engineering client or your engineering station or your design. With admission you have unlimited. So any number of designers can work on any number of projects at the same time. So that’s quite nice.
17:16
Laura
Yeah, so that’s where we build the projects. Cool, right. So to start off with, well, the second one, because I think the first bullet, I’m going to demo that a bit more.
17:26
Jaco
Okay.
17:27
Laura
That’s a lot easier to explain.
17:28
Jaco
Split containers.
17:29
Laura
Yeah. So second one, spit containers. Brand new container.
17:34
Lenny
That’s fun.
17:35
Laura
We love containers. Right? I’m just kidding. The people who know containers are like.
17:43
Jaco
Your wife loves containers.
17:48
Lenny
Yeah.
17:48
Laura
So guys, container is ideal for data, then screens. Why? Because as the name suggests, the container can split into two sections and it’s divided by a user adjustment slider. So that’s pretty cool. So just imagine you have a very limited space on a screen. You need to show quite a bit of information and it’s great, especially if you don’t have to show everything all the time. You give the user the opportunity and the capability to switch between views in one little container.
18:24
Jaco
This is a design container, by the way, in case you’re wondering, this is not a tupperware. There seems to be two trends where the one sort of follows having the minimum on the screen and only that’s relevant and applicable. And then you do still have the trend of people wanting to see as much as they possibly can one screen. So I think split containers will aid in that letter scenario.
18:51
Laura
Yeah, pretty cool. All right. And then they have added a find and replace menu option in the project browsers.
18:58
Lenny
This is very.
19:01
Laura
So it’s cool because they tell you it’s only in the project browser, but if you open it up, you can literally browse your whole project. Right. But this is pretty cool to find like views, transaction groups, windows templates, especially if you have a lot going on. And I know it’s like pretty much exactly the same as when you push control f you can find stuff you can replace and so forth. So that’s kind of the same concept. And then the last one is the added tag tooltips in the tag browser. So it shows the type of tag as well as a tag pass. In other words, you just hover over a tag, it brings up a little bit of information. Exactly.
19:47
Laura
And this minimizes the time that you actually have to spend to open up the tag editor, go and look at the tag type, or open up the tag to go and look for the tag pass. I’m not talking to you clever people that knows what tag type it is by just looking at the icon. Okay, we’ve got some new people coming in as well.
20:06
Lenny
But yeah, it really reduces a lot of custom properties.
20:13
Jaco
Do you want to talk through these quickly?
20:15
Laura
Yes, quickly. Script editor. Right. So guys, that’s where we script some python code for some additional capabilities that you want to add in a project. Now, the updated script editor is smarter and it’s more aware of DPI scaling, and it comes with a lot of quality of life improvements. Right. First up is our added inline find and replace feature. Prior to the update, the search function would appear via a modal or a little pop up dialogue. Let’s be honest, that was just getting in the way. Now with the new editor, you can perform a search without adding additional windows or obstructions to your workspace.
20:52
Jaco
That makes sense.
20:53
Laura
So two additional quality of life features are the code folding and the visible white space. Probably people are like, what are you talking about? The coders are sitting there like, yes.
21:04
Jaco
If you are a coder. And there’s a couple of. Is it hardcore? Is that the description for a coder? A hardcore, what do we call them, basement people? There are a couple of really phenomenal scripting and coding resources. Talent in our community, if you’re aware of this, let us know if you’re excited about this. I think you may definitely be.
21:25
Laura
So just real quick, the code folding guides that gives you the ability to collapse tasks and function bodies in the editor. And then it just pretty much means no more scrolling around for hours to find a specific function. And the visible white space is particularly useful for Python because you guys know we love indentation in Python. So it pretty much just reminds you to indent where you indent it. It actually makes it a lot more visible.
21:52
Lenny
The code folding, is there a right click collapse?
21:54
Laura
Oh yes, I’m going to show it to you. Yes. So that’s pretty cool as well. And then lastly he added autocomplete pop ups and parameter completion. Again, the pro scriptures are probably very happy about this one. So variable and parameter types now have type awareness but not type awareness as ides like in eclipse as we know them. This is unfortunately data types, but it’s still typo, right? So in various contexts, yes.
22:26
Lenny
Jaco.
22:30
Jaco
Parameter completion parameter.
22:34
Laura
Guys, there’s no more control spacing. Looking for a function, if you click on a function, the parameters disappear, you can’t remember what’s the data type, there’s no more of that. It autocompletes it. Yeah, it’s just absolutely amazing. It literally simplifies your scripting.
22:51
Jaco
Damien, hopefully we made your day.
22:55
Laura
Damien probably already knows about this. All right. And then lastly, guys, the tag browser and editor. Guys, it’s just a central location for interaction with all of your types of tags, especially in the designer. Right now, the tag browse trees. First off, I know a few people know about this. They used to constantly collapse, right? Requiring the user to repeatedly have to expand them over and over. So now the initial opening of the browser selects the tag that you are currently editing so it doesn’t just close everything up again. And then they also brought back category panes. This is cool, especially in the tag editor. So normally all properties were visible. That’s making the view overwhelming and harder to focus on.
23:46
Laura
Now the category panes group related properties together, so it just makes it a lot more, a lot easier to focus on the important parts. If you want to do some alarming, you can just click on the alarming category and all your properties are there that you need, no more need. I know it isn’t 7.9, it’s still there. They just took it out for a while and people correct. And then persistence of categories, guys, that’s also just when you make any type of edit to a tag property, you guys can remember. If you edit something, it literally takes away all the properties, it brings you a new screen and sometimes you can’t even remember where you’re at the moment. So they took all of that away.
24:26
Laura
Now if you edit on something, it’s a brand new little pop up screen, so you know exactly what’s going on and where you are. So they integrated the edit panes a lot better, which is cool, right?
24:37
Jaco
Are you going to show us?
24:38
Laura
Yes. And then that tag report tool, I’m going to demo that, you guys.
24:42
Lenny
Okay, should I stop my share now?
24:46
Laura
Yes, I need to be able to share my screen.
24:52
Lenny
Stop my share.
24:55
Jaco
There you go.
24:59
Laura
You disabled participant screen share.
25:09
Lenny
Whoopsie. Let me, here we go.
25:12
Jaco
Should be able to now.
25:13
Laura
Okay, let them see. Right.
25:20
Jaco
There we go.
25:21
Lenny
Cool.
25:22
Laura
Everyone see my little perspective? Cool, great. Okay, so real quick I’m going to jump through the designer to the gateway to the demo. So first thing is, guys, it’s the tag tooltip. So if I hover over a tag, it tells me what tag it is and the tag parts as well. And then. Sorry, let me just minimize that. There we go then. Also, like I mentioned, you can just type in control f. It’s like, wait, I’ve got a Mac. So it’s command f. Okay, so here’s that little find and replace window that I spoke about. So you can pretty much go and find anything in your project. So if you type in home and find it, you’ll see it brings up. Okay, these are the matches that I found. And obviously you can replace it. So everyone knows how find and replace works.
26:12
Laura
All right, so this is nice. Brand new feature. This is cool. And then the tag report tool. So you can go and click on the little buttons here in your tag browser, those three little ellipses on the side, and you’ll see they have an added tag report tool there.
26:29
Jaco
Is that a burger menu or is that something else?
26:31
Lenny
Burger?
26:31
Laura
No, burger menu is a live screen. This is like a little donut menu. Right? So here you can actually go ahead and search for all of your tags and you can create a little tag report tool. So you can search according to tag path, the quality, the tag type. Oh my goodness, are you going to stop? The types of tags, the traits, these are cool. So you can check which have alarms on it, which has history enabled. Guys, this is amazing. If you’re working with thousands of tags and you want to see where’s your alarms. Amazing. And then you can also search according to udts. So ancestors.
27:10
Tebello
All right, that’s very comprehensive.
27:13
Laura
Yeah, that is pretty cool, right? Let me check. I’m going to search just real quick for some good quality tags. And then if I search it will bring up all of my tags that are in good quality. And let’s say, for example, now I want to see which has alarms configured.
27:31
Lenny
Boom.
27:35
Laura
So you also have the option to go and export this report so it can save it as a CSV file, which is nice. You also have the option to change to configure some report columns. So if you don’t want to see the path and attack type and quality, you have options for pretty much look at all that. The data source, the dead band mode engineering units. This is very thorough, which is pretty cool.
27:59
Tebello
Can already see myself using this to find the bad quality tags and to.
28:03
Lenny
Check if the tag names are correct.
28:05
Laura
That’s usually a problem as well. And you can also go ahead and save the report. So if I save the report, right. Apparently I already have one. That’s fine. You can also then load a saved report that you have. Cool.
28:21
Jaco
That’s where you are.
28:23
Laura
Yeah. So that is a brand new tag report tool. Very excited for the cross reference. And then we also have the tag editor that has updated a bit. So if you double click on a tag or you right click and say edit tag. There we go. They updated this. This is what I meant by they now made categories. They kind of like limited it to the side, which is cool. The category panes. And just like I mentioned, if you go ahead, let’s say we go to the metadata, to the value. And this is an expression tag. And I actually now want to go and edit this. It’s a pop up. The screen doesn’t just slide to the right and you can’t see what you’re editing anymore. So this is very cool. So they changed that as well. And then our script editor. Right.
29:15
Laura
So this is very fun. So if you now go ahead, let me just add the little script over here. For those of you guys that are familiar with this, you’ll see the description and parameters have been removed from your script. Little script pay. It’s also now collapsible, which is great. And another cool thing is if you go and click on a parameter, it automatically pops it in now. And this is what I mean by the autocompletion.
29:46
Lenny
Right.
29:47
Laura
So you can see it now in action over there. This is pretty cool. So yeah, absolutely love that.
29:54
Lenny
Nice. Thank you.
29:55
Jaco
Probably a little bit technical for some, but definitely helpful for those that have been asking for a couple of incentives.
30:02
Laura
And just real quick, the spacing and indenting. This is what I meant by you can see the little visible white space. So if I click on whatever it might be, there’s no more control spacing. Right. You just put in the little full stop and it gives you all of the information. So there’s the parameters on t shirt. Stunning, stunning, stunning. Absolutely love this. So for Lenny with that Os, you can right click here, you can fold all. It’s got the options toggle the current fold, collapse everything, collapse all the folds and obviously expand all the folds. You also have options for the autocompletion. So you can decide what you want to do, parameter systems or not. And this is also cool where you can enable or disable your tablines in white space.
30:57
Jaco
Does make for a lot cleaner interface.
30:59
Laura
Definitely. So. Absolutely love this. Just real quick, the perspective branding. Sorry guys, they added it now to the bottom of the config page. So there’s the branding customization. You can decide what colors you want, you can upload your icons, your favicons and so on. And this is pretty much what it would look like. So real quick, it looks nice even in the login screen as well. And apparently can’t remember my own password. There we go. Cool. So here’s that split container that I mentioned so you can move your views as you want.
31:42
Jaco
That is very.
31:43
Tebello
Can you also do a vertical split.
31:45
Laura
So you can do vertical as well?
31:49
Jaco
Very nice. For a split between table data as.
31:53
Lenny
You add there, for example.
31:55
Laura
So this is the branding. Very fun.
31:58
Lenny
Very.
32:01
Jaco
I know it is a big design implementation consideration, the ability to do branding on your application to have a more familiar look and feel.
32:11
Laura
Yeah, cool.
32:12
Lenny
Laura, thank you very much.
32:13
Jaco
Sorry to rush, but that was helpful. If you have any questions on any of these updates, please let us know.
32:21
Laura
Going to take over.
32:22
Lenny
Yeah. Thank you very much.
32:29
Jaco
All right, let’s look.
32:35
Laura
Mr. President, thank you for the demonstration.
32:39
Lenny
Okay.
32:39
Jaco
On the canary side, what’s new?
32:41
Lenny
Cool.
32:42
Tebello
I don’t know if everyone knows, but on Tuesday the 30th we had a brand new release .2 update of the 21 release. 22.1 release. It was a buck release fix or buck fix release. Major points. We have our external performance improved for historians that have greater than 60,000 eggs. If you have very large instances. You used to have performance issues. Now that has been resolved. And our OPC collector, previously we’ve had an issue where it was your tag group would disappear when you add your tags from OxXL. Now that has been fixed. So that should come as a much needed update.
33:23
Jaco
And this was this past Tuesday?
33:25
Tebello
Yes, this past Tuesday.
33:26
Jaco
So when we talk about. That’s what we’re talking about when we’re saying Tuesday’s update.
33:33
Lenny
Okay.
33:33
Tebello
And then we have a few enhancements with regards to Iiot. MPTT collector now supports ordered messages. So what this means for you is when you’re defining the integrity for a message, you can now choose, I believe it’s option one on the quality of service which allows you to basically control the messaging integrity as well.
33:59
Lenny
Cool.
33:59
Jaco
Okay.
33:59
Tebello
And then also this is just a general 2.1 upgrade or update that is now we can send JSON payloads over MQTT that we can fix as also you can actually edit them and then.
34:14
Jaco
Send them in your MQTT functionality.
34:19
Lenny
It’s fairly comprehensive method of defining your.
34:23
Jaco
Palettes and MQTT is becoming a very popular protocol as well.
34:31
Tebello
Then with regards to actually administering or administrating the historian, this is major to me, but the history blocks, you can now define where to send old ones instead of having to delete them after you make changes or you now don’t need them anymore. So that improves how you manage your storage. So easier to use if you have another drive that you’re sending them to can point that out. And then now the obsolete tags in your historian have a different mark for you to see. You can easily identify them in your data file to see which ones are actually not in use anymore.
35:08
Jaco
That’s very helpful. That’s very helpful.
35:12
Tebello
A few quality of life enhancements. Now Axiom allows you to manually define the time for your report. So when you run a report, you can manually say from this time to this time. Previously it was just a time block that you were defining. Yes. Now you have greater precision over the time that you can use, if you’re not aware.
35:32
Jaco
For Canary, Axiom is the client tool for web based client tool for those.
35:36
Lenny
Of you that are wondering what Axiom is.
35:41
Jaco
And views.
35:42
Tebello
So if you remember that views are actually the hierarchy you have for your tags and your structure.
35:48
Lenny
Right?
35:49
Tebello
So that we have a new rule that we’ve added that allows you to copy your tags from one part of the model to another so they can easily use the same configuration and then just suggest it for the new part.
36:01
Jaco
That’s a big time saving.
36:05
Lenny
Okay.
36:05
Tebello
We are also expecting a new release for the 2.2 in October, November this year. And the major update for that is the SQL collector that now allows you to bring in data from an oracle database. What that allows you to do is if you have other business systems that have been relying on that uses Oracle, now you can bring in that data into Canary and then eventually share it with other parts as well. So you have access for more SQL.
36:35
Jaco
Technologies as your data source. Oracle is a popular system.
36:41
Lenny
Okay.
36:42
Tebello
And then axiom, I know a few.
36:44
Laura
Customers has asked about data freaks.
36:50
Jaco
Data enthusiasts.
36:52
Lenny
Yes.
36:53
Tebello
Now the data enthusiasts will be able to add data table control on their application so that you can see the data in tabular format. And then obviously that gives you more options for presenting that data depending on what message you’re trying to get across.
37:10
Jaco
Most of us are data, are obsessed with data, as we should be in our world, in our industry, and many people do, they don’t want to see actual table data in their views. So that’s very nice. Yes.
37:20
Tebello
And then for those of you who are using our read API. So for those of you who do not know what the read API is. So Canary allows you to read data in the Canary historian and share it with the party system using the rest APIs or the rest scheme of defining data. We’re adding get annotations. Now what that will allow you is to take comments and comments made to plans that are associated with specific tags and then take them to third party systems. So that will allow you at least to have better collaboration with your teams. If there’s another application that relies on comments or has a space where you want to add those comments, I like.
37:56
Jaco
That because very often there’s context in those annotations that’s super important. That’s very critical. Outside of then looking at those annotations, you can now share it the same way that you would share some of your other information by API, especially when.
38:10
Tebello
There’S something critical that happened and now someone left a comment, you can now see why would that happen?
38:14
Jaco
Right, okay.
38:16
Tebello
And then we are sunsetting a few, or canary sunsetting a few features. But that is only because there are other ways of doing it that are better and those will be sunset in version 23 coming next year. That is the had server, the mirror service and trend link. For those of you asking what is going to happen with the mirroring service? Are we losing mirroring?
38:40
Lenny
No.
38:40
Tebello
So what is recommended now is to use dual logging or you can use proxies to actually send the data to different historians. So you’re not losing out on anything, we’re just giving you the better way of doing it.
38:50
Jaco
It’s actually a more flexible configuration option that you have now as far as your architecture goes. Yes.
38:56
Tebello
And of course if you still want to use these, the last supported version will be 22.
39:01
Lenny
Okay, fantastic. Okay.
39:03
Tebello
An exciting one here. We have native control collectors for these brands and protocols. This is coming through the updated service. It’s a new service that will be added right now, is available upon request. But what this allows you to do is now you install that updated service that wherever you have your close to.
39:25
Jaco
Your source possible, doesn’t get any closer than this. Yes.
39:30
Tebello
And then with that you can actually get the data directly from your protocol or from your modbus server or device or your PLC without having to.
39:38
Jaco
Very helpful, yes.
39:40
Tebello
And that’s especially helpful where you don’t have an OPC server so you can get that connected directly.
39:46
Lenny
Okay.
39:47
Jaco
Was it Debbie?
39:48
Lenny
Thank you very much. Thank you. Appreciate that.
39:50
Jaco
Right, Lenny, we are going to have a chat about the analytics continuum. See, I’ve got that on the second.
39:56
Lenny
Let me quickly share my screen.
39:59
Jaco
Thank you, guys. And again, if you have any questions on any of the pointer topics that you’ve just seen, please let us know. But we will also share a couple.
40:09
Lenny
Of links with you after. All right, thanks, everybody. Yeah, we’re going to talk a little bit about the analytics continuum. Now. You’ll notice on the flow side, we kind of beginning to call flow analytics hub. I know in the past we spoke about flow as the information hub. Some people will think it’s a reporting or a dashboarding tool, but in essence, what it is it creates analytics hub, a very good foundation for analytics that you can then share via various means out to other systems people. Whoever needs to have this analytics information available. Now, the analytics continuum, very big word, as we mentioned for a Friday. What does the word continuum actually mean? Time traveling. So what is a continuum? A continuum is a series of elements that gradually changes over that series. Okay? So temperature is a continuum.
41:12
Lenny
The degrees change, very small increments, but the results can be from freezing to boiling. Your school education is a continuum. You started at kindergarten and hopefully you ended up at Matrik.
41:27
Laura
All right?
41:28
Lenny
Gradual steps.
41:29
Jaco
Matrix was the best.
41:32
Lenny
That is a continuum. All right, so analytics is exactly the same. Analytics is a continuum means it’s different elements in a series, and those elements change as you go higher into the range of.
41:44
Jaco
It’s really critical point because it is absolutely fluid. It is fluid. It’s almost a natural progression from purely viewing a KPI as a fixed parameter at a specific time value at a certain time. This is fluid in understanding the impact of that fluidity over time.
42:02
Lenny
Right. In high essence, analytics is then the practice of using your data. But to make informed decisions of that data and depending where you want to sit on the continuum, it’s going to depend on what type of analytics you can actually perform or deny those decisions.
42:18
Jaco
Correct. Based on your permission that you are.
42:21
Lenny
All right, so remember we said at the end, the comrades must listen. So listen properly now. All right? What are these series? The events for the analysis continuum. All right? So right at the bottom, we’ve got what we call descriptive analytics. This is where we ask the question, what happened? All right, by the virtue of the word happened, it means it’s historical. All right, so we look at historical data. We look at what was the production for the last shift? What was my water usage for the last shift? And with those two, I can then work out what was the intensity for the last shift. The key here is that if you have truly descriptive analytics, it means it’s repeatable and it’s automated. So that means no running around trying to collate things.
43:07
Lenny
To get this answer, it must be automatically done in a solution and provided by that system automatically in the back end. The next step, then into the continuum is then when we go a little bit higher, where we now look at diagnostic analytics. So why is it happening? I know my water usage was high for the past shift.
43:29
Laura
Why?
43:30
Lenny
Okay, now the biggest thing here is that you need to now add context to this historical information.
43:36
Jaco
And why is it, for example, higher on a Wednesday morning as opposed to on shift every Wednesday morning?
43:43
Lenny
Why is it higher when I run product ABC, whatever the case is. So the more context you can add, obviously, the better correlation potentially you can get with the questions around this KPI. And very importantly here is that we want to try and find correlations between things. And very important is that human entrance commentary can play a massive role in adding that additional context to this topic.
44:09
Jaco
And I love that kind of interaction with the system, because the context that we get from people, humans, in these processes and these applications every day, that’s the gold mine very often, which a system can’t give you that.
44:24
Lenny
Correct. All right. Now, after we understand why it’s happening, then we can go into this whole role of predictive. So we’ve got all this context, we understand a little bit of that historical information. So now we can start doing predictives and projections. What’s likely to happen? Again, we now talk about historical data, but now we can do a little bit of modeling on that data. We can apply regression, we can project.
44:51
Jaco
It into the future.
44:52
Lenny
We can do a little bit of forecasting of that data. And that then gives us this whole world of predictive analytics. And then to matriculate or to get to the end of your analytics continuum, we’ve got prescriptive. Right?
45:03
Laura
Of course.
45:04
Lenny
And this is now. Okay, I understand I’m going to have a failure in three months time. What potentially can I do now to avert that type of failure in three months time? Change the bearing, change the glance? Well, whatever the blue maintenance now. But that’s, like I said, it’s the holy grail. This takes that tangible knowledge of the humans into account as well. Me and you always joke about it, that women on the site where he just listens to the compressor and say there’s something wrong. That type of knowledge, you need to be able now to institutionalize into these systems, correct?
45:42
Laura
Right.
45:42
Lenny
So that’s continuum. That is the four steps to get to your analytic process. Now, let’s put them a little bit on the side here. What we want to do, though, is we want to be able to build a foundation of analytics. And my advice here is just don’t go straight into machine learning, AI predictive. I’m going to go right into predictive without having the foundation to actually do your descriptive and diagnostic analytics as well. Because what we see is that you have to start building this up. You can’t skip a few grades in school. I know in the olden days, someone’s parents probably skipped severe. Whatever the case is, I was never.
46:24
Jaco
Eligible for that for some reason.
46:27
Lenny
So don’t skip the continuity.
46:30
Jaco
It is that foundation is super critical because that foundation includes the quality of where the data comes from and what.
46:36
Lenny
It means and where you’re going to send it to.
46:38
Jaco
Exactly.
46:38
Laura
So you can’t predict on something you don’t know anything about.
46:41
Lenny
Correct. All right. But other than that is that we actually find that a lot of business cases and business value can be derived by the first two layers in the quantity. If you can institutionalize your descriptive analytics and you can get all of this context in one place to do your diagnostic analytics, you can solve a lot of business cases. And then obviously you can advise or go into the predictive and prescriptive pillars of that. All right, now, again, listen carefully. Listen properly. How do we do it at flow? So what does flow do? Flow creates this analytics foundation. And there’s five steps that we take to actually build out this foundation to give you a proper diagnostic and descriptive analytics platform. Right? So how do we do it? First of all, we model it. Right?
47:33
Lenny
So we will go and create a model of your KPIs, of your plant. We will then connect to the various data sources to create and pull the data in from. We will transform from those multiple data sources the analytics that you require in one single source now. And if you have that one single source that’s now transformed, calculated, aggregated, we can then either visualize it within flow or very importantly. And my most excited feature is about bridging it out to other systems. And this is the five steps that how we at flow go on to creating this analytics hub. All right, so how do we do it? A little bit more detail. What we do first is we create a model that’s an asset structure or hierarchy of your KPIs, your plants, your assets.
48:27
Lenny
And very important at this point is that, do you see in my model where the data is coming? No, it’s not relevant. So we decouple our model or the analytics hub or the KPI model from the actual sources where it’s coming from. It should not actually matter where this data is coming from because we have the capability to merge these different silos of information into one place. And very importantly, we have to be able to add context, right? So I know it sounds like a silly thing, but we do have the concept of adding context of time. That’s the first piece of context that we add to data that can be what is a day, what is a shift, and we add that type of context with what we call calendars to the data as well.
49:14
Lenny
That’s very, in essence, the first step of modeling the data. Then we can worry about connecting to different data sources. Now we’ve got different classes, types of data that we can connect to. The first one is what we call our real time class, right? So from version six upwards in flow, we added the capability for real time buffering. So what that means is flow will be able to buffer data from a real time source. Currently we support MQTT and MPC UA, and we will buffer that data for seven days within flow to then perform the higher level analytics, do the aggregation, and store the results inside of our database. It’s a rolling seven days and we can get it from these real time systems.
49:57
Lenny
The second type of floss of data source that we can connect to is what we call our time series historians or industrial historians. Obviously canary, ignition. We can connect to those. Jaco, you mentioned that article that canary has about all the other types of historians or database solutions coming into it. We can connect to influx we saw as one of those freeware ones that’s pushing down. So we can connect to that. And we kind of classify it as either the NoSQL type of historian like canary, the SQL based ones like ignition, and then there’s also a few cloud options where we can get the data from as well. And then the last class of data source that we can connect is, we call it our relational or transactional databases.
50:41
Lenny
This now includes typical Microsoft SQL, Oracle, postgres, MySQL, and this is for those applications that has data in the ERP.
50:50
Jaco
System, limbs, and very popular, correct?
50:55
Laura
All right.
50:56
Lenny
Now when we connect to all of these things, we now need to be able to transform the data. Sorry, before that, very important, we spoke about this. The manual entry capability, need to be able to have that manual context from unum to do that, add that manual caption. It could even be used to correct.
51:14
Jaco
The data that’s in two flavors. It’s either somebody selecting something at that time, or it’s in the form of a csv excel, it’s captured data, it’s.
51:23
Lenny
Form in our web. We serve it as a web browser. And at that point in time you go and capture the data available for.
51:30
Jaco
Us and you can define some of those.
51:32
Lenny
Correct?
51:33
Tebello
Can I be able to just use flow with only form data?
51:36
Lenny
And do you, can a lot of people start with that? That’s why I said the model is the important one.
51:41
Jaco
And Tebby, if it’s just to replace somebody running around with a clipboard, that’s already a big win.
51:47
Laura
They can now run around with a.
51:49
Jaco
Tablet and not waste the time trying to put all of that together afterwards.
51:56
Lenny
All right, I’m going to spend some time on the transformation part because this is where the secret source is. And we made quite a few changes in the latest version of flow regarding this. Now let’s take an example. I’ve got data from a historian, tag Canary historian, and that stores the data as a series of points. It’s the timestamp value and quality. We call this in flow a tag. Right? We might also have a SQL source, might be a schedule that I’m getting from my ERP. I’m writing a SQL query and I also get that amount of data into flow as a sequence of points. Now I need to perform analytics on top of these two tags coming in. As an example, I want to know while I was scheduled to run, how much energy did I use? Scheduled to run comes from mes.
52:50
Lenny
While running is the run feedback I get. That’s easterizing my story. So what we do is we take those two sources, we create tags in flow. And what we do is we do a point normalization on those two sources. Very critical. We do not lose data by doing this. So we do a row based calculation for each one of the points with zero data loss. And now what I can do is I can create what we call a calculator tag. All right, so now I know exactly while I was running. Do I need to use the feedback that I’ve got inside of that historian? Examples of what I can use this for. I can now filter out bad quality.
53:33
Lenny
As an example, I can filter out a range, I can filter out a top and a bottom, or I can filter out based on a different data source, which is very important. All right, now the next thing that we will do is we can use this tag for two things. I can use the tag information to go and create events. All right, so I can go and specify, I want to trigger on a downtime event as an example. And I can utilize my tags to create periods or event periods and now have all of this additional metadata and context associated with this. What was the shift? What was the product? What was the operator on that shift? And we use our events to create that. We store the event then inside of flow, examples of events is batches, downtime events, or machine running states.
54:24
Lenny
Right. I can also use the calculator tag to now add the context of time. So if I now want to know what was my hourly usage while I was running? Well, we will then do what we call period normalization, and we will store those period normalized values inside of flow as well and call those measures. Obviously, that’s the one that everybody used to remember, and this can now be your total production, et cetera.
54:49
Tebello
So with those tags, it means that I’ve cleaned my data before I’ve even used it in flow.
54:53
Lenny
Correct. That’s the whole point. The whole point is that we’re extending the transformation capability by introducing this whole concept of calculator tag. Obviously, I can overlay my event context on top of that data as well. Then we can go and visualize it within flow. We do have our own dashboarding and reporting solution. I can potentially then use that. It’s in a browser. But the one that I’m most excited about is bridging this data up. Now, we can bridge it back to our sources, definitely. But more important for me is that I can go and bridge it out via our API. That’s what I’m going to demo here. And I can actually go and put it into my ScadA, my bi tools, or my machine.
55:37
Jaco
Learning the messaging is a very important aspect. Or the ability to share with messaging, because ultimately, the end of the day, it’s people. You want the people to get access to that information regardless of where they are.
55:48
Lenny
Correct. And not just from messaging perspective, but think about your operator. That’s the part I’m going to demo is how do I get this now into my Scada? Cool.
55:57
Jaco
All right, do a quick handout.
55:59
Lenny
Before we do that, let’s do a quick handout. What I’m going to demo is how to serve this information via our API. Even the tags we can go and pass through our API. I’ve got a special treat for the guys for our latest version of that. All right, let’s quickly do the questions. All right, so what did we learn? Okay, so flow is sponsoring this Regalia pack. It’s not for free. So here’s the questions. The first one is, can you mention the classification for analytics based on the continuum? That’s question.
56:35
Jaco
I’m going to say, can you name one of the four classifications analytics.
56:39
Lenny
That’s fair.
56:40
Jaco
Hey, whoa. Niels.
56:42
Laura
Goodness.
56:47
Jaco
Thank you very much.
56:49
Lenny
What is the five steps of how flow does this? As I said, it’s okay.
56:54
Jaco
I was going to say just name one of the five steps of the flow process. But looking, I’ll be surprised if Neil doesn’t give us all five again.
57:00
Laura
Neil said the Guys.
57:04
Jaco
Unfortunately, you can only win one other comrades would be upset. Is anybody else able to name one of the five steps of the flow process? Anybody else outside of Nielse? Well done, Nielse.
57:18
Lenny
By the way, I’m just going to continue running a bit. Cool.
57:23
Jaco
Let’s go for it.
57:24
Lenny
All right, so inside of flow, Zachariah.
57:27
Jaco
Fantastic. These are.
57:28
Lenny
Thank you very much. Model and connect.
57:30
Jaco
Well done, Zachariah.
57:31
Lenny
All right, this is the config tool. This is where I create the model. This is where I create the context of my KPIs. You’ll notice I’ve got a little bit of an Oee demo here. I’m working out what is the quality of what I’m producing. So good model count, bad model count, and I can work out a quality score. Now, how do I share this information out to the operator? I would like the operator to have an idea of what is this current quality score? Currently we’re sitting at 95%. How do I share this with the operator? How would you have done this in the past?
58:04
Jaco
And ideally it would mean potentially creating multiple. Before you even get to a digital view, you would have to potentially write something on a board, some way in a whiteboard probably. But for the real time, you would probably have to share multiple screens from multiple systems one single view that an operators have viewed, meaning you have to look at multiple screens.
58:24
Lenny
One of the problems might also be that some of your scadas can’t embed.
58:29
Jaco
Scada can’t necessarily embed the way that she does.
58:32
Lenny
Yeah. So how would I get that value of 99 into my Scada to share it with the operator? I’m very excited. This is something new in flow that we’ve added since version six is we now have our rest API. So if I expand my flow server here, I’ve got access to the API. If I right click on this guy, go to task, say open it will go and open up my documentation for the flow API. I can get the config, what was my calendar settings. I can get the data very importantly, I can get the data for the tag. This includes calculator tags. So if I go and browse here for a tag for an hour, and this is a calculated tag that I’ve built, I can actually get that information passed through right here from that.
59:29
Lenny
So not only the measures or the events that we store, I can actually get the calculated tags from its source and I can get that result back here. So here I’ve created a water usage calculation and there is the values for that. Just to show you guys, what I’ve done is I’ve created a calculation here to say how much water do I use, but only when a pump is running. Right. I’ve got my flow transmitter as a source. I’ve got my pump as an additional tag. These are now tags. We don’t store them, but we do this data cleansing on top of this before we even do the aggregate. So you notice there’s this big null value portion here. Well, that’s where the pump was not running.
01:00:12
Lenny
Okay, so if I pop this guy out here and I take the tag for the pump, which in my case sits in a complete different historian, if I look at this water data here for the pump and I add this, that zero there is where I’m not going to add the flow. Right. So I’m cleaning this out before I do the higher level aggregation. And very cool for me is that I can actually see this in the editor while I’m building out my calc. I can see what the result is. And this is the tag data that I can now serve out with that API.
01:00:47
Laura
That’s pretty cool. I don’t think a lot of people know you can compare values from different data sources.
01:00:54
Lenny
Okay. I don’t want to be stood up by Laura. So I just want to say that we also have the search API. I can go and search in my model, in my metrics, in my events. I want to go and search for, I don’t know, let’s search for the DLF.
01:01:12
Jaco
Is this like the Google of API searches?
01:01:15
Lenny
Yes, the Google. And if I go and look for the DLF. Okay, cool. Under my area site a, I’ve got.
01:01:23
Jaco
A folder called DLF and directly through your API. That’s very cool.
01:01:28
Lenny
All right, so I’ve done this through the API. So if I look here. Yeah, that’s true. Underneath my site area line, there’s a folder called we’ll look for all of these components within this.
01:01:39
Jaco
Right, very good.
01:01:41
Lenny
Now, part of the API is to get the data right, as I explained. How do I get this data now into a scalar like let’s say ignition as an example. Well I could have used the API and I could have coded it in ignition, but what we’ve done is we’ve actually created a model. I’m going to go and log in here into my.
01:02:07
Laura
Super secure, I can’t remember my password and yours is just like insane.
01:02:14
Lenny
All right, so underneath my models I’ve got the flow analytics. So I hook this up and I hook it up to my flow server. My flow server generates a bearer token for me that I can then use to connect to the specific flow instance that I’ve got here. What this will allow me to do is let’s create a new perspective screen here. So I’m going to create a very quick new perspective screen. I’m going to create it and let’s add time series chart onto it. Let’s add an area chart. Right now I could have created a binding on the data for this chart, and I probably could have written HTTP binding to the API, I could have written a script.
01:03:03
Lenny
But now because of the module, if I go to my tag history providers here, let’s say I want to go and look for it for every hour, and I want to look at historical data for the past 24 hours. So I want to give him a view of the past 24 hours of what his quality score is. Okay.
01:03:22
Laura
Yeah, you just have to click on.
01:03:23
Lenny
Okay, thanks Laura. All right, so now what I can do is I can actually browse for my tags. So underneath historical provider I’ve got DLF, this now browses that exact same flow model that I’ve got and I can go into the quality, sorry, this is a little as well, but I can go and expand this and I can see, there we go. There’s my quality for my specific calendar, which is called the production calendar and it’s an hourly KPI. There we go. I can give it an alias quality score, hit apply and look at that slow data inside of there just to compare it. If I do an inline component here as well. Inline frame, drop this guy in here.
01:04:20
Lenny
I’ve created this exact same chart inside of flow as well, so let’s copy the link to this guy, go back into ignition and just paste it here as well. There’s the exact same chart inside of flow showing the exact same data, and you can see that it’s exactly identical. All right, so if I have to leave it like this, you don’t know which one is even coming from ignition or which one is actually coming from flow.
01:04:52
Laura
This is amazing.
01:04:54
Lenny
I can use other components very quickly as well. Obviously, a lot of people will just want to see what the value is. So let’s do a very quick gauge here. And I can do exactly the same. I can bind to the value of this tag. History binding. Let’s go and bind to that same tag. Nice hierarchy that I’ve got here. I want to do that for. Let’s just say there’s some two points. Again, I’m going to do it for historical data. I’m going to go for the past 2 hours. Let’s just give it an alias, else I’m going to forget. So let’s call it quality as well. Right now the data comes back as a JSON packet from our API. So I’m going to change it to a document tag in here.
01:05:49
Lenny
And you’ll notice it gives me the timestamp as well as the value back. Obviously, I need to go and pass it to show that. So I’m just going to add an expression here. And in the expression transform. No, that’s the wrong one. I need to do the script transform. Sorry. There we go. And all I’m going to do is I’m just going to strip out that value portion. I don’t need the timestamp. And I call this signal quality. Just going to strip that out. There we go. I love this recognition that it shows the result of your script immediately.
01:06:28
Jaco
Immediately say apply.
01:06:29
Lenny
And there I’ve got my quality score coming. Slowly utilizing the ignition perspective component.
01:06:35
Laura
That’s amazing.
01:06:36
Jaco
Fantastic.
01:06:37
Lenny
Thank you. Thanks very much.
01:06:43
Jaco
It’s interesting.
01:06:46
Laura
You can just stop the screen.
01:06:50
Jaco
Thank you. So probably. Is that the wrong machine, is it? Could it be? Yeah. Yes.
01:07:03
Lenny
Look at that. All right.
01:07:05
Jaco
That was Lenny’s demo.
01:07:22
Lenny
Yes. So we are going to go to the ICc. We are exhibiting at the ICc conference in fulsome. So, yeah, we’re going to have a booth. We’re going to talk to some people.
01:07:32
Jaco
Excellent.
01:07:33
Lenny
We also have a spot in the virtual conference after our presentation. So we are forming part of the ICC.
01:07:39
Jaco
Okay, fantastic. So last point from us. Our next engagement, the ICC. If you’re not familiar, that is inductive automation’s annual event. It’s called the Ignition Community Conference. It is the x or ten. It is the 10th one this year. It is sort of a dual event, a two in one event. There is the in person event, which is in Folsom. Few of us are fortunate enough to have us visas, and we’ll be heading over there to experience the event. And then there is also the virtual event from the third to the 5 October, it is going to be really value packed. It is the first in person one pre Covid, but they have maintained the virtual one. But this year specifically is going to be a really valuable event. So three to 5 October virtual event. Make sure you register for that.
01:08:32
Jaco
There is no cost to register and we will share those links with you as well as a follow up to the email. Thank you very much for your time. We did go over there a little bit, but thank you for listening. Hopefully some of that was valuable and let us have your feedback please.
01:08:48
Lenny
Cool. Thank you very much everybody. I just want to say thanks for the LDF to invite me and proud to be a sponsor the flow side for this session. Cool.
01:08:59
Jaco
Fantastic. So we’ll follow up with that email where we talk about some of the new enhancements and features for ignition that we shared as well as on the canary sites. Really exciting stuff. And I’m sure Lenny will give us a couple of links to some valuable assets as well.
01:09:11
Lenny
Thank you.
01:09:12
Jaco
Thank you to you guys for putting it together.
01:09:14
Lenny
Awesome.
01:09:15
Laura
Thank you.
01:09:16
Jaco
Fantastic. I don’t see any questions from anybody. No questions from anybody online. All right, fantastic. Let us know if there’s anything that you think of afterwards. Otherwise we will see you in a couple of months time. Thank you.