Annual Buyers Guide

Smart Maintenance Management Focus Annual Buyers’ Guide 2020 www.pwemag.co.uk Plant & Works Engineering | 21 and bringing with it new challenges. Factories, as an example, are getting overloaded with data. Every machine, process and system on the factory floor will be generating data during the plant’s operation. However, these facilities have commonly become data rich, but information poor. For example, technologies allow plant managers to extract data on the condition of mechanical equipment, such as a motor. However, tracking huge reams of data on the condition of a motor will only go so far. You need to use data, for it to be useful. What’s the solution? Keep your data collection streamlined by investing in a data management system. Here, you gain insight into the data you care about and use it in accordance with your risk management plan. Velocity The second ‘V’ in the big data flood is velocity. Velocity refers to the accelerating speed at which data is being generated and the lag time between when data is generated and when it is accessible for decision making. Faster analysis leads to faster responses. However, today’s data is created at such a rate that it exceeds the capability of many existing systems. Consider the motor condition monitoring as an example — you may be tracking 500 vibration datapoints per second to check its performance, but if your vibration analysis system is only able to analyze 200 data points per second, you have a problem. Ultimately, you need an entire big data infrastructure that is capable of processing this data quickly. Variety As more technologies begin to generate data, this information is becoming more diverse. From vibration analysis and condition monitoring, to data from enterprise systems such as market trends and product lifecycle management (PLM), organizations are finding they need to integrate increasingly complex data types from an array of systems. This often requires the vertical integration of several different systems. Using this more complicated integration model, condition monitoring data could identify when an industrial part was showing signs of failure, then automatically cross check inventory data to see if a replacement part is in stock. If a replacement part isn’t available, this system could make even more intelligent decisions by automatically repurchasing the part using an enterprise resource planning (ERP) system. However, this is just one example. Plant managers may find themselves wanting to collect all data types, even data that isn’t useful to them and store it in archives for historical analysis. Businesses need to keep their risk mitigation goals at the forefront of their data collection, rather than collecting data for data’s sake. Veracity The final ‘V’ refers to Veracity, the reliability of a particular data type. The key issue here is that the other dimensions of big data — that’s volume, velocity and variety — challenge the capacity of many existing systems. Consider the replacement part example. The scenario sounds ideal, but how can you integrate the condition monitoring big data of a legacy motor, with the availability data of the parts supply if it needs replacing, when this data belongs to a third party? To achieve true veracity, your data infrastructure must be able to make intelligent decisions, without hitting a wall when it requires action outside of the factory walls. Don’t let your huge reams of data flood your facility. There’s no value in generating data for data’s sake. For the most effective analysis, make sure you have all four dimensions of big data being in place — volume, variety, velocity and veracity.

RkJQdWJsaXNoZXIy MjQ0NzM=