Key Data Sources
The UAF must account for a variety of historian vendors and technologies to ensure cross-platform compatibility and free the enterprise from historian vendor lock-in. This includes supporting:
Time Series Historians
- Enterprise Class Historians - AVEVA PI, Canary Labs, etc
- Site Historians - Wonderware, Proficy, Citect, FactoryTalk, DeltaV, etc.
- Open Source Historians - InfluxDB, Timescale, QuestDB, etc.
- SQL-based Historians - Ignition, VTScada, etc.
Manufacturing Systems and Solutions
Unifying data from these diverse systems necessitates the ability to connect to and pull data from each of them. This integration must encompass both established and legacy providers, as well as modern solutions, to offer comprehensive and flexible data management and operational capabilities. By doing so, organizations can achieve seamless data integration and efficient operations across their entire manufacturing environment.
- Laboratory Information Management Systems (LIMS) - STARLIMS, LabWare, Thermo Fisher Scientific SampleManager, etc.
- Manufacturing Execution Systems (MES) - Siemens Opcenter, Rockwell Automation FactoryTalk, AVEVA MES, Sepasoft, etc.
- Enterprise Resource Planning (ERP) - SAP, Oracle, etc.
- Computerized Maintenance Management Systems (CMMS) - Fiix CMMS, UpKeep, Maintenance Connection, etc.
- Enterprise Asset Management (EAM) - IBM Maximo, Infor, SAP, etc.
Real Time Data Capture
In many manufacturing environments, crucial real-time data is not currently being archived, which limits the ability to analyze and transform this information. To address this gap, the UAF should include the capability to collect and store real-time data, playing a role similar to a traditional data historian but specifically for data that is not already being stored. When connecting to a Unified Namespace (UNS) as a real-time source, it is essential to understand what data is already being historized elsewhere. By identifying and capturing only the data that is not currently stored, the UAF ensures that all relevant information is available for comprehensive analysis and transformation, thereby enhancing decision-making and operational efficiency. This approach prevents redundancy, optimizes storage, and ensures a complete and accurate data set for actionable insights.
Common technologies and protocols that should be supported include OPC servers, MQTT brokers (vanilla and Sparkplug), web APIs, Kafka streams, and other real time data sources.
SQL Databases
Transaction-based data is crucial in providing the necessary context to slice and interpret time series data effectively. This type of data helps in understanding the events and transactions that occur within the manufacturing process, offering a detailed view of operational activities and their impact. Incorporating transaction-based data into the UAF allows for a more comprehensive analysis, enabling better decision-making and insights.To support this integration, the UAF must be capable of connecting to and pulling data from various SQL databases. This includes widely-used technologies such as:
- Microsoft SQL Server - An open-source database that is popular for its performance, reliability, and ease of use.
- MYSQL - Siemens Opcenter, Rockwell Automation FactoryTalk, AVEVA MES, Sepasoft, etc.
- PostgreSQL - An advanced open-source database that supports complex queries and a wide range of data types.
- Oracle DB - Renowned for its advanced features, scalability, and strong security measures.
Manual Data Capture
Entering data, categorizing events, and capturing comments and context is vital in ensuring that all relevant information is included in the UAF, especially when certain data points are not automatically collected by sensors or systems. This type of data often includes critical insights from human observations, quality checks, or maintenance activities that provide additional context and depth to the automated data streams.
To effectively integrate manual data, the UAF should support various methods for capturing and categorizing this information. This can be achieved through web forms, mobile applications, and by importing CSV files.
Major Data Consumers
The value of data already transformed and heavily encoded with operational process knowledge and context cannot be overstated. For enterprises, such data is a gold mine, offering rich insights and actionable intelligence. The UAF ensures that this data is not only accessible but also integrated seamlessly into various systems, saving data teams significant time and effort. This feature is especially crucial for data teams overseeing multiple sites, as it enables them to make informed decisions quickly and efficiently, leveraging a unified, context-rich dataset. By structuring the data to match the existing schemas of these systems, the UAF facilitates smooth integration and usability.
Data lakes and warehouses
Standard connectivity methods will ensure that existing enterprise architectures and data strategies are met. This means ensuring solutions like AWS, Azure, Snowflake, Databricks, Google BigQuery, and Oracle can be fed data streams from the UAF.
Business Intelligence tools
Supports integration with standard BI tools such as PowerBI and Tableau to facilitate data visualization and business analytics.
Advanced analytics and ML/AI
Connects to advanced analytics platforms and machine learning and AI tools through standard methods, creating a plug-and-play environment with structured, contextualized, and even preprocessed data in wide table formats with normalized timestamps, greatly increasing the time to value for these projects.
Back to the UNS
Publishes processed and contextualized data back to the Unified Namespace (UNS) to maintain a continuous and updated data flow, ensuring the integrity and accuracy of the entire data ecosystem.