Skip to content

pradeepcdlm/resource-surveillance-commons

 
 

Repository files navigation

Resource Surveillance & Integration Engine (surveilr) for Local-first Edge-based Stateful Data Aggregation

Table of Contents

Introduction to the Resource Surveillance & Integration Engine (surveilr)

The Resource Surveillance & Integration Engine (surveilr) is a stateful data preparation and integration platform for multiple systems that need to integrate and operate on common data in a local-first, edge-based SQL-centric manner.

  • Stateful means that the surveilr is not just passing data between multiple sytems but allows storing the data in an opinionated universal schema with full SQL querying support.
  • SQL-centric means that data should be queryable and orchestratable via a SQL.
  • Local-first means that content should be prepared and processed locally before going to a cloud or central server.
  • Edge-based means that data should be handled as close to where the data is collected rather than everything being done centrally.

surveilr's stateful, local-first, edge-based, SQL-centric capabilities are ideal for complex data integration tasks. While it can support a wide range of integration needs, surveilr is particularly valuable for integrating clinical operations data, patient records, pharmacy, and billing information across multiple systems. Beyond these use cases, surveilr also plays a crucial role in data processing and analysis pipelines, supporting infrastructure assurance, audits, content assembler, and more.

surveilr leverages a wide range of technologies to facilitate seamless communication and data transfer. It offers an opinionated architecture with design guidelines for file placement in specific ingestion folders, utilizing technologies such as WebDAV, SFTP, and virtual printers. It supports multiple message formats, including HL7, FHIR, JSON, and XML, alongside file types like CSV, Excel, and custom formats. Additionally, the module can connect to customer database systems through Capturable Executables, enabling database-level exchanges. The processed data can then be integrated with the host or third-party systems for further analysis and reporting.

In addition to simplifying data exchange processes, surveilr's local-first, stateful, edge-based architecture helps reduce sensitive data exposure (HIPAA-compliance) by allowing data to be anonymized or deidentified before going to central servers.

The Integration Engine operates through a small, yet powerful and easy manageable application that resides on any device like a phone, workstations and laptop PCs, or servers in the cloud with highly secure environment. Its primary role is to collect data from the host systems (the system where the data originates), and securely transmit designated third-party systems, it also collects and processes data from third-party systems and securely transfers it to central systems for analysis and decision-making. This process is fundamental to ensuring that sensitive data is shared safely and efficiently across different platforms.

Resource Surveillance & Integration Engine (surveilr) features and use cases

The various functional components/layers of the Resource Surveillance Integration Engine (surveilr) are given below,

Components/Layers Details
Acquisition of Content (Data) Steps involves in the content acquisition i.e preparation of the files/data for ingestion. Inaddition to the local files, we can use technologies such as WebDAV, SFTP, AWS S3, Git and Virtual Printer capabilities are the part of this layer.
Advanced Data Acquisition Layer This layer helps users for enhanced data preparation using Capturable Executables(CEs). With the support of processing instructions (PI) Capturable Executables(CEs) further helps the user to prepare data with specific message standards like JSON,plain/text etc
Message format/standard Support The Resource Surveillance & Integration Engine supports a wide range of standards like HL7, FHIR, JSON, and XML but can also handle a variety of file types for exchange, such as CSV, Excel, or custom formats.
Stateful Ingestion Involves the steps of ingesting and processing data into a structured universal schema with full SQL querying support. surveilr leverages state tables to track the status of ingested files, ensuring that files are not re-ingested unless changes are detected. This approach prevents redundant ingestion, optimizes performance, and supports incremental updates. By maintaining the state of ingested data, surveilr ensures efficient data management, scalability, and transparency, while also providing a clear audit trail of the ingestion history.
Web UI for Viewing Ingested Data Locally Resource Surveillance & Integration Engine (surveilr) provides a Web UI component that allows users to view ingested data in real time and access data processed offline
Data Synchronization and Aggregation The Data Synchronization and Aggregation phase involves systematically synchronizing and aggregating the ingested and processed data into a central data store. This ensures the data is ready for advanced analytics, reporting, and integration with other supporting systems.

Resource Surveillance & Integration Engine (surveilr) Ingestion/Integration Pipeline

The Resource Surveillance & Integration Engine (surveilr) is designed to support both real-time and offline/batch processing. It ensures flexible and scalable data integration across various healthcare systems and other sectors. The ingestion/integration processes are divided into several key phases, each focused on maximizing performance, security, and reliability for both immediate and delayed data processing and querying.

surveilr’s local-first edge-based architecture allows for distributed, scalable processing. It handles both real-time data streams and larger batch processing loads without overwhelming the central infrastructure.

Real-Time and Offline Processing

  • Real-Time Processing: The system is capable of processing data as it arrives in real-time.
  • Offline/Batch Processing: It also supports offline/batch processing on data.

Improved Reliability

surveilr’s ability to store and batch-process data ensures continuous functionality, because of its edge-based nature, i.e., it processes data close to its origin.

Security and Compliance

Whether handling real-time data or batch uploads, surveilr maintains strict security protocols, anonymizing sensitive information before transmission and ensuring compliance with standards like HIPAA.

Separation of Content Acquisition and Stateful Ingestion

surveilr separates the data acquisition phase from the stateful ingestion process, allowing for modular control and optimized handling. This distinction enhances both real-time and offline batch processing, enabling the system to continuously and efficiently capture, ingest, and store data in a stateful manner.

Content Acquisition

Data is collected using several methods, including:

  • WebDAV: It enables data acquisition through WebDAV, allowing users to upload files and data directly into designated ingestion locations or folders. This facilitates seamless file transfers from various source systems or data origins.
  • Virtual Printers: Allows capturing print jobs (e.g., reports or documents) and storing them in designated ingestion folders for further processing.
  • API, Direct Uploads and local file systems: Supports both real-time API data feeds and batch uploads of bulk data from external sources, cloud storage (e.g., Amazon S3, SFTP, Git, IMAP, and Direct Messaging Services).

These methods ensure that surveilr can handle both on-demand, real-time data integration and batch processing for periodic uploads or delayed data preparation and ingestion.

Stateful Ingestion

Once acquired, data moves to the stateful ingestion layer, where it is processed, transformed, enriched, and stored in a universal schema. This layer supports both real-time ingestion for immediate analysis and offline processing for batch jobs.

File Ingestion

File ingestion in surveilr involves importing and processing files from a file system into a Resource Surveillance State Database (RSSD) for monitoring and analysis, a process known as "walking the filesystem." This method scans directories and files, extracting their metadata and content to be stored in the RSSD.

Preparing for Ingestion

Before starting the ingestion process, it's important to determine which files and directories will be processed. surveilr provides an option that simulates the process without making any changes, ensuring that only the desired files are ingested. This feature is particularly useful for previewing files in the current working directory or specified directories before actual ingestion.

Performing File Ingestions

Files can be ingested either from the current working directory (CWD) or specific directories using the -r flag. Users can also utilize regular expressions to target specific directories or files. Ingestion can be followed by displaying statistics on the ingested data using the --stats option, which provides insights into the volume and type of data processed.

Command examples include:

  • Ingesting all files from the CWD or specific directories.
  • Previewing files before ingestion.
  • Displaying statistics post-ingestion to analyze the ingested content.

Task Ingestion

Task ingestion in surveilr allows users to automate the execution of shell tasks and convert their outputs into structured JSON data, which is stored in the uniform_resource table of the Resource Surveillance State Database (RSSD).

surveilr task ingestion allows running one or more Deno Task Shell commands through STDIN, executing them sequentially. The outputs are formatted as JSON (or another specified format) and inserted into the database. Inputs can be either:

  • Simple Text: Non-JSONL text treated as an anonymous command string, executed with the assumption that the output will be in JSON format.
  • JSONL Format: Text in JSONL format, where each object contains specific attributes. The value of the key is executed as a command, and the output is stored using the key as an identifier.

IMAP Email Ingestion

IMAP Email Ingestion in surveilr enables the direct ingestion of IMAP emails into the Resource Surveillance State Database (RSSD). This feature automates the process of retrieving emails from a specified folder and batch size, converting them into structured JSON data, and storing them in the ur_ingest_session_imap_acct_folder_message table of the RSSD.

It works with any email provider that supports IMAP, and for Microsoft 365 users, surveilr offers integration guidance through the Microsoft Graph API, ensuring comprehensive email data access.

Data Synchronization and Aggregation

After ingestion, surveilr synchronizes and aggregates processed data to a central data store. This ensures that both real-time and batch data are available for further processing, analysis, and reporting, allowing central systems to perform advanced analytics or integrations with other systems.

Content Acquisition: Flexible Methods

surveilr supports multiple content acquisition methods, enabling data collection from various sources and modes:

  • WebDAV: Facilitates seamless file transfers from different source systems.
  • Virtual Printers: Captures print jobs for later processing.
  • API and Direct Uploads: Handles real-time API data feeds and batch uploads of bulk data from sources like S3 buckets, SFTP, Git, and IMAP.

Capturable Executables for Enhanced Data Preparation

In typical ingestion pipelines, surveilr ingests files stored in specific locations and creates the Resource Surveillance State Database (RSSD). When the source of ingestion is processed output from a data source, surveilr leverages Capturable Executables (CEs) to enhance data preparation. These executables tasks/safe scripts that produce the data and in a more special cases do data transformation and validation tasks as part of the data preparation pipeline.

Data Validation and Sanitization

CEs ensure that both real-time/offline data streams and batch files meet necessary quality standards. They perform tasks such as:

  • Transformation: Converts real-time and batch data into standardized formats.
  • Automation: Triggers real-time and offline tasks, ensuring efficient data processing.

File Naming Pattern for Capturable Executables (CEs)

Capturable Executables (CEs) are scripts that can be run with specific arguments or behaviors to produce output files, which are then captured and stored in the surveillance state database. These scripts are managed through Processing Instructions (PIs) embedded in their file names.

The naming pattern for CEs allows for flexible execution and capture of results produced by these scripts. This further enable more control to the user to meet specific data preparation requirement in the ingestion process.

The file name pattern mainly uses a regular expression, like surveilr[(?P[^]]*)] and a few common file name patterns and its meaning are given below,

<filename>.surverilr[json] – this pattern expect to produce JSON output as the result of capturable executable.

<filename>.surverilr[txt] – this pattern expect to produce plain text output as the result of capturable executable.

<filename>.surverilr[xyz] – this pattern is a general notation where [xyz] is the required output data format. Eg. yml,md, etc.

Stateful Ingestion

The Stateful Ingestion phase is a pivotal component of the data processing pipeline in surveilr. It is responsible for handling data once it has been acquired from various sources. In this phase, data is processed, enriched, and stored in a standardized format, ensuring that it is ready for both real-time and offline analysis. The stateful nature of this layer allows it to maintain the context and state of data throughout the ingestion process, supporting seamless and efficient data management.

Key Functions of Stateful Ingestion

  • Data Parsing: Converts raw data into a structured format that can be easily processed. This may include parsing different data formats such as JSON, XML, CSV) etc.
  • Validation: Checks data for correctness and completeness. Invalid or incomplete data is flagged for review or correction.
  • Transformation: Applies transformations to standardize data, such as converting units, normalizing values, or mapping fields to a common schema.
  • Enrichment: Enhances data by adding supplementary information or context. This might include merging data from multiple sources or applying business rules to add derived fields.

The capabilities of the Resource Surveillance & Integration Engine (surveilr) can be extended at multiple levels:

  • Content-Level Extensibility: surveilr allows for extensibility at the content level through capturable executables. These scripts generate data in various formats, such as JSON, plain text, or others, which are then ingested and stored in the surveilr state schema, specifically in the uniform_resource table of the Resource Surveillance State Schema.
  • SQL DDL-Level Extensibility: Capturable executables can also include SQL Data Definition Language (DDL) scripts to create tables or other SQLite database objects, offering extensibility at the database schema level.
  • Programmatic/Scriptable Extensibility with TypeScript Integration: surveilr supports programmatic extensibility through integrated TypeScript or Deno tasks. These scripts can be customized to generate the necessary output for ingestion, making surveilr highly adaptable to specific business requirements. This extensibility enables the development of custom logic and workflows to produce data tailored to unique functional needs.

Data Storage

Processed data is stored in a universal schema, a standardized format that ensures consistency across datasets. This schema, known as the Resource Surveillance State Database (RSSD), facilitates easy querying, analysis, and integration.

A Resource Surveillance State Database (RSSD)

The Resource Surveillance State Database (RSSD) is a versatile and independent data source created through surveilr, which can be utilized across various tools, services, applications, or integrated into data warehouses once generated. Its independence allows for flexible usage in any computing environment that supports SQLite, enabling seamless integration into different systems and workflows. RSSD is designed to automate the collection of evidence such as code, test results, emails, issues/tickets, and wikis, ensuring compliance with security, privacy, safety, and regulatory standards through machine attestation rather than manual processes.

  • To generate an RSSD, the surveilr ingest command can be used for file or task ingestion. For instance, navigating into a directory and running surveilr ingest files will create an RSSD in the target folder. Alternatively, using the -r flag allows ingestion of content from a specified directory without changing directories. The resulting resource-surveillance.sqlite.db file will contain several tables storing the state data.

  • In environments with multiple RSSDs, it is important to configure unique identifiers for each database, which can be achieved by appending unique elements such as hostnames to the database filename. This can be done either by setting an environment variable (SURVEILR_STATEDB_FS_PATH) or passing a unique identifier using the -d flag during ingestion. These methods help manage and merge RSSDs across environments, ensuring that each database is easily identifiable.

  • The individual RSSD can be merged when we need to aggregate the data for more detailed analysis and integrated into external application or use for data warehouses.

Stateful Context Management

surveilr ensures end-to-end orchestration of the ingestion pipeline, maintaining the context of data throughout its lifecycle. Key features include:

  • Session Management: Tracks data sessions for continuity.
  • State Management: Maintains the state of data throughout the ingestion process, preserving context and tracking changes. This is crucial for handling updates, managing version history, and ensuring data integrity.

Web UI for Viewing Ingested Data

surveilr provides a Web UI for users to view ingested data in real-time or access processed offline data. Key features include:

  • Data Visualization: The Web UI module enable easy configurable web components to view the ingested data locally.
  • Querying Capabilities: Full SQL querying support enables users to extract and analyze data from both real-time and batch ingestion sources.
  • Audit and Traceability: The UI includes tracking features that provide visibility into both real-time and batch data flows, ensuring compliance and security audits are easily conducted.

Data Synchronization and Aggregation

The Data Synchronization and Aggregation phase is the final and critical step in the surveilr pipeline. After data has been ingested and processed through various stages, this phase ensures that the data is systematically synchronized and aggregated into a central data store. This final step consolidates both real-time and batch-processed data, making it accessible for advanced analytics, reporting, and integration with other supporting systems.

Key Functions

  • Data Synchronization: Ensures real-time and batch data are consistently updated in the central repository.
  • Data Aggregation: Consolidates data from different sources into a unified structure.
  • Data Transformation: Applies transformations to standardize data formats, ensure consistency, and enrich the data (e.g., merging fields, calculating metrics). The data transformation also include anonymizing sensitive information before the transmission.

Advanced Analytics and Reporting

Consolidated data enables sophisticated analysis and reporting through:

  • Business Intelligence (BI): Data visualization tools and dashboards.
  • Predictive Analytics: Analyzes historical data to forecast trends.
  • Custom Reporting: Generates tailored reports based on user requirements.

Integration with Other Systems

surveilr facilitates data exchange across different systems through:

  • APIs: Provides APIs for external systems to access aggregated data.
  • Data Export: Exports data in formats like CSV and JSON.

Data Quality and Governance

surveilr maintains accuracy, consistency, and reliability through:

  • Data Validation: Ensures data correctness before synchronization.
  • Audit Trails: Tracks changes and updates to maintain transparency.
  • Compliance: Adheres to data privacy and regulatory standards such as GDPR and HIPAA.

Technical Summary

The following diagram depicts the overall architecture of the Resource Surveillance Integration Engine (surveilr) in the data acquisition, ingestion, transformation, visualization, aggregation, and synchronization pipeline.

Resource Surveillance & Integration Engine (surveilr) Dataflow diagram

Resource Surveillance & Integration Engine (surveilr) - Data Flow Overview

  • The surveilr operates as a local-first, edge-based, SQL-centric platform designed to integrate and process data from various sources and formats. It enables secure file acquisition through traditional OS file systems, cloud storage (e.g., AWS S3), web-based file management (WebDAV), and secure transfer protocols (SFTP). Additionally, it supports multi-source file acquisition and digital document conversion through virtual printers.

  • Ingestion pipelines handle structured and unstructured data, including formats such as JSON, XML, CSV, and industry-standard protocols like HL7, FHIR, and EDI. Data can be transformed into queryable formats like JSON and columnar storage formats (e.g., Parquet) to optimize for analytics and reporting.

  • surveilr’s flexible architecture allows integration with PLM systems like GitHub and Jira, ERP systems, and messaging platforms. Custom SQL tables, built-in SQL views, and transformation capabilities enable users to store, process, and analyze data across multiple databases, including SQLite, PostgreSQL, MySQL, and cloud-based systems like DuckDB.

  • The system features a web-based user interface (Surveilr Web UI) for SQL querying and content management, alongside a console that supports SQL navigation, orchestration, and auditing, making surveilr a powerful tool for scalable, secure, and efficient data integration and processing.

About

Reusable and common surveilr patterns

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • TypeScript 98.8%
  • Other 1.2%