As an engineer using PI View or SQL writers for a report, I would like option to remove data older than a certain cutoff. For example, a continuous view is used to generate data for a report on the last 8 hours. Data older than that are no longer needed for the report and should be removed to avoid unnecessarily large tables.
One workaround with Text and Hadoop writers is the Append Timestamp option, but there's no ability to remove data from other targets.41 votes
For supporters of this idea, what is the downstream client tool that consumes the data prepared by the PI Integrator? Is the concern about disk space or how to only display data for a specified time range?
For the former, this seems to be a different type of “rolling” reporting use case that we do not currently have a solution for. In the mean time, as Rory mentioned below, a stored procedure on SQL would work if this idea is intended for SQL writers.
For the latter, is it possible to create “rolling timespans” for your graphs, reports, tables, etc.? For example, here’s a video on how to create calculated measures for a table based on last day or month of data: youtube.com/watch?v=NevFWL2JoMU
As a user of PI Integrator for Business Analytics, I'd like to have it support high availability (HA).26 votes
Please specify what kind of high availability is requested. For PI Integrator for BA to run in a highly available fashion?
To support other applications that are highly available? If so, which ones?
As a PII4BA administrator I would like to specify a logging level so I can filter which logs I want to store and which ones I don't want to store in the database.19 votes
Is this an issue of limited SQL storage? We have a log retention policy that can be configured by the user in the CAST.UI.WindowsService.exe CONFIG file today. By default it only retains a maximum of 2 months of data AND no more than 100k rows AND no more than 1000MB of disk space. Similar retention policies can be created for the PIIntegratorStats database as well.
Let me know if this would satisfy the requirement. Or provide more information on how specifying the logging level would help. If you could, which logging levels would you keep vs. throw away and why?
Add the possibility of adding a target as a json file format8 votes
If you voted for this idea, did you mean JSON format in the Text File target, or to support JSON format in a new target or existing target? Currently we support JSON format for the Apache Kafka, Azure Event Hubs, Azure IoT Hub, Amazon Kinesis Data Streams, and Amazon S3 writers.
The current 'Update' feature backfills all assets matching the asset shape defined in the view.
However, when there is a need to back-fill only a small set of assets in a view there is no option to select such assets.
Adding a select option would prevent updating all assets in the view and user could select only those assets which require a back-fill6 votes
Thank you for your feedback! We would need more discussion with supporters to understand the problem.
How does the lack of this functionality impact your work? How often does this happen? How many assets are typically in your view? And how many assets are typically affected?
Does the new automatic updates through persistent signups (introduced in 2018) mitigate this problem at all?
Please feel free to email me or post a comment to help us understand the pain point.
- Don't see your idea?