Webserver for log processing
Logs processed by Logscraper are produced by two services that are part of one processing pipeline:
A service that reads new insights data from S3 Bucket (based on events sent from Kafka). Insights rules are applied to such data and result (in JSON format) is sent back into Kafka into different topic.
Incoming and outgoing Kafka topics are configurable, these can be even handled by different Kafka instances.
Source codes of this service are available at https://github.com/RedHatInsights/ccx-data-pipeline
Insights Results Aggregator is a service that provides Insight OCP data that are being consumed by OpenShift Cluster Manager. That data contain information about clusters status (especially health, security, performance) based on results generated by Insights rules engine. Insights OCP data are consumed from selected broker (usually Kafka), stored in a storage (that basically works as a cache) and exposed via REST API endpoints.
Source codes of this service are available at https://github.com/RedHatInsights/insights-results-aggregator
The whole data flow is described there
Two examples for aggregator logs and one for pipeline logs are given: https://github.com/cloudhubs/logscraper/blob/master/docs/Transformations.md