This file is not needed to be stored in git at all, remove it.
It will be added during publishing process by CI scripts and is only
needed for publishing to docs.openstack.org.
Add doc/build to gitignore file so that no new content will be checked
Add api/autoindex.rst to toctree to fix build.
Add usual build_sphinx stanza to setup.cfg so that building and
publishing in OpenStack CI works. These parameter are crucial.
Breaking down the aggregation into two stages.
The first stage aggregates raw metrics frequently and is
implemented as a Spark Streaming job which
aggregates metrics at a configurable time interval
(defaults to 10 minutes) and writes the intermediate
aggregated data, or instance usage data
to new "metrics_pre_hourly" kafka topic.
The second stage is implemented
as a batch job using Spark Streaming createRDD
direct stream batch API, which is triggered by the
first stage only when first stage runs at the
top of the hour.
Also enhanced kafka offsets table to keep track
of offsets from two stages along with streaming
batch time, last time version row got updated
and revision number. By default it should keep
last 10 revisions to the offsets for each
The monasca-transform is a new component in Monasca that
aggregates and transforms metrics.
monasca-transform is a Spark based data driven aggregation
engine which collects, groups and aggregates existing individual
Monasca metrics according to business requirements and publishes
new transformed (derived) metrics to the Monasca Kafka queue.
Since the new transformed metrics are published as any other
metric in Monasca, alarms can be set and triggered on the
transformed metric, just like any other metric.
Co-Authored-By: Flint Calvin <email@example.com>
Co-Authored-By: David Charles Kennedy <firstname.lastname@example.org>
Co-Authored-By: Ashwin Agate <email@example.com>
Implements: blueprint monasca-transform