OASIS Mailing List ArchivesView the OASIS mailing list archive below
or browse/search using MarkMail.

 


Help: OASIS Mailing Lists Help | MarkMail Help

tosca message

[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]


Subject: [OASIS Issue Tracker] (TOSCA-189) DEFER - Application Monitoring - Proposal (design proposal)


     [ https://issues.oasis-open.org/browse/TOSCA-189?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Chip Holden updated TOSCA-189:
------------------------------

    Assignee: Tomer Shtilman

> DEFER - Application Monitoring - Proposal (design proposal)
> -----------------------------------------------------------
>
>                 Key: TOSCA-189
>                 URL: https://issues.oasis-open.org/browse/TOSCA-189
>             Project: OASIS Topology and Orchestration Specification for Cloud Applications (TOSCA) TC
>          Issue Type: New Feature
>          Components: Profile-YAML
>            Reporter: Tomer Shtilman
>            Assignee: Tomer Shtilman
>
> When we consider monitoring performance of a Cloud, we can broadly classify it to 2 categories
> 1. Infrastructure/Hardware Monitoring -  This involves performance of the various infrastructure components in the cloud like Virtual Machines, Storage, Network etc. 
> E.g
> • CPU usage; total – all CPUs, per CPU, and delta between
> CPUs
> • Disk usage; total, free, used
> • Disk Latency
> • Percentage Busy
> • Percentage Ready
> • Memory; percentage used, swap activity
> • Network; bytes in/out
> ...
> 2.Application monitoring - In Calculating Application Performance we cannot go by the resources utilized by the application as in a cloud, applications move around and so the monitoring solution needs to track and map them.
> E.g Application Response Time - key metric in Application Performance management which actually calculates the time taken for the application to respond to user requests. 
> So just like we can detect deviations in application hardware performance we would like to so the same for application  : KPIs,response times, request statuses , or order throughput in order to allow us to be proactive with the business impact .
> With this application monitoring we can:
> - Understand the real-time performance of the cloud services from the end user’s perspective.
> - Gain visibility into your workload, even when you do not control the backing infrastructure.
> - Isolate problems and drill down to the root cause to immediately take action.
> - Define thresholds and create alerts
> We believe that TOSCA should recommends a monitoring service spec to be optionally implemented by TOSCA containers and provide a set of monitoring capabilities on the application workloads. 
> This is a crucial and basic capability of any application lifecycle management orchestrator.
> The idea is to simple allow the app developer to express in its service template the desired app KPIs to be collected and doing some dynamic reactions upon certain KPI’s threshold crossing.
> The monitoring engine applies the Sample Metric collection on the exposed software component endpoint interface
> In the example below you can see a simple db (software component) hosted on a compute, there is Sample Metric collected every minute on this software component, in addition there is an hourly aggregation based on this minutely sampling.
> tosca_definitions_version: tosca_simple_yaml_1_0
> monitoring_types:
> # Metric base type
>   tosca.monitoring.Metric:
>       derived_from: tosca.nodes.Root
>       description: The basic metric type all other TOSCA metric types derive from
>       properties:
>           polling_schedule:
>                 type: string
>           return_type:
>                 type: string
>           metric_unit:
>                 type: string
>           aggregation_method:
>                 type: string
>                 constraints:
>                     - valid_values: [SUM, AVG, MIN, MAX, COUNT]
>   
> # A single metric sample
>   tosca.monitoring.MetricSample:
>       derived_from: tosca.monitoring.Metric
>       description: A single metric sample,applicatio KPI, like CPU, MEMORY, etc.
>       properties:
>           node_state:
>                type: string
>                constraints:
>                     - valid_values: [RUNNING, CREATING, STARTING, TERMINATING, ..]
>       requirements: 
>                   #a sample metric requires an endpoint 
>            - endpoint: tosca.capabilities.Endpoint 
> #An aggregated metric
>   tosca.monitoring.AggregatedMetric:
>         derived_from: tosca.monitoring.Metric
>         description: An aggregated metric
>         properties:
>            # The time window in millis for aggregating the metric
>            msec_window:
>                 type: integer
>                 constraints:
>                    - greater_than: 0
>         requirements:
>            - basedonmetric: tosca.monitoring.Metric
> relationship_types:
> # a relationship between sample and endpoint
>    tosca.relationships.monitoring.EndPoint:
>        short_name: endpoint
>        derived_from: tosca.relationships.Root
>        valid_targets: [ tosca.capabilities.Endpoint ]
> #this is a relationship to enforce that aggregated metric is based on other sample/aggregate metric   
>     tosca.relationships.monitoring.BasedOnMetric:
>       short_name: basedonmetric
>       derived_from: tosca.relationships.DependsOn
>       valid_targets: [ alu.capabilities.Monitorable.MetricSample,alu.capabilities.Monitorable.AggregatedMetric ]
>                   
>                   
> node_templates:
> server1:
>     type: alu.nodes.Compute
>     properties:
>     ...
>     interfaces:
>       tosca.interfaces.node.lifecycle.Standard:
>         create: scripts/create.sh
>   oracle_db:
>     type: tosca.nodes.SoftwareComponent
>     requirements:
>       - host: server1
>      capabilities:
>          monitoring_endpoint:
>             type: tosca.capabilities.Endpoint
>             properties:
>                protocol: http
>                ...
> monitoring_templates:
> #single sample connects to the monitoring endpoint
> oracle_connections_per_minute_sampled:
>     type: tosca.monitoring.MetricSample
>     properties:
>       polling_schedule: 0 0/1 * 1/1 * ? *
>       return_type: integer
>                   # Defines the aggregation that is done over the instances of the tier
>       aggregation_method: SUM
>     #sampling (collecting the metric) is done through the endponint
>      requirements:
>         endpoint:       #based on proposal TOSCA-188
>         target: oracle_db.monitoring_endpoint
>         relationship: tosca.relationships.monitoring.EndPoint
> #aggregation over the sample, polled hourly
> oracle_connections_per_hour_aggregated:
>     type: tosca.monitoring.AggregatedMetric
>     properties:
>        polling_schedule: 0 0 0/1 1/1 * ? *
>        return_type: integer
>                    #  Defines the aggregation that is done for the metric over time
>        aggregation_method: AVG
>        msec_window: 3600000
>     requirements:
>         basedonmetric:  #based on proposal TOSCA-188
>            target: oracle_connections_per_minute_sampled
>            relationship: tosca.relationships.monitoring.BasedOnMetric
>                                   



--
This message was sent by Atlassian JIRA
(v6.2.2#6258)


[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]