Skip to end of metadata
Go to start of metadata

You are viewing an old version of this content. View the current version.

Compare with Current View Version History

« Previous Version 11 Next »

 

Quality Assurance

Please use following slide deck as a guide line of how to do a Quality Check on a brand new published dashboard:

 

If you notice anything strange on the dashboard, please consider the following :

  • Check Deployment of listeners and jobs:
    • Sensor Listener + Sensor Data Aggregation running?
    • MCS API Connector enabled and latest version?
  • Check config of sensors in Studio (Sensor Devices: correct Device ID, linked to Room or WP, linked to Zone etc)
  • Check config of floorplans in Studio (and IWMS if relevant)
  • Check locations master data in IWMS + config of locations in Location Grouping and Room Capacity
  • Check Dashboards Space category mapping in Studio > Settings > Dashboards
  • Check installation of sensors + customer environment (might be empty due to Covid or warmer due to climate)
  • If eg comfort values are off the charts, book a call for the BI team to check the incoming values (Incidents > Cobundu Incidents > Nature: Cobundu Dashboards)

 

Comfort Values visualized

Logic behind visualization: Average temperature value for room X (if 2 comfort sensors are linked to that room) = (10 °C +20 °C) / 2 = 15 °C  and this is how it will be exposed in the timeslots.

Check out the scenarios/use cases concerning the visualization of comfort data on Live Floorplan and in Dashboards on Live floorplan settings-page.

 

 

 

 

 

  • No labels