Difference between revisions of "Risk Register"

From MAGEEC
Jump to: navigation, search
(Sort table)
 
(8 intermediate revisions by 2 users not shown)
Line 8: Line 8:
 
! Work Package !! Description !! Likelihood !! Impact !! Risk !! Mitigation
 
! Work Package !! Description !! Likelihood !! Impact !! Risk !! Mitigation
 
|-
 
|-
| WP7 || The compiler infrastructure does not work for energy
+
| WP6 || Cannot measure per-function energy data.
| 5 || 3 || 15
+
| 6 || 2 || 12
| The system would be reconfigured for speed minimisation based on experience that a) this is possible (MILEPOST) and b) speed is a rough proxy for energy. However, this does not represent as significant advance on current knowledge, but does make it more accessible.
+
| Identified relatively late on in Q4. We are looking at approaches including: a) using cycle accurate modeling to apportion energy usage; 2) using performance pofiling as a proxy to apportion energy usage; 3) using static energy analysis to approtion energy usage; 4) using per-application energy usage as the per-function energy usage.
 
|-
 
|-
| WP9 || Compiler energy efficiency optimisation passes yield no measurable benefit
+
| All || Milestones not met
| 7 || 2 || 14
+
| 10 || 1 || 10
| High likelihood is as this hasn't been attempted before. Decouple from infrastructure development, allowing parallel development and use appropriate expert staff (Joern Rennecke).
+
| Milestones 6/3, 6/4, 7/1 and 7/2 very unlikely to be met (see milestones for details). Impact reduced to 1, since this relates to a non-core aspect of the project. Mitigations are listed against specific issues in this risk register and the project plan.  
|-
 
| WP4 || Unable to find sufficient benchmark and test programs
 
| 7 || 2 || 14
 
| Engagement with wider community to identify programs. Committing more resource to developing [http://www.cs.bris.ac.uk/Research/Micro/beebs.jsp BEEBS] and test programs (Bristol)
 
|-
 
| WP5 || Failure to identify viable features (technical issue)
 
| 4 || 3 || 12
 
| Fallback is to use the MILEPOST features. We have started engagement with the MILEPOST team to gain from their experience as much as possible.
 
 
|-
 
|-
| WP1, WP2 || Compiler infrastructure cannot be made generic
+
| WP3 || Lack of knowledge on programming the energy measurment board.
| 5 || 2 || 10
 
| Ensuring good interface design before building infrastructure. Note that non-fully generic solution still has value.
 
|-
 
| All || Loss of key personnel
 
 
| 5 || 2 || 10
 
| 5 || 2 || 10
| Ensure the consortium has multiple people skilled in each area. Note that the area of compiler optimisation (WP8) has one person covering this, but is a smaller part of the project.
+
| James Pallister is the only person who knows how to code the firmware of the energy measurement board. Mitigation: train up additonal engineers and document the software.
 
|-
 
|-
| WP10 || Failure to engage with open source communities.
+
| WP4 || Data Gather phase taking too long
 
| 5 || 2 || 10
 
| 5 || 2 || 10
| Early and open discussions with the relevant communities (already started). Appointment of project member (Andrew Back, AB Open) charged with driving this very time consuming activity.
+
| Software and board reliability issues now resolved. However number of tests still remains very large. Mitigation through Plackett-Burnham experimental design and use of large numbers of target platforms in parallel.
 
|-
 
|-
| All || Milestones not met
+
| WP3 || Need to calibrate measurements
| 5 || 2 || 10
+
| 3 || 3 || 9
| Regular meetings and prior experience in these fields. Manage interactions/dependencies between work packages and re-visit time allocations as more information arises.
+
| When using many measurement boards, we find that results vary depending on the actual chip, room temperature and other environmental factors. Mitigate by callibrating each test run with a standard benchmark.  
 
|-
 
|-
| All || Unable to recruit suitable staff
+
| WP7 || The compiler infrastructure does not work for energy
 
| 3 || 3 || 9
 
| 3 || 3 || 9
| Already mitigated through existing experience and staff from all consortium members. People pipeline from UoB ensures high quality candidates.
+
| The system would be reconfigured for speed minimisation based on experience that a) this is possible (MILEPOST) and b) speed is a rough proxy for energy. However, this does not represent as significant advance on current knowledge, but does make it more accessible. Work carried out thoughout this project by the team and James Pallister is increasingly showing that compiling for energy demonstrates benefits.
 
|-
 
|-
| WP4 || Data Gather phase taking too long due to board reliability
+
| WP4 || Unable to find sufficient benchmark, test programs and case study
 +
| 4 || 2 || 8
 +
| BEEBS v2 provides 80 tests and some data variants. Per-function datagathering should mean this is effectively a much larger data set. Three credible real-world case studies identified. Likelihood still left at 4, to reflect doubts over per-function gathering and desire for even more tests.
 +
|-
 +
| WP10 || Failure to engage with open source communities.
 +
| 3 || 2 || 6
 +
| Likelihood further reduced due to success of GNU Tools Cauldron presentation, which was reported more widely than just the GNU community.
 +
|-
 +
| All || Loss of key personnel
 
| 3 || 2 || 6
 
| 3 || 2 || 6
| Exploration of a smaller data set whilst full data gather occurs and redundancy of boards.
+
| This has happened in repsect of WP 7, due to illness, but this is a non-core part of the project. Embecosm has recruited more staff, and UoB work packages are almost complete, so further loss is unlikely to cause any problems.
 +
|-
 +
| WP9 || Compiler energy efficiency optimisation passes yield no measurable benefit
 +
| 2 || 2 || 4
 +
| Proof of concept implementation demonstrates that this does work. Likelihood placed at 2, to reflect the fact that this will not be implented in an actual compiler (see WP 7).
 +
|-
 +
| WP5 || Failure to identify viable features (technical issue)
 +
| 1 || 3 || 3
 +
| Fallback is to use the MILEPOST features. We have started engagement with the MILEPOST team to gain from their experience as much as possible. Likelihood reduces in Q2, since we are now using the MILEPOST features. Still some risk that we are unable to use any new features.
 +
|-
 +
| WP1, WP2 || Compiler infrastructure cannot be made generic
 +
| 1 || 2 || 2
 +
| Ensuring good interface design before building infrastructure. Note that non-fully generic solution still has value. Likelihood reduced in Q2, since implementation n progress appears to be completely viable for both GCC and LLVM via clean interfaces.
 +
|-
 +
| All || Unable to recruit suitable staff
 +
| 0 || 3 || 0
 +
| Risk deleted, due to nearing project completion.
 
|-
 
|-
 
| All || Failure of consortium to work together
 
| All || Failure of consortium to work together
| 2 || 2 || 4
+
| 0 || 2 || 0
 
| Consortium has already worked together.
 
| Consortium has already worked together.
 
|-
 
|-
| WP3 || Board v4 doesn't work
+
| WP3 || MAGEEC v2 board doesn't work
| 2 || 1 || 2
+
| 0 || 1 || 0
| Continue to use existing board. (v4 board is nice to have)
+
| Continue to use existing board. (MAGEEC v2 board is nice to have). Lowered in Q4 as we are successfully using v2 boards.
 
|}
 
|}
  
 
[[Category:Planning]]
 
[[Category:Planning]]

Latest revision as of 16:39, 4 September 2014

The following describes the risks to this project, their likelihood and impact on the project, as well as procedures for mitigating the risk.

For each risk, the likelihood of it occurring is rated on a scale of 1 (very unlikely) to 10 (certain to happen) and the impact of this occurring is rated on a scale of 1 (minor impact), 2 (significant impact), 3 (total wipe out of the project). Risk is calculated as the product of the two numbers.

A risk mitigation strategy is provided when the risk is greater than or equal to 10 or where the impact is 3.

Work Package Description Likelihood Impact Risk Mitigation
WP6 Cannot measure per-function energy data. 6 2 12 Identified relatively late on in Q4. We are looking at approaches including: a) using cycle accurate modeling to apportion energy usage; 2) using performance pofiling as a proxy to apportion energy usage; 3) using static energy analysis to approtion energy usage; 4) using per-application energy usage as the per-function energy usage.
All Milestones not met 10 1 10 Milestones 6/3, 6/4, 7/1 and 7/2 very unlikely to be met (see milestones for details). Impact reduced to 1, since this relates to a non-core aspect of the project. Mitigations are listed against specific issues in this risk register and the project plan.
WP3 Lack of knowledge on programming the energy measurment board. 5 2 10 James Pallister is the only person who knows how to code the firmware of the energy measurement board. Mitigation: train up additonal engineers and document the software.
WP4 Data Gather phase taking too long 5 2 10 Software and board reliability issues now resolved. However number of tests still remains very large. Mitigation through Plackett-Burnham experimental design and use of large numbers of target platforms in parallel.
WP3 Need to calibrate measurements 3 3 9 When using many measurement boards, we find that results vary depending on the actual chip, room temperature and other environmental factors. Mitigate by callibrating each test run with a standard benchmark.
WP7 The compiler infrastructure does not work for energy 3 3 9 The system would be reconfigured for speed minimisation based on experience that a) this is possible (MILEPOST) and b) speed is a rough proxy for energy. However, this does not represent as significant advance on current knowledge, but does make it more accessible. Work carried out thoughout this project by the team and James Pallister is increasingly showing that compiling for energy demonstrates benefits.
WP4 Unable to find sufficient benchmark, test programs and case study 4 2 8 BEEBS v2 provides 80 tests and some data variants. Per-function datagathering should mean this is effectively a much larger data set. Three credible real-world case studies identified. Likelihood still left at 4, to reflect doubts over per-function gathering and desire for even more tests.
WP10 Failure to engage with open source communities. 3 2 6 Likelihood further reduced due to success of GNU Tools Cauldron presentation, which was reported more widely than just the GNU community.
All Loss of key personnel 3 2 6 This has happened in repsect of WP 7, due to illness, but this is a non-core part of the project. Embecosm has recruited more staff, and UoB work packages are almost complete, so further loss is unlikely to cause any problems.
WP9 Compiler energy efficiency optimisation passes yield no measurable benefit 2 2 4 Proof of concept implementation demonstrates that this does work. Likelihood placed at 2, to reflect the fact that this will not be implented in an actual compiler (see WP 7).
WP5 Failure to identify viable features (technical issue) 1 3 3 Fallback is to use the MILEPOST features. We have started engagement with the MILEPOST team to gain from their experience as much as possible. Likelihood reduces in Q2, since we are now using the MILEPOST features. Still some risk that we are unable to use any new features.
WP1, WP2 Compiler infrastructure cannot be made generic 1 2 2 Ensuring good interface design before building infrastructure. Note that non-fully generic solution still has value. Likelihood reduced in Q2, since implementation n progress appears to be completely viable for both GCC and LLVM via clean interfaces.
All Unable to recruit suitable staff 0 3 0 Risk deleted, due to nearing project completion.
All Failure of consortium to work together 0 2 0 Consortium has already worked together.
WP3 MAGEEC v2 board doesn't work 0 1 0 Continue to use existing board. (MAGEEC v2 board is nice to have). Lowered in Q4 as we are successfully using v2 boards.