-
Operational Excellence
-
- Resources have identified owners
- Processes and procedures have identified owners
- Operations activities have identified owners responsible for their performance
- Team members know what they are responsible for
- Mechanisms exist to identify responsibility and ownership
- Mechanisms exist to request additions, changes, and exceptions
- Responsibilities between teams are predefined or negotiated
-
- Executive Sponsorship
- Team members are empowered to take action when outcomes are at risk
- Escalation is encouraged
- Communications are timely, clear, and actionable
- Experimentation is encouraged
- Team members are encouraged to maintain and grow their skill sets
- Resource teams appropriately
- Diverse opinions are encouraged and sought within and across teams
-
- Use version control
- Test and validate changes
- Use configuration management systems
- Use build and deployment management systems
- Perform patch management
- Implement practices to improve code quality
- Share design standards
- Use multiple environments
- Make frequent, small, reversible changes
- Fully automate integration and deployment
-
Security
-
- Evaluate and implement new security services and features regularly
- Automate testing and validation of security controls in pipelines
- Identify and prioritize risks using a threat model
- Keep up-to-date with security recommendations
- Keep up-to-date with security threats
- Identify and validate control objectives
- Secure account root user and properties
- Separate workloads using accounts
-
- Analyze public and cross-account access
- Manage access based on life cycle
- Share resources securely with a third party
- Reduce permissions continuously
- Share resources securely within your organization
- Establish emergency access process
- Define permission guardrails for your organization
- Grant least privilege access
- Define access requirements
-
- Build a program that embeds security ownership in workload teams
- Centralize services for packages and dependencies
- Manual code reviews
- Automate testing throughout the development and release lifecycle
- Train for application security
- Regularly assess security properties of the pipelines
- Deploy software programmatically
- Perform regular penetration testing
-
-
Reliability
-
- How do you ensure sufficient gap between quotas and maximum usage to accommodate failover?
- How do you automate quota management?
- How do you monitor and manage service quotas?
- How do you accommodate fixed service quotas and constraints through architecture?
- How do you manage service quotas and constraints across accounts and Regions?
- How do you manage service quotas and constraints?
- How do you build a program that embeds reliability into workload teams?
-
- How do you enforce non-overlapping private IP address ranges in all private address spaces?
- How do you prefer hub-and-spoke topologies over many-to-many mesh?
- How do you ensure IP subnet allocation accounts for expansion and availability?
- How do you provision redundant connectivity between private networks in the cloud and on-premises environments?
- How do you use highly available network connectivity for workload public endpoints?
-
- Monitor end-to-end tracing of requests through your system
- Conduct reviews regularly
- Analytics
- Automate responses (Real-time processing and alarming)
- Send notifications (Real-time processing and alarming)
- Define and calculate metrics (Aggregation)
- Monitor End-to-End Tracing of Requests Through Your System
-
- Monitor all components of the workload to detect failures
- Fail over to healthy resources
- Automate healing on all layers
- Rely on the data plane and not the control plane during recovery
- Use static stability to prevent bimodal behavior
- Send notifications when events impact availability
- Architect your product to meet availability targets and uptime service level agreements (SLAs)
-
-
Cost Optimization
-
- Establish ownership of cost optimization
- Establish a partnership between finance and technology
- Establish cloud budgets and forecasts
- Implement cost awareness in your organizational processes
- Monitor cost proactively
- Keep up-to-date with new service releases
- Quantify business value from cost optimization
- Report and notify on cost optimization
- Create a cost-aware culture
-
- Perform cost analysis for different usage over time
- Analyze all components of this workload
- Perform a thorough analysis of each component
- Select components of this workload to optimize cost in line with organization priorities
- Perform cost analysis for different usage over time
- Select software with cost effective licensing
-
-
Performance
-
- Learn about and understand available cloud services and features
- Evaluate how trade-offs impact customers and architecture efficiency
- Use guidance from your cloud provider or an appropriate partner to learn about architecture patterns and best practices
- Factor cost into architectural decisions
- Use policies and reference architectures
- Use benchmarking to drive architectural decisions
- Use a data-driven approach for architectural choices
-
- Use purpose-built data store that best support your data access and storage requirements
- Collect and record data store performance metrics
- Evaluate available configuration options for data store
- Implement Strategies to Improve Query Performance in Data Store
- Implement data access patterns that utilize caching
-
- Understand how networking impacts performance
- Evaluate available networking features
- Choose appropriate dedicated connectivity or VPN for your workload
- Use load balancing to distribute traffic across multiple resources
- Choose network protocols to improve performance
- Choose your workload's location based on network requirements
- Optimize network configuration based on metrics
-
- Establish key performance indicators (KPIs) to measure workload health and performance
- Use monitoring solutions to understand the areas where performance is most critical
- Define a process to improve workload performance
- Review metrics at regular intervals
- Load test your workload
- Use automation to proactively remediate performance-related issues
- Keep your workload and services up-to-date
-
-
Sustainability
-
- Optimize geographic placement of workloads based on their networking requirements
- Align SLAs with sustainability goals
- Optimize geographic placement of workloads based on their networking requirements
- Stop the creation and maintenance of unused assets
- Optimize team member resources for activities performed
- Implement buffering or throttling to flatten the demand curve
-
- Optimize software and architecture for asynchronous and scheduled jobs
- Remove or refactor workload components with low or no use
- Optimize areas of code that consume the most time or resources
- Optimize impact on devices and equipment
- Use software patterns and architectures that best support data access and storage patterns
- Remove unneeded or redundant data
- Use technologies that support data access and storage patterns
- Use policies to manage the lifecycle of your datasets
- Use shared file systems or storage to access common data
- Back up data only when difficult to recreate
- Use elasticity and automation to expand block storage or file system
- Minimize data movement across networks
-
- Articles coming soon
Make informed decisions to deploy systems and changes
Making Informed Decisions to Deploy Systems and Changes
Making informed decisions about deploying systems and changes is crucial for maintaining the reliability and stability of your workload. By establishing processes for both successful and unsuccessful changes, teams can minimize risk and handle potential failures effectively. Pre-mortems, risk evaluations, and compliance checks help ensure that all aspects of the deployment are thoroughly considered before going live.
Plan for Both Successful and Unsuccessful Changes
Establish processes for handling both successful and unsuccessful changes. Planning for success includes having clear deployment procedures and validation steps. Planning for failure means having rollback mechanisms, mitigation steps, and communication protocols in place. By preparing for both outcomes, teams can reduce risks and ensure that issues are handled efficiently.
Use Pre-Mortems to Anticipate Failure
Conduct pre-mortem exercises to simulate a potential failure before deploying changes. In a pre-mortem, the team imagines that the deployment has failed and works backward to identify potential causes and mitigation strategies. This proactive exercise helps uncover vulnerabilities, anticipate failure scenarios, and develop procedures to mitigate risks. Pre-mortems make deployments safer by encouraging teams to plan for worst-case situations.
Evaluate Benefits and Risks of Deployment
Before deploying any change, evaluate the potential benefits and risks associated with it. Assess whether the change will enhance performance, improve reliability, or introduce new features that benefit users. Simultaneously, evaluate potential risks, including the possibility of downtime, degraded performance, or security vulnerabilities. Understanding both sides helps teams make informed decisions about whether and how to proceed with a change.
Verify Compliance with Governance Requirements
Ensure that all changes comply with governance, regulatory, and organizational standards. This includes verifying that changes adhere to security controls, data handling policies, and any industry-specific compliance requirements. By verifying compliance, teams can avoid issues related to regulatory violations or governance failures that could lead to costly consequences.
Establish Decision-Making Frameworks
Create decision-making frameworks that include approval processes, change reviews, and go/no-go criteria. Decision-making frameworks help ensure that only changes that are well-vetted and low-risk are deployed to production environments. This provides a structured way to assess readiness, mitigate risks, and validate compliance before executing a change.
Supporting Questions
- How do pre-mortem exercises help anticipate failure before deploying changes?
- What processes are in place for handling both successful and unsuccessful changes?
- How are changes evaluated for risks and compliance before being deployed?
Roles and Responsibilities
Change Manager
Responsibilities:
- Conduct pre-mortem exercises to anticipate potential failure scenarios and develop mitigation strategies.
- Evaluate the benefits and risks of proposed changes and decide whether they are ready for deployment.
Compliance Officer
Responsibilities:
- Verify that changes comply with governance and regulatory requirements, ensuring adherence to security and industry standards.
- Conduct compliance reviews before changes are deployed to production environments.
Release Manager
Responsibilities:
- Implement decision-making frameworks, including go/no-go criteria for deployment.
- Ensure that both successful and unsuccessful outcomes are planned for, with rollback mechanisms and procedures in place.
Artifacts
- Pre-Mortem Report: A report summarizing the outcomes of pre-mortem exercises, including identified risks, potential failure causes, and mitigation strategies.
- Risk Assessment Document: A document evaluating the risks and benefits of deploying a change, including the potential impact and mitigation measures.
- Compliance Checklist: A checklist used to verify that all changes comply with governance, regulatory, and security requirements before being deployed.
Relevant AWS Tools
Change Management and Compliance Tools
- AWS Systems Manager Change Manager: Helps manage changes to workloads by automating the change approval process and ensuring compliance with governance requirements.
- AWS Config: Tracks configuration changes and verifies compliance with governance standards, helping ensure that changes meet policy requirements.
Monitoring and Risk Evaluation Tools
- Amazon CloudWatch: Monitors system performance and health, providing data that helps evaluate the risks associated with deploying changes.
- AWS Trusted Advisor: Provides recommendations for optimizing AWS environments, including cost, security, and performance improvements, which are useful for evaluating the potential benefits of changes.
Collaboration and Decision-Making Tools
- AWS Systems Manager OpsCenter: Provides a central hub for managing and reviewing operational data, allowing teams to assess incidents and changes in real time.
- Amazon Chime: Facilitates meetings for decision-making processes, such as pre-mortem discussions, risk assessments, and change review sessions.