Cyber security and the growth of untrusted infrastructure and hybrid workforces
Cyber security and the growth of untrusted infrast...
Cyber security and the growth of untrusted infrastructure and hybrid workforces
Today’s mass usage of cloud services presents an enormous challenge for security teams from highly regulated industries. As financial enterprises shift their focus to DevSecOps, security capabilities and skillsets are now expected to be embedded as part of the DNA of every application and platform. This is triggering a cultural shift in the adoption of cloud native security practices in financial services firms.
We spoke to Synechron’s and Citihub Digital’s SMEs about cloud native security and what it means in a post-pandemic world where the workforce operates in the ‘new normal’ whilst malicious parties take advantage of the vulnerabilities.
As enterprises pivot to building business critical systems on untrusted public cloud infrastructure, how will cyber risk be balanced with the simultaneous shift to a hybrid workforce on untrusted devices?
Eklove Mohan, Synechron: ‘Untrusted’ is the perception but the reality is ‘unawareness’. Look at any of the cybercrimes and you will see that the cause of the breach is mostly due to ‘misconfiguration’. Cloud has a ‘shared responsibility’ security model. The Cloud Service Provider’s (CSP) responsibility is to provide infrastructure and ensure security ‘of’ the cloud but not ‘in’ the cloud, i.e., the application and data have to be secured by the customer with either the tools provided by the cloud provider or a third party.
Some of the key factors that should be considered to balance cyber risk include:
- Utilize ‘least privilege’ – By default, all access to cloud resources is denied. You only enable a resource if it is required by an application.
- Use certified images – Restrict the use of public images. Ensure that the security team within the organization tightens (blocks ports, installs certified versions of tools/applications, disables telnet, etc.) and certifies the OS images used by the application. This will ensure consistency, and if a flaw is identified, it gets fixed for all.
- Select cloud native solutions over third-party tools – Most organizations prefer to keep things cloud agnostic. Hence, opt for third-party tools. However, if the organizations have pretty much decided on a cloud provider, it is preferred that they go with their cloud native services. We have seen that in terms of security, the implementation/patches by CSPs are generally more up to date when compared to hosted third-party tools.
- Choose virtual desktop services – One of the biggest headaches for any organization is protecting their own assets from misuse. With physical devices, like laptops/desktops, there is always a chance of loss or theft. A malicious user many not easily be able to log in to the stolen device, but they canaccess data from the hard disk easily. With virtual desktops, the company’s data is not sitting on the machine. Also, virtual desktops are connected to the organization’s active directory, hence the security is centralized and controlled by an elite set of security professionals within the organization.
Are static and dynamic code scanning tools enough to assure security principles in the Software Development Life Cycle (SDLC)?
Eklove Mohan, Synechron: When it comes to enterprise security, code scanning alone will not ensure applications and data are secured. Static and dynamic code scanning tools will just help to ensure that you have the basic scenarios covered and the hacker has to sweat a little to get into the application/system. In the recent past, organizations have understood the fact that they may never be able to totally protect their application or infrastructure from being hacked. Therefore, what they are focusing on is how quickly an exploit can be identified and then remediated or blocked. Due to this, we have seen a series of tools being used during the SDLC phase which includes: Runtime Application Self Protection (RASP), User and Entity Behavior Analytics (UEBA), Web Application Firewall (WAF) shielding, obfuscation, etc. Consider it like an onion which has all these layers of security and at the core is the application. To get to the core, the hacker needs to peel all these layers. This Defense-In-Depth (DND) technique ensures that either the hacker gives up (unlikely) or the organization catches them before they can get to the core and takes preventive actions before the damage is done.
We’ve been evangelizing DevOps for several years now. Why don’t we see an order of magnitude impact on value delivery? Are Agile DevOps teams bottlenecked by production change gates and risk assurance? Where is the chain of trust failing and why?
Abhilash Panickar, Synechron: Software release management practices in most large enterprises continue to rely on a traditional, manual and multi-layered set of change management processes. One would expect that as a result, outcomes become predictable, risks are reduced, release cycles are consistent, etc. However, the reality is that most of these big bang releases typically result in extended downtimes, system outages, support calls and eventually a scramble to roll things back to bring systems online before the start of business hours. It might appear that the lack of DevOps adoption is to blame for this predicament, but it is 2021 and many enterprises have already jumped on the DevOps bandwagon due to its overwhelming popularity as an industry trend. Often the issue points to a lack of maturity in the DevOps processes resulting in insufficient reliability and predictability of assurances that SRE’s and Operations demand for service quality. A lack of maturity in DevOps investment often results in a hybrid approach in most organizations where deployment pipelines and high levels of automation is applied for releases in the lower environments, but with a traditional, manual release process for Production environments.
How do we go about mitigating some of these issues and challenges? Organizations need to move past the experimentation and isolated targeted adoption phases into a more comprehensive approach towards adopting the DevOps model and strengthening it with the principles behind the ‘Continuous Delivery’ approach to bring more reliability and trust into the automated, low touch release and change management processes. It is a proven model and many global organizations have successfully implemented it at scale. Most of the technology is now available either in open source or as off-the-shelf capabilities in different cloud platforms.
Some of the specific principles include:
- DevOps culture which looks at automating every step of the release process
- Automated integration and smoke tests
- Many small incremental releases versus one big-bang release
- Deployment architecture and release processes that limit the ‘Blast Radius’
- Dial-Up Weighted Routing – Release new versions to a smaller userbase and then dial-up/rollback as needed – 5% > 10% > 25% > 50% > 75% > 100% users
- A/B Testing
- Automated, metrics-driven change management gates that control release execution based on a set of acceptable quality guardrails
- Blue/Green, Red/Black for instant cut-over and rollbacks
- Synthetic transactions to uncover potential issues automatically coupled with self-healing/recovery capabilities
- Smart operational monitoring with anomaly detection using ML
How are one-time security & compliance assurance gates failing the enterprise in both agility and risk assurance? And why is now the time to prioritize fixing their shortcomings?
Abhilash Panickar, Synechron: As organizations become more agile with ever shorter release cycles (weeks/days versus monthly/quarterly), the security and compliance model hasn’t changed very much – often still relying on manual reviews and change approval checkpoints prior to the release.. The Security and Compliance domain has traditionally been looked upon as a specialized skillset managed centrally by an independent team; vertically focused, rather than being decentralized across the different teams in the enterprise. To be effective in the rapidly changing IT landscape which has embraced the API-First, Cloud-First, Automate-Everything mentality, Security and Compliance processes within the organization need to be adapted and embedded in the SDLC, as well providing a more continuous form of evaluation and verification instead of a one-off review before every release.
What are some of the changes that enterprises can adopt to become more effective?
- Embrace Security First along with API and Cloud First — Security and compliance should not be an afterthought. Instead, teams should embrace the shift-left mentality and make it an integral part of the team’s responsibility – ‘design it, build it, run it’. The teams can be supported by the horizontal Security COE team to provide the necessary guidelines and guardrails for them to be successful.
- Implement automation – Automate compliance testing on a continuous basis by embedding security and compliance testing tools in the build and deployment pipelines. DevSecOps is the industry term for this unified approach that brings together operations, security and app dev in a collaborative framework.
- Augment with manual reviews – Not every threat vector can be analyzed and verified using the automation tools. Use cyber scenario based test exercises (table-top exercises, Cyber assessments, etc.) alongside funcitional testing tools to strengthen the tools-driven compliance process.
- Collect and analyze everything – Implement a continuous monitoring program driven by real environmental data to provide the required information necessary to support decision making and compliance assessment.
- Enable security trainings – Security is a continuously evolving landscape requiring periodic upskilling/reskilling to make the teams aware of the emerging and on-going threats.
How to determine if your enterprise is prepared for the inevitability of cyber exploits? Why have your investments in high availability and fault tolerance not prepared you for recovery from Cyber scenarios and exploits?
Chris Zanelli, Citihub Digital: Cyber threat and exploit scenarios differ from traditional disasters in several key ways, often requiring additional recovery capabilities not handled by application failover or traditional disaster recovery site failover strategies. The key challenge to recovering from cyber exploits lies in the fact that Cyber scenarios have complex, atypical patterns that present themselves with high variability across:
- the time of the exploit
- the time the impact is realized
- the time of detection
Given that each of these critical timeframes can vary significantly, it poses a significant challenge to answering the two fundamental questions to structure recovery:
- When was the last known good state?
- How do I recover to that last known good state?
Executing a point in time recovery against the moving parts of infrastructure, data sources, data schemas, application code and configurations can be quite complex and goes far beyond how enterprises organize responses to disaster scenarios and structure their change management capabilities.
How should enterprises protect business critical and sensitive data? Must enterprises maintain air-gapped backups that are protected from active compromise? How can an enterprise prepare itself against latent attacks and protect data integrity?
Chris Zanelli, Citihub Digital: Cyber Security and Recovery from cyber exploits requires a high level of confidence that the desired recovery state is ‘known to be good’ from both an operational perspective and from a data integrity perspective. As traditional resiliency and disaster recovery has focused on minimizing downtime and outages through active replication and geographically diverse ‘active/active’ sites, what we have learned about cyber exploits is that they aggressively seek to move laterally across infrastructure and data.
Highly available systems with near-real-time replication increase the likelihood of malware, ransomware, or other exploit code replicating itself across your distributed environments. Cyber exploits tend to exhibit latency in two critical timeframes which impede identification and response:
- Latency between the point of impact and remediation, which represents your actual customer/business exposure
- Latency between the time-of-exploit and the time-of-impact, which represents a period of unknown exposure
As the drift between the time-of-exploit and the time-of-recovery increases, not only does the potential for data loss increase but the likelihood of successful recovery also diminishes greatly. Successful recovery of data assets becomes a function of how long you retain air-gapped backups for your critical data sets.
As an enterprise’s data infrastructure is a massive target for bad actors, how can you ensure that you know where your all your golden source data is and how to recover it?
Chris Zanelli, Citihub Digital: As large enterprises seek to enhance their capabilities around air-gapped backups, data retention, and point-in-time recovery it is imperative to ensure that the investments that are being made and the priorities placed around data recovery are centered first and foremost around golden source and authoritative data sets. A well thought out Cyber Resiliency program must spend a significant amount of time making sure that the most critical and hard-to-replace data sets are protected first, whilst data sets that can be rebuilt or derived from golden-source copies can be a secondary priority.
The benefits of establishing this level of data governance and lineage will be significant not only for enhancing recovery from cyber exploits, but will also be significant in establishing enterprise data management capabilities across critical data sets and the lineage of data re-use across the organization. Organizations looking to reduce duplication of data assets, eliminate the costs of data duplication, enable better data warehousing, reduce their compliance footprint, and improve overall data quality will also find significant improvements in their Cyber Security posture.
How can enterprises start to understand the impacts of Cyber threat scenarios on their lines of business? Is now the time to enhance a BCP strategy for Cyber Resiliency if it’s centered on executing site failover patterns only?
Chris Zanelli, Citihub Digital: As mentioned previously, traditional resiliency and disaster recovery efforts have focused on minimizing downtime and outages through active replication and failover to geographically diverse ‘active/active’ sites. While ensuring a rapid recovery to localized failures, this strategy can be counter-productive in a cyber exploit
Today’s exposures and cyber scenarios require the enterprise to invest both in High Availability and full stack, point-in-time recovery solutions. The capabilities employed for Disaster Recovery need to ensure there is an adequate data retention period and full stack recovery plan on hand for point-in-time recovery that will pre-date the initial exploit. Businesses that fail to invest in this area will be highly exposed to ransomware and extended outages caused by the inability to identify and restore to a known good state, and quickly resume services in the face of an outage or an ongoing, public exploit.
Incorporating Cyber scenarios in your Disaster Recovery planning, testing, education, and other exercises is crucial to understanding your ongoing capability gaps and execution maturity, and ultimately reinforcing what the mechanics of cyber recovery execution require from both tooling and expertise perspectives.
JOINING FORCES is a blog series aimed to showcase the synergies between Synechron’s and Citihub Digital’s SMEs.
Synechron’s October 2020 acquisition of Citihub Digital allows for multiple complementary offerings across the combined firm. This acquisition expands Synechron’s existing digital, consulting, and technology capabilities across the global financial services and insurance industry’s landscape. Together, the firms provide a targeted focus on digital transformation, architecture/operating model and application modernization, cloud enablement, critical cybersecurity, and other strategic business solutions.
About our SMEs
Chris Zanelli, Associate Partner, Citihub Digital | Chris is a versatile and well-rounded technologist with 15+ years’ experience in the Financial Technology Services industry with expertise in both commercial technology offerings and building end-to-end enterprise technology platforms. He has a proven track record of high impact delivery that emphasizes business outcomes through product ownership, agile ways of working, data management, and software development automation. Coupled with an ORIE background with focus in linear and non-linear programming and Neural Networks, he brings practical experience in scaled delivery to the emerging applications of Artificial Intelligence and Machine Learning.
Eklove Mohan, Sr Director – Technology, Synechron | Eklove plays a variety of roles in Innovation, Research & Development, Cloud Initiatives, DevOps, Technology Evaluation, mentoring and training of the young IT generation. During his time at Synechron, he has had extensive exposure and experience working with cutting-edge technology through the delivery of tactical and strategic solutions to clients. He also has had the opportunity to work with some of the most inspirational thought leaders in the industry.
Abhilash Panickar, Sr Director – Technology, Synechron | As a software architect, Abhilash spends most of his time analyzing business requirements, articulating solutions, driving the decision-making process for key stakeholders and working closely with development teams to translate design and vision into reality. He has developed expertise in implementing proprietary solutions and building apps using various BPM and SOA platforms in On-Premises, Cloud and Hybrid deployment models using different development methodologies as well as different architecture & governance frameworks.
FT Names Citihub Digital as one of 2021’s Leading Management Consultants
“We’re thankful to be recognized for the 4th successive year for our leadership in financial ser...
Designing a Next Generation Pipeline for Cloud Native Application Compliance
How do financial services firms ensure that cloud native apps remain secure and compliant to various...
Breaking Stereotypes, Bolstering Diversity
Find out how Associate Partner Russ Simon is breaking stereotypes and what Citihub Digital is doing ...
Assessing and Governing Data Quality in the Enterprise
How can enterprises assess and improve data quality, and how can they use governance to drive the on...
Essential Capabilities for building Defense in Depth for Public and Hybrid Cloud
The explosive increase in Cloud Services consumed across the industry, coupled with a dramatic incre...
Cyber Resilience is more than Detect and Prevent, it’s also about Respond and Recover
Joining forces with Synechron, our Risk & Cyber Security SMEs Graham Fletcher and Gavin Wilson share...
Digital Transformations in the age of COVID-19
Joining forces with Synechron, our Enterprise Transformation SMEs offer their perspectives on why a ...
Meeting your Data Obligations in the Cloud After Brexit
Joining forces with Synechron, our RegTech SMEs – Bob Mudhar & Anand Chandra - talk about the the ...
How to set up a Test Coverage threshold in Go and Github
Senior Consultant Luis Carrazana enumerates the steps needed to implement a Test Coverage Threshold...