Data protection in the cloud

Issue 1 2023 Infrastructure


Rick Vanover.

The ‘Global DataSphere’ is exploding in size. IDC predicts that by 2026, the amount of data in the world will have doubled again. While most enterprises have digitised their operations, they continue to add more strategic workloads and create more and more data. So, as the amount of data enterprises have to deal with grows exponentially, moving to the cloud based on an elaborated strategy offers significant benefits like scalability, flexibility and cost-effective storage.

But can this go on forever? Gartner expects total worldwide end-user spending on public cloud services will reach a record $592 billion this year, a 21% increase from 2022. The Cloud Security Alliance (CSA) reported that 96% of companies say they have insufficient security for sensitive cloud data, so across the board we have a long way to go on this journey. Here are three best practices for enterprises to protect their data in the cloud.

1. Know your data

The first step to solving any problem is to know what you are dealing with. Before you can protect anything, you need to know who is storing what, and where. Is everyone in the business using the same accounts? To make sure this is done right, IT teams often need to play detective or go on a journey of discovery across the business. To find these threads, it is often necessary to look through finances and collect invoices for cloud costs across the organisation.

When brought together, the amount of data kept by most enterprises, whether it was migrated over from on-premises or originally stored in the cloud, is vast. Humans are natural hoarders, and the digital world is no exception. While the ‘virtual garage’ of the cloud can store endless boxes of data, locating everything is only half the battle. In order to know what data is mission-critical and sensitive, you will need to classify it.

Automated data classification engines can help you sort and organise, so you are not blindly trying to protect everything to the nth degree. Once you know exactly what you have stored on the cloud (and where) only then can you start looking at how this data is secured.

As organisations face a fairly low barrier of entry to move data to the cloud, teams may not have prioritised the security and network processes that are required; if the migration happened too fast this can easily be the case. Similarly, because the cloud is a completely different environment to secure, things are often missed. There are many new service types that do not always exist on-premises and many of these need to be protected and recovered in the case of attacks or outages. Examples of these include code in cloud storage, applications that leverage other cloud services, and APIs provided in the cloud.

2. Know your responsibilities

A key issue is enterprises often not realising exactly what they are responsible for, regarding security and data protection in the cloud. There is a big gap in awareness of the shared responsibility model on which cloud security is built. This means they assume the provider is responsible for certain security measures, when in reality it is their job. While it does depend on the cloud provider, typically the provider is responsible for the security of the infrastructure and the physical facilities that host it. Securing applications, data and access to the environment, however, is the responsibility of the customer.

In practice, this means enterprises need to ensure they have backups of all critical and sensitive data stored in the cloud, in case of breaches or outages. The best practice is to have multiple backups in different locations (e.g. one on-premises as well as a cloud copy) and have copies of data across different mediums, with at least one copy kept offsite, offline and immutable – even better yet, all three.

The other core security responsibility that lies with the enterprise is controlling access and privileges. If every user of your cloud has access to ‘God Mode’, any breach is going to be devastating. Likewise, if you are using a single account to do multiple functions like protection and provisioning.

The best practice is to ensure multiple accounts are used across the business, using access and identity management correctly across accounts and subscriptions, so you can easily remove the failure domain, in the case of a security breach. At a user level, ensure the principle of least privilege is followed across the cloud environment, so that people only have access to the resources and environments they need.

3. Keep it cost-effective

In all likelihood, putting the previous two principles into place will be a significant project for most enterprises. The good news is the initial heavy lift to do so will not be required again on the same scale. However, to keep the cloud environment healthy and cost-effective long term, it is important to have cloud data hygiene processes in place.

Ensure you have a proper data lifecycle process. Without it, the good work done initially will become ineffective and expensive over time, with the business paying to store and protect the wrong data in the wrong ways. Data needs to be on the right storage platform in the cloud – and this will change during its lifecycle. For example, it might move from block resource to object storage to archive storage. The costs associated with these are variable, so make sure you are not storing (or backing up) data in inefficient ways.

This is one small part of avoiding eventual ‘bill shock’ for cloud computing and storage costs. Beyond simple data, costs are API costs, data egress (transfer) and more. I always recommend enterprises have an established ‘cloud economic model’ that they follow to prevent costs from piling up and to ensure spending matches expectations. To use a real-life analogy, if you leave a light on or forget to cancel a subscription you no longer use, your monthly bills will be higher than expected. If this happens across an enterprise cloud environment, the total tally can be eye-watering.

As enterprises' (and the world’s) amount of stored data continues to grow over the next five years, the cloud is going to be a vital piece of the puzzle in managing this. Enterprises need to look beyond just storing and protecting their data, and look at ways to utilise it and unlock value for their business and their customers. Doing this requires re-factoring for greater agility, but this will also mean the business is prepared for the ‘whatever’. Cloud computing is nothing if not dynamic, and will continue to evolve, with best practise bound to change. If enterprises become data-centric now, on both the cloud and on-premises, they will be ready for whatever the future throws their way.




Share this article:
Share via emailShare via LinkedInPrint this page



Further reading:

VPS hosting set to dominate in 2025
Infrastructure
SME market growth and the increasing need for a digital footprint are pushing VPS growth in South Africa, especially since it is now perceived as a viable business tool, scalable by nature, with improved performance.

Read more...
Threats, opportunities and the need for post-quantum cryptography
AI & Data Analytics Infrastructure
The opportunities offered by quantum computing are equalled by the threats this advanced computer science introduces. The evolution of quantum computing jeopardises the security of any data available in the digital space.

Read more...
Highest capacity ePMR HDDs
Infrastructure Products & Solutions
Western Digital has announced that it is now shipping the world’s highest capacity UltraSMR HDD with up to 32TB leveraging the time-tested, reliable energy-assisted PMR (ePMR) recording technology for hyperscalers, CSPs and enterprises.

Read more...
Axis introduces ACS Edge and cloud storage
Axis Communications SA Surveillance Infrastructure Products & Solutions
Axis Communications has launched two new solutions within the AXIS Camera Station ecosystem, AXIS Camera Station Edge (ACS Edge) and AXIS Camera Station Cloud Storage (ACS Cloud Storage).

Read more...
Autonomous healing systems are the future
Infrastructure Information Security AI & Data Analytics
Autonomous healing software, an emerging technology, is gaining traction for its potential to transform how organisations manage software maintenance, security, and system performance.

Read more...
Navigating a connected, AI-driven future at SATNAC 2024
Infrastructure AI & Data Analytics IoT & Automation
The 2024 Southern Africa Telecommunication Networks and Applications Conference concluded its 26th edition with a call to harness AI to drive positive change across the continent. Moreover, students from Wits, North West and Pretoria universities won the best research paper awards.

Read more...
Cost-effective and reliable remote connectivity
Agriculture (Industry) Integrated Solutions Infrastructure
Companies that operate in hard-to-connect areas now have access to reliable connectivity due to a collaboration between MTN South Africa, Vox and Tarana technology.

Read more...
Data resilience in the age of AI
Infrastructure AI & Data Analytics
The discourse around AI has reached a fever pitch, but this ‘age of AI’ we have entered is just a chapter in a story that has been going on for years – digital transformation.

Read more...
Is cybersecurity broken?
Information Security Infrastructure
Companies are spending large amounts on cybersecurity, yet cyber threats continue to persist and thrive. Security executives are under tremendous pressure, and companies are questioning the wisdom of their security budgets.

Read more...
Data resilience for companies of all sizes
Technews Publishing Infrastructure Products & Solutions
Nakivo offers backup, recovery and replication solutions for all companies, but is focused on small- to medium-sized businesses where customers need affordable solutions that are simple to use and reliable.

Read more...