Cloud backup and disaster recovery - Cloud like filing cabinet, hosting or database with folders isolated on white background.

Creating a Successful Cloud Backup and Disaster Recovery Plan

Cloud backup and disaster recovery is easy to set up. But how can you be sure these provisions will actually work when the worst happens, and you encounter a genuine disaster?

As always, success is dependent on planning, the better your plan, the less risk of failure.

Audit your systems and data

Effective backup (and recovery) relies on having at least one copy of all of your data duplicated off-site to protect against loss of the original. As you begin designing a backup and DR plan, there are three questions to answer:

1.  What data assets do we have?

2.  Where is the data stored?

3.  What tools do we need to backup that information to the cloud?

As your business adopts more cloud services, defining what you have and where it is located is the first step to defining a workable plan. As well as the on-premise file shares, you probably have information stored in several SaaS services, including Office 365 and OneDrive. You may also have hosted email (Outlook 365) – anything stored there will also need to be included in your provisions.

You will then need to assess the available technologies to find one that can backup all of your data from every location. Do you need to install local agents? Can backup and restore operations be completed from a single console? Are there any permissions issues you need to address when connecting to third-party services?

Once you have these answers, you are in a position to select a cloud backup provider and to begin backing up to their hosted service.

Prioritise your systems and data

During a disaster, your business must be able to resume skeleton operations as quickly as possible. This means prioritising applications and data so that they are recovered in order of importance to the running of your business.

You should also look to automate as much of the recovery process as possible, freeing your IT staff to focus on other tasks (there will be plenty during a show-stopping outage).

Test your recovery plan

The worst time to discover your disaster recovery plan doesn’t work is during a disaster. Any kind of oversight could have significant consequences – which is why testing is so important.

By testing the DR plan in advance, you can confirm that:

    • All of your data is being backed up correctly
    • Your cloud backup archives are usable
    • Your order of restore does match your operational priorities.
    • Recovery completes within the acceptable timeframes.

You can then refine your DR plan, tweaking, optimising and re-aligning the various elements as required. Your cloud backup and disaster recovery plan must then be reviewed and tested regularly to ensure it continues to meet your needs.

Define – Prioritise – Test

Much like any other DR plan, an effective cloud backup strategy has three elements; define, prioritise and test. So long as your plan addresses them all, your business systems should be ready to face any outage.

To learn more about cloud backup and how to build a disaster recovery plan that meets the needs of your business, please get in touch.

cloud backup strategy

Do you need to get physical with a cloud backup strategy?

Virtualising backup with the cloud is powerful, effective and extremely safe. But just because data is now being archived off-site does not mean that hardware can be completely removed from your backup strategy.

In fact, physical hardware may still have an extremely important role to play in your cloud backup strategy.

1. Export by hard drive

The initial speed of a cloud backup may take weeks to complete as you transfer terabytes of data offsite. The actual time taken will depend on network and broadband speeds. Without careful traffic management, the uploads may negatively impact day-to-day operations too.

The process can be accelerated by shipping physical drives to the backup provider so that the data can be copied locally. This will be exponentially quicker – and arguably more secure – than trying to upload over the internet.

2. Restore by hard drive

Restoring from cloud archives is just as important – and fraught with the same difficulties. Speed of recovery will be limited by available internet bandwidth and download speeds.

For downloads that can be sized in gigabytes, online recovery will probably be acceptable. But for a disaster recovery scenario which involves a large amount of data, the speed of transfer is critical.

In the same way that physical hard drives can accelerate seeding of backups, they can also be employed to speed up recovery. If you plan to make cloud backup your principal method of data recovery, check to see if your service has the option of shipping physical disks.

3. Cloud as backup

The issue of time to recovery is of critical importance. Knowing that a complete dataset may take days to recover from the internet, it may be that the cloud is best deployed as a secondary backup.

In this scenario, your existing systems provide real-time services for instant recovery, while periodic (daily / weekly / monthly) backups are replicated to the cloud. Maintaining physical backups on-site minimises time to recovery, while off-site backups help to maintain the integrity and ensure that data is always recoverable.

4. Local servers for recovery testing

You know that your data is always protected when using cloud backup services – but how do you go about recovering it? Keeping spare physical servers will allow you to test your recovery protocols and ensure that they deliver against business needs.

For best results, keep at least one example of each bare metal server to ensure everything works correctly.

5. Physical recovery documentation

Modern business is driven by digital data – but there will always be a place for hard copy records in certain circumstances. In the case of disaster recovery, you must maintain physical, off-line copies of the information required to brings systems back online.

Records must include the recovery action plan, applications and serial numbers. And don’t forget to include contact details for the individual who holds the administrative passwords required for recovery and reconfiguration.

The future is hybrid

Until available bandwidth increases exponentially, there will always be a place for physical assets in your backup regime. The trick is knowing where to divide the load between local and cloud.

WTL offer a range of cloud based solutions. that can extend the rigour of your on-premise backup without without compromising control, visibility, or auditability.

For more assistance in defining a cloud backup strategy that delivers the reliability, speed and security your business demands, please give us a call.

Moving to the Cloud

The Less Scary Road to the Moving to the Cloud

Cloud adoption is set to become a computing norm – even for companies that have until now rejected these technologies. But as hosted software (G Suite, Office 365, Salseforce.com etc) gather pace, most have been unable to completely avoid cloud services.

Much of the discussion around cloud migration suggests that it is a ‘big bang’, all-or-nothing play, with the whole data centre being shifted to the cloud. Although possible (in theory) not all workloads belong in the cloud.

Cloud migration doesn’t have to be ‘big bang’

Many cloud operators give the impression that adoption is not only inevitable but that all of your systems will eventually be hosted. And the sooner this transition takes place, the better.

The reality is that this is your business and your infrastructure, and you are fully justified in moving at your own pace. For various reasons (unfamiliarity, security concerns, uncertainty etc) you have resisted major cloud adoption projects – so it makes sense to maintain a cautious roll-out.

Cloud migration on your terms

One way to maintain control of the process and the speed at which you move is to bring cloud technologies into your data centre first. Using platforms like VMware vCloud Suite, Microsoft Hyper-V virtualisation or OpenStack.

Deploying a private cloud allows your business to migrate applications and workloads, learning how the concepts and technologies apply to your business. At the same time, you can take advantage of automation and self-service to accelerate your IT operations to deliver a better quality of service to your in-house users.

This approach can be more expensive than going with one of the large platforms like AWS, Azure or Google Cloud. With cloud in-house, however, you retain full control of the process so you can migrate applications and servers at your own pace. This makes the transition more manageable and lays the groundwork for when you do decide to migrate to a lower-cost public cloud provider.

Re-engineering workloads for the cloud

One of the key benefits of cloud platforms is their elastic pricing model – you only pay for what you use. However, simply moving your virtual servers into the cloud is not efficient.

Your on-premise systems are configured to run 24x7x365 because there is no reason to let them spin down. But in the cloud where every resource is billable – CPU cycles, RAM, storage etc – you pay for running servers, even when they are not being accessed.

The major cloud platforms allow you to set servers to spin down automatically overnight for instance, helping to reduce costs. However, these servers are themselves considered relatively heavyweight.

The future of operating in the cloud lies in containerisation. This technology breaks applications into blocks that can be created and destroyed automatically according to demand. Unlike a virtual server, the container is a much smaller package, containing nothing but your application code and the libraries required to run it; there is no operating system or additional applications, helping to minimise the number of resources used – and therefore costs.

With a private cloud, you can begin the process of re-engineering and optimising for the cloud before moving to a public cloud platform. This will help to contain costs when you do finally migrate and simplify the process of transition.

To learn more about the moving to the cloud and how to simplify the transition to the cloud, please get in touch.

Cloud Partner

How to choose a cloud partner – 13 factors to consider

If the cloud is an integral part of your IT strategy, choosing the right cloud partner is one of the most important decisions you must make. Here are 13 factors to consider that will help you chose a cloud partner that is right for your business.

1. Technology stack

To contain costs, maximise ROI and accelerate deployment, you want a provider whose technology stack is compatible with your own. As well as using the same hypervisor technology, their hardware and connectivity needs to be reliable, fast and resilient. The platform must outperform your own.

2. Public or Private

Public cloud platforms offer multi-tenancy solutions to share resources and contain costs. Private cloud offers dedicated infrastructure for each client. Although carefully configured to prevent data passing between tenants’ instances, public cloud may not be suitable for your most sensitive workloads. Make sure any prospective supplier offers the right cloud formats for your workloads.

3. Management responsibilities

Maintaining the underlying infrastructure and connectivity is the responsibility of the provider – but what else do they offer? Can you spin up your own instances or allocate virtual resources, or is that handled by the provider too? Understanding these responsibilities will also help to understand the ongoing costs and staffing requirements.

4. Infrastructure transparency

By ceding control of the underlying infrastructure, cloud services do not offer the same degree of visibility as your on-premise data centre. Make sure that you understand the tools required to deliver business value and whether they can be used on the platform being evaluated – or if the provider offers an equivalent toolset.

5. The route to the cloud

There are many routes to the cloud – you need to pick the option that balances speed, future-readiness and cost. You need to understand the data and workloads that need to be migrated, and how it will be moved. Does the provider offer a managed migration service to help plug skills and knowledge gaps within your own team?

6. Third-party networking support

Pure cloud deployments are extremely rare – you will probably continue to run at least some of your operations in-house. You need to understand whether the provider offers support for the third party network technologies that connect on-premise with the cloud, and to what degree. Will your provider assume all the risk, some, or none at all?

7. Physical hosting options

If you operate any form of hosted infrastructure or co-located systems, it makes sense to consolidate them where possible. If your processes require physical hardware of any kind, you need to know if your partner is willing to host it.

8. Automating the cloud

Your on-premise systems use automation to streamline processes and reduce manual workload – are there similar options available in the cloud too? What are your options for automation and reporting? Is there an API available allowing you to connect third-party tools – or to develop your own?

9.  Disaster recovery provisions

Cloud platforms are resilient by design – but the risk of data loss remains. How is your data protected at the virtual and infrastructure levels? Can you configure and control backup frequencies? And importantly, how is data recovered when required? Does the provider offer disaster recovery as a service (DRaaS) and what are their SLAs?

10. Cloud failover mechanisms

Cloud outages are rare, but the risk of ransomware infection, accidental deletion and data corruption are almost as common as in your on-premise data centre. User error is still a threat wherever your systems are hosted. Your ideal provider needs to provide a way to recover data and operations within minutes – including cloud-to-cloud fail-over if there is a significant ransomware/data corruption event.

11. Compliance and security

Your business has a legal duty to protect data from loss or theft – and to ensure it is stored securely in an approved territory (normally the UK and EU). As you assess the various options available you need to know how responsibility is divided between parties, the tools and services (if any) that are available to maintain compliance and the physical location of your data to ensure it is not being stored illegally.

12. Technical support provisions

The provider’s technical support offering takes on additional importance when you have outsourced key elements of your IT estate. You need to be sure about the level of service you will receive, including how issues are reported, investigated and resolved. You must also ascertain what the service level parameters are to ensure you have the level of coverage appropriate to your workloads.

13. How to track costs

The pay-as-you-use cloud computing model ensures that you never invest in over-capacity for your on-premise data centre. But it also makes budget over-spend far easier, particularly given the complexities of some cloud billing models. Your ideal provider will have transparent pricing, cost models and tools that allow you to track your spend and balance resource requirements with budget constraints.

Every factor is important

13 factors might seem like a lot – but every one of them is important to consider when chosing a cloud partner. After all, this is a strategic partnership that should last years. The time and effort invested in conducting this due diligence will help to smooth your cloud adoption projects – and ensure you get exactly what you need to grow your business.

To learn more about the WTL offering and how we can help you build a flexible, future-ready platform in the cloud, please get in touch.

Oracle Cloud at Customer

All the power of the public cloud in your data centre

Private cloud technologies have been instrumental in helping businesses build fully scalable databases. As these platforms age, attention turns to the next evolution; systems must continue to deliver improved performance and ROI, so the next round of investments is crucial.

The fact that your business has been using a private database cloud indicates that your workloads have been deemed too sensitive for public alternatives. This does mean that you have missed out on some of the additional benefits available when using public cloud though.

With Oracle Exadata Cloud at Customer, the functionality and feature gap is narrowing. Which means that the next generation of private database cloud can deliver more for less. How much less? Oracle suggests a 47% reduction in the total cost of operations for starters, followed by a 256% return on investment over five years.

Simplified licensing

One of the biggest headaches associated with the private cloud has been licensing. Pooled resources allow for instant scaling according to need and application, but perpetual licenses are a fixed commodity – and you must have enough to cover usage at all times. This means buying additional licenses ‘just in case’ to meet occasional spikes in demand.

With Exadata Cloud at Customer (ExaCC), licenses can be activated and released in line with demand – just like popular public cloud services. Your private cloud database can quickly scale from 2 to 400 cores and back again. You only pay for what you use, when you use it – and you never risk having the wrong number of licenses again.

Reduced management overheads

An in-house private cloud typically relies on a collection of technologies held together by custom scripts and management tools. Although they “work”, there is a lot of manual integration required to keep everything running.

Upgrading to ExaCC dramatically reduces management overheads. As part of the service, Oracle assumes full responsibility for managing infrastructure, reducing the burden on your IT team. Oracle clients report their management processes are 69% more efficient with EXaCC. This approach also ensures you are unaffected by the ongoing cloud skills shortage.

Cloud scale-out

Data storage and processing needs continue to increase year on year, placing an additional burden on the in-house infrastructure. You existing private cloud is still constrained by a finite pool of underlying infrastructure resources.

ExaCC provides optional automated scale-out capabilities, seamlessly linking on-premise private cloud with the hosted public cloud to extend capabilities for specified workloads. This allows you to increase resource usage beyond the constraints of your private database cloud as and when required, also billed on a pay-as-you-use basis.

Enhanced security

One of the key reasons for choosing a private cloud is the degree of control you maintain over applications and data. ExaCC helps to further de-risk the operating environment by bundling Oracle ‘defence in depth’ tools as part of the service.

Security safeguards are built into both hardware and software, with additional encryption included in the database engine. And it’s all a standard part of the ExaCC service.

Just scratching the surface

As public cloud feature sets continue to evolve, there’s no reason why the private cloud has to be left behind. Oracle Exadata Cloud at Customer brings many of the very best features to the local data centre – and delivers incredible benefits in the process.

To learn more about ExaCC and how you can generate an ROI on your private cloud investments in as little as six months, please get in touch.