VDrO v.4 – Run a DR plan

This is the last article about how to integrate the Continuous Data Protection (CDP)  technology (available from VBR v.11) and VDrO v.4 (former VAO).

In this part, we are going to see what happens when an orchestration plan is launched.

Yes, I wrote the word “see” because I created a short video showing the tasks that are automatically completed when a Disaster Recovery is occurring.

If you need more details about how to set up the environment, please read the previous articles.

Let me know if videos and youtube platform are a good way to expose technological valuable topics.

Thx for reading and watching and take care

VDrO v.4 – Create a DR plan

C. Create an Orchestration Plan

The DR plan is a sum of more Orchestration plans. This article is going to explain how to create them.

Just a small and important note before continuing: it’s mandatory to have already completed the steps described in the last article.

Let’s start!

From the main menu of the VAO server select the Manage button.

Now click on New as shown in picture 1.

Picture 1

The easy wizard is going to ask to choose a scope.

In our example, we use Linux-CDP as shown in picture 2.

Picture 2

Now fill up the plan info with the Plan Name, the description, the contact name of the plan (picture 3),

Picture 3

Select the type of Plans. In this article choose the CDP replica (picture 4)

Picture 4

In the next step please check if the correct “VM Group” appears.
If it doesn’t, it’s necessary to go back to the setup phase (please read the previous article) and fix the issue.

In our example, it appears correctly (Ubuntu-CDP) as shown in the next two pictures (5 and 6).

Picture 5

Picture 6

The next step shows the VM Recovery options (picture 7).

It gives the operator control of the plan. For example, stopping the plan if something goes wrong.

Picture 7

In “New VM Template” menu the VAO user can add additional steps to the orchestration process; for example, starting the CDP replica job first and shut down the source VM after (Picture 8)

Picture 8

Tips I: I created a customized script to change the IP Address of the VM.

Tips II: it’s possible to set up the access credential directly from this page by clicking the button on the bottom of the page. It is available for Windows VM only.

The next step defines the RTO & RPO.

The most important thing to remember here is that the RPO has to be equal or major than the RPO set in the CDP replica job (picture 9).

https://lnx.gable.it/wp-content/uploads/2021/03/vao-cdp-45.jpgPicture 9

The last steps define when the plan report shall be automatically generated (picture 10) and if the readiness check has to run at the end of every single wizard (recommended option) (picture 11).

Picture 10

Picture 11

The result is shown in picture 12

Picture 12

The next article is going to be a video to see VDrO in action.

Take care and see you soon

Digital Trasformation & Data Mobility

If Cloud has been the most used word in the last five years, the words that have been buzzing the IT world in the last five months are Digital Transformation

From Wikipedia:

Digital Transformation (DT or DX) is the adoption of digital technology to transform services and businesses, through replacing non-digital or manual processes with digital processes or replacing older digital technology with newer digital technology”.

Or: Digital Transformation must help companies to be more competitive through the fast deployment of new services always aligned with business needs.

Note 1: Digital transformation is the basket, technologies to be used are the apples, services are the means of transport, shops are clients/customers.

1. Can all the already existing architectures work for Digital Transformation?

  • I prefer to answer rebuilding the question with more appropriate words:

2. Does Digital transformation require that data, applications, and services move from and to different architectures?

  • Yes, this is a must and It is called Data Mobility

Note 2: Data mobility regards the innovative technologies able to move data and services among different architectures, wherever they are located.

3. Does Data-Mobility mean that the services can be independent of the below Infrastructure?

  • Actually, it is not completely true; it means that, despite nowadays there is not a standard language allowing different architecture/infrastructure to talk to each other, the Data-mobility is able to get over this limitation.

4. Is it independent from any vendors?

  • When a standard is released all vendors want to implement it asap because they are sure that these features will improve their revenue. Currently, this standard doesn’t still exist.

    Note 3: I think the reason is that there are so many objects to count, analyze, and develop that the economical effort to do it is at the moment not justified

    5. Is already there a Ready technology “Data-Mobility”?

    The answer could be quite long but, to do short a long story, I wrote the following article that is composed of two main parts:

  • Application Layer (Container – Kubernetes)
  • Data Layer (Backup, Replica)

Application Layer – Container – Kubernetes

Using Local Persistent Volumes with Container Engine for Kubernetes | IaaS Blog - Oracle Cloud Infrastructure News

In the modern world, services are running in a virtual environment (VMware, Hyper-V, KVM, etc).

There are still old services that run on legacy architecture (Mainframe, AS400 ….), (old doesn’t mean that they are not updated but just they have a very long story)

In the next years, the services will be run in a special “area” called “container“.

The container runs on Operating System and can be hosted in a virtual/physical/Cloud architecture.

Why containers and skills on them are so required?

There are many reasons and I’m listing them in the next rows.

  1. The need of IT Managers is to move data among architectures in order to improve resilience and lower costs.
  2. The Container technology simplifies the developer code writing because it has a standard widely used language.
  3. The services ran on the container are fast to develop, update and change.
  4. The container is de facto a new standard that has a great advantage. It gets over the obstacle of missing standards among architectures (private, hybrid, and public Cloud).

A deep dive about point d.

Any company has its own core business and in the majority of cases, it needs IT technology.

Any size of the company?
Yes, just think about your personal use of the mobile phone, maybe to book a table at the restaurant or buying a ticket for a movie. I’m also quite sure it will help us get over the Covid threat.

This is the reason why I’m still thinking that IT is not a “cost” but a way to get more business and money improving efficiency in any company.

Are there specif features to allow the data-mobility in the Kubernetes environment?

Yes and for giving you a great example please have a look at the Kasten K10 product because it has many and advanced features (The topic will be well covered in the next articles because it is one of my great working passions).

Data-LayerCloud Backup Restore Icona - Download gratuito, PNG e vettoriale

What about services that can’t be containerized?

Is there a simple way to move data among different architectures?

Yes, that’s possible using copies of the data of VMs, Physical Servers.

In this business scenario, it’s important that the software can create Backup/Replicas wherever the workloads are located.

Is it enough?

From my point of view, it is not. The software has to be able to restore data intra-architectures.

For example, a customer can need to restore some on-premises workloads of his VMware architecture in a public cloud,  or restore a backup of a VM located in a public cloud to a Hyper-V on-premises environment.

In other words, working with Backup/Replica and restore in a multi-cloud environment.

The next pictures show the Data Process.

I called it “The cycle of Data” because leveraging from a copy it is possible to freely move data from and to any Infrastructure (Public, hybrid, private Cloud).

Pictures 1 and 2 are just examples of the data-mobility concept. They can be modified by adding more platforms.

The starting point of Picture 1 is a backup on-premises that can be restored on-premises and on-cloud. Picture 2 shows backup of a public cloud workload restored on cloud or on-premises.

It’s an open circle where data can be moved around platforms.

Note 4: A good suggestion is to use data-mobility architecture to set up a cold disaster recovery site (cold because data used to restore site are backup).

Picture 1

Picture 2

There is one more point to complete this article and it is the Replica feature.

Note 5: For Replica I intend the way to create a mirror of the production workload. Comparing to backup, in this scenario the workload can be switched-on without any restore operation because it is already written in the language of the host-hypervisor.

The main scope of replica technology is to create a hot Disaster Recovery site.

More details about how to orchestrate DR are available on this site at the voice Veeam Availability Orchestrator (Now Veeam Disaster Recovery Orchestrator)

The replica can be developed with three different technologies: 

  • Lun/Storage replication
  • I/O split
  • Snapshot based

I’m going to cover those scenarios and kasten k10 business cases in future articles.

That’s all for today folks.

See you soon, and take care.