In the guest blog below, originally written by Gerardo Dada from Data Core we learn about software-defined storage including its role in disaster recovery and improved efficiency. This article explains why software-defined storage is essential in the evolution of the data centre.
You can find Gerardo’s full blog below, as well as a link to the original at the bottom.
The line between ‘traditional’ data centers and the cloud is increasingly blurring; many data centers today primarily consist of virtualized resources in a co-location environment. As part of this continuing evolution, IT will require more flexibility and freedom than ever before.
Software-defined infrastructure is defined by four main characteristics:
- Pooled resources that eliminate silos, delivered as a service
- Virtualization: hardware abstraction and vendor-agnostic
- Automation: software-based provisioning, control, and reporting, via API or management interface.
- Services that enhance capabilities
In many IT environments today, storage teams spend up to 50% of their time migrating data. Once you have adopted software-defined storage, migrations become a thing of the past. Data is moved seamlessly, leveraging machine learning-assisted auto-tiering across storage arrays to dynamically optimize performance and make best use of the performance profile of each array. Capacity planning is simpler, as the system manages capacity at a pool level and allows thin-provisioning to maximize utilization.
The modern data center also requires data to be reliable and available all the time. This usually means incorporating storage technologies that support synchronous in local and metro clusters, asynchronous replication for disaster recovery, and continuous data protection (which is similar to a time machine that is able to undo any damage from ransomware attacks).
Those who have gone through a disaster or a storage failure understand that the most important aspect of ensuring availability is the recovery process—and it should be instantaneous and automatic, should not require human intervention, resulting in zero impact for users and applications. Once the problem is averted and the failed storage system is back online, it is ideal to have zero-touch fall back and re-build to go back to the original state and be ready for a future failure.
As IT departments look to reap the benefits of the software-defined data center, the performance, the freedom and flexibility advantages of software-defined storage will be increasingly realized – in addition to performance and availability. This will help them spend less time on repetitive tasks and expand the technology to cover more of their IT footprint, including additional workloads or data centers.
Originally posted by Gerardo Dada here:
https://www.datacore.com/blog/software-defined-storage-is-critical-for-modernizing-the-data-center/