|
General
|
|
|
- Fully Supported
- Limitation
- Not Supported
- Information Only
|
|
Pros
|
- + Extensive platform support
- + Extensive data protection capabilities
- + Flexible deployment options
|
- + Flexible architecture
- + Broad range of hardware support
- + Built for performance
|
- + Flexible architecture
- + Extensive platform support
- + Several Microsoft integration points
|
|
Cons
|
- - No native data integrity verification
- - Dedup/compr not performance optimized
- - Disk/node failure protection not capacity optimized
|
- - Limited data protection capabilities
- - No stretched clustering
- - No dedup capabilities
|
- - Minimal data protection capabilities
- - No Quality-of-service mechanisms
- - No native encryption capabilities
|
|
|
|
Content |
|
|
|
WhatMatrix
|
WhatMatrix
|
WhatMatrix
|
|
|
|
Assessment |
|
|
|
Name: SANsymphony
Type: Software-only (SDS)
Development Start: 1998
First Product Release: 1999
NEW
DataCore was founded in 1998 and began to ship its first software-defined storage (SDS) platform, SANsymphony (SSY), in 1999. DataCore launched a separate entry-level storage virtualization solution, SANmelody (v1.4), in 2004. This platform was also the foundation for DataCores HCI solution. In 2014 DataCore formally announced Hyperconverged Virtual SAN as a separate product. In May 2018 integral changes to the software licensing model enabled consolidation because the core software is the same and since then cumulatively called DataCore SANsymphony.
One year later, in 2019, DataCore expanded its software-defined storage portfolio with a solution especially for the need of file virtualization. The additional SDS offering is called DataCore vFilO and operates as scale-out global file system across distributed sites spanning on-premises and cloud-based NFS and SMB shares.
Recently, at the beginning of 2021, DataCore acquired Caringo and integrated its know how and software-defined object storage offerings into the DataCore portfolio. The newest member of the DataCore SDS portfolio is called DataCore Swarm and together with its complementary offering SwarmFS and DataCore FileFly it enables customers to build on-premises object storage solutions that radically simplify the ability to manage, store, and protect data while allowing multi-protocol (S3/HTTP, API, NFS/SMB) access to any application, device, or end-user.
DataCore Software specializes in the high-tech fields of software solutions for block, file, and object storage. DataCore has by far the longest track-record when it comes to software-defined storage, when comparing to the other SDS/HCI vendors on the WhatMatrix.
In April 2021 the company had an install base of more than 10,000 customers worldwide and there were about 250 employees working for DataCore.
|
Name: PowerFlex
Type: Software-only (SDS)
Development Start: Early 2011
First Product Release: dec 2012
NEW
ScaleIO was founded early 2011 and began to ship its first Software Defined Storage (SDS) solution, Elastic Converged Storage (ECS), late 2012. In June 2013, ScaleIO was acquired by EMC. Early 2016 EMC ScaleIO introduced its first Hyper Converged Infrastructure (HCIS) solution, ScaleIO Node, with the ECS platform at its core. In september 2016 ScaleIO Node was re-branded to ScaleIO Ready Node when the server hardware changed from Quantum to Dell. In march 2018 the ScaleIO product family was re-branded to VxFlex OS and VxFlex Ready Node respectively. In june 2020 the VxFlex OS product family was re-branded to PowerFlex.
Customer install base and number of employees are unknown at this time.
|
Name: StarWind Virtual SAN
Type: SDS
Development Start: 2003
First Product Release: 2011
StarWind Software is a privately held company which started in 2008 as a spin-off from Rocket Division Software, Ltd. (founded in 2003). It initially provided free Software Defined Storage (SDS) offerings to early adopters in 2009. Sometime in 2011 the company released its product Native SAN (later rebranded to Virtual SAN). In 2015 StarWind executed a successful 'pivot shift' from software-only company to become a hardware vendor and brought Hyper-Convergence from the Enterprise level to SMB and ROBO. Apart of HCA solutions, StarWind keeps focus on developing and improving its Virtual SAN solution. In 2018, StarWind released Virtual SAN for vSphere - an SDS solution specifically aimed at VMware vSphere environments.
In March 2020 the company had a StarWind Virtual SAN install base of more than 4,500 customers worldwide. In June 2019 there were more than 250 employees working for StarWind.
|
|
|
GA Release Dates:
SSY 10.0 PSP12: jan 2021
SSY 10.0 PSP11: aug 2020
SSY 10.0 PSP10: dec 2019
SSY 10.0 PSP9: jul 2019
SSY 10.0 PSP8: sep 2018
SSY 10.0 PSP7: dec 2017
SSY 10.0 PSP6 U5: aug 2017
.
SSY 10.0: jun 2014
SSY 9.0: jul 2012
SSY 8.1: aug 2011
SSY 8.0: dec 2010
SSY 7.0: apr 2009
.
SSY 3.0: 1999
NEW
10th Generation software. DataCore currently has the most experience when it comes to SDS/HCI technology, when comparing SANsymphony to other SDS/HCI platforms.
SANsymphony (SSY) version 3 was the first public release that hit the market back in 1999. The product has evolved ever since and the current major release is version 10. The list includes only the milestone releases.
PSP = Product Support Package
U = Update
|
GA Release Dates:
PowerFlex 3.5: jun 2020
VxFlex OS 3.0.1.1: jan 2020
VxFlex OS 3.0.1: sep 2019
VxFlex OS 3.0: mar 2019
VxFlex OS 2.6.1.1: jan 2019
VxFlex OS 2.6.1: oct 2018
VxFlex OS 2.6: may 2018
VxFlex OS 2.5: feb 2018
VxFlex OS 2.0.1.4: oct 2017
VxFlex OS 2.0.1.3: mar 2017
VxFlex OS 2.0: mar 2016
VxFlex OS 1.32: may 2015
VxFlex OS 1.31: dec 2014
VxFlex OS 1.30: sep 2014
VxFlex OS 1.20: oct 2013
VxFlex OS 1.10: dec 2012
NEW
6th Generation software. PowerFlexs feature set is getting richer, but the platform still lacks some advanced functionality that competing SDS/HCI offerings provide.
|
StarWind VSAN for vSphere Release Dates:
VSAN build 13170: oct 2019
VSAN build 12859: feb 2019
VSAN build 12658: dec 2018
VSAN build 12533: sep 2018
StarWind VSAN for Hyper-V Release Dates:
VSAN build 13279: oct 2019
VSAN build 13182: aug 2019
VSAN build 12767: feb 2019
VSAN build 12658: nov 2018
VSAN build 12585 oct 2018
VSAN build 12393: aug 2018
VSAN build 12166: may 2018
VSAN build 12146: apr 2018
VSAN build 11818: dec 2017
VSAN build 11456: aug 2017
VSAN build 11404: jul 2017
VSAN build 11156: may 2017
VSAN build 11071 may 2017
VSAN build 10927 apr 2017
VSAN build 10914 apr 2017
VSAN build 10833: apr 2017
VSAN build 10811 mar 2017
VSAN build 10799: mar 2017
VSAN build 10695: feb 2017
VSAN build 10547: jan 2017
VSAN build 9996: aug 2016
VSAN build 9980 aug 2016
VSAN build 9781: jun 2016
VSAN build 9611 jun 2016
VSAN build 9052: may 2016
VSAN build 8730: nov 2015
VSAN build 8716 nov 2015
VSAN build 8198 jun 2015
VSAN build 7929 apr 2015
VSAN build 7774 feb 2015
VSAN build 7509 de 2014
VSAN build 7471 dec 2014
VSAN build 7354 nov 2014
VSAN build 7145
VSAN build 6884
Version 8 Release 10 StarWind software.
|
|
|
|
Pricing |
|
|
Hardware Pricing Model
Details
|
N/A
SANsymphony is sold by DataCore as a software-only solution. Server hardware must be acquired separately.
The entry point for all hardware and software compatibility statements is: https://www.datacore.com/products/sansymphony/tech/compatibility/
On this page links can be found to: Storage Devices, Servers, SANs, Operating Systems (Hosts), Networks, Hypervisors, Desktops.
Minimum server hardware requirements can be found at: https://www.datacore.com/products/sansymphony/tech/prerequisites/
|
N/A
For the PowerFlex software-only solution, server hardware must be acquired separately.
Dell EMC does not maintain a Hardware Compatibility List (HCL) with supported hardware for PowerFlex implementations.
For guidance on proper hardware configurations, Dell EMC provides hardware requirements and reference architectures (SPEX).
|
N/A
StarWind Virtual SAN is sold by StarWind as a software-only solution. Server hardware must be acquired separately.
|
|
Software Pricing Model
Details
|
Capacity based (per TB)
NEW
DataCore SANsymphony is licensed in three different editions: Enterprise, Standard, and Business.
All editions are licensed per capacity (in 1 TB steps). Except for the Business edition which has a fixed price per TB, the more capacity that is used by an end-user in each class, the lower the price per TB.
Each edition includes a defined feature set.
Enterprise (EN) includes all available features plus expanded Parallel I/O.
Standard (ST) includes all Enterprise (EN) features, except FC connections, Encryption, Inline Deduplication & Compression and Shared Multi-Port Array (SMPA) support with regular Parallel I/O.
Business (BZ) as entry-offering includes all essential Enterprise (EN) features, except Asynchronous Replication & Site Recovery, Encryption, Deduplication & Compression, Random Write Accelerator (RWA) and Continuous Data Protection (CDP) with limited Parallel I/O.
Customers can choose between a perpetual licensing model or a term-based licensing model. Any initial license purchase for perpetual licensing includes Premier Support for either 1, 3 or 5 years. Alternatively, term-based licensing is available for either 1, 3 or 5 years, always including Premier Support as well, plus enhanced DataCore Insight Services (predictive analytics with actionable insights). In most regions, BZ is available as term license only.
Capacity can be expanded in 1 TB steps. There exists a 10 TB minimum per installation for Business (BZ). Moreover, BZ is limited to 2 instances and a total capacity of 38 TB per installation, but one customer can have multiple BZ installations.
Cost neutral upgrades are available when upgrading from Business/Standard (BZ/ST) to Enterprise (EN).
|
Per TB
Editions:
Basic
Enterprise (add-on)
The PowerFlex Basic license includes: High Availability, Self-Healing, Scale-out architecture, Support for 1,000s of nodes, Elasticity, Hyper Convergence, Automated performance tuning, Any Drive support (SSD/PCIe/HDD), Automatic Cluster Growth, Asymmetric nodes support – heterogeneous support of different server brands / OSes. Advanced Management – includes vSphere plugin, REST API, OpenStack Support.
The PowerFlex Enterprise license includes: QoS, Data Obfuscation, Snapshots, Auto tiering (Flash caching) – use XtremCache with PowerFlex, RAM caching, Fault Sets, Thin provisioning, Fine Granularity (FG) storage pools.
EMC uses a tiered pricing structure per physical device capacity TB for both the Basic and Enterprise licenses. The more capacity that is purchased, the cheaper the price per TB.
|
Hyper-V: Per Node + Per TB
vSphere: Per Node (storage capacity included)
There are two StarWind Virtual SAN editions: StarWind Virtual SAN for Hyper-V and StarWind Virtual SAN for vSphere. StarWind Virtual SAN for Hyper-V is licensed per node + per amount of HA storage provisioned by StarWind Virtual SAN. StarWind Virtual SAN for vSphere is licensed per the node. There the amount of HA storage provisioned by StarWind Virtual SAN is always unlimited.
HA = Highly Available
|
|
Support Pricing Model
Details
|
Capacity based (per TB)
Support is always provided on a premium (24x7) basis, including free updates.
More information about DataCores support policy can be found here:
http://datacore.custhelp.com/app/answers/detail/a_id/1270/~/what-is-datacores-support-policy-for-its-products
|
Per TB
Subscriptions: Basic, Enhanced, Premium
Basic: Response based on severity level within 9-5 basis plus 24x7 remote support.
Enhanced: Basic support plus next business day onsite.
Premium: Options include four-hour onsite responses as well as post-warranty maintenance.
|
Per Node
StarWind provides three editions of StarWind Virtual SAN Support:
1. Standard Support for business days, business hours and up to 4 hours response time
2. Premium Support covers 24x7x365 support and up to 1 hour response time
3. Proactive Support provides Premium Support and monitors the health of the system and proactively notifies about potential issues on the hardware level, hypervisor level and StarWind software-level.
|
|
|
Design & Deploy
|
|
|
|
|
|
|
Design |
|
|
Consolidation Scope
Details
|
Storage
Data Protection
Management
Automation&Orchestration
DataCore is storage-oriented.
SANsymphony Software-Defined Storage Services are focused on variable deployment models. The range covers classical Storage Virtualization over Converged and Hybrid-Converged to Hyperconverged including a seamless migration between them.
DataCore aims to provide all key components within a storage ecosystem including enhanced data protection and automation & orchestration.
|
Compute
Storage
Dell EMC is stack-oriented, whereas the PowerFlex platform itself is heavily storage-focused.
With the PowerFlex platform Dell EMC aims to provide key generic components within a Private Cloud ecosystem.
|
Storage
Management
StarWind Virtual SAN consolidates storage from different servers by replicating and presenting it over the iSCSI protocol as a single pool.
|
|
|
1, 10, 25, 40, 100 GbE (iSCSI)
8, 16, 32, 64 Gbps (FC)
The bandwidth required depends entirely on the specifc workload needs.
SANsymphony 10 PSP11 introduced support for Emulex Gen 7 64 Gbps Fibre Channel HBAs.
SANsymphony 10 PSP8 introduced support for Gen6 16/32 Gbps ATTO Fibre Channel HBAs.
|
1, 10, 25, 40, 100 GbE
PowerFlex supports ethernet connectivity using SFP+ or Base-T. Dell EMC recommends at least 10GbE to avoid the network becoming a performance bottleneck.
|
1, 10, 25, 40, 100 GbE
StarWind requires at least three dedicated network interfaces:
- one for StarWind Synchronization
- one for iSCSI traffic/Heartbeat)
- one for Management/Heartbeat
At least one Heartbeat interface must be on a separate network adapter and redundant.
For iSCSI and synchronization, minimum 1 GbE of bandwidth and latency under 5ms is a requirement.
|
|
Overall Design Complexity
Details
|
Medium
DataCore SANsymphony is able to meet many different use-cases because of its flexible technical architecture, however this also means there are a lot of design choices that need to be made. DataCore SANsymphony seeks to provide important capabilities either natively or tightly integrated, and this keeps the design process relatively simple. However, because many features in SANsymphony are optional and thus can be turned on/off, in effect each one needs to be taken into consideration when preparing a detailed design.
|
High
PowerFlex is able to meet many different use-cases because of its flexible technical architecture, however this also means there are a lot of design choices that need to be made. In addition PowerFlex does not encompass many native data protection capabilities and data services. A complete solution design therefore requires the presence of multiple technology platforms.
|
Medium
StarWind Virtual SAN is able to meet many different use-cases because of its flexible technical architecture, however this also means there are a lot of design choices that need to be made. Also, StarWind Virtual SAN does not encompass many native data protection capabilities and data services. A complete solution design therefore requires the presence of multiple technology platforms.
|
|
External Performance Validation
Details
|
SPC (Jun 2016)
ESG Lab (Jan 2016)
SPC (Jun 2016)
Title: 'Dual Node, Fibre Channel SAN'
Workloads: SPC-1
Benchmark Tools: SPC-1 Workload Generator
Hardware: All-Flash Lenovo x3650, 2-node cluster, FC-connected, SSY 10.0, 4x All-Flash Dell MD1220 SAS Storage Arrays
SPC (Jun 2016)
Title: 'Dual Node, High Availability, Hyper-converged'
Workloads: SPC-1
Benchmark Tools: SPC-1 Workload Generator
Hardware: All-Flash Lenovo x3650, 2-node cluster, FC-interconnect, SSY 10.0
ESG Lab (Jan 2016)
Title: 'DataCore Application-adaptive Data Infrastructure Software'
Workloads: OLTP
Benchmark Tools: IOmeter
Hardware: Hybrid (Tiered) Dell PowerEdge R720, 2-node cluster, SSY 10.0
|
ESG Lab (oct 2016)
ESG Lab (Oct 2016)
Title: 'Optimize Hyper-converged Infrastructure with Dell EMC ScaleIO Software-defined Storage'
Workloads: Generic, Oracle OLTP
Benchmark Tools: FIO, SLOB
Hardware: Hybrid Intel servers, 4-8 node cluster, ScaleIO 2.0.0
|
N/A
No StarWind Virtual SAN validated test reports have been published in 2016/2017/2018/2019.
|
|
Evaluation Methods
Details
|
Free Trial (30-days)
Proof-of-Concept (PoC; up to 12 months)
SANsymphony is freely downloadble after registering online and offers full platform support (complete Enterprise feature set) but is scale (4 nodes), capacity (16TB) and time (30 days) restricted, what all can be expanded upon request. The free trial version of SANsymphony can be installed on all commodity hardware platforms that meet the hardware requirements.
For more information please go here: https://www.datacore.com/try-it-now/
|
Free Download (forever)
Proof-of-Concept (PoC)
PowerFlex block storage software is available for free and for an unlimited time, without any capacity restrictions. The free version of PowerFlex is restricted to non-production usage and as such does not contain any maintenance/support.
|
Community Edition (forever)
Trial (up to 30 days)
Proof-of-Concept
There are 2 ways for end-user organizations to evaluate StarWind:
1. StarWind Free. A free version of the software-only Virtual SAN product can be downloaded for the StarWind website. The free version has full functionality but StarWind Management Console works only in the monitoring mode without the ability to create or manage StarWind. All the management is performed via PowerShell and set of script templates.
The free version is intended to be self-supported or community-supported on public discussion forums. This
2. StarWind Trial. The trial version has full functionality and all the management capabilities for StarWind Management Console and is limited to 30 days but can be prolonged if required.
|
|
|
|
Deploy |
|
|
Deployment Architecture
Details
|
Single-Layer
Dual-Layer
Single-Layer = servers function as compute nodes as well as storage nodes.
Dual-Layer = servers function only as storage nodes; compute runs on different nodes.
Single-Layer:
- SANsymphony is implemented as virtual machine (VM) or in case of Hyper-V as service layer on Hyper-V parent OS, managing internal and/or external storage devices and providing virtual disks back to the hypervisor cluster it is implemented in. DataCore calls this a hyper-converged deployment.
Dual-Layer:
- SANsymphony is implemented as bare metal nodes, managing external storage (SAN/NAS approach) and providing virtual disks to external hosts which can be either bare metal OS systems and/or hypervisors. DataCore calls this a traditional deployment.
- SANsymphony is implemented as bare metal nodes, managing internal storage devices (server-SAN approach) and providing virtual disks to external hosts which can be either bare metal OS systems and/or hypervisors. DataCore calls this a converged deployment.
Mixed:
- SANsymphony is implemented in any combination of the above 3 deployments within a single management entity (Server Group) acting as a unified storage grid. DataCore calls this a hybrid-converged deployment.
|
Single-Layer
Dual-Layer
There are two main PowerFlex components to consider:
- Storage Data Server (SDS) that is used to present storage volumes
- Storage Data Client (SDC) that is used to access storage volumes that are presented
Therefore PowerFlex can be setup in two ways:
1. By installing both the SDS and the SDC component on the same server, creating a hyper-converged single-layer configuration.
2. By installing the SDS and SDC component on separate servers, creating a traditional dual-layer configuration.
The flexibility of the PowerFlex platform allows the two configurations to be mixed within the same environment.
|
Single-Layer
Dual-Layer (secondary)
Single-Layer = servers function as compute nodes as well as storage nodes.
Dual-Layer = servers function only as storage nodes; compute runs on different nodes.
In a Single-Layer architecture, the servers with Virtual SAN software function as compute nodes as well as storage node; StarWind performs storage synchronous replication
In a Dual-Layer architecture, StarWind Virtual SAN software replicates data in active-active mode between the dedicated storage servers and provides HA storage for use by separate compute nodes that do not have StarWind Virtual SAN installed.
|
|
Deployment Method
Details
|
BYOS (some automation)
BYOS = Bring-Your-Own-Server-Hardware
DataCore SANsymphony is made easy by providing a very straightforward implementation approach.
|
BYOS (some automation)
BYOS = Bring-Your-Own-Server-Hardware
|
BYOS (some automation)
|
|
|
Workload Support
|
|
|
|
|
|
|
Virtualization |
|
|
Hypervisor Deployment
Details
|
Virtual Storage Controller
Kernel (Optional for Hyper-V)
The SANsymphony Controller is deployed as a pre-configured Virtual Machine on top of each server that acts as a part of the SANsymphony storage solution and commits its internal storage and/or externally connected storage to the shared resource pool. The Virtual Storage Controller (VSC) can be configured direct access to the physical disks, so the hypervisor is not impeding the I/O flow.
In Microsoft Hyper-V environments the SANsymphony software can also be installed in the Windows Server Root Partition. DataCore does not recommend installing SANsymphony in a Hyper-V guest VM as it introduces virtualization layer overhead and obstructs DataCore Software from directly accessing CPU, RAM and storage. This means that installing SANsymphony in the Windows Server Root Partition is the preferred deployment option. More information about the Windows Server Root Partition can be found here: https://docs.microsoft.com/en-us/windows-server/administration/performance-tuning/role/hyper-v-server/architecture
The DataCore software can be installed on Microsoft Windows Server 2019 or lower (all versions down to Microsoft Windows Server 2012/R2).
Kernel Integrated, Virtual Controller and VIB are each distributed architectures, having one active component per virtualization host that work together as a group. All three architectures are capable of delivering a complete set of storage services and good performance. Kernel Integrated solutions reside within the protected lower layer, VIBs reside just above the protected kernel layer, and Virtual Controller solutions reside in the upper user layer. This makes Virtual Controller solutions somewhat more prone to external actions (eg. most VSCs do not like snapshots). On the other hand Kernel Integrated solutions are less flexible because a new version requires the upgrade of the entire hypervisor platform. VIBs have the middle-ground, as they provide more flexibility than kernel integrated solutions and remain relatively shielded from the user level.
|
vSphere: Virtual Storage Controller + kernel module
Hyper-V/KVM: OS Drivers and packages
VMware vSphere: The PowerFlex Virtual Machine (SVM) is deployed as a pre-configured Virtual Machine on top of each server that acts as a part of the PowerFlex storage solution and commits its internal storage to the shared resource pool. The Virtual Storage Controller (VSC) has direct access to the physical disks, so the hypervisor is not impeding the I/O flow.
Hyper-V/KVM: The PowerFlex components can be installed and configured on multiple nodes from one central server via a web client by using PowerFlex Installation Manager (IM). IM has a REST API that enables install, extend, and uninstall functionalities.
Kernel Integrated, Virtual Controller and VIB are each distributed architectures, having one active component per virtualization host that work together as a group. All three architectures are capable of delivering a complete set of storage services and good performance. Kernel Integrated solutions reside within the protected lower layer, VIBs reside just above the protected kernel layer, and Virtual Controller solutions reside in the upper user layer. This makes Virtual Controller solutions somewhat more prone to external actions (eg. most VSCs do not like snapshots). On the other hand Kernel Integrated solutions are less flexible because a new version requires the upgrade of the entire hypervisor platform. VIBs have the middle-ground, as they provide more flexibility than kernel integrated solutions and remain relatively shielded from the user level.
|
Virtual Storage Controller (vSphere)
User-Space (Hyper-V)
StarWind VSAN for vSphere is deployed by downloading an OVF template of a preconfigured Linux-based VM. It should be deployed on each VMware vSphere node that takes part in StarWind replication. The OVF contains StarWind SDS stack pre-configured and pre-installed.
StarWind VSAN for Hyper-V is deployed by downloading the latest build of StarWind. The installation process will automatically install the required components. The build should be installed on each node that will take part in StarWind replication.
|
|
Hypervisor Compatibility
Details
|
VMware vSphere ESXi 5.5-7.0U1
Microsoft Hyper-V 2012R2/2016/2019
Linux KVM
Citrix Hypervisor 7.1.2/7.6/8.0 (XenServer)
'Not qualified' means there is no generic support qualification due to limited market footprint of the product. However, a customer can always individually qualify the system with a specific SANsymphony version and will get full support after passing the self-qualification process.
Only products explicitly labeled 'Not Supported' have failed qualification or have shown incompatibility.
|
VMware vSphere ESXi 6.5-6.7U3*
Microsoft Hyper-V 2012R2/2016/2019**
Citrix Hypervisor 7.1.2/7.6/8.0./8.1 (XenServer)
NEW
PowerFlex supports all major hypervisor platforms.
*At this time PowerFlex 3.5 only supports VMware vSphere 7 in dual-layer configurations (SDC core component). Support for VMware vSphere 7 (SDS core component) is planned in a 2H 2020 release of PowerFlex Manager.
**PowerFlex only supports Microsoft Windows Server + Hyper-V dual-layer configurations (SDC core component). Single-layer deployments (SDS core component) are not supported.
SDC =Storage Data Client
SDS =Storage Data Server
|
VMware vSphere ESXi 6.0-6.7
Microsoft Hyper-V 2012-2019
StarWind is actively working on supporting KVM.
|
|
Hypervisor Interconnect
Details
|
iSCSI
FC
The SANsymphony software-only solution supports both iSCSI and FC protocols to present storage to hypervisor environments.
DataCore SANsymphony supports:
- iSCSI (Switched and point-to-point)
- Fibre Channel (Switched and point-to-point)
- Fibre Channel over Ethernet (FCoE)
- Switched, where host uses Converged Network Adapter (CNA), and switch outputs Fibre Channel
|
PowerFlex
PowerFlex uses a proprietary block storage and metadata protocol over TCP/IP. It is NOT iSCSI due to PowerFlex’s distributed nature.
|
iSCSI
NFS
SMB3
iSCSI is the native StarWind VSAN storage protocol, as StarWind VSAN provides block-based storage.
NFS can be used as the storage protocol in VMware vSphere environments by leveraging the File Server role that the Windows OS provides.
SMB3 can be used as the storage protocol in Microsoft Hyper-V environments by leveraging the File Server role that the Windows OS provides.
In both VMware vSphere and Microsoft Hyper-V environments, iSCSI is used as a protocol to provide block-level storage access. It allows consolidating storage from multiple servers providing it as highly available storage to target servers. In the case of vSphere, VMware iSCSI Initiator allows connecting StarWind iSCSI devices to ESXi hosts and further create datastores on them. In the case of Hyper-V, Microsoft iSCSI Initiator is utilized to connect StarWind iSCSI devices to the servers and further provide HA storage to the cluster (i.e. CSV).
In virtualized environments In-Guest iSCSI support is still a hard requirements if one of the following scenarios is pursued:
- Microsoft Failover Clustering (MSFC) in a VMware vSphere environment
- A supported MS Exchange 2013 Environment in a VMware vSphere environment
Microsoft explicitely does not support NFS in both scenarios.
|
|
|
|
Bare Metal |
|
|
Bare Metal Compatibility
Details
|
Microsoft Windows Server 2012R2/2016/2019
Red Hat Enterprise Linux (RHEL) 6.5/6.6/7.3
SUSE Linux Enterprise Server 11.0SP3+4/12.0SP1
Ubuntu Linux 16.04 LTS
CentOS 6.5/6.6/7.3
Oracle Solaris 10.0/11.1/11.2/11.3
Any operating system currently not qualified for support can always be individually qualified with a specific SANsymphony version and will get full support after passing the self-qualification process.
SANsymphony provides virtual disks (block storage LUNs) to all of the popular host operating systems that use standard disk drives with 512 byte or 4K byte sectors. These hosts can access the SANsymphony virtual disks via SAN protocols including iSCSI, Fibre Channel (FC) and Fibre Channel over Ethernet (FCoE).
Mainframe operating systems such as IBM z/OS, z/TPF, z/VSE or z/VM are not supported.
SANsymphony itself runs on Microsoft Windows Server 2012/R2 or higher.
|
Microsoft Windows Server
Linux Distributions
IBM AIX
NEW
Supported Windows versions (SDC core component only):
Windows Server 2012R2/2016/2019
Supported Linux versions:
RHEL 6.9/6.10/7.5/7.6/7.7/7.8/8.0/8.1/8.2
CentOS 6.9/6.10/7.5/7.6/7.7/7.8/8.0/8.1/8.2
Oracle Linux 6.9/6.10/7.5/7.6/7.7
SLES 11SP4/12SP4/12SP5/15/15SP1
Ubuntu 16.04.6/18.04.2/18.04.3
Supported AIX versions (SDC core component only):
AIX 7.2 TL3
SDC =Storage Data Client
|
Microsoft Windows Server 2012/2012R2/2016/2019
StarWind VSAN provides highly available storage over iSCSI between the StarWind nodes and can additionally share storage over iSCSI to any OS that supports the iSCSI protocol.
|
|
Bare Metal Interconnect
Details
|
iSCSI
FC
FCoE
|
Block Device Driver
The PowerFlex Data Client (SDC) component is installed on application servers that are going to consume storage.
The PowerFlex Data Server (SDS) component is installed on storage servers that are used to contribute local storage to the shared resource pool.
|
iSCSI
|
|
|
|
Containers |
|
|
Container Integration Type
Details
|
Built-in (native)
DataCore provides its own Volume Plugin for natively providing Docker container support, available on Docker Hub.
DataCore also has a native CSI integration with Kubernetes, available on Github.
|
Built-in (native)
Dell EMC provides its own software plugins for container support (both Docker and Kubernetes).
|
N/A
StarWind Virtual SAN relies on the container support delivered by the hypervisor platform.
|
|
Container Platform Compatibility
Details
|
Docker CE/EE 18.03+
Docker EE = Docker Enterprise Edition
|
Docker EE 1.12+
Mesos
Docker EE = Docker Enterprise Edition
|
Docker CE 17.06.1+ for Linux on ESXi 6.0+
Docker EE/Docker for Windows 17.06+ on ESXi 6.0+
Docker CE = Docker Community Edition
Docker EE = Docker Enterprise Edition
|
|
Container Platform Interconnect
Details
|
Docker Volume plugin (certified)
The DataCore SDS Docker Volume plugin (DVP) enables Docker Containers to use storage persistently, in other words enables SANsymphony data volumes to persist beyond the lifetime of both a container or a container host. DataCore leverages SANsymphony iSCSI and FC to provide storage to containers. This effectively means that the hypervisor layer is bypassed.
The Docker SDS Volume plugin (DVP) is officially 'Docker Certified' and can be downloaded from the Docker Hub. The plugin is installed inside the Docker host, which can be either a VM or a Bare Metal Host connect to a SANsymphony storage cluster.
For more information please go to: https://hub.docker.com/plugins/datacore-sds-volume-plugin
The Kubernetes CSI plugin can be downloaded from GitHub. The plugin is automatically deployed as several pods within the Kubernetes system.
For more information please go to: https://github.com/DataCoreSoftware/csi-plugin
Both plugins are supported with SANsymphony 10 PSP7 U2 and later.
|
Docker Volume Plugin (certified)
The 'REX-Ray for PowerFlex' plugin is a block volume plugin that connects containers to persistent storage served by PowerFlex.
The 'REX-Ray for PowerFlex' plugin is officially 'Docker Certified' and can be downloaded from the online Docker Store.
|
Docker Volume Plugin (certified) + VMware VIB
vSphere Docker Volume Service (vDVS) can be used with VMware vSAN, as well as VMFS datastores and NFS datastores served by VMware vSphere-compatible storage systems.
The vSphere Docker Volume Service (vDVS) installation has two parts:
1. Installation of the vSphere Installation Bundle (VIB) on ESXi.
2. Installation of Docker plugin on the virtualized hosts (VMs) where you plan to run containers with storage needs.
The vSphere Docker Volume Service (vDVS) is officially 'Docker Certified' and can be downloaded from the online Docker Store.
The StarWind HA VMFS (datastore) can be used for deploying containers just as on common VMFS datastore.
|
|
Container Host Compatibility
Details
|
Virtualized container hosts on all supported hypervisors
Bare Metal container hosts
The DataCore native plug-ins are container-host centric and as such can be used across all SANsymphony-supported hypervisor platforms (VMware vSphere, Microsoft Hyper-V, KVM, XenServer, Oracle VM Server) as well as on bare metal platforms.
|
Virtualized container hosts on all supported hypervisors
Bare Metal container hosts
The PowerFlex native plugins are container-host centric and as such can be used across all PowerFlex-supported hypervisor platforms (VMware vSphere, Microsoft Hyper-V, Linux KVM and Citrix XenServer) as well as on bare metal platforms.
|
Virtualized container hosts on VMware vSphere hypervisor
Because the vSphere Docker Volume Service (vDVS) and vSphere Cloud Provider (VCP) are tied to the VMware Sphere platform, they cannot be used for bare metal hosts running containers.
|
|
Container Host OS Compatbility
Details
|
Linux
All Linux versions supported by Docker CE/EE 18.03+ or higher can be used.
|
CentOS
CoreOS
Debian
Red Hat
Ubuntu
'REX-Ray for PowerFlex' has been qualified for the mentioned Linux operating systems.
Container hosts running the Windows OS are not (yet) supported.
Supported CoreOS versions (SDC core component only):
CoreOS 2191.5.0
|
Linux
Windows 10 or 2016
Any Linux distribution running version 3.10+ of the Linux kernel can run Docker.
vSphere Storage for Docker can be installed on Windows Server 2016/Windows 10 VMs using the PowerShell installer.
|
|
Container Orch. Compatibility
Details
|
Kubernetes 1.13+
|
Kubernetes 1.6+
|
Kubernetes 1.6.5+ on ESXi 6.0+
|
|
Container Orch. Interconnect
Details
|
Kubernetes CSI plugin
The Kubernetes CSI plugin provides several plugins for integrating storage into Kubernetes for containers to consume.
DataCore SANsymphony provides native industry standard block protocol storage presented over either iSCSI or Fibre Channel. YAML files can be used to configure Kubernetes for use with DataCore SANsymphony.
|
PowerFlex-CSI Plugin
VxFlex OS 3.0 introduced support for the PowerFlex CSI plugin based on the Kubernetes Container Storage Interface (CSI) specification. The PowerFlex-CSI plugin is leveraged to provision and manage persistent volumes in Kubernetes v1.13 and later.
Before CSI volume plugins were “in-tree” meaning their code was part of the core Kubernetes code and shipped with the core Kubernetes binaries. Storage vendors wanting to add support for their storage system to Kubernetes (or even fix a bug in an existing volume plugin) were forced to align with the Kubernetes release process. In addition, third-party storage code caused reliability and security issues in core Kubernetes binaries and the code was often difficult (and in some cases impossible) for Kubernetes maintainers to test and maintain. CSI is 'out-of-tree' meaning that with CSI, third-party storage providers can write and deploy plugins exposing new storage systems in Kubernetes without ever having to touch the core Kubernetes code. This gives Kubernetes users more options for storage and makes the system more secure and reliable.
|
Kubernetes Volume Plugin
vSphere Cloud Provider (VCP) for Kubernetes allows Pods to use enterprise grade persistent storage. VCP supports every storage primitive exposed by Kubernetes:
- Volumes
- Persistent Volumes (PV)
- Persistent Volumes Claims (PVC)
- Storage Class
- Stateful Sets
Persistent volumes requested by stateful containerized applications can be provisioned on vSAN, VVol, VMFS or NFS datastores.
|
|
|
|
VDI |
|
|
VDI Compatibility
Details
|
VMware Horizon
Citrix XenDesktop
There is no validation check being performed by SANsymphony for VMware Horizon or Citrix XenDesktop VDI platforms. This means that all versions supported by these vendors are supported by DataCore.
|
VMware Horizon
Citrix XenDesktop
Dell EMC has not published any recent PowerFlex Reference Architecture whitepapers for both VMware Horizon and Citrix XenDesktop platforms.
|
VMware Horizon
Citrix XenDesktop
Although StarWind supports both VMware and Citrix VDI deployments on top of StarWind VSAN HA storage, there is currently no specific documentation available.
|
|
|
VMware: 110 virtual desktops/node
Citrix: 110 virtual desktops/node
DataCore has not published any recent VDI reference architecture whitepapers. The only VDI related paper that includes a Login VSI benchmark dates back to december 2010. There a 2-node SANsymphony cluster was able to sustain a load of 220 VMs based on the Login VSI 2.0.1 benchmark.
|
VMware: unknown
Citrix:unknown
Dell EMC has not published any recent PowerFlex Reference Architecture whitepapers for both VMware Horizon and Citrix XenDesktop platforms.
|
VMware: up to 260 virtual desktops/node
Citrix: up to 220 virtual desktops/node
The load bearing numbers are based on approximate calculations of VDI infrastructure that StarWind Virtual SAN can support.
There are no LoginVSA benchmark numbers on record for StarWind Virtual SAN as of yet.
|
|
|
Server Support
|
|
|
|
|
|
|
Server/Node |
|
|
Hardware Vendor Choice
Details
|
Many
SANsymphony runs on all server hardware that supports x86 - 64bit.
DataCore provides minimum requirements for hardware resources.
|
Many
PowerFlex does not have a HCL and instead provides minimum requirements for hardware resources.
|
Many
StarWind Virtual SAN is a hardware-agnostic solution and does not have a strict HCL or supported hardware platforms.
|
|
|
Many
SANsymphony runs on all server hardware that supports x86 - 64bit.
DataCore provides minimum requirements for hardware resources.
|
Many
PowerFlex does not have a HCL and instead provides minimum requirements for hardware resources.
|
Many
StarWind Virtual SAN is a hardware-agnostic solution and does not have a strict HCL or supported hardware platforms.
|
|
|
1, 2 or 4 nodes per chassis
Note: Because SANsymphony is mostly hardware agnostic, customers can opt for multiple server densities.
Note: In most cases 1U or 2U building blocks are used.
Also Super Micro offers 2U chassis that can house 4 compute nodes.
Denser nodes provide a smaller datacenter footprint where space is a concern. However, keep in mind that the footprint for other datacenter resources such as power and heat and cooling is not necessarily reduced in the same way and that the concentration of nodes can potentially pose other challenges.
|
1, 2 or 4 nodes per chassis
Because PowerFlex is mostly hardware agnostic, customers can opt for multiple server densities.
Note: In most cases 1U or 2U building blocks are used.
Also Super Micro offers 2U chassis that can house 4 compute nodes.
Denser nodes provide a smaller datacenter footprint where space is a concern. However, keep in mind that the footprint for other datacenter resources such as power and heat and cooling is not necessarily reduced in the same way and that the concentration of nodes can potentially pose other challenges.
|
1, 2 or 4 nodes per chassis
StarWind Virtual SAN is a hardware-agnostic solution and does not have strict HCL or supported hardware platforms.
|
|
|
Yes
DataCore does not explicitly recommend using different hardware platforms, but as long as the hardware specs are somehow comparable, there is no reason to insist on one or the other hardware vendor. This is proven in practice, meaning that some customers run their productive DataCore environment on comparable servers of different vendors.
|
Yes
PowerFlex allows for mixing different server hardware in a single solution; also PowerFlex allows different volume types (HDD-only, Hybrid, All-Flash) to exist within a single solution.
|
Yes
Although StarWind does not recommend assymetric configurations (mixing nodes with different CPUs, storage type and networks in same replica) for Virtual SAN environments, if a customer understands the possible issues with performance (any solution that provides storage active-active replication performs with the speed of the slowest component), different servers can be mixed in a single solution.
|
|
|
|
Components |
|
|
|
Flexible
Minimum hardware requirements need to be fulfilled.
For more information please go to: https://www.datacore.com/products/sansymphony/tech/compatibility/
|
Flexible
Dell EMC provides minimal hardware requirements in its PowerFlex documentation.
|
Flexible
Minimum hardware requirements need to be fulfilled.
For more information please go to:
https://www.starwindsoftware.com/system-requirements
|
|
|
Flexible
|
Flexible
Dell EMC provides minimal hardware requirements in its PowerFlex documentation.
|
Flexible
Minimum hardware requirements need to be fulfilled.
For more information please go to:
https://www.starwindsoftware.com/system-requirements
|
|
|
Flexible
Minimum hardware requirements need to be fulfilled.
For more information please go to: https://www.datacore.com/products/sansymphony/tech/compatibility/
|
Flexible
PowerFlex supports magnetic (HDD) and solid-state disks (SSD), as well as flash PCI Express (PCIe) cards.
|
Flexible
Minimum hardware requirements need to be fulfilled.
For more information please go to:
https://www.starwindsoftware.com/system-requirements
|
|
|
Flexible
Minimum hardware requirements need to be fulfilled.
For more information please go to: https://www.datacore.com/products/sansymphony/tech/compatibility/
|
Flexible
PowerFlex supports:
• 1,10, 25, 40, 100 gigabit networks;
• IP-over-InfiniBand networks.
The use of dual-port network interface cards is recommended.
Management and data storage traffic can be performed across the same IP network or across separate IP networks.
|
Flexible
Minimum hardware requirements need to be fulfilled.
For more information please go to:
https://www.starwindsoftware.com/system-requirements
|
|
|
NVIDIA Tesla
AMD FirePro
Intel Iris Pro
DataCore SANsymphony supports the hardware that is on the hypervisor HCL.
VMware vSphere 6.5U1 officially supports several GPUs for VMware Horizon 7 environments:
NVIDIA Tesla M6 / M10 / M60
NVIDIA Tesla P4 / P6 / P40 / P100
AMD FirePro S7100X / S7150 / S7150X2
Intel Iris Pro Graphics P580
More information on GPU support can be found in the online VMware Compatibility Guide.
Windows 2016 supports two graphics virtualization technologies available with Hyper-V to leverage GPU hardware:
- Discrete Device Assignment
- RemoteFX vGPU
More information is provided here: https://docs.microsoft.com/en-us/windows-server/remote/remote-desktop-services/rds-graphics-virtualization
The NVIDIA website contains a listing of GRID certified servers and the maximum number of GPUs supported inside a single server.
Server hardware vendor websites also contain more detailed information on the GPU brands and models supported.
|
NVIDIA Tesla
AMD FirePro
Intel Iris Pro
PowerFlex supports the hardware that is on the hypervisor HCL.
VMware vSphere 6.5U1 officially supports several GPUs for VMware Horizon 7 environments:
NVIDIA Tesla M6 / M10 / M60
NVIDIA Tesla P4 / P6 / P40 / P100
AMD FirePro S7100X / S7150 / S7150X2
Intel Iris Pro Graphics P580
More information on GPU support can be found in the online VMware Compatibility Guide.
Windows 2016 supports two graphics virtualization technologies available with Hyper-V to leverage GPU hardware:
- Discrete Device Assignment
- RemoteFX vGPU
More information is provided here: https://docs.microsoft.com/en-us/windows-server/remote/remote-desktop-services/rds-graphics-virtualization
The NVIDIA website contains a listing of GRID certified servers and the maximum number of GPUs supported inside a single server.
Server hardware vendor websites also contain more detailed information on the GPU brands and models supported.
|
NVIDIA Tesla
AMD FirePro
Intel Iris Pro
StarWind Virtual SAN supports the hardware that is on the hypervisor HCL.
|
|
|
|
Scaling |
|
|
|
CPU
Memory
Storage
GPU
The SANsymphony platform allows for expanding of all server hardware resources.
|
CPU
Memory
Storage
GPU
|
CPU
Memory
Storage
GPU
StarWind Virtual SAN allows expansion of all server hardware resources.
|
|
|
Storage+Compute
Compute-only
Storage-only
Storage+Compute: In a single-layer deployment existing SANsymphony clusters can be expanded by adding additional nodes running SANsymphony, which adds additional compute and storage resources to the shared pool. In a dual-layer deployment both the storage-only SANsymphony clusters and the compute clusters can be expanded simultaneously.
Compute-only: Because SANsymphony leverages virtual block volumes (LUNs), storage can be presented to hypervisor hosts not participating in the SANsymphony cluster. This is also beneficial to migrations, since it allows for online storage vMotions between SANsymphony and non-SANsymphony storage platforms.
Storage-only: In a dual-layer or mixed deployment both the storage-only SANsymphony clusters and the compute clusters can be expanded independent from each other.
|
Compute+storage
Compute-only (SDC)
Storage-only
PowerFlex supports both single-layer (hyper-converged) and dual-layer architectures, as well as a mix of both. This flexibility allows the platforms Storage Data Servers (SDSs) to present storage volumes to servers that have the Storage Data Client (SDC) component installed.
Storage+Compute: Existing PowerFlex clusters can be expanded by adding additional PowerFlex nodes that have both SDS and SDC components installed, which adds additional compute and storage resources to the shared pool.
Compute-only: Existing PowerFlex clusters can be expanded by adding additional PowerFlex nodes that only have the SDC component installed, which adds additional compute resources to the shared pool.
Storage-only: Existing PowerFlex clusters can be expanded by adding additional PowerFlex nodes that only have the SDS component installed, which adds additional storage resources to the shared pool.
|
Storage+Compute
Compute-only
Storage-only
In case of Storage-only scale out, StarWind Virtual SAN Storage Nodes will be based on Windows Server bare metal with no hypervisor software installed eg. VMware ESXi. StarWind software will be running as a Windows-native application and providing storage to the hypervisor hosts.
|
|
|
1-64 nodes in 1-node increments
There is a maximum of 64 nodes within a single cluster. Multiple clusters can be managed through a single SANsymphony management instance.
|
3-512 storage nodes in 1-node increments
PowerFlex can be used in two ways:
1. By installing the SDS and SDC components on the same server, creating a hyper-converged single-layer configuration.
2. By installing the SDS and SDC components on separate servers, creating a traditional 2-layer configuration.
These configurations can be mixed within the same environment.
The maximum number of SDSs per system is 512 at this time.
|
2-64 nodes in 1-node increments
The 64-node limit applies to both VMware vSphere and Microsoft Hyper-V environments.
|
|
Small-scale (ROBO)
Details
|
2 Node minimum
DataCore prevents split-brain scenarios by always having an active-active configuration of SANsymphony with a primary and an alternate path.
In the case SANsymphony servers are fully operating but do not see each other, the application host will still be able to read and write data via the primary path (no switch to secondary). The mirroring is interrupted because of the lost connection and the administrator is informed accordingly. All writes are stored on the locally available storage (primary path) and all changes are tracked. As soon as the connection between the SANsymphony servers is restored, the mirror will recover automatically based on these tracked changes.
Dual updates due to misconfiguration are detected automatically and data corruption is prevented by freezing the vDisk and waiting for user input to solve the conflict. Conflict solutions could be to declare one side of the mirror to be the new active data set and discarding all tracked changes at the other side, or splitting the mirror and merge the two data sets into a 3rd vDisk manually.
|
3 Node minimum
PowerFlexs smallest deployment contains 3 nodes.
|
1 Node minimum
StarWind Virtual SAN can be used as standalone iSCSI target providing storage from one node (no HA). For HA two nodes are required. StarWind Virtual SAN can be further scaled by adding more storage or new nodes into the storage cluster.
|
|
|
Storage Support
|
|
|
|
|
|
|
General |
|
|
|
Block Storage Pool
SANsymphony only serves block devices to the supported OS platforms.
|
Block Pool
PowerFlex only serves block devices as storage volumes to the supported OS platforms.
The internal PowerFlex Metadata Manager (MDM) holds cluster-wide component mapping.
|
Block Storage Pool
StarWind VSAN only serves block devices as storage volumes to the supported OS platforms.
The underlying storage is first aggregated with hardware or software RAID. Then, the storage is replicated by StarWind at the block level across 2 or 3 nodes and further provided as a single pool (single StarWind virtual device) or as multiple pools (multiple StarWind devices).
|
|
|
Partial
DataCores core approach is to provide storage resources to the applications without having to worry about data locality. But if data locality is explicitly requested, the solution can partially be designed that way by configuring the first instance of all data to be stored on locally available storage (primary path) and the mirrored instance to be stored on the alternate path (secondary path). Furthermore every hypervisor host can have a local preferred path, indicated by the ALUA path preference.
By default data does not automatically follow the VM when the VM is moved to another node. However, virtual disks can be relocated on the fly to other DataCore node without losing I/O access, but this relocation takes some time due to data copy operations required. This kind of relocation usually is done manually, but we allow automation of such tasks and can integrate with VM orchestration using PowerShell for example.
Whether data locality is a good or a bad thing has turned into a philosophical debate. Its true that data locality can prevent a lot of network traffic between nodes, because the data is physically located at the same node where the VM resides. However, in dynamic environments where VMs move to different hosts on a frequent basis, data locality in most cases requires a lot of data to be copied between nodes in order to maintain the physical VM-data relationship. The SDS/HCI vendors today that choose not to use data locality, advocate that the additional network latency is negligible.
|
None
Whether data locality is a good or a bad thing has turned into a philosophical debate. Its true that data locality can prevent a lot of network traffic between nodes, because the data is physically located at the same node where the VM resides. However, in dynamic environments where VMs move to different hosts on a frequent basis, data locality in most cases requires a lot of data to be copied between nodes in order to maintain the physical VM-data relationship. The SDS/HCI vendors today that choose not to use data locality, advocate that the additional network latency is negligible.
|
Partial
StarWinds core approach is to keep data close (local) to the VM in order to avoid slow data transfers through the network and achieve the highest performance the setup can provide. The solution is designed to store the first instance of all data on locally available storage (primary path) and the mirrored instance to be stored on the alternate path (secondary path). Furthermore every hypervisor host can have a local preferred path, indicated by the ALUA path preference. Data does not automatically follow the VM when the VM is moved to another node.
Whether data locality is a good or a bad thing has turned into a philosophical debate. Its true that data locality can prevent a lot of network traffic between nodes, because the data is physically located at the same node where the VM resides. However, in dynamic environments where VMs move to different hosts on a frequent basis, data locality in most cases requires a lot of data to be copied between nodes in order to maintain the physical VM-data relationship. The SDS/HCI vendors today that choose not to use data locality, advocate that the additional network latency is negligible.
|
|
|
Direct-attached (Raw)
Direct-attached (VoV)
SAN or NAS
VoV = Volume-on-Volume; The Virtual Storage Controller uses virtual disks provided by the hypervisor platform.
|
Direct-attached (Raw, RAID, File)
|
Direct-attached (Raw)
SAN or NAS
Direct-attached: StarWind can take control of formatted disks (NTFS). Also, StarWind software can present RAW unformatted disks over SCSI Pass Through Interface that enables remote initiator clients to use any type of a hard drive (PATA/SATA/RAID).
External SAN/NAS Storage: SAN/NAS can be connected over Ethernet connectivity and can be used for StarWind Virtual SAN as soon as they are provided as block storage (iSCSI). In case it is required to replicate data between NAS systems, there should be 2 NAS systems connected to nodes that will be used for StarWind Virtual SAN.
|
|
|
Magnetic-only
All-Flash
3D XPoint
Hybrid (3D Xpoint and/or Flash and/or Magnetic)
NEW
|
Magnetic-Only
Hybrid
All-Flash
|
Magnetic-only
Hybrid (Flash+Magnetic)
All-Flash
|
|
Hypervisor OS Layer
Details
|
SD, USB, DOM, SSD/HDD
|
SD, USB or DOM
VxFlex 3.0 and up cannot be deployed on the 32 GB SATADOM boot device that was sold in the
first generation of Dell EMC ScaleIO/VxRack Nodes and VxRack Flex (a hardware solution
based on Quanta servers).
DOM = Disk On Module
|
SD, USB, DOM, SSD/HDD
|
|
|
|
Memory |
|
|
|
DRAM
|
DRAM
|
DRAM
DRAM can be used for caching in a write-back or write-through mode. Additionally, it can be used for creating StarWind RAM disks.
For further information, please visit:
https://www.starwindsoftware.com/high-performance-ram-disk-emulator
|
|
|
Read/Write Cache
DataCore SANsymphony accelerates reads and writes by leveraging the powerful processors and large DRAM memory inside current generation x86-64bit servers on which it runs. Up to 8 Terabytes of cache memory may be configured on each DataCore node, enabling it to perform at solid state disk speeds without the expense. SANsymphony uses a common cache pool to store reads and writes in.
SANsymphony read caching essentially recognizes I/O patterns to anticipate which blocks to read next into RAM from the physical back-end disks. That way the next request can be served from memory.
When hosts write to a virtual disk, the data first goes into DRAM memory and is later destaged to disk, often grouped with other writes to minimize delays when storing the data to the persistent disk layer. Written data stays in cache for re-reads.
The cache is cleaned on a first-in-first-out (FiFo) basis. Segment overwrites are performed on the oldest data first for both read- and write cache segment requests.
SANsymphony prevents the write cache data from flooding the entire cache. In case the write data amount runs above a certain percentage watermark of the entire cache amount, then the write cache will temporarily be switched to write-through mode in order to regain balance. This is performed fully automatically and is self-adjusting, per virtual disk as well as on a global level.
|
Read Cache
PowerFlex calls the use of DRAM as caching layer Read RAM Cache (rmcache). This is a tunable feature: rmcache can be enabled or disabled per volume.
Writes are only buffered in the host memory for Read after Write caching. One way to achieve Write buffering is to use Raid controllers (e.g. LSI, PMC etc.) that have battery backup for write buffering.
|
Read/Write Cache
StarWind Virtual SAN accelerates reads and writes by leveraging conventional RAM.
The memory cache is filled up with data mainly during write operations. During read operations, data enters the cache only if the latter contains either empty memory blocks or the lines that were allocated for these entries earlier and have not been fully exhausted yet.
StarWind VSAN supports two Memory (L1 Cache) Policies:
1. Write-Back, caches writes in DRAM only and acknowledges back to the originator when complete in DRAM.
2. Write-Through, caches writes in both DRAM and underlying storage, and acknowledges back to the originator when complete in the underlying storage.
This means that exclusively caching writes in convential memory is optional. When the Write-Through policy is used, DRAM is used primarily for caching reads.
To change the cache size, first, the StarWind service should be stopped and change cache and then start the service and then repeat the same process on the partner node. This allows keeping VMs up and running during cache changes.
In the majority of use cases, there is no need to assign L1 cache for all-flash storage arrays.
Note: In case of using the Write-Back policy for DRAM, UPS units have to be installed to ensure the correct shutdown of StarWind Virtual SAN nodes. If a power outage occurs, this will prevent the loss of cached data. The UPS capacity must cover the time required for flushing the cached data to the underlying storage.
|
|
|
Up to 8 TB
The actual size that can be configured depends on the server hardware that is used.
|
Configurable
The read cache size can be set anywhere between 128MB and 300GB per SDS server.
|
Configurable
The size of L1 cache should be equal to the amount of the average working data set.
There are no default or maximum values for RAM cache as such. The maximum size that can be assigned for StarWind RAM cache is limited by the available RAM to the system; also, you need to make sure that other applications running will have enough of RAM for their operations.
Additionally, the total amount of L1 cache assigned influences the time required for system shutdown so overprovisioning of the L1 cache amount can cause StarWind service interruption and the loss of cached data. The minimum size assigned for RAM cache in either write-back or write-through mode is 1MB. However, StarWind recommends assigning a StarWind RAM cache size that matches the size of the working data set.
|
|
|
|
Flash |
|
|
|
SSD, PCIe, UltraDIMM, NVMe
|
SSD, PCIe, UltraDIMM, NVMe
|
SSD, NVMe
|
|
|
Persistent Storage
SANsymphony supports new TRIM / UNMAP capabilities for solid-state drives (SSD) in order to reduce wear on those devices and optimize performance.
|
Read Cache
Write Buffer (Flash Storage Pools)
Storage Tier
PowerFlex calls the use of Flash as caching layer Read Flash Cache (rfcache). This is a tunable feature: rfcache can be enabled or disabled. Rfcache is used to increase read performance and buffers writes to increase the performance of Read-after-Write I/Os.
|
Read/Write Cache (hybrid)
Persistent storage (all-flash)
StarWind Virtual SAN supports a single Flash (L2 Cache) Policy:
1. Write-Through, caches writes in both Flash (SSD/NVMe) and the underlying storage, and acknowledges back to the originator when complete in the underlying storage.
With the write-through policy, new blocks are written both to cache layer and the underlying storage synchronously. However, in this mode, only the blocks that are read most frequently are kept in the cache layer, accelerating read operations. This is the only mode available for StarWind L2 cache.
In the case of Write-Through cached data does not need to be offloaded to the backing store when a device is removed or the service is stopped.
In the majority of use cases, if L1 cache is already assigned, there is no need to configure L2 cache.
A StarWind Virtual SAN solution with L2 cache configured should have more RAM available, since L2 cache needs 6.5 GB of RAM per 1 TB of L2 cache to store metadata.
|
|
|
No limit, up to 1 PB per device
The definition of a device here is a raw flash device that is presented to SANsymphony as either a SCSI LUN or a SCSI disk.
|
0 - 64 devices per node
Flash devices are not mandatory in a PowerFlex solution.
PowerFlex supports a maximum of 64 devices (disks) per SDS server. This includes both HDD and Flash devices. In VMware vSphere environments the maximum is 59 devices (disks) per SDS server.
|
No limitations
The definition of a device here is a raw flash device that is presented to Virtual SAN as either a SCSI LUN or a SCSI disk.
|
|
|
|
Magnetic |
|
|
|
SAS or SATA
SAS = 10k or 15k RPM = Medium-capacity medium-speed drives
SATA = NL-SAS = 7.2k RPM = High-capacity low-speed drives
In this case SATA = NL-SAS = MDL SAS
|
SAS or SATA
SAS = 10k or 15k RPM = Medium-capacity medium-speed drives
SATA = NL-SAS = 7.2k RPM = High-capacity low-speed drives
|
SAS or SATA
|
|
|
Persistent Storage
|
Write Buffer (HDD Storage Pools)
Storage Tier
|
Persistent Storage
|
|
Magnetic Capacity
Details
|
No limit, up to 1 PB (per device)
The definition of a device here is a raw flash device that is presented to SANsymphony as either a SCSI LUN or a SCSI disk.
|
0 - 64 devices per node
PowerFlex supports a maximum of 64 devices (disks) per SDS server. This includes both HDD and Flash devices. In VMware vSphere environments the maximum is 59 devices (disks) per SDS server.
PowerFlex 2.6 supports devices with a capacity up to 8TB.
|
No limitations
|
|
|
Data Availability
|
|
|
|
|
|
|
Reads/Writes |
|
|
Persistent Write Buffer
Details
|
DRAM (mirrored)
If caching is turned on (default=on), any write will only be acknowledged back to the host after it has been succesfully stored in DRAM memory of two separate physical SANsymphony nodes. Based on de-staging algorithms each of the nodes eventually copies the written data that is kept in DRAM to the persistent disk layer. Because DRAM outperforms both flash and spinning disks the applications experience much faster write behavior.
Per default, the limit of dirty-write-data allowed per Virtual Disk is 128MB. This limit could be adjusted, but there has never been a reason to do so in the real world. Individual Virtual Disks can be configured to act in write-through mode, which means that the dirty-write-data limit is set to 0MB so effectively the data is directly written to the persistent disk layer.
DataCore recommends that all servers running SANsymphony software are UPS protected to avoid data loss through unplanned power outages. Whenever a power loss is detected, the UPS automatically signals this to the SANsymphony node and write behavior is switched from write-back to write-through mode for all Virtual Disks. As soon as the UPS signals that power has been restored, the write behavior is switched to write-back again.
|
Flash/HDD
The persistent write buffer depends on the type of the storage pool (Flash or HDD).
|
DRAM (mirrored)
Flash Layer (SSD, NVMe)
|
|
Disk Failure Protection
Details
|
2-way and 3-way Mirroring (RAID-1) + opt. Hardware RAID
DataCore SANsymphony software primarily uses mirroring techniques (RAID-1) to protect data within the cluster. This effectively means the SANsymphony storage platform can withstand a failure of any two disks or any two nodes within the storage cluster. Optionally, hardware RAID can be implemented to enhance the robustness of individual nodes.
SANsymphony supports Dynamic Data Resilience. Data redundancy (none, 2-way or 3-way) can be added or removed on-the-fly at the vdisk level.
A 2-way mirror acts as active-active, where both copies are accessible to the host and written to. Updating of the mirror is synchronous and bi-directional.
A 3-way mirror acts as active-active-backup, where the active copies are accessible to the host and written to, and the backup copy is inaccessible to the host (paths not presented) and written to. Updating of the mirrors active copies is synchronous and bi-directional. Updating of the mirrors backup copy is synchronous and unidirectional (receive only).
In a 3-way mirror the backup copy should be independent of existing storage resources that are used for the active copies. Because of the synchronous updating all mirror copies should be equal in storage performance.
When in a 3-way mirror an active copy fails, the backup copy is promoted to active state. When the failed mirror copy is repaired, it automatically assumes a backup state. Roles can be changed manually on-the-fly by the end-user.
DataCore SANsymphony 10.0 PSP9 U1 introduced System Managed Mirroring (SMM). A multi-copy virtual disk is created from a storage source (disk pool or pass-through disk) from two or three DataCore Servers in the same server group. Data is synchronously mirrored between the servers to maintain redundancy and high availability of the data. System Managed Mirroring (SMM) addresses the complexity of managing multiple mirror paths for numerous virtual disks. This feature also addresses the 256 LUN limitation by allowing thousands of LUNs to be handled per network adapter. The software transports data in a round robin mode through available mirror ports to maximize throughput and can dynamically reroute mirror traffic in the event of lost ports or lost connections. Mirror paths are automatically and silently managed by the software.
The System Managed Mirroring (SMM) feature is disabled by default. This feature may be enabled or disabled for the server group.
With SANsymphony 10.0 PSP10 adds seamless transition when converting Mirrored Virtual Disks (MVD) to System Managed Mirroring (SMM). Seamless transition converts and replaces mirror paths on virtual disks in a manner in which there are no momentary breaks in mirror paths.
|
1 Replica (2N)
+ opt. Hardware RAID
PowerFlex uses replicas to protect data within the cluster. In addition, hardware RAID can be implemented to enhance the robustness of individual nodes.
Replicas: Before any write is acknowledged to the host, it is synchronously replicated on an adjacent node. All nodes in the cluster participate in replication. This means that with 2N one instance of the data that is written is stored on one node and another instance of that data is stored on a different node in the cluster. This happens in a fully distributed manner, in other words, there is no dedicated partner node. When a disk fails, it is marked offline and data is read from another instance instead. At the same time data rebuilds of the associated replicas is initiated in order to restore the desired protection level (2N). All nodes participate in the rebuild task. When the failed disk is replaced/revived, the disk is repurposed to resume its original role.
|
1-2 Replicas (2N-3N)
+ Hardware RAID (1, 5, 10)
StarWind Virtual SAN replicates the storage to protect data within the cluster. In addition, hardware or software RAID is implemented to enhance the robustness of individual nodes.
Replicas+Hardware RAID: Before any write is acknowledged to the host, it is synchronously replicated to one or two designated partner nodes. This means that with 2N one instance of data that is written is stored on the local node and another instance of that data is stored on the designated partner node in the cluster. With 3N one instance of data that is written is stored on the local node, and two other instances of that data are stored on designated partner nodes in the cluster. When a physical disk fails, hardware RAID maintains data availability.
|
|
Node Failure Protection
Details
|
2-way and 3-way Mirroring (RAID-1)
DataCore SANsymphony software primarily uses mirroring techniques (RAID-1) to protect data within the cluster. This effectively means the SANsymphony storage platform can withstand a failure of any two disks or any two nodes within the storage cluster. Optionally, hardware RAID can be implemented to enhance the robustness of individual nodes.
SANsymphony supports Dynamic Data Resilience. Data redundancy (none, 2-way or 3-way) can be added or removed on-the-fly at the vdisk level.
A 2-way mirror acts as active-active, where both copies are accessible to the host and written to. Updating of the mirror is synchronous and bi-directional.
A 3-way mirror acts as active-active-backup, where the active copies are accessible to the host and written to, and the backup copy is inaccessible to the host (paths not presented) and written to. Updating of the mirrors active copies is synchronous and bi-directional. Updating of the mirrors backup copy is synchronous and unidirectional (receive only).
In a 3-way mirror the backup copy should be independent of existing storage resources that are used for the active copies. Because of the synchronous updating all mirror copies should be equal in storage performance.
When in a 3-way mirror an active copy fails, the backup copy is promoted to active state. When the failed mirror copy is repaired, it automatically assumes a backup state. Roles can be changed manually on-the-fly by the end-user.
DataCore SANsymphony 10.0 PSP9 U1 introduced System Managed Mirroring (SMM). A multi-copy virtual disk is created from a storage source (disk pool or pass-through disk) from two or three DataCore Servers in the same server group. Data is synchronously mirrored between the servers to maintain redundancy and high availability of the data. System Managed Mirroring (SMM) addresses the complexity of managing multiple mirror paths for numerous virtual disks. This feature also addresses the 256 LUN limitation by allowing thousands of LUNs to be handled per network adapter. The software transports data in a round robin mode through available mirror ports to maximize throughput and can dynamically reroute mirror traffic in the event of lost ports or lost connections. Mirror paths are automatically and silently managed by the software.
The System Managed Mirroring (SMM) feature is disabled by default. This feature may be enabled or disabled for the server group.
With SANsymphony 10.0 PSP10 adds seamless transition when converting Mirrored Virtual Disks (MVD) to System Managed Mirroring (SMM). Seamless transition converts and replaces mirror paths on virtual disks in a manner in which there are no momentary breaks in mirror paths.
|
1 Replica (2N)
+ opt. Hardware RAID
PowerFlex uses replicas to protect data within the cluster. In addition, hardware RAID can be implemented to enhance the robustness of individual nodes.
Replicas: Before any write is acknowledged to the host, it is synchronously replicated on an adjacent node. All nodes in the cluster participate in replication. This means that with 2N one instance of the data that is written is stored on one node and another instance of that data is stored on a different node in the cluster. This happens in a fully distributed manner, in other words, there is no dedicated partner node. When a disk fails, it is marked offline and data is read from another instance instead. At the same time data rebuilds of the associated replicas is initiated in order to restore the desired protection level (2N). All nodes participate in the rebuild task. When the failed disk is replaced/revived, the disk is repurposed to resume its original role.
|
1-2 Replicas (2N-3N)
Replicas: Before any write is acknowledged to the host, it is synchronously replicated to one or two designated partner nodes. This means that with 2N one instance of data that is written is stored on the local node and another instance of that data is stored on the designated partner node in the cluster. With 3N one instance of data that is written is stored on the local node, and two other instances of that data are stored on designated partner nodes in the cluster.
|
|
Block Failure Protection
Details
|
Not relevant (usually 1-node appliances)
Manual configuration (optional)
Manual designation per Virtual Disk is required to accomplish this. The end-user is able to define which node is paired to which node for that particular Virtual Disk. However, block failure protection is in most cases irrelevant as 1-node appliances are used as building blocks.
SANsymphony works on an N+1 redundancy design allowing any node to acquire any other node as a redundancy peer per virtual device. Peers are replacable/interchangable on a per Virtual Disk level.
|
Fault Sets
Block failure protection can be achieved by assigning nodes in the same appliance to different Fault Sets.
Fault Sets: When using Fault Sets, one instance of the data is kept within the local Fault Set and another instance of the data is kept within another Fault Set. By applying Fault Sets, rack failure protection can be achieved as well.
|
Not relevant (usually 1-node appliances)
In a 3-node cluster, StarWind can provide 3-way mirroring, allowing the cluster to withstand a failure of two nodes without loosing data accessibility. In a cluster of more than 3 nodes, a grid architecture can be configured to withstand the failure of two nodes.
|
|
Rack Failure Protection
Details
|
Manual configuration
Manual designation per Virtual Disk is required to accomplish this. The end-user is able to define which node is paired to which node for that particular Virtual Disk.
|
Fault Sets
Block failure protection can be achieved by assigning nodes in the same appliance to different Fault Sets.
Fault Sets: When using Fault Sets, one instance of the data is kept within the local Fault Set and another instance of the data is kept within another Fault Set. By applying Fault Sets, rack failure protection can be achieved as well.
|
N/A
|
|
Protection Capacity Overhead
Details
|
Mirroring (2N) (primary): 100%
Mirroring (3N) (primary): 200%
+ Hardware RAID5/6 overhead (optional)
|
Replica (2N): 100% + Hardware RAID overhead (optional)
|
Replica (2N) + RAID1/10: 200%
Replica (2N) + RAID5: 125-133%
Replica (3N) + RAID1/10: 300%
Replica (3N) + RAID5: 225-233%
|
|
Data Corruption Detection
Details
|
N/A (hardware dependent)
SANsymphony fully relies on the hardware layer to protect data integrity. This means that the SANsymphony software itself does not perform Read integrity checks and/or Disk scrubbing to verify and maintain data integrity.
|
Disk scrubbing (software)
In-flight integrity checks
Persistent integrity checks
NEW
The Background Device Scanner constantly searches for, and fixes, device errors before they can affect the system, thus providing additional data reliability. The scanner runs in the background, not interrupting other Storage Pool activities (such as adding and removing volumes). When scanning is enabled for a Storage Pool, the scanner seeks out corrupted sectors in the devices in that pool. The scanner also provides SNMP reporting about errors found.
In-flight checksum protection is provided for data reads and writes. This feature addresses errors that change the payload during the transit through the PowerFlex system. PowerFlex protects data in-flight by calculating and validating the checksum value for the payload at both ends. The checksum protection mode can be applied per Storage Pool.
Persistent checksum protection is provided for all data that is stored in Fine Granularity (FG) storage pools by default. This cannot be changed. Fine Granularity (FG) layout saves checksum data before and after processing to guarantee data integrity (compressed or not). There are also system checksums for metadata.
PowerFlex 3.5 adds enhancements to:
- Fine Granularity (FG) data layout: metadata cache for higher FG performance,
- Medium Granularity (MG) data layout: checksum,
- Sub-device error handling for improved resiliency.
|
N/A (hardware dependent)
StarWind Virtual SAN fully relies on the hardware layer to protect data integrity. This means that the StarWind software itself does not perform Read integrity checks and/or Disk scrubbing to verify and maintain data integrity.
|
|
|
|
Points-in-Time |
|
|
|
Built-in (native)
|
Built-in (native)
NEW
PowerFlex has native snapshot capabilities. These snapshot capabilities include the support of Consistency Groups. This assures that snapshots of different volumes within the same group that are taken together at exactly the same time.
VxFlex OS 3.0 introduced FIne Granularity (FG) storage pools. Fine Granularity (FG) snapshots are Redirect on Write (RoW) instead of Copy on Write (CoW) that is used in Medium Granularity (MG) storage pools.
PowerFlex 3.5 introduces Secure Snapshots. These storage snapshots cannot be deleted and thus enable compliance with the financial and healthcare industry.
|
N/A
StarWind Virtual SAN does not have native snapshot capabilities. Hypervisor-native snapshot capabilities can be leveraged instead.
|
|
|
Local + Remote
SANsymphony snapshots are always created on one side only. However, SANsymphony allows you to create a snapshot for the data on each side by configuring two snapshot schedules, one for the local volume and one for the remote volume. Both snapshot entities are independent and can be deleted independently allowing different retention times if needed.
There is also the capability to pair the snapshot feature along with asynchronous replication which provides you with the ability to have a third site long distance remote copy in place with its own retention time.
|
Local
|
N/A
StarWind Virtual SAN does not have native snapshot capabilities. Hypervisor-native snapshot capabilities can be leveraged instead.
|
|
Snapshot Frequency
Details
|
1 Minute
The snapshot lifecycle can be automatically configured using the integrated Automation Scheduler.
|
1 minute
VxFlex OS 3.0 introduced snapshot policy management.
Snapshots can be created every x minutes/hours/days.
Snapshot retention is based on the number of existing snapshots (retain last x snapshots).
The PowerFlex multi-level snapshot structure can have consist of up to six retention levels, where the 1st level having the most frequent snapshots.
Every Volume Tree (V-Tree) can have up to 127 snapshots next to the root volume. Policy-managed snapshots can create up to 60 snapshots per root volume.
|
N/A
StarWind Virtual SAN does not have native snapshot capabilities. Hypervisor-native snapshot capabilities can be leveraged instead.
|
|
Snapshot Granularity
Details
|
Per VM (Vvols) or Volume
With SANsymphony the rough hierarchy is: physical disk(s) or LUNs -> Disk Pool -> Virtual Disk (=logical volume).
Although DataCore SANsymphony uses block-storage, the platform is capable of attaining per VM-granularity if desired.
In Microsoft Hyper-V environments, when a VM with vdisks is created through SCVMM, DataCore can be instructed to automatically carve out a Virtual Disk (=storage volume) for every individual vdisk. This way there is a 1-to-1 alignment from end-to-end and snapshots can be created on the VM-level. The per-VM functionality is realized by installing the DataCore Storage Management Provider in SCVMM.
Because of the per-host storage limitations in VMware vSphere environments, VVols is leveraged to provide per VM-granularity. DataCore SANsymphony Provider v2.01 is certified for VMware ESXi 6.5 U2/U3, ESXi 6.7 GA/U1/U2/U3 and ESXi 7.0 GA/U1.
|
Per VM (Vvols) or Volume
NEW
Although Dell EMC PowerFlex uses block-storage, the platform is capable of attaining per VM-granularity by leveraging VMware Virtual Volumes (Vvols).
VxFlex OS 3.0.1 introduced VVols certification for VMware ESXi 6.0-7.0U1 through the Dell Storage VASA 2.0 Provider.
|
N/A
StarWind Virtual SAN does not have native snapshot capabilities. Hypervisor-native snapshot capabilities can be leveraged instead.
|
|
|
Built-in (native)
DataCore SANsymphony incorporates Continuous Data Protection (CDP) and leverages this as an advanced backup mechanism. As the term implies, CDP continuously logs and timestamps I/Os to designated virtual disks, allowing end-users to restore the environment to an arbitrary point-in-time within that log.
Similar to snapshot requests, one can generate a CDP Rollback Marker by scripting a call to a PowerShell cmdlet when an application has been quiesced and the caches have been flushed to storage. Several of these markers may be present throughout the 14-day rolling log. When rolling back a virtual disk image, one simply selects an application-consistent or crash-consistent restore point from just before the incident occurred.
|
External
PowerFlex does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
External
StarWind Virtual SAN does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
|
|
Local or Remote
All available storage within the SANsymphony group can be configured as targets for back-up jobs.
|
N/A
PowerFlex does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
N/A
StarWind Virtual SAN does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
|
|
Continuously
As Continuous Data Protection (CDP) is being leveraged, I/Os are logged and timestamped in a continous fashion, so end-users can restore to virtually any-point-in-time.
|
N/A
PowerFlex does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
N/A
StarWind Virtual SAN does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
|
Backup Consistency
Details
|
Crash Consistent
File System Consistent (Windows)
Application Consistent (MS Apps on Windows)
By default CDP creates crash consistent restore points. Similar to snapshot requests, one can generate a CDP Rollback Marker by scripting a call to a PowerShell cmdlet when an application has been quiesced and the caches have been flushed to storage.
Several CDP Rollback Markers may be present throughout the 14-day rolling log. When rolling back a virtual disk image, one simply selects an application-consistent, filesystem-consistent or crash-consistent restore point from (just) before the incident occurred.
In a VMware vSphere environment, the DataCore VMware vCenter plug-in can be used to create snapshot schedules for datastores and select the VMs that you want to enable VSS filesystem/application consistency for.
|
N/A
PowerFlex does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
N/A
StarWind Virtual SAN does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
|
Restore Granularity
Details
|
Entire VM or Volume
With SANsymphony the rough hierarchy is: physical disk(s) or LUNs -> Disk Pool -> Virtual Disk (=logical volume).
Although DataCore SANsymphony uses block-storage, the platform is capable of attaining per VM-granularity if desired.
In Microsoft Hyper-V environments, when a VM with vdisks is created through SCVMM, DataCore can be instructed to automatically carve out a Virtual Disk (=storage volume) for every individual vdisk. This way there is a 1-to-1 alignment from end-to-end and snapshots can be created on the VM-level. The per-VM functionality is realized by installing the DataCore Storage Management Provider in SCVMM.
Because of the per-host storage limitations in VMware vSphere environments, VVols is leveraged to provide per VM-granularity. DataCore SANsymphony Provider v2.01 is VMware certified for ESXi 6.5 U2/U3, ESXi 6.7 GA/U1/U2/U3 and ESXi 7.0 GA/U1.
When configuring the virtual environment as described above, effectively VM-restores are possible.
For file-level restores a Virtual Disk snapshot needs to be mounted so the file can be read from the mount. Many simultaneous rollback points for the same Virtual Disk can coexist at the same time, allowing end-users to compare data states. Mounting and changing rollback points does not alter the original Virtual Disk.
|
N/A
PowerFlex does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
N/A
StarWind Virtual SAN does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
|
Restore Ease-of-use
Details
|
Entire VM or Volume: GUI
Single File: Multi-step
Restoring VMs or single files from volume-based storage snapshots requires a multi-step approach.
For file-level restores a Virtual Disk snapshot needs to be mounted so the file can be read from the mount. Many simultaneous rollback points for the same Virtual Disk can coexist at the same time, allowing end-users to compare data states. Mounting and changing rollback points does not alter the original Virtual Disk.
|
N/A
PowerFlex does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
N/A
StarWind Virtual SAN does not provide any backup/restore capabilities of its own. Therefore it relies on existing 3rd party data protection solutions.
VMwares free-of-charge backup software that comes with any vSphere license, VMware vSphere Data Protection (VDP), has been declared End-of-Availability and is not supported for VMware vSphere 6.7 and up.
|
|
|
|
Disaster Recovery |
|
|
Remote Replication Type
Details
|
Built-in (native)
DataCore SANsymphonys remote replication function, Asynchronous Replication, is called upon when secondary copies will be housed beyond the reach of Synchronous Mirroring, as in distant Disaster Recovery (DR) sites. It relies on a basic IP connection between locations and works in both directions. That is, each site can act as the disaster recovery facility for the other. The software operates near-synchronously, meaning that it does not hold up the application waiting on confirmation from the remote end that the update has been stored remotely.
|
Built-in (native)
NEW
PowerFlex 3.5 introduces native asynchronous replication for dual layer configurations where the PowerFlex hosts serve a storage-only role. Native asynchronous replication is not supported on single-layer (compute+storage) deployments.
|
Built-in (native; stretched clusters only)
External
StarWind Virtual SAN does not have any remote replication capabilities of its own. Stretched Clustering with synchronous replication is the exception.
Therefore in non-stretched setups it relies on external remote replication mechanisms like the ones natively available in the hypervisor platform (VMware vSphere Replication, Microsoft Hyper-V Replica) or any 3rd party remote replication application (eg. Zerto VR, Veeam VM replica, Azure Site Recovery).
vSphere Replication requires the deployment of virtual appliances. No specific integration exists between StarWind VSAN and VMware vSphere VR.
|
|
Remote Replication Scope
Details
|
To remote sites
To MS Azure Cloud
On-premises deployments of DataCore SANsymphony can use Microsoft Azure cloud as an added replication location to safeguard highly available systems. For example, on-premises stretched clusters can replicate a third copy of the data to MS Azure to protect against data loss in the event of a major regional disaster. Critical data is continuously replicated asynchronously within the hybrid cloud configuration.
To allow quick and easy deployment a ready-to-go DataCore Cloud Replication instance can be acquired through the Azure Marketplace.
MS Azure can serve only as a data repository. This means that VMs cannot be restored and run in an Azure environment in case of a disaster recovery scenario.
|
To remote sites
NEW
Replication occurs between two PowerFlex systems, connected by WAN, and designated as peer systems. Storage Data Replicators (SDR) are responsible for processing all I/Os of replication volumes. The SDR includes a journal.
|
VR: To remote sites, To VMware clouds
HR: To remote sites, to Microsoft Azure (not part of Windows Server 2019)
VMware vSphere Replication (VR): VMware vSphere Replication allows for replication of VMs to a different vSphere cluster on a remote site or to any supported VMware Cloud Service Provider (vSphere Replication to Cloud). This includes VMware on AWS and VMware on IBM Cloud.
Hyper-V Replica (HR): Hyper-V Replica is an integral part of the Hyper-V role. This feature enables block-level log-based replication of an active source VM to a passive destination VM located on another Hyper-V server or to Microsoft Azure (requires Azure Site Recovery, which is a paid external service, i.e. not part of Windows Server 2019).
Because Hyper-V Replica operates on the hypervisor layer, it is storage agnostic. This means that on one site you can have Hyper-V 2019 on SAN, whereas on the other site you can have Hyper-V 2019 on S2D.
|
|
Remote Replication Cloud Function
Details
|
Data repository
All public clouds can only serve as data repository when hosting a DataCore instance. This means that VMs cannot be restored and run in the public cloud environment in case of a disaster recovery scenario.
In the Microsoft Azure Marketplace there is a pre-installed DataCore instance (BYOL) available named DataCore Cloude Replication.
BYOL = Bring Your Own License
|
|
VR: DR-site (VMware Clouds)
HR: DR-site (Azure)
VMware vSphere Replication (VR): Because VMware on AWS and VMware on IBM Cloud are full vSphere implementations, replicated VMs can be started and run in a DR-scenario.
|
|
Remote Replication Topologies
Details
|
Single-site and multi-site
Single Site DR = 1-to-1
Multiple Site DR = 1-to many, many-to 1
|
Single-site
NEW
Single Site DR = 1-to-1
Multiple Site DR = 1-to many, many-to 1
|
VR: Single-site and multi-site
HR: Single-site and chained
Single Site DR = 1-to-1
Multiple Site DR = 1-to many, many-to 1
Hyper-V Replica (HR): Besides 1-to-1 replications Hyper-V Replica allows for extended (chained) replication. A VM can be replicated from a primary host to a secondary host, and then be replicated from the secondary host to a third host. Please note that it is not possible to replicate from the primary host directly to the second and the third (1-to-many).
|
|
Remote Replication Frequency
Details
|
Continuous (near-synchronous)
SANsymphony Asynchronous Replication is not checkpoint-based but instead replicates continuously. This way data loss is kept to a minimum (seconds to minutes). End-users can inject custom consistency checkpoints based on CDP technology which has no minimum time slot/frequency.
|
30 seconds (Asynchronous)
NEW
PowerFlex 3.5 leverages journal-based replication to achieve very low Recovery Point Objective (RPO) and ensure minimal data loss. PowerFlex 3.5 supports a replication frequency of 30 seconds (minimum) up to 1 hour (maximum).
To ensure RPO compliance, PowerFlex replicates at least twice for every RPO period. For example, setting RPO to one minute means that PowerFlex can immediately return to operation at the target system with only one minute’s worth of data loss. In order to achieve an RPO of one minute, replication takes place at least every 30 seconds. RPO compliance is calculated according to when the application data arrives at the target journal. Achieving RPO is the most important consideration in replication management. The user-defined RPO is the goal of replication. Even so, RPO compliance is not guaranteed. The system reports RPO compliance for each Replication Consistency Group (RCG).
|
VR: 5 minutes (Asynchronous)
HR: 30 seconds (Asynchronous)
SW: Continuous (Stretched Cluster)
Hyper-V Replica (HR): With Hyper-V Replica replication frequency can be set to 30 seconds, 5 minutes, or 15 minutes on a per-VM basis.
|
|
Remote Replication Granularity
Details
|
VM or Volume
With SANsymphony the rough hierarchy is: physical disk(s) or LUNs -> Disk Pool -> Virtual Disk (=logical volume).
Although DataCore SANsymphony uses block-storage, the platform is capable of attaining per VM-granularity if desired.
In Microsoft Hyper-V environments, when a VM with vdisks is created through SCVMM, DataCore can be instructed to automatically carve out a Virtual Disk (=storage volume) for every individual vdisk. This way there is a 1-to-1 alignment from end-to-end and snapshots can be created on the VM-level. The per-VM functionality is realized by installing the DataCore Storage Management Provider in SCVMM.
Because of the per-host storage limitations in VMware vSphere environments, VVols is leveraged to provide per VM-granularity. DataCore SANsymphony Provider v2.01 is VMware certified for ESXi 6.5 U2/U3, ESXi 6.7 GA/U1/U2/U3 and ESXi 7.0 GA/U1.
|
VM (Vvols) or Volume
NEW
PowerFlex replicates volume pairs. A volume pair consists of one volume at the source system and one volume at the target system. Replicated volumes at the target are limited to read-only access.
|
VR: VM
HR: VM
Both vSphere Replication (VR) and Hyper-V Replica operate on the VM level.
|
|
Consistency Groups
Details
|
Yes
SANsymphony provides the option to use Virtual Disk Grouping to enable end-users to restore multiple Virtual Disks to the exact same point-in-time.
With SANsymphony the rough hierarchy is: physical disk(s) or LUNs -> Disk Pool -> Virtual Disk (=logical volume).
|
Yes
NEW
The Replication Consistency Group (RCG) is a logical container for volumes whose application data need to be replicated consistency to each other.
|
VR: No
HR: No
|
|
|
VMware SRM (certified)
DataCore provides a certified Storage Replication Adapter (SRA) for VMware Site Recovery Manager (SRM). DataCore SRA 2.0 (SANsymphony 10.0 FC/iSCSI) shows official support for SRM 6.5 only. It does not support SRM 8.2 or 8.1.
There is no integration with Microsoft Azure Site Recovery (ASR). However, SANsymphony can be used with the control and automation options provided by Microsoft System Center (e.g. Operations Manager combined with Virtual Machine Manager and Orchestrator) to build a DR orchestration solution.
|
N/A
NEW
PowerFlex does not provide a Storage Replication Adapter (SRA) for VMware SRM implementations at this time.
|
N/A
StarWind Virtual SAN does not have a-synchronous native remote-replication capabilities and does not provide a Storage Replication Adapter (SRA) for VMware SRM implementations.
|
|
Stretched Cluster (SC)
Details
|
VMware vSphere: Yes (certified)
DataCore SANsymphony is certified by VMware as a VMware Metro Storage Cluster (vMSC) solution. For more information, please view https://kb.vmware.com/kb/2149740.
|
N/A
Dell EMC formally does not support PowerFlex clusters that are stretched across data centers.
|
vSphere: Yes
Hyper-V: Yes
StarWind Virtual SAN supports active-active Stretched Clustering which leverages native synchronous block-level replication.
|
|
|
2+sites = two or more active sites, 0/1 or more tie-breakers
Theoretically up to 64 sites are supported.
SANsymphony does not require a quorum or tie-breaker in stretched cluster configurations, but can be used as an optional component. The Virtual Disk Witness can provide a tie-breaker role if for instance redundant inter site paths are not implemented. The tie-breaker node (server or device) must be other than the two nodes presenting a virtual disk. Access to the Virtual Disk Witness is leading for storage node behavior.
There are 3 ways to configure the stretched cluster without any tie-breakers:
1. Default: in a split-brain scenario both sides stay active allowing upper infrastructure layers (OS/database/application) to make a decision (eg. clustering principles). In any case SANsymphony prevents a merge when there is a risk to data integrity, and the end-user has to make the choice on how to proceed next (which side is true)
2. Select one side to go inaccessible
3. Select both sides to go inaccessible.
|
N/A
Dell EMC formally does not support PowerFlex clusters that are stretched across data centers.
|
vSphere: 2+ sites = minimum of two active sites + optional tie-breaker in 3rd site
Hyper-V: 2+ sites = minimum of two active sites + optional tie-breaker in 3rd site
|
|
|
<=5ms RTT (targeted, not required)
RTT = Round Trip Time
In truth the user/app with the least tolerated write latency defines the acceptable RTT or distance. In practice
|
N/A
Dell EMC formally does not support PowerFlex clusters that are stretched across data centers.
|
<=5ms RTT
|
|
|
<=32 hosts at each active site (per cluster)
The maximum is per cluster. The SANsymphony solution can consist of multiple stretched clusters with a maximum of 64 nodes each.
|
N/A
Dell EMC formally does not support PowerFlex clusters that are stretched across data centers.
|
No set maximum number of nodes
|
|
SC Data Redundancy
Details
|
Replicas: 1N-2N at each active site
DataCore SANsymphony provides enhanced stretched cluster availability by offering local fault protection with In Pool Mirroring. With In Pool Mirroring you can choose to mirror the data inside the local Disk Pool as well as mirror the data across sites to a remote Disk Pool. In the remote Disk Pool data is then also mirrored. All mirroring happens synchronously.
1N-2N: With SANsymphony Stretched Clustering, there can be either 1 instance of the data at each site (no In Pool Mirroring) or 2 instances of the data a each site (In Pool RAID-1 Mirroring).
|
N/A
Dell EMC formally does not support PowerFlex clusters that are stretched across data centers.
|
Replicas: 0-3 Replicas at each active site
|
|
|
Data Services
|
|
|
|
|
|
|
Efficiency |
|
|
Dedup/Compr. Engine
Details
|
Software (integration)
NEW
SANsymphony provides integrated and individually selectable inline deduplication and compression. In addition, SANsymphony is able to leverage post-processing deduplication and compression options available in Windows 2016/2019 as an alternative approach.
|
Software (Limited)
VxFlex OS 3.0 introduced data efficiency capabilities by adding inline compression when using the new Fine Granularity (FG) data layout that uses storage allocation units of 4KB instead of the already existing Medium Granularity (MG) that uses storage allocation units of 1MB.
Fine Granularity (FG) Storage Pools require nodes with NVDIMMs and SSD or NVMe storage media. The current release is currently supported on Linux-based systems only.
VxFlex OS 3.0 does not have deduplication capabilities.
|
vSphere: Software (native)
Hyper-V: Software (integration)
StarWind Virtual SAN for vSphere provides inline deduplication and compression at the software level. In Hyper-V environments, post-process deduplication and compression can be used by leveraging Windows Server OS native capabilities.
|
|
Dedup/Compr. Function
Details
|
Efficiency (space savings)
Deduplication and compression can provide two main advantages:
1. Efficiency (space savings)
2. Performance (speed)
Most of the time deduplication/compression is primarily focussed on efficiency.
|
Efficiency
VxFlex OS 3.0 introduced data efficiency capabilities by adding inline compression when using the new Fine Granularity (FG) data layout that uses storage allocation units of 4KB instead of the already existing Medium Granularity (MG) that uses storage allocation units of 1MB.
VxFlex OS 3.0 compression allows better efficiency in thin-provisioned volumes and snapshots.
|
Efficiency (space savings)
Deduplication and compression can provide two main advantages:
1. Efficiency (space savings)
2. Performance (speed)
Most of the time deduplication/compression is primarily focussed on efficiency.
|
|
Dedup/Compr. Process
Details
|
Deduplication: Inline (post-ack)
Compression: Inline (post-ack)
Deduplication/Compression: Post-Processing (post process)
NEW
Deduplication can be performed in 4 ways:
1. Immediately when the write is processed (inline) and before the write is ackowledged back to the originator of the write (pre-ack).
2. Immediately when the write is processed (inline) and in parallel to the write being acknowledged back to the originator of the write (on-ack).
3. A short time after the write is processed (inline) so after the write is acknowleged back to the originator of the write - eg. when flushing the write buffer to persistent storage (post-ack)
4. After the write has been committed to the persistent storage layer (post-process).
The first and second methods, when properly integrated into the solution, are most likely to offer both performance and capacity benefits. The third and fourth methods are primarily used for capacity benefits only.
DataCore SANSymphony 10 PSP12 and above leverage both inline deduplication and compression, as well as post process deduplication and compression techniques.
With inline deduplication incoming writes first hit the memory cache of the primary host and are replicated to the cache of a secondary host in an un-deduplicated state. After the blocks have been written to both memory caches, the primary host acknowledges the writes back to the originator. Each host then destages the written blocks to the persistent storage layer. During destaging, written blocks are deduplicates and/or compressed.
Windows Server 2019 deduplication is performed outside of IO path (post-processing) and is multi-threaded to speed up processing and keep performance impact minimal.
|
Deduplication: N/A
Compression: inline (post-ack)
Deduplication and Compression can be performed in 4 ways:
1. Immediately when the write is processed (inline) and before the write is ackowledged back to the originator of the write (pre-ack).
2. Immediately when the write is processed (inline) and in parallel to the write being acknowledged back to the originator of the write (on-ack).
3. A short time after the write is processed (inline) so after the write is acknowleged back to the originator of the write - eg. when flushing the write buffer to persistent storage (post-ack)
4. After the write has been committed to the persistent storage layer (post-process).
The first and second methods, when properly integrated into the solution, are most likely to offer both performance and capacity benefits. The third and fourth methods are primarily used for capacity benefits only.
VxFlex OS 3.0 inline compression has been designed to provide capacity benefits and happens post-ack.
|
vSphere: Inline
Hyper-V: Post-Processing
Deduplication can be performed in 4 ways:
1. Immediately when the write is processed (inline) and before the write is ackowledged back to the originator of the write (pre-ack).
2. Immediately when the write is processed (inline) and in parallel to the write being acknowledged back to the originator of the write (on-ack).
3. A short time after the write is processed (inline) so after the write is acknowleged back to the originator of the write - eg. when flushing the write buffer to persistent storage (post-ack)
4. After the write has been committed to the persistent storage layer (post-process).
The first and second methods, when properly integrated into the solution, are most likely to offer both performance and capacity benefits. The third and fourth methods are primarily used for capacity benefits only.
StarWind Virtual SAN for vSphere inline deduplication works in the following manner:
1. In the initial phase, any blocks that consist entirely of zeros are identified and recorded only in metadata.
2. In the second phase, the incoming data is processed to determine whether it is redundant data (data that has been written before) or not. The redundancy of this data is checked through metadata maintained by the kernel module. Any block of data that is found to be redundant will not be written out. Instead, metadata will be updated to point to the original copy of the block already stored on media.
3. Once the initial and second phases are completed, compression is applied to the remaining individual data blocks. The compressed data blocks are then packed together into fixed length (4KB) blocks and stored on media.
Windows Server 2019 deduplication is performed outside of IO path (post-processing) and is multi-threaded to speed up processing and keep performance impact minimal.
|
|
Dedup/Compr. Type
Details
|
Optional
NEW
By default, deduplication and compression are turned off. For both inline and post-process, deduplication and compression can be enabled.
For inline deduplication and compression the feature can be turned on per node. The entire node represents a global deduplication domain. Deduplication and compression work across pools and across vDisks. Individual pools can be selected to participate in capacity optimization. Either deduplication or compression or both can be selected per individual vDisk. Pools can host both capacity optimized and non-capacity optimized vDisks at the same time. The optional capacity optimization settings can be added/changed/removed during operation for each vDisk.
For post-processing the feature can be enabled per pool. All vDisks in that pool would be deduplicated and compressed. Each pool is an independent deduplication domain. This means only data in the pool is capacity optimized, but not across pools. Additionally, for post-processing capacity optimization can be scheduled so admins can decide when deduplication should run.
With SANsymphony the rough hierarchy is: physical disk(s) or LUNs -> Disk Pool -> Virtual Disk (=logical volume).
|
Optional
Inline compression is turned off by default on Fine Granularity (FG) Pools, and can be enabled by checking the Enable Compression check-box.
|
Optional
vSphere: By default inline data deduplication is turned off and can be enabled.
Hyper-V: By default post-process deduplication and compression are turned off. Deduplication and compression can be enabled for selected volumes, either manually or scheduled.
|
|
Dedup/Compr. Scope
Details
|
Persistent data layer
|
Persistent Storage
VxFlex inline compression is applied to the persistent storage layer, including storage-based snapshots.
|
Persistent data layer
StarWind Virtual SAN for vSphere inline deduplication is performed on the persistant storage layer (XFS partition with inline deduplication enabled on which StarWind devices are located).
Microsoft Windows Server native deduplication and compression can be set for StarWind virtual devices formatted as NTFS.
Windows Server 2019 Deduplication only happens in the persistent data layer and not in the cache. The cache is not accessible from the file system and so deduplication cannot be applied to it.
|
|
Dedup/Compr. Radius
Details
|
Pool (post-processing deduplication domain)
Node (inline deduplication domain)
NEW
With SANsymphony the rough hierarchy is: physical disk(s) or LUNs -> Disk Pool -> Virtual Disk (=logical volume).
For inline deduplication and compression raw physical disks are added to a capacity optimization pool. The entire node represents a global deduplication domain. Deduplication and compression work across pools and across vDisks. Individual pools can be selected to participate in capacity optimization.
The post-processing capability provided through Windows Server 2016/2019 is highly scalable and can be used with volumes up to 64 TB and files up to 1 TB in size. Data deduplication identifies repeated patterns across files on that volume.
|
Volume
Inline compression is enabled on the V-Tree level. A V-Tree (short for volume tree) is the structure comprised of a volume and the snapshots resulting from that volume.
|
Volume
StarWind Virtual SAN for vSphere inline deduplication is scalable up to 256TB volume size.
Windows Server 2019 deduplication is highly scalable and can be used with volumes up to 64TB and files up to 4TB in size. Data deduplication identifies repeated patterns across files on that volume.
|
|
Dedup/Compr. Granularity
Details
|
4-128 KB variable block size (inline)
32-128 KB variable block size (post-processing)
NEW
With inline deduplication and compression, the data is organized in 128 KB segments. Depending on the optimization setting, a write into such a segment first gets compressed (when compression is selected) and then a hash is generated. If the hash is unique, the 128 KB segment is written back and the hash is added to the deduplication hash-table. If the hash is not unique, the segment is referenced in the deduplication hash table and discarded. The smallest chunk in the segment can be 4 KB.
For post-processing the system leverages deduplication in Windows Server 2016/2019, files within a deduplication-enabled volume are segmented into small variable-sized chunks (32–128 KB), duplicate chunks are identified, and only a single copy of each chunk is physically stored.
|
4 KB fixed block size
PowerFlex inline compression uses 4K fixed block segments.
|
vSphere: 4KB fixed block size
Hyper-V: 32-128 KB variable block size
Inline deduplication present in StarWind VSAN for vSphere stored deduplicated and compressed block in the 4k block size.
By leveraging deduplication in Windows Server 2019, files within a deduplication-enabled volume are segmented into small variable-sized chunks (32–128 KB), duplicate chunks are identified, and only a single copy of each chunk is physically stored.
|
|
Dedup/Compr. Guarantee
Details
|
N/A
Microsoft provides the Deduplication Evaluation Tool (DDPEVAL) to assess the data in a particular volume and predict the dedup ratio.
|
N/A
At this time Dell EMC does no guarantee a minimum savings. Compression gains will vary per workload and use case. Dell EMC states up to 10x more storage capacity can be gained by leveraging PowerFlex inline compression.
|
N/A
|
|
|
Full (optional)
Data rebalancing needs to be initiated manually by the end-user. It depends on the specific use case and end-user environment if this makes sense. When end-users want to isolate new workloads and corresponding data on new nodes, data rebalancing is not used.
|
Full
Data is automatically redistributed evenly across all disks in the cluster when a disk is either added or removed.
Data migrations are performed in a many-to-many fashion.
There is no user-intervention required for any the redistribution activities.
Data Rebalancing can be throttled by changing the priority policy. The default policy for rebalance: Favor Application I/O. Alternative policies are: No Limit, Limit Concurrent I/O, Dynamic Bandwidth Throttling.
|
Full (optional)
Data rebalancing needs to be initiated manually. When a new StarWind Virtual SAN node is added, a StarWind Support Engineer assists with replicating data to the new partner.
|
|
|
Yes
DataCore SANsymphonys Auto-Tiering is a real-time intelligent mechanism that continuously positions data on the appropriate class of storage based on how frequently the data is accessed. Auto-Tiering leverages any combination of Flash and traditional disk technologies, whether it is internal or array based, with up to 15 different storage tiers that can be defined.
As more advanced storage technologies become available, existing tiers can be modified as necessary and additional tiers can be added to further diversify the tiering architecture.
|
N/A
Although PowerFlexs storage architecture can host multiple persistent storage tiers (pool with magnetic devices next to a pool with flash devices) within a single solution, it does not provide any intelligent and automatic means to distribute data across these pools of storage bases on certain criteria.
|
Partial (integration; optional)
StarWind Virtual SAN can leverage the data tiering capabilities available in the Windows Server 2016/2019 OS (Storage Spaces). This is the case for Hyper-V and VMware vSphere environments.
|
|
|
|
Performance |
|
|
|
vSphere: VMware VAAI-Block (full)
Hyper-V: Microsoft ODX; Space Reclamation (T10 SCSI UNMAP)
DataCore SANsymphony iSCSI and FC are fully qualified for all VMware vSphere VAAI-Block capabilities that include: Thin Provisioning, HW Assisted Locking, Full Copy, Block Zero
Note: DataCore SANsymphony does not support Thick LUNs.
DataCore SANsymphony is also fully qualified for Microsoft Hyper-V 2012 R2 and 2016/2019 ODX and UNMAP/TRIM.
Note: ODX is not used for files smaller than 256KB.
VAAI = VMware vSphere APIs for Array Integration
ODX = Offloaded Data Transfers
UNMAP/TRIM support allows the Windows operating system to communicate the inactive block IDs to the storage system. The storage system can wipe these unused blocks internally.
|
vSphere: VMware VAAI-Block (partial)
The VxFlex OS 2.6 supported VAAI features are: Atomic Test & Set (ATS), Zero Blocks/Write Same, Thin Provisioning in ESXi 5.x and later hosts, Block Delete in ESXi 5.x and later hosts
ScaleIO 1.32 and later were not listed anymore in the VMware Compatibility Matrix because of the proprietary nature of the storage protocol/SDC (non-iSCSI).
VAAI = VMware APIs for Array Integration
|
vSphere: VMware VAAI-Block (full)
Hyper-V: Microsoft ODX; UNMAP/TRIM
StarWind Virtual SAN iSCSI is fully qualified for all VMware vSphere VAAI-Block capabilities that include: Thin Provisioning, HW Assisted Locking, Full Copy, Block Zero.
StarWind Virtual SAN is also fully qualified for Microsoft Hyper-V 2016 and 2019 ODX, as well as TRIM (for SATA SSDs) and UNMAP (for SAS SSDs).
Note: ODX is not used for files smaller than 256KB.
VAAI = VMware vSphere APIs for Array Integration
ODX = Offloaded Data Transfers
UNMAP/TRIM support allows the Windows operating system to communicate the inactive block IDs to the storage system. The storage system can wipe these unused blocks internally.
|
|
|
IOPs and/or MBps Limits
QoS is a means to ensure specific performance levels for applications and workloads. There are two ways to accomplish this:
1. Ability to set limitations to avoid unwanted behavior from non-critical clients/hosts.
2. Ability to set guarantees to ensure service levels for mission-critical clients/hosts.
SANsymphony currently supports only the first method. Although SANsymphony does not provide support for the second method, the platform does offer some options for optimizing performance for selected workloads.
For streaming applications which burst data, it’s best to regulate the data transfer rate (MBps) to minimize their impact. For transaction-oriented applications (OLTP), limiting the IOPs makes most sense. Both parameters may be used simultaneously.
DataCore SANsymphony ensures that high-priority workloads competing for access to storage can meet their service level agreements (SLAs) with predictable I/O performance. QoS Controls regulate the resources consumed by workloads of lower priority. Without QoS Controls, I/O traffic generated by less important applications could monopolize I/O ports and bandwidth, adversely affecting the response and throughput experienced by more critical applications. To minimize contention in multi-tenant environments, the data transfer rate (MBps) and IOPs for less important applications are capped to limits set by the system administrator. QoS Controls enable IT organizations to efficiently manage their shared storage infrastructure using a private cloud model.
More information can be found here: https://docs.datacore.com/SSV-WebHelp/quality_of_service.htm
In order to achieve consistent performance for a workload, a separate Pool can be created where selected vDisks are placed. Alternatively 'Performance Classes' can be assigned to differentiate between data placement of multiple workloads.
|
IOPs and/or MBps Limits (maximums)
QoS is a means to ensure specific performance levels for applications and workloads. There are two ways to accomplish this:
1. Ability to set limitations to avoid unwanted behavior from non-critical clients/hosts.
2. Ability to set guarantees to ensure service levels for mission-critical clients/hosts.
PowerFlex currently supports only the first method.
PowerFlex Limiter: Users can adjust the amount of IOPs and/or bandwidth (MBps) that one SDC can generate for a volume. These parameters are configured using the PowerFlex CLI (Command Line Interface), and the REST interface, on a per client/per volume basis.
|
N/A
Quality-of-Service (QoS) is a means to ensure specific performance levels for applications and workloads. There are two ways to accomplish this:
1. Ability to set limitations to avoid unwanted behavior from non-critical VMs.
2. Ability to set guarantees to ensure service levels for mission-critical VMs.
StarWind Virtual SAN currently does not offer any QoS mechanisms.
|
|
|
Virtual Disk Groups and/or Host Groups
SANsymphony QoS parameters can be set for individual hosts or groups of hosts as well as for groups of Virtual Disks for fine grained control.
In a VMware VVols (=Virtual Volumes) environment a vDisk corresponds 1-to-1 to a virtual disk (.vmdk). Thus virtual disks can be placed in a Disk Group and a QoS Limit can then be assigned it. DataCore SANsymphony Provider v2.01 has VVols certification for VMware ESXi 6.5 U2/U3, ESXi 6.7 GA/U1/U2/U3 and ESXi 7.0 GA/U1.
In Microsoft Hyper-V environments, when a VM with vdisks is created through SCVMM, DataCore can be instructed to automatically carve out a Virtual Disk (=storage volume) for every individual vdisk. This way there is a 1-to-1 alignment from end-to-end and QoS Limits can be applied on the virtual disk level. The 1-to-1 allignment is realized by installing the DataCore Storage Management Provider in SCVMM.
|
Per client and volume
QoS parameters are configured on a per client/per volume basis.
|
N/A
Quality-of-Service (QoS) is a means to ensure specific performance levels for applications and workloads. There are two ways to accomplish this:
1. Ability to set limitations to avoid unwanted behavior from non-critical VMs.
2. Ability to set guarantees to ensure service levels for mission-critical VMs.
StarWind Virtual SAN currently does not offer any QoS mechanisms.
|
|
|
Per VM/Virtual Disk/Volume
With SANsymphony the rough hierarchy is: physical disk(s) or LUNs -> Disk Pool -> Virtual Disk (=logical volume).
In SANsymphony 'Flash Pinning' can be achieved using one of the following methods:
Method #1: Create a flash-only pool and migrate the individual vDisks that require flash pinning to the flash-only pool. When using a VVOL configuration in a VMware environment, each vDisk represents a virtual disk (.vmdk). This method guarantees all application data will be stored in flash.
Method #2: Create auto-tiering pools with at least 1 flash tier. Assign the Performance Class “Critical” to the vDisks that require flash pinning and place them in the auto-tiering pool. This will effectively and intelligently put as much of the data that resides in the vDisk in the flash tier as long that the flash tier has enough space available. Therefore this method is on a best-effort basis and dependent on correct sizing of the flash tier(s).
Methods #1 and #2 can be uses side-by-side in the same DataCore environment.
|
Per Volume
When creating a volume select a flash storage pool. Data that are placed on the volume exist in Flash only.
|
N/A
|
|
|
|
Security |
|
|
Data Encryption Type
Details
|
Built-in (native)
SANsymphony 10.0 PSP9 introduced native encryption when running on Windows Server 2016/2019.
|
N/A
PowerFlex does not have native data encryption capabilities. Adding data encryption capabilities requires either 1st party or 3rd party security software.
|
N/A
StarWind Virtual SAN does not have native data encryption capabilities. Adding data encryption capabilities requires encryption storage hardware. Alternatively, software options such as Microsoft Bitlocker or vSphere Virtual Machine Encryption can be leveraged.
VMware Virtual Machine Encryption requires VMware vSphere
Enterprise Plus or VMware vSphere Platinum licences.
|
|
Data Encryption Options
Details
|
Hardware: Self-encrypting drives (SEDs)
Software: SANsymphony Encryption
Hardware: In SANsymphony deployments the encryption data service capabilities can be offloaded to hardware-based SED offerings available in server- and storage solutions.
Software: SANsymphony provides software-based data-at-rest encryption that is XTS-AES 256bit compliant.
|
Hardware: Self-encrypting drives (SEDs)
Software: Dell EMC CloudLink
NEW
Hardware: In PowerFlex deployments the encryption data service capabilities can be offloaded to hardware-based SED offerings available in server- and storage solutions.
Software: Dell EMC CloudLink uses dm-crypt, a native Linux encryption package, to secure SDS devices. A proven high-performance volume encryption solution, dm-crypt is widely implemented for Linux machines.
The CloudLink Agent can be deployed on physical and
virtual Linux PowerFlex Storage Data Servers (SDSs) and supports fully converged, two-layer, and mixed configurations. In VMware vSphere deployments the CloudLink Agent software is installed in the PowerFlex Storage Virtual Machine (SVM) that runs on each ESXi host that contributes storage.
PowerFlex 3.5 supports CloudLink 6.9 and 7.0.
Dell EMC has not yet published interoperability test reports validating the use of one or more third-party software-based encryption solutions with its PowerFlex platform.
|
Hardware: Self-encrypting drives (SEDs)
Software: Microsoft BitLocker Drive Encryption; VMware Virtual machine Encryption
Hardware: In StarWind Virtual SAN deployments the encryption data service capabilities can be offloaded to hardware-based SED offerings available in server- and storage solutions.
Software: Microsoft Bitlocker provides software encryption on standalone and cluster based NTFS or ReFS(v2) volumes. Cluster volumes (CSV) encryption support was added in Windows 2012 Server.
Microsoft BitLocker uses the Advanced Encryption Standard (AES) encryption algorithm with either 128-bit or 256-bit keys. It is generally recommended to use 256-bit keys because of their superior strength.
Bitlocker can be used to encrypt the local storage, encrypt Cluster Shared Volumes created on StarWind HA devices where VMs are located or encrypt VMs with Trusted Platform Module.
As for ESXi hosts, Bitlocker can be used to encrypt the Windows Server VMs storage (only non-bootable partitions). However, this does not provide full security. Alternatively, vSphere Virtual Machine Encryption can be used.
|
|
Data Encryption Scope
Details
|
Hardware: Data-at-rest
Software: Data-at-rest
Hardware: SEDs provide encryption for data-at-rest; SEDs do not provide encryption for data-in-transit.
Software: SANsymphony provides encryption for data-at-rest; it does not provide encryption for data-in-transit. Encryption can be enabled per individual virtual disk.
|
Hardware: Data-at-rest
Software: Data-at-rest
Hardware: SEDs provide encryption for data-at-rest; SEDs do not provide encryption for data-in-transit.
Software: Dell EMC CloudLink provides encryption for data-at-rest; Dell EMC CloudLink does not provide encryption for data-in-transit.
|
Hardware: Data-at-rest
Software: Data-at-rest
Hardware: SEDs provide encryption for data-at-rest; SEDs do not provide encryption for data-in-transit.
Software: Microsoft BitLocker provides encryption for data-at-rest as well as data-in-transit during live migration of a VM; VMware Virtual Machine Encryption provides encryption for data-at-rest.
|
|
Data Encryption Compliance
Details
|
Hardware: FIPS 140-2 Level 2 (SEDs)
Software: FIPS 140-2 Level 1 (SANsymphony)
FIPS = Federal Information Processing Standard
FIPS 140-2 defines four levels of security:
Level 1 > Basic security requirements are specified for a cryptographic module (eg. at least one Approved algorithm or Approved security function shall be used).
Level 2 > Also has features that show evidence of tampering.
Level 3 > Also prevents the intruder from gaining access to critical security parameters (CSPs) held within the cryptographic module.
Level 4 > Provides a complete envelope of protection around the cryptographic module with the intent of detecting and responding to all unauthorized attempts at physical access.
|
Hardware: FIPS 140-2 Level 2 (SEDs)
Software: FIPS 140-2 Level 1 (Dell EMC CloudLink)
FIPS = Federal Information Processing Standard
FIPS 140-2 defines four levels of security:
Level 1 > Basic security requirements are specified for a cryptographic module (eg. at least one Approved algorithm or Approved security function shall be used).
Level 2 > Also has features that show evidence of tampering.
Level 3 > Also prevents the intruder from gaining access to critical security parameters (CSPs) held within the cryptographic module.
Level 4 > Provides a complete envelope of protection around the cryptographic module with the intent of detecting and responding to all unauthorized attempts at physical access.
|
Hardware: FIPS 140-2 Level 2 (SEDs)
Software: FIPS 140-2 Level 1 (BitLocker; VMware Virtual Machine Encryption)
Microsoft BitLocker has been validated for Federal Information Processing Standard (FIPS) 140-2 in March 2018.
FIPS = Federal Information Processing Standard
FIPS 140-2 defines four levels of security:
Level 1 > Basic security requirements are specified for a cryptographic module (eg. at least one Approved algorithm or Approved security function shall be used).
Level 2 > Also has features that show evidence of tampering.
Level 3 > Also prevents the intruder from gaining access to critical security parameters (CSPs) held within the cryptographic module.
Level 4 > Provides a complete envelope of protection around the cryptographic module with the intent of detecting and responding to all unauthorized attempts at physical access.
|
|
Data Encryption Efficiency Impact
Details
|
Hardware: No
Software: No
Hardware: Because data encryption is performed at the end of the write path, storage efficiency mechanisms are not impaired.
Software: Because data encryption is performed at the end of the write path, storage efficiency mechanisms are not impaired.
|
Hardware: No
Software: No
Hardware: Because data encryption is performed at the end of the write path, storage efficiency mechanisms are not impaired.
Software: Because data encryption is performed at the end of the write path, storage efficiency mechanisms are not impaired.
|
Hardware: No
Software: No (Bitlocker); No (VMware Virtual Machine Encryption)
Hardware: Because data encryption is performed at the end of the write path, storage efficiency mechanisms are not impaired.
Software: Microsoft BitLocker can be used to provide whole-disk encryption on a deduplicated disk since BitLocker sits at the end of the write path. VMware Virtual Machine Encryption encrypts tge data on the host before it is written to storage, thus negatively impacting backend storage features such as deduplication and compression. However, Microsoft post-process deduplication is executed at the filesystem layer.
|
|
|
|
Test/Dev |
|
|
|
Yes
Support for fast VM cloning via VMware VAAI and Microsoft ODX.
|
No
|
No
StarWind VSAN is an SDS solution. The cloning functionality is not related to storage capabilities and is performed on the hypervisor level.
|
|
|
|
Portability |
|
|
Hypervisor Migration
Details
|
Hyper-V to ESXi (external)
ESXi to Hyper-V (external)
VMware Converter 6.2 supports the following Guest Operating Systems for VM conversion from Hyper-V to vSphere:
- Windows 7, 8, 8.1, 10
- Windows 2008/R2, 2012/R2 and 2016
- RHEL 4.x, 5.x, 6.x, 7.x
- SUSE 10.x, 11.x
- Ubuntu 12.04 LTS, 14.04 LTS, 16.04 LTS
- CentOS 6.x, 7.0
The VMs have to be in a powered-off state in order to be migrated across hypervisor platforms.
Microsoft Virtual Machine Converter (MVMC) supports conversion of VMware VMs and vdisks to Hyper-V VMs and vdisks. It is also possible to convert physical machines and disks to Hyper-V VMs and vdisks.
MVMC has been offcially retired and can only be used for converting VMs up to version 6.0.
Microsoft System Center Virtual Machine Manager (SCVMM) 2016 also supports conversion of VMs up to version 6.0 only.
|
Hyper-V to ESXi (external)
ESXi to Hyper-V (external)
VMware Converter 6.2 supports the following Guest Operating Systems for VM conversion from Hyper-V to vSphere:
- Windows 7, 8, 8.1, 10
- Windows 2008/R2, 2012/R2 and 2016
- RHEL 4.x, 5.x, 6.x, 7.x
- SUSE 10.x, 11.x
- Ubuntu 12.04 LTS, 14.04 LTS, 16.04 LTS
- CentOS 6.x, 7.0
The VMs have to be in a powered-off state in order to be migrated across hypervisor platforms.
Microsoft Virtual Machine Converter (MVMC) supports conversion of VMware VMs and vdisks to Hyper-V VMs and vdisks. It is also possible to convert physical machines and disks to Hyper-V VMs and vdisks.
MVMC has been offcially retired and can only be used for converting VMs up to version 6.0.
Microsoft System Center Virtual Machine Manager (SCVMM) 2016 also supports conversion of VMs up to version 6.0 only.
|
Hyper-V to ESXi (external)
ESXi to Hyper-V (external)
StarWind V2V Converter is a StarWind proprietary tool that can be leveraged for virtual-to-virtual (V2V) use cases as well as physical-to-virtual (P2V) use cases. The tool supports all major VM formats: VHD/VHDX, VMDK, QCOW2, and StarWind native IMG. Both the source and target VM copies exist simultaneously because the conversion procedure is more like a cloning process than a replacement. As a convenient side effect, StarWind V2V Converter basically creates a backup copy of the VMs, making the process even safer.
When converting the VM to VHDX format, StarWind V2V Converter enables the activation of Windows Repair Mode. This way the virtual machine will automatically adapt to the given hardware environment and negate any compatibility problems.
|
|
|
|
File Services |
|
|
|
Built-in (native)
SANsymphony delivers out-of-box (OOB) file services by leveraging Windows native SMB/NFS and Scale-out File Services capabilities. SANsymphony is capable of simultaneously handling highly-available block and file level services.
Raw storage is provisioned from within the SANsymphony GUI to the Microsoft file services layer, similar to provisioning Storage Spaces Volumes to the file services layer. This means any file services configuration is performed from within the respective Windows service consoles e.g. quotas.
More information can be found under: https://www.datacore.com/products/features/high-availability-nas-cluster-file-sharing.aspx
|
N/A
PowerFlex does not provide any file serving capabilities of its own.
Inside a Guest VM all native file service features of the Microsoft Windows and/or Linux operating system can be leveraged to host network shares.
Linux requires Samba Server components to provide SMB file shares.
Depending on the OS of the Guest VM providing file services, quotas can been set on the share or the filesystem level.
|
vSphere: N/A
Hyper-V: Built-in (native)
StarWind Virtual SAN for Hyper-V delivers out-of-box (OOB) file services by leveraging Windows native SMB/NFS and Scale-out File Services capabilities. StarWind Virtual SAN is capable of simultaneously handling highly-available block and file level services.
StarWind virtual devices formatted as NTFS volumes are provisioned to the Microsoft file services layer. This means any file services configuration is performed from within the respective Windows service consoles e.g. quotas.
|
|
Fileserver Compatibility
Details
|
Windows clients
Linux clients
Because SANsymphony leverages Windows Server native CIFS/NFS and Scale-out File services, most Windows and Linux clients are able to connect.
|
N/A
PowerFlex does not provide any file serving capabilities of its own.
Inside a Guest VM all native file service features of the Microsoft Windows and/or Linux operating system can be leveraged to host network shares.
Linux requires Samba Server components to provide SMB file shares.
Depending on the OS of the Guest VM providing file services, quotas can been set on the share or the filesystem level.
|
Windows clients
Linux clients
Because StarWind Virtual SAN leverages Windows Server native CIFS/NFS and Scale-out File services, most Windows and Linux clients are able to connect.
|
|
Fileserver Interconnect
Details
|
SMB
NFS
Because SANsymphony leverages Windows Server native CIFS/NFS and Scale-out File services, Windows Server platform compatibility applies:
SMB versions1,2 and 3 are supported, as are NFS versions 2, 3 and 4.1.
|
N/A
PowerFlex does not provide any file serving capabilities of its own.
Inside a Guest VM all native file service features of the Microsoft Windows and/or Linux operating system can be leveraged to host network shares.
Linux requires Samba Server components to provide SMB file shares.
Depending on the OS of the Guest VM providing file services, quotas can been set on the share or the filesystem level.
|
SMB
NFS
Because StarWind Virtual SAN leverages Windows Server native CIFS/NFS and Scale-out File services, Windows Server platform compatibility applies:
SMB versions1,2 and 3 are supported, as are NFS versions 2, 3 and 4.1.
|
|
Fileserver Quotas
Details
|
Share Quotas, User Quotas
Because SANsymphony leverages Windows Server native CIFS/NFS and Scale-out File services, all Quota features available in Windows Server can be used.
|
N/A
PowerFlex does not provide any file serving capabilities of its own.
Inside a Guest VM all native file service features of the Microsoft Windows and/or Linux operating system can be leveraged to host network shares.
Linux requires Samba Server components to provide SMB file shares.
Depending on the OS of the Guest VM providing file services, quotas can been set on the share or the filesystem level.
|
Share Quotas, User Quotas
Because StarWind Virtual SAN leverages Windows Server native CIFS/NFS and Scale-out File services, all Quota features available in Windows Server can be used.
|
|
Fileserver Analytics
Details
|
Partial
Because SANsymphony leverages Windows Server native CIFS/NFS, Windows Server built-in auditing capabilities can be used.
|
N/A
PowerFlex does not provide any file serving capabilities of its own.
Inside a Guest VM all native file service features of the Microsoft Windows and/or Linux operating system can be leveraged to host network shares.
Linux requires Samba Server components to provide SMB file shares.
Depending on the OS of the Guest VM providing file services, quotas can been set on the share or the filesystem level.
|
Partial
Because StarWind Virtual SAN leverages Windows Server native CIFS/NFS, Windows Server built-in auditing capabilities can be used.
|
|
|
|
Object Services |
|
|
Object Storage Type
Details
|
N/A
DataCore SANsymphony does not provide any object storage serving capabilities of its own.
|
N/A
PowerFlex does not provide any object storage serving capabilities of its own.
|
N/A
StarWind Virtual SAN does not provide any object storage serving capabilities of its own.
|
|
Object Storage Protection
Details
|
N/A
DataCore SANsymphony does not provide any object storage serving capabilities of its own.
|
N/A
PowerFlex does not provide any object storage serving capabilities of its own.
|
N/A
StarWind Virtual SAN does not provide any object storage serving capabilities of its own.
|
|
Object Storage LT Retention
Details
|
N/A
DataCore SANsymphony does not provide any object storage serving capabilities of its own.
|
N/A
PowerFlex does not provide any object storage serving capabilities of its own.
|
N/A
StarWind Virtual SAN does not provide any object storage serving capabilities of its own.
|
|
|
Management
|
|
|
|
|
|
|
Interfaces |
|
|
GUI Functionality
Details
|
Centralized
SANsymphonys graphical user interface (GUI) is highly configurable to accommodate individual preferences and includes guided wizards and workflows to simplify administration. All actions available from the GUI may also be scripted with PowerShell Commandlets to orchestrate workflows with other tools and applications.
|
Centralized
NEW
PowerFlex management, capacity monitoring, performance monitoring and efficiency reporting is performed through the PowerFlex HTML-5 based management web interface (GUI).
|
Centralized
StarWind Web-based Management provides the ability to administrate the StarWind Virtual SAN infrastructure from any remote location using any HTML 5 web console.
|
|
|
Single-site and Multi-site
|
Single-system
The PowerFlex GUI enables viewing the entire system, and then drilling down to various elements.
|
Single-site and Multi-site
From within the StarWind Web Management Console servers from different clusters and sites can be added.
|
|
GUI Perf. Monitoring
Details
|
Advanced
SANsymphony has visibility into the performance of all connected devices including front-end channels, back-end channels, cache, physical disks, and virtual disks. Metrics include Read/write IOPs, Read/write MBps and Read/Write Latency at all levels. These metrics can be exported to the Windows Performance Monitoring (Perfmon) utility where other server parameters are being tracked.
The frequency at which performance metrics can be captured and reported on is configurable, real-time down to 1 second intervals and long term recording at 2 minutes granularity.
When a trend analysis is required, an end-user can simply enable a recording session to capture metrics over a longer period of time.
|
Basic
In VxFlex OS 3.0.1 and up Volumes/SDCs can be queried for I/O latency metrics.
|
Basic
StarWind Management Console can be used to view basic storage performance characteristics.
StarWind Management Console provides:
- counters on a per-host level: CPU/RAM load, CPU load, RAM load, Total IOPS, Total Bandwidth
- counters on a per-device level: Read Bandwidth, Write Bandwidth, Total Bandwidth, Total IOPS
- selectable are Server with all targets (StarWind devices) or each separate StarWind device
- retention time of performance information is last 24 hours
- refresh rate of performance metrics is every 30 seconds
|
|
|
VMware vSphere Web Client (plugin)
VMware vCenter plug-in for SANsymphony
SCVMM DataCore Storage Management Provider
Microsoft System Center Monitoring Pack
DataCore offers deep integration with VMware vSphere and Microsoft Hyper-V, as well as their respective systems management tools, vCenter and System Center.
SCVMM = Microsoft System Center Virtual Machine Manager
|
VMware vSphere Web Client (plugin)
NEW
The PowerFlex VMware vSphere plug-in enables performing basic provisioning and maintenance within a VMware environment. In addition, the plug-in provides a wizard to deploy PowerFlex in the VMware environment.
The PowerFlex 3.5 vSphere plug-in does not support VMware vSphere 7. Support for VMware vSphere 7 is planned in a H2 2020 release of PowerFlex Manager.
In dual-layer configurations vSphere 7 can be configured as the compute layer, with the storage layer being a supported operating system version for PowerFlex 3.5.
The vSphere plug-in in PowerFlex 3.5 continues to support vSphere versions 6.5 and 6.7. No further development of the vSphere plug-in is planned.
|
vSphere: Web Client (plugin)
Hyper-V: SCVMM 2016 (add-in)
The StarWind vCenter plug-in provides a 1-to-1 representation of the StarWind Web Console inside the VMware vSphere Web Client. It does not add additional StarWind Virtual SAN-related actions to any Web Client menus.
StarWind Virtual SAN for vSphere environments can be additionally managed from the VM web console providing storage and networking management capabilities as well as performance metrics.
|
|
|
|
Programmability |
|
|
|
Full
Using DataCores native management console, Virtual Disk Templates can be leveraged to populate storage policies. Available configuration items: Storage profile, Virtual disk size, Sector size, Reserved space, Write-trough enabled/disabled, Storage sources, Preferred snapshot pool, Accelerator enabled/disabled, CDP enabled/disabled.
Virtual Disk Templates integrate with System Center Virtual Machine Manager (SCVMM), VMware Virtual Volumes (VVol) and OpenStack. Virtual Disk Templates are also fully supported by the REST-API allowing any third-party integration.
Using Virtual Volumes (VVols) defined through DataCore’s VASA provider, VMware administrators can self-provision datastores for virtual machines (VMs) directly from their familiar hypervisor interface. This is possible even for devices in the DataCore pool that don’t natively support VVols and never will, as SANsymphony can be used as a storage-virtualization layer for these devices/solutions. DataCore SANsymphony Provider v2.01 has VVols certification for VMware ESXi 6.5 U2/U3, ESXi 6.7 GA/U1/U2/U3 and ESXi 7.0 GA/U1.
Using Classifications and StoragePools defined through DataCore’s Storage Management Provider, Hyper-V administrators can self-provision virtual disks and pass-through LUNS for virtual machines (VMs) directly from their familiar SCVMM interface.
|
Partial
PowerFlex has implemented VMwares vSphere API for Storage Awareness (VASA). PowerFlexs VASA enables control of PowerFlex storage to be handled by the vCenter administrator. PowerFlexs VASA uses Virtual Volumes (VVols) to enable VMs hosted on VMware vSphere to have their storage mapped directly to storage in PowerFlex.
PowerFlexs VASA automatically assigns VMware storage policies based on keywords
that appear in the PowerFlex Storage Pool names.
PowerFlexs VASA runs on a separate VM that needs to be deployed.
Dell EMC PowerFlex is currently not listed in the VMware Compatibility Guide for VVOL support.
|
Partial (Protection)
|
|
|
REST-APIs
PowerShell
The SANsymphony REST-APIs library includes more than 200 new representational state transfer (REST) operations, so automation can be leveraged more extensively. RESTful interfaces are used by products such as Lenovo XClarity, Cisco Embedded Resource Manager and Dell OpenManage to manage infrastructure in the enterprise.
SANsymphony provides its own Powershell cmdlets.
|
REST-APIs
PowerShell
sCLI
The PowerFlex CLI (sCLI) enables performing the entire set of configure, maintain, and monitor activities in a PowerFlex system.
|
REST-APIs (through Swordfish)
PowerShell
StarWind Virtual SAN does not offer native REST APIs (functionality still under development). Alternatively, StarWind over Swordfish API can be configured if programming is a requirement.
For more information, please view: https://www.starwindsoftware.com/resource-library/starwind-swordfish-provider/
|
|
|
OpenStack
OpenStack: The SANsymphony storage solution includes a Cinder driver, which interfaces between SANsymphony and OpenStack, and presents volumes to OpenStack as block devices which are available for block storage.
Datacore SANsymphony programmability in VMware vRealize Automation and Microsoft System Center can be achieved by leveraging PowerShell and the SANsymphony specific cmdlets.
|
OpenStack
EMC ViPR Controller + ViPR vRealize Orchestrator Plug-in
OpenStack: The PowerFlex elastic storage solution includes a Cinder driver, which interfaces between PowerFlex and OpenStack, and presents volumes to OpenStack as block devices which are available for block storage. It also includes an OpenStack Nova driver, for handling compute and instance volume related operations. The PowerFlex driver executes the volume operations by communicating with the backend PowerFlex MDM through the PowerFlex REST Gateway.
|
N/A
|
|
|
Full
The DataCore SANsymphony GUI offers delegated administration to secondary users through fine-grained Role-based Access Control (RBAC). The administrator is able to define Virtual Disk ownership as well as privileges associated with that particular ownership. Owners must have Virtual Disk privileges in an assigned role in order to perform operations on the virtual disk. Access can be very refined. For example, one owner may have the privilege to create a snapshot of a virtual disk, but not have the ability to serve or unserve the same virtual disk. Privilege sets define the operations that can be performed. For instance, in order for an owner to perform snapshot, rollback, or replication operations, they would require those privilege sets in an assigned role.
|
N/A (not part of PowerFlex license)
PowerFlex does not provide any end-user self service capabilities of its own.
A self service portal enables end-users to access a portal where they can provision and manage VMs from templates, eliminating administrator requests or activity.
Self-Service functionality can be enabled by leveraging EMC ViPR Controller or VMware vRealize Automation (vRA). Both require a separate license.
|
N/A
|
|
|
|
Maintenance |
|
|
|
Unified
All storage related features and functionality are built into the DataCore SANsymphony platform. The consolidation means, that only one product needs to be installed and upgraded and minimal dependencies exist with other software.
Integration with 3rd party systems (e.g. OpenStack, vSphere, System Center) are delivered seperately but are free-of-charge.
|
Partially Distributed
For a number of features and functions PowerFlex relies on other components that need to be installed and upgraded next to the core PowerFlex platform. Examples are EMC RecoverPoint and EMC ViPR Controller. As a result some dependencies exist with other software.
|
Partially Distributed
For a number of features and functions StarWind Virtual SAN relies on other components that need to be installed and upgraded next to the core Windows platform. Examples are backup/restore and advanced management software. As a result some dependencies exist with other software.
|
|
SW Upgrade Execution
Details
|
Rolling Upgrade (1-by-1)
Each SANsymphony update is packaged in an installation Wizard which contains a fully guided upgrade process. The upgrade process checks all system requirements and performs a system health before starting the upgrade process and before moving from one node to the next.
The user can also decide to upgrade a SANsymphony cluster manually and follow all steps that are outlined in the Release Notes.
|
Rolling Upgrade (1-by-1)
The PowerFlex Installation Manager (IM) can be used to automatically upgrade an entire Fault Set in a non-disruptive manner.
|
Manual Upgrade (1-by-1)
When updating StarWind VSAN, the latest build can be downloaded and executed. StarWind will automatically update all the components of the SDS stack on a single node. The update process has to be launched manually on every node in a VSAN cluster in sequentially.
|
|
FW Upgrade Execution
Details
|
Hardware dependent
Some server hardware vendors offer rolling upgrade options with their base software or with a premium software suite. With some other server vendors, BIOS and Baseboard Management Controller (BMC) updated have to be performed manually and 1-by-1.
DataCore provides integrated firmware-control for FC-cards. This means the driver automatically loads the required firmware on demand.
|
Hardware dependent
Some server hardware vendors offer rolling upgrade options with their base software or with a premium software suite. With some other server vendors, BIOS and Baseboard Management Controller (BMC) updated have to be performed manually and 1-by-1.
|
Manual Upgrade (1-by-1 )
To perform drivers updates, a remote session with a StarWind Support engineer is scheduled who will perform all the updates. Firmware and driver updates are first verified for any possible issues.
To keep the VMs and applications running during maintenance, StarWind performs the following steps:
1. Check that StarWind devices are synchronized on both nodes and that all iSCSI connections are active.
2. Move VMs inside the cluster to the partner node.
3. In case the process presumes several restarts, put StarWind service into a disabled state or manual start on the required node.
4. Install the required updates/drivers.
5. Restart the node if required.
6. In case required, shut down the node and perform a firmware update.
|
|
|
|
Support |
|
|
Single HW/SW Support
Details
|
No
With regard to DataCore SANsymphony as a software-only offering (SDS), DataCore does not offer unified support for the entire solution. This means storage software support (SANsymphony) and server hardware support are separate.
|
No
With regard to PowerFlex as a software-only offering, EMC does not offer unified support for the entire solution. This means storage software support (PowerFlex ECS) and server hardware support are separate.
|
Yes (optional)
StarWind Virtual SAN can optionally be supplied with StarWind ProActive Support to preventively resolve issues on hardware, hypervisor, and the StarWind SDS stack. Additionally, StarWind offers Managed Services covering not only SDS stack but also applications and services running on top of it and near it.
|
|
Call-Home Function
Details
|
Partial (HW dependent)
With regard to DataCore SANsymphony as a software-only offering (SDS), DataCore does not offer call-home for the entire solution. This means storage software support (SANsymphony) and server hardware support are separate.
|
Partial (HW dependent)
With regard to PowerFlex as a software-only offering (SDS), EMC does not offer call-home for the entire solution. This means storage software support (PowerFlex ECS) and server hardware support are separate.
During PowerFlex installation, depending on the operating system, either the PowerFlex Installation Manager (IM) web client or the PowerFlex VMware Deployment Wizard can be used to configure the Call Home feature, which controls the alert severity threshold used for alert reporting.
VxFlex OS 3.0.1 introduces email alerting using a customer based smtp email server alternative to the
standard (E)SRS service.
|
Yes
StarWind ProActive Support combines both the 'Call Home' functionality and Predictive Analytics.
StarWind ProActive Support service notifies the StarWind Support team about any issues occurring on the node where StarWind Virtual SAN is installed covering storage, networking, compute and software layers. This is achieved by StarWind Agents running on each StarWind node and collecting the metrics from the servers and StarWind software.
|
|
Predictive Analytics
Details
|
Partial
Capacity Management: DataCore SANsymphony Analysis and Reporting supports depletion monitoring of the capacity, complements pool space threshold warnings by regularly evaluating the rate of capacity consumption and estimating when space will be depleted. The regularly updated projections give you a chance to add more storage to the pool before you run out of storage. It also helps you do a better job of capacity planning with fewer surprises. To help allocate costs, especially in private cloud and hosted cloud services, SANsymphony generates reports quantifying the storage resources consumed by specific hosts or groups of hosts. The reports tally several parameters.
Health Monitoring: A combination of system health checks and access to device S.M.A.R.T. (Self-Monitoring, Analysis and Reporting Technology) alerts help to isolate performance and disk problems before they become serious.
DataCore Insight Services (DIS) offers additional capabilites including log-analytics for predictive failure analysis and actionable insights - including hardware.
DIS also provides predictive capacity trend analysis in order to pro-actively warn about licensing limitations being reached within x days and/or disk pools running out of capacity.
|
N/A
|
Partial
StarWind ProActive support is capable of Predictive Analytics by analyzing abnormal patterns on storage, networking, compute and software layers. This allows StarWinds Support team to automatically receive notifications as to possible issues that might occur on the servers.
Additionally, any other pattern that caused issues in one environment is analyzed and integrated into the ProActive Support database allowing the service to notify the StarWind Support team about the same patterns occurring in other clients environments.
|
|