This Site Is Depreciated
  • This is the depreciated Version of Veeam Backup & Replication Best Practices
  • Contacting Veeam Software
  • DNS Resolution
  • Veeam Backup Server
    • Deployment Method
    • Backup Server Placement
    • Sizing and System Requirements
    • Veeam Backup & Replication Database
    • Protecting Veeam Backup & Replication Configuration
  • Veeam Enterprise Manager
    • vCloud Director Self Service Portal
  • Search Server and Indexing
  • Proxy Servers - Introduction
    • Proxy - VMware vSphere
      • Transport Modes
        • Direct Storage Access
        • Virtual Appliance Mode
        • Network Mode
        • Backup from Storage Snapshots
          • NetApp Data ONTAP integration
          • Nimble Storage integration
        • Selecting a Transport Mode
      • Sizing a Backup Proxy
    • Proxy - Microsoft Hyper-V
    • Proxy - Nutanix AHV
  • Primary and secondary Storage BPs
    • HPE 3PAR VMs disks considerations
    • DellEMC Data Domain advanced scalability
  • Backup Repository
    • Repository Types
      • SMB
      • Deduplication Appliances
      • Integration specifics
      • Windows Server Deduplication
      • Object Storage
    • Repository Planning
      • Sizing
      • Per VM Backup Files
      • Scale-out Backup Repository
        • Capacity Tier
    • Repository HA
  • WAN Acceleration
    • Analyzing Wan Acceleration Workload
    • Comparing WAN Acceleration Modes
    • Sizing For WAN Acceleration
    • Sizing Targets for WAN Accereration Relationship
    • Deployments For WAN Acceleration
    • Is WAN Acceleration Right For me
  • Tape Support
    • Tape Deployments
    • Tape Media Information
    • Tape Config Requirements
    • Tape Parallel Processing
    • Tape Virtual Full
    • Tape Writing to Tape
    • Tape Restores
  • Veeam Explorers
  • Interaction with VMware vSphere
  • Interaction with Microsoft Hyper-V
  • Job Configuration
    • Backup Methods
    • Encryption
    • Deduplication and Compression
    • Backup Job
    • Backup Copy Job
    • Replication Job
    • Application-Aware Image Processing
  • Data Verification Using Virtual Labs
  • Overview of Applications Support
    • Active Directory
    • Microsoft Exchange
    • Microsoft SQL Server
    • Microsoft SharePoint Server
    • Oracle Database
    • MySQL
    • IBM Notes/Domino
    • SAP HANA
  • POC Guide
    • Assessment
    • Accelerated Evaluation
    • Enhanced Evaluation
      • Workshop Example
      • Preparation
      • Automation
  • Infrastructure Hardening
    • Segmentation using Zones
    • Hardening Backup Repository - Linux
    • Hardening Backup Repository - Windows
  • Backup & Replication Anatomy
    • Backup
    • VM Restore
    • Instant VM Recovery
    • Windows File-Level Restore
    • Replication
  • Networking Diagrams
    • Backup Server
    • Proxy Server
    • Repository Server
    • Storage Integration
    • Data Validation
  • Application-aware Image Processing
  • Enterprise Manager
  • Sizing Summary
Powered by GitBook
On this page
  • 3PAR useful facts
  • Additional resources

Was this helpful?

  1. Primary and secondary Storage BPs

HPE 3PAR VMs disks considerations

PreviousPrimary and secondary Storage BPsNextDellEMC Data Domain advanced scalability

Last updated 4 years ago

Was this helpful?

Restoring a VM with thin disks leads to a painful slow restore on HPE 3PAR arrays. Usually this condition happens only when restoring backups taken from another infrastructure to a new one, based on HPE 3PAR. Leaving restore options with all the default value, the original format of the virtual disks is kept and if the restored VMs had originally thin provisioned disks, it will be restored with the same format. HPE best practices’ regarding 3PAR is to provision VMs using think disks.

3PAR useful facts

3PAR has a built-in chip, ASIC, used for features like zero-detection, RAID calculation and dedupe/compression:

ASIC can eliminate zeros inline, avoiding useless writes on a virtual volume (LUN):

When a thin-provisioned VM is deleted, the underlying array is not informed that the data is no longer in use (this is an expected behavior in vSphere – that’s how it works) and, when writes occur, the previously allocated blocks must be first purged. There is a manual procedure, described here, to manually trigger the unmap of these blocks. What is not working as expected with 3PAR arrays is that when a write in thin provision mode occurs on previously written blocks, ASIC cannot intercept zeroes and an unexpected data format leads to poor write performances. The same condition would not happen if you restore thin provisioned disks on newly created 3PAR virtual volume.

When restoring VMs on a 3PAR based infrastructure, force the restore in thick (lazy zeroed works as well – and will also save some time) if source VM were created with thin disks. All the zeroes will be automatically detected, skipped and never written. There will much less writes and less I/O flowing to the array.

Additional resources

Why the right architecture matters with vSphere
HPE 3PAR StoreServ Storage and VMware vSphere 6.5 best practices
HPE 3PAR VMware ESX/ESXi Implementation Guide
HPE 3PAR StoreServ Architecture
3par asic
3par asic