0% found this document useful (0 votes)
35 views

Powerflex Specification Sheet

Spec power flex

Uploaded by

Rubayat Matin
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
35 views

Powerflex Specification Sheet

Spec power flex

Uploaded by

Rubayat Matin
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 16

Specification Sheet

Dell PowerFlex
The ultimate software-defined infrastructure

PowerFlex Family
PowerFlex software-defined infrastructure enables broad consolidation across the data center, encompassing almost any
type of workload and deployment topology. Its software-first architecture enables automation and programmability of the
complete infrastructure stack. It provides scalability, performance, and resiliency, enabling effortless adherence to
stringent workload SLAs. As a universal infrastructure platform, PowerFlex combines compute and high-performance
software-defined storage resources in a managed, unified fabric for both block and file. Available in flexible consumption
options (rack, appliance, custom nodes, or in the public cloud), it enables various deployment architectures: independent
compute and storage (two-layer), HCI (single-layer), or a mixture of the two. PowerFlex is ideal for high performance
applications and databases, building an agile private/hybrid cloud, or consolidating resources in heterogeneous
environments. To learn about the business value and benefits organizations have achieved by using PowerFlex to run
and manage their important business workloads, please read this white paper.

SPECIFICATION SHEET 1
Selected Definitions
System – A PowerFlex system is the collection of entities managed by the Metadata Management (MDM) cluster.

MDM – Metadata Manager. A highly-available storage management cluster that resides alongside other software
components within the system but sits outside the data path and supervises storage cluster health and
configuration. It coordinates rebalancing and rebuilding/reprotecting data as changes occur in the system.

Protection Domain – A protection domain is a logical entity that consists of a group of SDSs that provide data protection
for each other. Each SDS belongs to one (and only one) protection domain. By definition, each protection domain
is a unique set of SDSs. Protection domains can be added during installation and modified post-installation.

Storage Pool - A storage pool is a set of physical storage devices within a protection domain. Each storage device
belongs to one (and only one) storage pool. A volume is distributed over all devices residing in the same storage
pool.

SDS – Storage Data Server. A software service, running on a node that contributes disks to the storage cluster. Working
together, several SDSs abstract local storage, maintain storage pools, and present volumes to the SDCs. Each
SDS node is a fault unit, and the distributed mesh-mirror copies of data are never placed on the same fault unit.

SDC – Storage Data Client. A client kernel driver that provides front-end volume access to operating systems,
applications, or hypervisors. It presents PowerFlex volumes as local block devices. The SDC maintains peer-to-
peer connections to every SDS managing a storage pool. It translates between the proprietary PowerFlex data
transport protocol and block SCSI commands.

Device – Local, direct attached block storage (DAS) in a node that is managed by an SDS and is contributed to a storage
pool.

Volume – Analogous to a LUN, a volume is a subset of a storage pool’s capacity presented by an SDC as a local block
device. A volume’s data is evenly distributed across all disks comprising a storage pool, according to the data
layout selected for that storage pool.

MG – A “medium granularity” data layout on the storage disks comprising a storage pool. This is the original storage pool
option and provides very high performance.

FG - A “fine granularity” data layout on the storage disks comprising a storage pool. This storage pool option is designed
for space efficiency, especially with heavy snapshot use. It requires the use of NVDIMMs and enables
PowerFlex’s inline compression features.

Fault Set – A collection of SDSs that are managed together as a single fault unit. When employed, the distributed mesh-
mirror copies of data are never placed within the same fault set.

SDR – Storage Data Replicator. A software service that lives alongside the SDS and other services and facilitates
asynchronous replication activities between remote PowerFlex systems. The SDR implements journal shipping,
coordinating both the collection of writes into source-side journals and the application of received writes to
volumes on the target side.

SDT – Storage Data Target. Enables NVMe initiator clients to map and use PowerFlex volumes using the NVMe/TCP
protocol. The SDT software service translates between the NVMe and proprietary PowerFlex protocols, supports
discovery services, and manages client host connections.

Specification Sheet 2
System Limits
PowerFlex supports the following system limits in virtue of the software capabilities. Note that reaching some limits will
preclude reaching others. (For example, although the max volume size is 1PB, creating very large volumes will preclude
creating the max number of volumes in a Protection Domain – 32,768 – because the total size of all volumes in a storage
pool is 4PB.) Under some configurations and consumption choices, these limits may differ due to the node, networking
hardware, or management tools being employed.

For complete listing of product limits, look for the Dell PowerFlex 4.5.x Technical Overview at the link provided.

PowerFlex Item Product Limit

System Raw Capacity 16 PB


Minimum: 240 GB, Maximum: 8 TB
Device size (Maximum 15.36 TB for SSDs on medium
granularity storage pools)
Volume Size Minimum: 8 GB, Maximum: 1 PB

Maximum filesystem partitions per volume 15

Maximum total number of volumes and snapshots in system 131,072 a

Maximum total number of volumes and snapshots in protection


32,768
domain
Maximum total number of volumes and snapshots per storage pool 32,768

Maximum number of snapshots per source/root volume 126


160 TB (medium granularity)
Maximum raw capacity per SDS
128 TB (fine granularity)
Maximum SDCs per system 2048

Maximum SDSs per system 512 a

Maximum SDSs per protection domain 128 a

Maximum devices (drives) per SDS server 64 (includes any NVDIMM devices)

Maximum devices per protection domain 8192

Maximum devices per storage pool 300

Total size of all volumes per storage pool 4PB

Maximum volumes that can be mapped to a single SDC 1024

System over provisioning factor 5x net/usable capacity per MG layout

Fine-granularity maximum compression 10x raw capacity

Maximum storage pools per system 1024

Maximum storage pools per protection domain 64

Maximum fault sets per protection domain 64

Maximum Snapshot Policies per system 1000

Specification Sheet 3
PowerFlex Item Product Limit

Maximum number of snapshots a snapshot policy can be defined to


60
retain (not including locked snapshots)
Maximum volumes per local Consistency Group (snapshot) 1024

Maximum number of volume-to-SDC mappings per system 262,143

Maximum user accounts 256

Maximum number of concurrent logged-in management clients


128
(GUI/REST/CLI)
a If more are needed, contact Customer Support

Flexible Deployment Topologies


PowerFlex’s extreme flexibility meets the diverse and rapidly evolving needs of modern enterprises, offering
unprecedented choice for customers to architect their mission-critical IT environments. Mix and match storage, compute,
and HCI nodes in a dynamic deployment, scaling storage and compute resources together or independently, one node at
a time, as needs dictate.

The functional character of a node is determined primarily by the installation/presence of software services running on a
node. However, PowerFlex nodes are configured and purchased as “storage,” “compute,” or “HCI/hyperconverged”
nodes. This reflects the type and quantity of resources in the node, ensuring that resources are suited to the expected
usage. For example, storage nodes have less RAM and compute nodes usually have no capacity disks in them.

Specification Sheet 4
PowerFlex Consumption Options
With PowerFlex, you have choice and flexibility in how you choose to consume the PowerFlex architecture:

• PowerFlex rack is a fully engineered system with integrated networking. It is designed to simplify deployment
and accelerate time to value.

• PowerFlex appliance is a flexible solution with a small starting point and massive scale potential. PowerFlex
appliance provides a broad choice of supported networking with either full or partial network automation.

• PowerFlex custom nodes have the same performance and scale potential but leave the network management
and hardware life-cycling up to the user.

• Public cloud – powering APEX Block Storage. This is a supported software-only deployment of the software-
defined storage layer on recommended compute instances (with attached storage) in Amazon Web Services or
Microsoft Azure. Only the MG data layout and “independent (2-layer)” block architectures are supported. Fault
Sets may be used to distribute the cluster across multiple Availability Zones, thereby improving resiliency even to
disruptions with an AZ. Native asynchronous replication may be used to migrate data between cloud and on-
premises PowerFlex systems, or to establish cloud-based BC/DR data protection schemes.

PowerFlex is also available with OpEx-based consumption options with APEX Custom Solutions. Customers can choose
between APEX Flex on Demand and APEX Datacenter Utility based on their unique requirements.

Node Options and Specifications

PowerFlex R660 PowerFlex R760 PowerFlex R7625 PowerFlex R6625


Chassis 1 RU 2 RU 1 RU
CPU technology 4th Gen Intel Xeon 4th Gen AMD EPYC
CPU sockets Two
CPU cores (total) 16 - 112 32 - 192
CPU frequency 1.8 GHz - 3.7 GHz 2.20 GHz – 4.15 GHz
RAM 256GB - 4TB RDIMM 256GB - 4TB RDIMM

Maximum storage 76TB SAS 154TB* SAS


diskless
capacity (raw TB) 154TB* NVMe 154TB* NVMe

Drive bays 10 x 2.5” 24 x 2.5” diskless

Persistent memory Yes No

Boot solution 480GB (RAID1) “BOSS-N1”

Nvidia GPU options A2, L4 H100, A100, A40, A30, A16, A2, L40, L4 A2, L4

Network Nvidia ConnectX-6 OCP and PCIe


connectivity
Broadcom 57414 OCP and PCIe
(standard 4x25Gb)

Management port iDRAC 9 Out of Band Management

* PowerFlex version 4 or greater required for 154TB, otherwise maximum is 128TB

Specification Sheet 5
PowerFlex R650 PowerFlex R750 PowerFlex R7525 PowerFlex R6525
Chassis 1 RU 2 RU 1 RU
CPU technology 3rd Gen Intel Xeon 3rd Gen AMD EPYC
CPU sockets Two
CPU cores (total) 16 - 80 16 - 128
CPU frequency 2.00 GHz - 3.60 GHz 2.00 GHz - 3.70 GHz
RAM 256 GB - 8 TB 256 GB - 4 TB
76TB SAS 154TB* SAS
Maximum storage
38TB SATA 92TB SATA diskless
capacity (raw TB)
154TB* NVMe 154TB* NVMe
Drive bays 10 x 2.5” 24 x 2.5” diskless

NVDIMM support Yes No

Boot solution 480 GB SATA M.2 (RAID1) “BOSS-S2”

Nvidia GPU options A2, T4 A100, A40, A30, A16, A10, A2, T4, L40 A2, T4

Network Nvidia ConnectX-5 OCP and PCIe


connectivity
Broadcom 57414 OCP and PCIe
(standard 4x25Gb)

Management port iDRAC 9 Out of Band Management

* PowerFlex version 4 or greater required for 154TB, otherwise maximum is 128TB

PowerFlex R640 PowerFlex R740xd PowerFlex R840


Chassis 1 RU 2 RU
CPU technology 2nd Gen Intel Xeon
CPU sockets Two Four
CPU cores (total) 8 - 56 16 - 112
CPU frequency 2.1 GHz - 3.8 GHz 2.1 GHz - 3.8 GHz
RAM 96 GB - 3072 GB 384 GB - 6144 GB
76TB SAS 154TB* SAS
Maximum storage
38TB SATA 92TB SATA
capacity (raw TB)
76TB NVMe 154TB* NVMe
Drive bays 10 x 2.5” 24 x 2.5”

NVDIMM support Yes† Yes

Boot solution 240 GB SATA M.2 (RAID1) “BOSS”

Nvidia GPU options T4 A100, A40, A30, A16, A10, T4 -

Network Mellanox ConnectX-4 rNDC


connectivity Mellanox ConnectX-4
(standard 4x25Gb) Mellanox ConnectX-6

Management port iDRAC 9 Out of Band Management

* PowerFlex version 4 or greater required for 154TB, otherwise maximum is 128TB



R640 does not support both NVMe and NVDIMM together

Specification Sheet 6
Consolidation: OS, Hypervisor, Platform Support

The platform supports a broad range of operating environments – bare metal operating systems, hypervisors, and
container platforms – simultaneously with a unified infrastructure platform and management. By allowing users to flexibly
mix these architectures in a single deployment, PowerFlex enables you to deploy, scale, and evolve all your applications
to meet your business objectives.

Selected OS/Hypervisor Support

PowerFlex Item Product Support


ESXi 6.7, ESXi-7.0
ESXi 7.0 Update 3f (minimum for NVMe/TCP)
Windows Server 2016, 2019, 2022 + Hyper-V
XenServer 7.x CU2 LTSR
Citrix Hypervisor 8.x
RHEL 7.9, 8.4, 8.5, 8.6, 8.7, 8.8, 9.0, 9.1, 9.2
Storage Data Client CentOS 7.9, CentOS Stream 8.x, Stream 9.x
SLES 12 SP5, 15 SP3, 15 SP4, 15 SP5
Oracle Linux 7.9, 8.5, 8.6, 8.7, 8.8, 9.0, 9.1, 9.2 – with RH or UEK Kernels (+ KVM)
IBM AIX 7.2 TL5, IBM AIX 7.3 TL0
Ubuntu 18.04.6 LTS and earlier
Ubuntu 20.04.2 LTS and earlier
RHEL CoreOS (when using PowerFlex SDC container for CSI driver)
ESXi 6.7 and 7.0* (only with PowerFlex Manager – rack and appliance)
RHEL 7.9, 8.4, 8.5, 8.6, 8.7, 8.8, 9.0, 9.1, 9.2
CentOS 7.9, CentOS Stream 8.x, Stream 9.x
SLES 12 SP5, 15 SP3, 15 SP4, 15 SP5
Storage Data Server
PowerFlex EmbeddedOS (Linux)*
Oracle Linux 7.9, 8.5, 8.6, 8.7, 8.8, 9.0, 9.1, 9.2 – with RH or UEK Kernels (+ KVM)
Ubuntu 18.04.6 LTS and earlier
Ubuntu 20.04.2 LTS and earlier
* Only ESXi 7 and the PowerFlex EmbeddedOS are fully managed and life-cycled by PowerFlex Manager

Specification Sheet 7
PowerFlex Software Features and Functions
PowerFlex offers many enterprise data services. For example:

• Snapshots – read/write or read-only snapshots; snapshot scheduling; and secure/immutable snapshots.

• Compression – inline compression is enabled when using the fine-granularity data layout for storage pools.

• Native Asynchronous Replication – PowerFlex includes native async replication capabilities between
PowerFlex clusters – up to 5 in any arbitrary topology. Note: individual volumes are replicated to only 1 target.

PowerFlex Replication Item Product Limit


Number of destination systems for replication 4
Maximum number of SDR per system 128
Maximum number of Replication Consistency Group (RCG) 1024
Maximum number of Volume Pairs per RCG 1024
Maximum replicated Volume Pairs per system 32,000
Maximum number of remote protection domains 8
Maximum number of copies per RCG 1
Recovery Point Objective (RPO) Minimum: 15 seconds, Maximum: 1 hour
Maximum replicated volume size 64 TB

PowerFlex File Services


PowerFlex File Services run on a set of PowerFlex File Controllers. PowerFlex File Controller nodes, also known as File
Nodes, are diskless nodes that are clustered together and host the NAS Server containers. The NAS Servers in turn host
the tenant namespaces, with their individual security policies and file systems. The File Controller Nodes extend the
functionality of a PowerFlex cluster that supplies the underlying block storage. A PowerFlex volume is mapped to each
file system that is presented by the NAS Servers. The volumes, and thus the NAS filesystems, may be dynamically scaled
in the background. In-line compression is supported when the volume storage pool is FG. All major protocols are
supported.

New in 4.5, multiple NAS servers can be brought together under a Global Namespace, allowing for over 8PB of filesystem
space to be shared via SMB (with DFS), NFS v4, or both, though a single IP address and namespace.

PowerFlex Item Product Limit New in 4.5


Maximum NAS cluster size (number of nodes) 16 (must be an even number)
Minimum NAS cluster size (number of nodes) 2
256 TB
Maximum file system size
(Minimum 8 GB)
Maximum number of file systems 4,096 (256*16N) 16,384 (1024*16N)
Maximum number of NAS servers 512 2048
Maximum file snapshots 2,500 57,000
Maximum number of file systems per NAS server 125

Specification Sheet 8
Maximum number of file systems plus mounted snaps per
1,500
NAS server
Maximum NFS servers per system 512
Maximum SMB servers per system 512
Maximum SMB shares per node 10,000
Maximum SMB shares per system 160,000
Maximum NFS exports per node 5,000
Maximum NFS exports per system 80,000
Maximum tree quotas per file system 8,191
Maximum file names per directory 10 million
Maximum sub-directories/files per directory 10 million
Maximum number of home directories 40,000
Maximum SMB TCP connections 128,000
Maximum NFS TCP connections 128,000
Maximum TCP connections per system 153,600
Maximum unique ACLs per file system 4 million
Maximum directories per file system > 10 billion
Maximum open files/directories 512,000
Maximum files per file system 32 billion

PowerFlex File Features

Feature Description
Supported Protocols NFS v3/v4, SMB (CIFS) v2/v3, FTP, SFTP, and NDMP
User quotas and Tree quotas
Extend/shrink file system (space reclaim)
File System Operations
File system read/write snapshots
Single / Global Namespace option
Data Reduction Inline compression when used with FG storage pools
Data Protection 3-way NMDP support for backup
CAVA - Common Antivirus Agent for SMB Clients
Security CEPA - Common Event Publishing Agent (version 4.5+)
D@RE with PowerFlex Enterprise Encryption and KeyStore
SRS/ESE (Call Home)
Serviceability Alerts
Data collection aka “native audit log”
UI and REST API
Management and Monitoring CloudIQ Integration
SNMP v2 and v3 support

Specification Sheet 9
Example of data path communication between clients, NAS servers, and block storage backend.

File Controller Node Options

Component Model CPU Cores RAM(GB) NIC (GbE) Local Storage (GB)

Small Node PowerFlex R650S Intel Xeon 2x5317 2x12 (24) 128 4 x 25 480GB BOSS M.2

Medium Node PowerFlex R650M Intel Xeon 2x6346 2x16 (32) 256 4 x 25 480GB BOSS M.2

4 x 25
Large Node PowerFlex R650L Intel Xeon 2x6348 2x28 (56) 256 or 480GB BOSS M.2
4 x 100

Data Access Protocols


In addition to the file access protocols, listed above, PowerFlex supports two block protocols. The primary transport
protocol is a proprietary TCP-based protocol that efficiently moves data between the Storage Data Servers (SDSs) and
Storage Data Clients (SDCs), as well as among the contributing SDSs. The architecture includes native multipathing
between the SDC and all SDSs that host volume data. The SDC translates this to a subset of the standard SCSI
commands, for consumption by operating systems, hypervisors, and applications that can access raw block devices.

Specification Sheet 10
Example of SDC – SDS communication with SDC installed in ESXi.

PowerFlex 4.0 also introduced support for NVMe/TCP, allowing for the consumption of PowerFlex volumes without
installing the proprietary kernel driver. Support for NVMe/TCP is facilitated by the Storage Data Target (SDT) service,
which runs on nodes also running the SDS service. The SDT translates between the system’s native PowerFlex protocol
and NVMe commands. It also functions as a discovery service for client initiators.

NVMe/TCP requires kernels that contain native support for the protocol. In VMware, this is ESXi 7.0 Update 3f or later. It
is also available as a Tech Preview in supported Linux Distributions: RHEL 8.6 and later, SLES 15 SP3 and later, Ubuntu
22.04.

Example of NVMe/TCP communication with PowerFlex storage with ESXi.

Specification Sheet 11
NVMe/TCP Limits

PowerFlex Item Product Limit


Maximum volumes mapped to a single NVMe host (Linux) 1024
32 (in ESXi 7.0)
Maximum volumes mapped to a single NVMe host (ESXi)
256 (in ESXi 8.0)
Maximum NVMe hosts connected to system 1024 (included in total SDCs per system)
Maximum SDTs per protection domain 128
Minimum SDTs per protection domain 2*
Maximum SDTs per system 512
8 (in ESXi 7.0u3)
Maximum paths in multipathing driver per volume
32 (in ESXi 8.0u1)
Maximum connections per host per protection domain 16
Maximum NVMe host connections (I/O controllers) per SDT 512
Maximum NVMe host connections (I/O controllers) per
65,519
system
Maximum I/O controller queue depth 128†
Maximum I/O controller queues 32†
Maximum volume-to-host mappings (SDC/NVMe) per system 262,143
* Using minimum SDTs may block the ability to reach maximum NVMe hosts.
† Number of queues + queue depth is automatically negotiated on connection.

PowerFlex Manager (PFxM)


PowerFlex Manager is the M&O software layer that further enables ITOM automation and LCM capabilities for hardware
and networking. Starting with PowerFlex 4.0, the unified PowerFlex Manager brings together three separate components
used in previous releases – PowerFlex Manager, the core PowerFlex UI, and the PowerFlex gateway UI. The new
PowerFlex Manager and UI runs as containerized services in a distributed Kubernetes platform

PowerFlex Manager offers standards-based open APIs and custom Ansible modules, making it simple to integrate with
third party tools and custom workflows. Further, when paired with Dell CloudIQ, PowerFlex leverages an AI/ML-based
approach to infrastructure monitoring and management, ensuring simplicity and consistency at scale.

Specification Sheet 12
PowerFlex Clustering, Scaling and Management

Min Nodes Per Cluster 4 Storage Only nodes minimum (6 or more recommended)
(Two-layer Configuration) 1 to 3 Compute Only nodes (depending on host OS)

Min Nodes Per


Cluster (HCI 4 HCI Nodes minimum (6 or more recommended)
Configuration)

Scaling Increments 1 Node (HCI, Compute Only or Storage Only) †

EmbeddedOS Jump Server 16GB RAM, 4 vCPU, 500GB storage


PowerFlex Secure Connect Gateway 4GB RAM, 2 vCPU, 16GB storage
Management Node PowerFlex Management VMs (3x) 32GB RAM, 16 vCPU, 650GB storage (each)
PowerFlex Enterprise Encryption and
Requirements‡ KeyStore (optional) 6GB RAM, 4 vCPU, 64GB storage
(Supplied as virtual machine images)

* In 2-layer environments where existing compute nodes are to be utilized or compute nodes are running an operating system not
supported by PowerFlex Manager, the minimum requirement is for four storage nodes only.

A single node is the minimum scaling required to expand an existing Storage Pool. Creation of a net new Storage Pool requires the
additionof a minimum of 3 Storage or HCI Nodes.

New PowerFlex appliance deployments include a single-node management controller (with an option for three-node for larger
systems). New PowerFlex integrated rack deployments include a three-node or four-node management controller cluster. These
PowerFlex Management Controller options are ESXi based.

PowerFlex Manager: Supported Switches

PowerFlex Manager Supported Switches


Cisco Nexus 3172TQ, Cisco Nexus 31108TC-V, Cisco Nexus 92348GC-X,
Management Switches*
Dell S4148T-ON
Cisco Nexus 3132QX, Cisco Nexus 3164Q, Cisco Nexus 93180YC-EX,
Cisco Nexus 93180YC-FX and 93180YC-FX3, Cisco Nexus 93240YC-FX2, Cisco Nexus
Access or Leaf Switches
N93360YC-FX2, Dell S5048F-ON, Dell S5248F-ON, Dell S5296F-ON‡, Dell S5224F-ON‡,
Dell S4148F-ON‡

Cisco Nexus 9236C, Cisco Nexus 9336C-FX2, Cisco Nexus 9364C-GX, Cisco Nexus 9364C-GX,
Aggregation or Spine Switches
Dell S5232F-ON

* For PowerFlex appliance, the management switch can be “bring your own”.

Appliance only

Specification Sheet 13
Power and Dimensions

PowerFlex R660 PowerFlex R760 PowerFlex R6625 PowerFlex R7625


700 W Titanium 700 W Titanium
700 W Titanium 700 W Titanium
800 W Platinum 800 W Platinum
Fully redundant 800 W Platinum 800 W Platinum
1400 W Platinum 1400 W Platinum
power supplies 1100 W Titanium 1100 W Titanium
1800 W Titanium 1800 W Titanium
(100-240Vac) 1400 W Platinum 1400 W Platinum
2400 W Platinum 2400 W Platinum
1800 W Titanium 1800 W Titanium
2800 W Titanium 2800 W Titanium
Redundant
8 6 8 6
cooling fans
H 42.8 mm 86.8 mm 42.8 mm 86.8 mm
Physical W 434 mm 434 mm 434 mm 434 mm
dimensions D 823 mm 772 mm 751 mm 700 mm
Wgt 21.2 kg 35.3 kg 21.2 kg 24.6 kg

PowerFlex R650 PowerFlex R750 PowerFlex R6525 PowerFlex R7525


800 W 800 W 800 W
Fully redundant 1100 W
1100 W 1100 W 1100 W
power supplies 1400 W
1400 W 1400 W 1400 W
(100-240Vac) 2400 W
1100 W (48-60Vdc) 2400 W 1100 W (48-60Vdc)

Redundant
8 6 8 6
cooling fans
H 42.8 mm 86.8 mm 42.8 mm 86.8 mm
Physical W 434 mm 434 mm 434 mm 434 mm
dimensions D 751 mm 700 mm 751 mm 700 mm
Wgt 21.2 kg 35.3 kg 21.2 kg 24.6 kg

PowerFlex R640 PowerFlex R740xd PowerFlex R840


750 W 1100 W
Fully redundant 1600 W
1100 W 1600 W
power supplies 2000 W
1600 W 2000 W
(100-240Vac) 2400 W
1100 W (48Vdc) 2400 W

Redundant
8 6 6
cooling fans
H 42.8 mm 86.8 mm 86.8 mm
Physical W 434 mm 434 mm 434 mm
dimensions D 734 mm 679 mm 679 mm
Wgt 21.9 kg 28.1 kg 28.1 kg

Specification Sheet 14
Environmental and Certificates

PowerFlex R660 PowerFlex R760 PowerFlex R6625 PowerFlex R7625


Ambient operating
10°C to 35°C 10°C to 35°C 10°C to 35°C 10°C to 35°C
temperature (A2)
Storage
-40°C to 65°C -40°C to 65°C -40°C to 65°C -40°C to 65°C
temperature range
Operating relative
humidity (non- 8% to 80% 8% to 80% 8% to 80% 8% to 80%
condensing)
Operating altitude
3048m 3048m 3048m 3048m
with no deratings

PowerFlex R650 PowerFlex R750 PowerFlex R6525 PowerFlex R7525


Ambient operating
10°C to 35°C 10°C to 35°C 10°C to 35°C 10°C to 35°C
temperature (A2)
Storage
-40°C to 65°C -40°C to 65°C -40°C to 65°C -40°C to 65°C
temperature range
Operating relative
humidity (non- 8% to 80% 8% to 80% 8% to 80% 8% to 80%
condensing)
Operating altitude
3048m 3048m 3048m 3048m
with no deratings

PowerFlex R640 PowerFlex R740xd PowerFlex R840


Ambient operating
10°C to 35°C 10°C to 35°C 10°C to 35°C
temperature (A2)
Storage
-40°C to 65°C -40°C to 65°C -40°C to 65°C
temperature range
Operating relative
humidity (non- 10% to 80% 10% to 80% 10% to 80%
condensing)
Operating altitude
3048m 3048m 3048m
with no deratings

Specification Sheet 15
Statement of Compliance
Dell Information Technology Equipment is compliant with all currently applicable regulatory requirements for
Electromagnetic Compatibility, Product Safety, and Environmental Regulations where placed on market.

Detailed regulatory information and verification of compliance is available at the Dell Regulatory Compliance website.
https://www.dell.com/REGULATORY_COMPLIANCE

Learn more about Contact us for Follow us for


PowerFlex and visit feedback and PowerFlex
the Info Hub requests news

© 2024 Dell Inc. or its subsidiaries. All Rights Reserved. Dell Technologies, Dell, EMC, Dell EMC and
Specification Sheet
other trademarks are trademarks of Dell Inc. or its subsidiaries. Other trademarks may be trademarks 16
of their respective owners.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy