He Leading Parallel Luster File Sy Stem: WWW - Beegfs.io
He Leading Parallel Luster File Sy Stem: WWW - Beegfs.io
www.thinkparq.com
www.beegfs.io
ABOUT BEEGFS
What is BeeGFS Sy stem Architecture
BeeGFS offers maximum performance and scalability on various BeeGFS supports a w ide range of Linux distributions such
levels. It supports distributed file contents with flexible as RHEL/Fedora, SLES/OpenSuse or Debian/Ubuntu as well as a
striping across storage servers on a per-file or per-directory basis wide range of Linux kernels from ancient 2.6.18 up to the
as well as distributed metadata. latest vanilla.
BeeGFS is optimized especially for use in environments w here The storage services run on top of an ex isting local
performance matters to provide: filesy stem (such as xfs, zfs or others) using the normal POSIX
interface and clients and servers can be added to an ex isting
� Best in class client throughput: 8 GB/s with only a single sy stem w ithout dow ntime.
process streaming on a 100GBit network, while a few
streams can fully saturate the network. BeeGFS supports multiple netw orks and dynamic failover in
� Best in class metadata performance: Linear scalability case one of the network connections is down.
through dynamic metadata namespace partitioning.
� Best in class storage throughput: BeeGFS servers allow
flexible choice of underlying file system to perfectly fit the
given storage hardware.
BeeGFS client and server components can also run on the same
physical machines. Thus, BeeGFS can turn a compute rack
into a cost-efficient conv erged data processing and
shared storage unit, eliminating the need for external storage
resources and providing simplified management.
BEEGFS ON DEMAND
Max imum Usability BeeOND
The BeeGFS serv er components are userspace daemons, BeeOND (BeeGFS on demand) allows on the fly creation of a
while the client is a nativ e kernel module that does not complete parallel file system instance on a given set of hosts
require any patches to the kernel itself. All BeeGFS components with just one single command.
can be installed and updated without even rebooting the
machine.
BeeGFS was designed for extreme scalability. In a testbed1 In the same testbed system with 20 servers, each equipped
with 20 servers and up to 640 client processes (32x the with a single node local performance of 1332 MB/s (write) and
number of metadata servers), BeeGFS delivers a sustained file 1317 MB/s (read), and 160 client processes, BeeGFS
creation rate of more than 500,000 creates per second, demonstrates linear scaling to a sustained throughput of 25
making it possible to create one billion files in as little time as GB/s - which is 94.7% of the maximum theoretical local write
about 30 minutes. and 94.1% of the maximum theoretical local read throughput.
539.7 25.2
600 30
read 24.8
creates
write
25
500
Thousand Creates / second
20
400
GB/s 15
300
10
200
5
100
20
20 18
18 16
16 14
14 12
12 0 10
0 10
8 8
6 6
4 4
1
2 # Metadata Servers 2 # Storage Servers
1
1
Bechmark Sy stem:
20 servers with 2x Intel Xeon X5660 @ 2.8 GHz, 48GB RAM running a Scientific Linux 6.3, Kernel 2.6.32-
279. Each server is equipped with 4x Intel 510 Series SSD (RAID 0) running ext4 as well as QDR
Infiniband. Tests performed using BeeGFS version 2012.10.
LICENSING MODEL
Free to use & Open source User Comments
Free Updates
“Now under heavy load, our large BeeGFS system
is performing well - bioinfo users are seeing >2x
Ready-to-use binary packages & speedup in their apps, with no more hotspots
Source Code
when hitting common index files. Unlike the
previous system, BeeGFS does not lock up under
Enterprise Features (High-availability, heavy load and even under heavy load,
quota, ACLs) interactive use remains zippy. The only complaint
is how long it's taking us to move data off the
Next Business Day Service Level old system.“
Agreements for Support Harry Mangalam, UC Irvine, USA
Life Sciences
ThinkParQ GmbH
Trippstadter Str. 110 Phone: +49 631 277 576 300 www.thinkparq.com
67663 Kaiserslautern, Germany sales@thinkparq.com www.beegfs.io