# Grid Storage

Grid Storage offered by EGI HTC providers

## What is it?

Grid storage enables storage of files in a fault-tolerant and scalable environment, and sharing it with distributed teams. Your data can be accessed through multiple protocols, and can be replicated across different providers to increase fault-tolerance. Grid storage gives you complete control over what data you share, and with whom you share the data.

The main features of grid storage:

• Access highly-scalable storage from anywhere
• Control the data you share
• Organise your data using a flexible, hierarchical structure

Grid storage file access is based on the gridFTP and WebDav/HTTP protocols, together with XRootD and legacy SRM (under deprecation at some of the endpoints).

Several grid storage implementations are available in the EGI Cloud, the most common being:

## Endpoint discovery

The grid storage endpoints that are available to a user’s Virtual Organizations are discoverable via the EGI Information System (BDII).

The lcg-infosites command can be used to obtain VO-specific information on existing grid storages, using the following syntax:

$lcg-infosites --vo voname -[v] -f [site name] [option(s)] [-h| --help] [--is BDII] For example, to list the Storage Elements (SEs) available to the biomed VO, we could issue the following command:$ lcg-infosites --vo biomed  se

Avail Space(kB)  Used Space(kB)  Type  SE
------------------------------------------
280375465082             n.a  SRM   ccsrm.ihep.ac.cn
10995116266              11  SRM   cirigridse01.univ-bpclermont.fr

## Access from the command line

Access to grid storage via a command line interface (CLI) requires users to obtain a valid X.509 user VOMS proxy. Please refer to the Check-in documentation for more information.

The CLI widely used to access grid-storage is gfal2, which is available for installation both on RHEL and Debian compatible systems.

In particular, gfal2 provides an abstraction layer on top of several storage protocols (XRootD, WebDAV, SRM, gsiftp, etc), offerint a convenient API that can be used over different protocols.

The gfal2 CLI can be installed as follows (for RHEL compatible systems):

domatest
gb
SSE-demo
test
tpctest

$gfal-mkdir gsiftp://dcache-door-doma01.desy.de/dteam/test ### Copy a local file$ gfal-copy test.json gsiftp://dcache-door-doma01.desy.de/dteam/test
Copying file:///root/Documents/test.json   [DONE]  after 0s

$gfal-copy gsiftp://prometheus.desy.de/VOs/dteam/public-file gsiftp://dcache-door-doma01.desy.de/dteam/test Copying gsiftp://prometheus.desy.de/VOs/dteam/public-file [DONE] after 3s ### Download a file to a local folder$ gfal-copy gsiftp://prometheus.desy.de/VOs/dteam/public-file /tmp
Copying gsiftp://prometheus.desy.de/VOs/dteam/public-file   [DONE]  after 0s

### Delete a file

\$ gfal-rm gsiftp://dcache-door-doma01.desy.de/dteam/test/public-file
gsiftp://dcache-door-doma01.desy.de/dteam/test/public-file      DELETED

## Access via EGI Data Transfer

The EGI Data Transfer service provides mechanisms to optimize the transfer of files between EGI Online Storage endpoints. Both a graphical user interface (GUI) and command line interfaces (CLI) are available to perform bulk movement of data. Please check out the related documentation for more information.

## Integration with Data Management frameworks

Grid storage access, most of the time, is hidden from users by the integration with the Data Management Frameworks (DMFs) used by Collaborations and Experiments.

For example, the EGI Workload Manager provides a way to efficiently access grid storage endpoints in order to read/store files, and to catalogue the existing file and related metadata.

When running computation via the EGI Workload Manager, users do not actually access the storage directly. However, users can retrieve the output of the computation once it has been stored on the grid.