Using nearline storage

From Alliance Doc
Jump to navigation Jump to search
Other languages:

Nearline is a tape-based filesystem intended for inactive data. Datasets which you do not expect to access for months are good candidates to be stored in /nearline.

Restrictions and best practices

Size of files

Retrieving small files from tape is inefficient, while extremely large files pose other problems. Please observe these guidelines when storing files in /nearline:

  • Files smaller than ~10GB should be combined into archive files (tarballs) using tar or a similar tool.
  • Files larger than 4TB should be split in chunks of 1TB using the split command or a similar tool.
  • DO NOT SEND SMALL FILES TO NEARLINE, except for indexes (see Create an index below).

Do not compress your data

There is no need to compress the data in /nearline since the tape archive system automatically performs compression using specialized circuitry. If your data is already compressed, it can be copied to /nearline without any issues.

Use tar or dar

Use tar or dar to create an archive file.

Keep the source files in their original filesystem. Do NOT copy the source files to /nearline before creating the archive.

Create the archive directly in /nearline. This does not require additional storage space and is more efficient than creating the archive in /scratch or /project and copying it to /nearline later.

If you have hundreds of gigabytes of data, the tar options -M (--multi-volume) and -L (--tape-length) can be used to produce archive files of suitable size. If you are using dar, you can similarly use the -s (--slice) option.

Create an index

When you bundle files, it becomes inconvenient to find individual files. To avoid having to restore an entire large collection from tape when you only need one or a few of the files from this collection, you should make an index of all archive files you create. Create the index as soon as you create the collection. For instance, you can save the output of tar with the verbose option when you create the archive, like this:

Question.png
[name@server ~]$ tar cvvf /nearline/def-sponsor/user/mycollection.tar /project/def-sponsor/user/something > /nearline/def-sponsor/user/mycollection.index

If you've just created the archive (again using tar as an example), you can create an index like this:

Question.png
[name@server ~]$ tar tvvf /nearline/def-sponsor/user/mycollection.tar > /nearline/def-sponsor/user/mycollection.index

Index files are an exception to the rule about small files on /nearline: it's okay to store them in /nearline.

No access from compute nodes

Because data retrieval from /nearline may take an uncertain amount of time (see How it works below), we do not permit reading from /nearline in a job context. /nearline is not mounted on compute nodes.

Use a data-transfer node if available

Creating a tar or dar file for a large volume of data can be resource-intensive. Please do this on a data-transfer node (DTN) on clusters that have them. On clusters that have no DTN nodes, use a login node.

Use a terminal multiplexer

Archiving large file collections can take several hours or even days. Your SSH session might be interrupted before your program finishes, or you might want to close your session, keep your program running in the background, and come back to it later. Run tar or dar in a terminal multiplexer such as tmux to manage these issues.

Use dar in non-interactive mode

When used in a terminal, dar runs in interactive mode and asks to confirm certain operations. When used without an attached terminal, dar runs in non-interactive mode and assumes a “no” answer to any questions. We recommend to explicitly disable interactivity with dar -Q. This is especially useful when running dar in an unattended terminal multiplexer. See Dar for additional information.

Why /nearline?

Tape as a storage medium has these advantages over disk and solid-state (SSD) media.

  1. Cost per unit of data stored is lower.
  2. The volume of data stored can be easily expanded by buying more tapes.
  3. Energy consumption per unit of data stored is effectively zero.

Consequently we can offer much greater volumes of storage on /nearline than we can on /project. Also, keeping inactive data off of /project reduces the load and improves its performance.

How it works

  1. When a file is first copied to (or created on) /nearline, the file exists only on disk, not tape.
  2. After a period (on the order of a day), and if the file meets certain criteria, the system will copy the file to tape. At this stage, the file will be on both disk and tape.
  3. After a further period the disk copy may be deleted, and the file will only be on tape.
  4. When such a file is recalled, it is copied from tape back to disk, returning it to the second state.

When a file has been moved entirely to tape (that is, when it is virtualized) it will still appear in the directory listing. If the virtual file is read, it will take some time for the tape to be retrieved from the library and copied back to disk. The process which is trying to read the file will block while this is happening. This may take from less than a minute to over an hour, depending on the size of the file and the demand on the tape system.

Transferring data from /nearline

While transferring data with Globus or any other tool, the data that was on tape gets automatically restored on disk upon reading it. Since tape access is relatively slow, each file restoration can hang the transfer for a few seconds to a few minutes. Therefore, users should expect longer transfer times from /nearline.

For an overview of the state of all files saved on /nearline, some clusters support the following command:

Question.png
[name@server ~]$ diskusage_report --nearline --per_user --all_users

The different Location's are:

  • On disk and tape: this data is available on disk.
  • Modified, will be archived again: the newest version of the data is on disk.
  • Archiving in progress: the data is being copied or moved to tape.
  • On tape: the data is only on tape.

Then, you can determine whether or not a given file has been moved to tape or is still on disk using the lfs hsm_state command. "hsm" stands for "hierarchical storage manager".

#  Here, <FILE> is only on disk.
$ lfs hsm_state <FILE>
<FILE>:  (0x00000000)

# Here, <FILE> is in progress of being copied to tape.
$ lfs hsm_state <FILE>
<FILE>: [...]: exists, [...]

# Here, <FILE> is both on the disk and on tape.
$ lfs hsm_state <FILE>
<FILE>: [...]: exists archived, [...]

# Here, <FILE> is on tape but no longer on disk.  There will be a lag when opening it. 
$ lfs hsm_state <FILE>
<FILE>: [...]: released exists archived, [...]

You can explicitly force a file to be recalled from tape without actually reading it with the command lfs hsm_restore <FILE>.

Cluster-specific information

/nearline is only accessible as a directory on login nodes and on DTNs (Data Transfer Nodes).

To use /nearline, just put files into your ~/nearline/PROJECT directory. After a period of time (24 hours as of February 2019), they will be copied onto tape. If the file remains unchanged for another period (24 hours as of February 2019), the copy on disk will be removed, making the file virtualized on tape.

If you accidentally (or deliberately) delete a file from ~/nearline, the tape copy will be retained for up to 60 days. To restore such a file contact technical support with the full path for the file(s) and desired version (by date), just as you would for restoring a backup. Note that since you will need the full path for the file, it is important for you to retain a copy of the complete directory structure of your /nearline space. For example, you can run the command ls -R > ~/nearline_contents.txt from the ~/nearline/PROJECT directory so that you have a copy of the location of all the files.

/nearline service similar to that on Béluga.

/nearline service similar to that on Béluga.

/nearline service similar to that on Béluga.

HPSS is the /nearline service on Niagara.
There are three methods to access the service:

1. By submitting HPSS-specific commands htar or hsi to the Slurm scheduler as a job in one of the archive partitions; see the HPSS documentation for detailed examples. Using job scripts offers the benefit of automating /nearline transfers and is the best method if you use HPSS regularly. Your HPSS files can be found in the $ARCHIVE directory, which is like $PROJECT but with /project replaced by /archive.

2. To manage a small number of files in HPSS, you can use the VFS (Virtual File System) node, which is accessed with the command salloc --time=1:00:00 -pvfsshort. Your HPSS files can be found in the $ARCHIVE directory, which is like $PROJECT but with /project replaced by /archive.

3. By using Globus for transfers to and from HPSS using the endpoint computecanada#hpss. This is useful for occasional usage and for transfers to and from other sites.