July 7, 2017.
On Wednesday, July 12, the Livermore Computing Data Storage Group will release a new version of HTAR on all open (OCF) and secure (SCF) production machines. HTAR—a utility that combines a flexible file-bundling tool with fast parallel access to high performance storage—allows users to store and selectively retrieve very large sets of files efficiently. This update to HTAR allows specifying entries to exclude during creation of new archive files.


FTP (File Transfer Protocol) is an industry-standard protocol and user interface for transferring files between computer systems by means of a series of interactive commands. FTP involves a local client (software you execute to send or receive files) and a remote server (software elsewhere that responds to instructions from your client to accept or deliver files).

HTAR ("HPSS Tape Archiver") is an LC-designed TAR-like utility program that makes TAR-compatible archive (library) files but with High Performance Storage System (HPSS) support and enhanced archive-management features.

This manual provides a basic guide for generating output and transferring files on LC computers. It outlines naming and locating tips for local network printers and introduces local tools (such as LISTPQ) for discovering and taking advantage of LC's standardized print-queue names.

Both TapeFIS and OWL (one-way link) FIS are available on the iSNSI network. iSNSI TapeFIS is bidirectional and requires Derivative Classifier (DC) intervention for file transfers from Pinot to the OCF. Login examples are provided below. OWL  transfers are significantly faster but are unidirectional for transfers from CZ and RZ to the SCF.

This page includes a table that helps determine what is needed to access classified (SCF), unclassified Collaboration Zone (OCF-CZ), and unclassified Restricted Zone (OCF-RZ) Livermore Computing (LC) systems.

IBM Spectrum LSF is a batch scheduler that allows users to run their jobs on Livermore Computing’s (LC) Sierra (link is external) (CORAL) high performance computing (HPC) clusters. IBM Cluster System Management (CSM) is the resource manager for the Sierra systems.

At Lawrence Livermore National Laboratory and increasingly more places, very large displays are being used in theatres for scientific visualization. These displays contain thousands of pixels in both X and Y, which combine to total several million pixels. To utilize these displays, very large movies are created. These movies, which are greater than about 1 million pixels, may play back unevenly due to the inability to move them off disk fast enough to render at the optimal frame rate, usually 30 fps.

In addition to our Tutorials, users of Livermore Computing (LC) systems have access to several types of user information and documentation. Questions may be directed to the LC Hotline via e-mail or via telephone at (925) 422-4532. 

Slurm is a combined batch scheduler and resource manager that allows users to run their jobs on Livermore Computing’s (LC) high performance computing (HPC) clusters.  This document describes the process for submitting and running jobs under the Slurm Workload Manager.