Using CVMFS » History » Version 2
Welcome to MINOS CVMFS!
CERN Virtual Machine File System is a distributed disk system for providing an experiment's code and libraries to interactive
node and grids worldwide. It is used by CMS and Atlas and well as most experiments at FNAL.
The code manager copies a code release to a CVMFS work space and "publishes" it. This process examines the code, compresses it,
and inserts it in a database. The original database is called the tier 0 copy. Remote sites may support tier 1 copies of the
database, synced to the tier 0.
The user's grid job sees a CVMFS disk mounted and containing a copy of the experiment's code, which can be accessed in any way
the code would be accessed on a standard disk. The disk is actually a custom nfs server with a small ( ~8 GB) local cache on
the node and a backend that sends file requests to a squid web cache. The squid may get its data from the tier 1 database, if
available, or from the tier 0. As a practical matter, most grid jobs do not access much in a release, usually just a small set
of shared object libraries, and these end up cached on the worker node, or on the squid, thereby avoiding a long-distance
CVMFS is efficienct only for distributing code and small data files which are required by a large number of nodes on the grid.
On the other hand, datasets, such as event data files, are many files which are each sent to only one node during a grid job.
CVMFS is not efficienct for this type of data distribution or for this sort of data volume. Data files should be distributed
through dCache, which is designed to deliver each file to one node, and to handle the data volume. A single large file which
is to be distributed to all nodes also need to be avoided since it would would churn or overflow the small local caches.
Examples of this sort of file are the Genie flux files or a large analysis fit template library.
The limitations of CVMFS can be found here (don't worr too much about this unless you maintain CVMFS)