Handling large collections of files: Difference between revisions

Jump to navigation Jump to search
no edit summary
No edit summary
No edit summary
Line 1: Line 1:
{{Draft}}
{{Draft}}


In certain domains it is common to have to manage very large collections - meaning hundreds of thousands or more - of files, which individually are often though not always fairly small, e.g. less than a few hundred kilobytes. In these cases, a problem naturally arises from storing such data on Compute Canada clusters due to the filesystem quotas that limit the number of distinct filesystem objects to 500K for the project space (by default) and 1M for the scratch space in most instances. So how can a user or group of users store these necessary data sets on the cluster? In this page we will present a variety of different solutions and workarounds, each of which has its own pros and cons, and allow you as a reader to judge for yourself which is the optimal approach for you.  
In certain domains, notably machine learning, it is common to have to manage very large collections of files, meaning hundreds of thousands or more.  The individual files may be fairly small, e.g. less than a few hundred kilobytes. In these cases, a problem arises due to [[/Storage_and_file_management#Filesystem_quotas_and_policies|filesystem quotas]] on Compute Canada clusters that limit the number of filesystem objects. So how can a user or group of users store these necessary data sets on the cluster? In this page we will present a variety of different solutions, each with its own pros and cons, so you may judge for yourself which is an appropriate one for you.  
<!-- This text should not appear -->
<!-- This text should not appear -->


Bureaucrats, cc_docs_admin, cc_staff
2,879

edits

Navigation menu