|
|
(54 intermediate revisions by 7 users not shown) |
Line 1: |
Line 1: |
| <languages /> | | <languages /> |
| <translate> | | <translate> |
| = Report an issue = <!--T:1-->
| |
| * Please report issues to the [[Technical Support|technical support]] team.
| |
|
| |
|
| = Shared issues = <!--T:2-->
| | <!--T:15--> |
| * The status page at http://status.computecanada.ca/ is not updated automatically yet, so may lag in showing current status.
| | Problems that affect many users and are being investigated, such as a cluster-wide malfunction or outage, are on [https://status.alliancecan.ca/ the Alliance Status page]. This "Known issues" page describes problems that affect many users but that may take some time to repair, or are not planned for repair at this time. Problems that only affect a specific software package are described on the wiki page for that software package. |
|
| |
|
| == Scheduler issues == <!--T:6--> | | ==Report an issue== <!--T:1--> |
| * Interactive jobs started via <tt>salloc</tt> require X11 forwarding to be enabled when you connect to the cluster. On Linux and MacOS, you can typically enable X11 forwarding by adding the option <tt>-X</tt> or <tt>-Y</tt> to your <tt>ssh</tt> command.
| | Please report issues to the [[Technical Support|technical support]] team. |
| **If you do not have X11 forwarding capabilities, you can use the command <tt>srun --pty bash</tt> as the command to be run by <tt>salloc</tt>. For example: <tt>salloc --time=1:00:00 --ntasks=1 srun --pty bash</tt> ([[User:Mboisson|Maxime Boissonneault]] ([[User talk:Mboisson|talk]]) 12:26, 13 December 2017 (UTC))
| |
| * The CC Slurm configuration encourages whole-node jobs. When appropriate, users should request whole-node rather than per-core resources. See [[Job_scheduling_policies#Whole_nodes_versus_cores;|Job Scheduling - Whole Node Scheduling]].
| |
| * By default, the job receives environment settings from the submitting shell. This can lead to irreproducible results if it's not what you expect. To force the job to run with a fresh-like login environment, you can submit with <tt>--export=none</tt> or add <tt>#SBATCH --export=NONE</tt> to your job script.
| |
|
| |
|
| == Quota and filesystem problems == <!--T:7--> | | ==Shared issues== <!--T:2--> |
| === Quota errors on /project filesystem ===
| | The [https://status.alliancecan.ca/ status page] is updated manually, so there may be a delay between when a problem begins and when it is posted to the status page. |
| * This topic has been moved to [[Frequently Asked Questions]].
| |
|
| |
|
| === Nearline === <!--T:10--> | | ===Scheduler issues=== <!--T:6--> |
| * Nearline capabilities are not yet available; see https://docs.computecanada.ca/wiki/National_Data_Cyberinfrastructure for a brief description of the intended functionality.
| |
| ** July 17 update: still not working. If you need your nearline RAC2017 quota, contact [[Technical Support|technical support]].
| |
|
| |
|
| === Missing symbolic links to project folders === <!--T:11-->
| | <!--T:14--> |
| * Upon login to the new clusters, symbolic links are not always created in the user's account, as described in [[Project layout]]. If this is the case, please verify that your access to the cluster is enabled on this page [https://ccdb.computecanada.ca/services/resources https://ccdb.computecanada.ca/services/resources].
| | No known issues. |
|
| |
|
| = Cedar only = <!--T:3--> | | ===Quota and filesystem issues=== <!--T:7--> |
| Nothing to report at this time.
| |
|
| |
|
| = Graham only = <!--T:4--> | | ====Missing project folder==== <!--T:11--> |
| * /home is on an NFS appliance that does not support ACLs, so setfacl/getfacl doesn't work there.
| | Upon creation of a new account for a Principal Investigator, the [[Project layout|<code>/project</code>]] storage space might not be allocated until the next business day. |
| ** Workaround: use the /project or /scratch filesystems instead.
| |
| ** Might be resolved by an update or reconfiguration.
| |
| * <span style="color: red; text-decoration: line-through;"> diskusage_report (and alias 'quota') do not report on Graham /home </span> (FIXED as of 2017-11-27)
| |
| * Compute nodes cannot access Internet
| |
| ** Solution: Contact [[Technical Support|technical support]] to request exceptions to be made; describe what you need to access and why.
| |
|
| |
|
| <!--T:12--> | | ==Cluster-specific issues== <!--T:17--> |
| * Crontab is not offered on Graham.
| | |
| | ===Béluga=== <!--T:16--> |
| | No known issues. |
| | |
| | ===Cedar=== <!--T:3--> |
| | No known issues. |
| | |
| | ===Graham=== <!--T:4--> |
| | Graham's /scratch is often slow; it will be replaced soon. |
| | |
| | ===Narval=== <!--T:18--> |
| | No known issues. |
|
| |
|
| = Other issues = <!--T:5-->
| |
| #Modules don't work for shells other than bash(sh) and tcsh.
| |
| #*Workaround: (this appears to work but not tested extensively)
| |
| #**<tt>source $LMOD_PKG/init/zsh</tt>
| |
| #**<tt>source $LMOD_PKG/init/ksh</tt>
| |
| </translate> | | </translate> |
Problems that affect many users and are being investigated, such as a cluster-wide malfunction or outage, are on the Alliance Status page. This "Known issues" page describes problems that affect many users but that may take some time to repair, or are not planned for repair at this time. Problems that only affect a specific software package are described on the wiki page for that software package.
Report an issue
Please report issues to the technical support team.
Shared issues
The status page is updated manually, so there may be a delay between when a problem begins and when it is posted to the status page.
Scheduler issues
No known issues.
Quota and filesystem issues
Missing project folder
Upon creation of a new account for a Principal Investigator, the /project
storage space might not be allocated until the next business day.
Cluster-specific issues
Béluga
No known issues.
Cedar
No known issues.
Graham
Graham's /scratch is often slow; it will be replaced soon.
Narval
No known issues.