Accessing the Cluster

Shell Access

The Anselm cluster is accessed by SSH protocol via login nodes login1 and login2 at address The login nodes may be addressed specifically, by prepending the login node name to the address.

Login address Port Protocol Login node 22 ssh round-robin DNS record for login1 and login2 22 ssh login1 22 ssh login2

The authentication is by the private key


Please verify SSH fingerprints during the first logon. They are identical on all login nodes:

29:b3:f4:64:b0:73:f5:6f:a7:85:0f:e0:0d:be:76:bf (DSA) d4:6f:5c:18:f4:3f:70:ef:bc:fc:cc:2b:fd:13:36:b7 (RSA)

Private key authentication:

On Linux or Mac, use

local $ ssh -i /path/to/id_rsa

If you see warning message "UNPROTECTED PRIVATE KEY FILE!", use this command to set lower permissions to private key file.

local $ chmod 600 /path/to/id_rsa

On Windows, use PuTTY ssh client.

After logging in, you will see the command prompt:

                       /\                  | |
                      /  \   _ __  ___  ___| |_ __ ___
                     / /\ \ | '_ \/ __|/ _ \ | '_ ` _ \
                    / ____ \| | | \__ \  __/ | | | | | |
                   /_/    \_\_| |_|___/\___|_|_| |_| |_|


Last login: Tue Jul 9 15:57:38 2013 from
[username@login2.anselm ~]$

Example to the cluster login:


The environment is not shared between login nodes, except for shared filesystems.

Data Transfer

Data in and out of the system may be transferred by the scp and sftp protocols. (Not available yet.) In case large volumes of data are transferred, use dedicated data mover node for increased performance.

Address Port Protocol 22 scp, sftp 22 scp, sftp 22 scp, sftp 22 scp, sftp

The authentication is by the private key


Data transfer rates up to 160MB/s can be achieved with scp or sftp.

1TB may be transferred in 1:50h.

To achieve 160MB/s transfer rates, the end user must be connected by 10G line all the way to IT4Innovations and use computer with fast processor for the transfer. Using Gigabit ethernet connection, up to 110MB/s may be expected. Fast cipher (aes128-ctr) should be used.


If you experience degraded data transfer performance, consult your local network provider.

On linux or Mac, use scp or sftp client to transfer the data to Anselm:

local $ scp -i /path/to/id_rsa my-local-file
local $ scp -i /path/to/id_rsa -r my-local-dir


local $ sftp -o IdentityFile=/path/to/id_rsa

Very convenient way to transfer files in and out of the Anselm computer is via the fuse filesystem sshfs

local $ sshfs -o IdentityFile=/path/to/id_rsa mountpoint

Using sshfs, the users Anselm home directory will be mounted on your local computer, just like an external disk.

Learn more on ssh, scp and sshfs by reading the manpages

$ man ssh
$ man scp
$ man sshfs

On Windows, use WinSCP client to transfer the data. The win-sshfs client provides a way to mount the Anselm filesystems directly as an external disc.

More information about the shared file systems is available here.

Connection Restrictions

Outgoing connections, from Anselm Cluster login nodes to the outside world, are restricted to following ports:

Port Protocol
22 ssh
80 http
443 https
9418 git


Please use ssh port forwarding and proxy servers to connect from Anselm to all other remote ports.

Outgoing connections, from Anselm Cluster compute nodes are restricted to the internal network. Direct connections form compute nodes to outside world are cut.

Port Forwarding

Port Forwarding From Login Nodes


Port forwarding allows an application running on Anselm to connect to arbitrary remote host and port.

It works by tunneling the connection from Anselm back to users workstation and forwarding from the workstation to the remote host.

Pick some unused port on Anselm login node (for example 6000) and establish the port forwarding:

local $ ssh -R

In this example, we establish port forwarding between port 6000 on Anselm and port 1234 on the By accessing localhost:6000 on Anselm, an application will see response of The traffic will run via users local workstation.

Port forwarding may be done using PuTTY as well. On the PuTTY Configuration screen, load your Anselm configuration first. Then go to Connection->SSH->Tunnels to set up the port forwarding. Click Remote radio button. Insert 6000 to Source port textbox. Insert Click Add button, then Open.

Port forwarding may be established directly to the remote host. However, this requires that user has ssh access to

$ ssh -L 6000:localhost:1234


Port number 6000 is chosen as an example only. Pick any free port.

Port Forwarding From Compute Nodes

Remote port forwarding from compute nodes allows applications running on the compute nodes to access hosts outside Anselm Cluster.

First, establish the remote port forwarding form the login node, as described above.

Second, invoke port forwarding from the compute node to the login node. Insert following line into your jobscript or interactive shell

$ ssh  -TN -f -L 6000:localhost:6000 login1

In this example, we assume that port forwarding from login1:6000 to has been established beforehand. By accessing localhost:6000, an application running on a compute node will see response of

Using Proxy Servers

Port forwarding is static, each single port is mapped to a particular port on remote host. Connection to other remote host, requires new forward.


Applications with inbuilt proxy support, experience unlimited access to remote hosts, via single proxy server.

To establish local proxy server on your workstation, install and run SOCKS proxy server software. On Linux, sshd demon provides the functionality. To establish SOCKS proxy server listening on port 1080 run:

local $ ssh -D 1080 localhost

On Windows, install and run the free, open source Sock Puppet server.

Once the proxy server is running, establish ssh port forwarding from Anselm to the proxy server, port 1080, exactly as described above.

local $ ssh -R 6000:localhost:1080

Now, configure the applications proxy settings to localhost:6000. Use port forwarding to access the proxy server from compute nodes as well.

Graphical User Interface

VPN Access

  • Access to IT4Innovations internal resources via VPN.