[ Skip to the content ]

Institute of Formal and Applied Linguistics Wiki


[ Back to the navigation ]

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
gpu [2017/03/16 16:45]
kocmanek
gpu [2017/03/16 17:10]
kocmanek [Basic commands]
Line 43: Line 43:
      
  
 +===== How to use cluster =====
 +
 +In this section will be explained how to use cluster properly. 
 +==== TensorFlow Environment ====
 +
 +Majority people at UFAL use TensorFlow. To start using it you need to create python virtual environment (virtualenv or use Anaconda for it). Into the environment you must place TensorFlow. The TF is either in CPU or GPU version.
 +
 +  pip install tensorflow
 +  pip install tensorflow-gpu
 +  
 +You can use prepared environment by adding into your ~/.bashrc
 +
 +  export PATH=/a/merkur3/kocmanek/ANACONDA/bin:$PATH
 +
 +And then you can activate your environment:
 +
 +  source activate tf1
 +  source activate tf1cpu
 +
 +This environment have TensorFlow 1.0 and all necessary requirements for NeuralMonkey.
 +
 +==== Using cluster ====
 +
 +Rule number one, always use the GPU queue (never log in machine by ssh). Always use qsub or qsubmit with proper arguments.
 +
 +For testing and using the cluster interactively you can use qrsh (this should not be used for long running experiments since the console is not closed on the end of the experiment). Following command will assign you a GPU and creates interactive console.
 +
 +  qrsh -q gpu.q -l gpu=1 -pty yes bash
 +  
 +For running experiments you must use qsub command:
 +
 +  qsub -q gpu.q -l gpu=1,gpu_cc_min3.5=1,gpu_ram=2G WHAT_SHOULD_BE_RUN
 +  
 +Cleaner way to use cluster is with /home/bojar/tools/shell/qsubmit
 +
 +  qsubmit --gpumem=2G --queue="gpu.q" WHAT_SHOULD_BE_RUN
 +  
 +==== Basic commands ====
 +
 +  lspci
 +    # is any such hardware there?
 +  nvidia-smi
 +    # more details, incl. running processes on the GPU
 +    # nvidia-* are typically located in /usr/bin
 +  watch nvidia-smi
 +    # For monitoring GPU activity in a separate terminal (thanks to Jindrich Libovicky for this!)
 +  nvcc --version
 +    # this should tell CUDA version
 +    # nvcc is typically installed in /usr/local/cuda/bin/
 +  theano-test
 +    # dela to vubec neco uzitecneho? :-)
 +    # theano-* are typically located in /usr/local/bin/
 +  /usr/local/cuda/samples/1_Utilities/deviceQuery/deviceQuery
 +    # shows CUDA capability etc.
 ===== Performance tests ===== ===== Performance tests =====
  

[ Back to the navigation ] [ Back to the content ]