Version 8 (modified by 13 years ago) (diff) | ,
---|
UFO Server
System
- Host name: ufosrv1.ka.fzk.de
- Interfaces: eth0 (10 GBit), eth1 (upper-right socket)
- eth0: dhcp
- eth1: 141.52.111.135/22
- Gateway: 141.52.111.208 (via dhcp)
- Name server: 141.52.111.248 (via dhcp)
- Running services
- SSH on ports 22 and 24
- NX server over SSH
- VirtualGL (OpenGL forwarding)
Hardware
- Display connected to integrated video card (Matrox G200)
- 2 x Xeon X5650 / Intel X58 / 96 GB DDR3
- System drives: 2 x Hitachi 2TB SATA2
- Areca ARC-1880 Raid Controller (x16 slot)
- 16 x 2 TB Hitachi HUA722020ALA330 in external Areca Enclosure
- 4 x 256 Crucial RealSSD C300
- External PCIe 2.0 x16 (x16 slot)
- External GPU box from One Stop Systems
- 4 x NVIDIA GeForce? GTX580
- 2 x NVIDIA GeForce? GTX580 (x16 slots)
- Intel 82598EB 10GBit Ethernet (x4 slot)
- Silicon Software CameraLink? FrameGrabber? MicroEnable? IV Full (PCIe 1.0 x4 slot)
- Free slots
- PCI express: x4 and x1
- Storage: 2 x SSD in the main server case
Areca Raid Configuration
- A single Areca-1880 controller handles both external storage box with SATA hard drives and internal SSD cache. Only a pair of system hard drives are connected to the SATA controller integrated in the motherboard.
- 16 x Hitachi 2TB SATA hard drives in the external enclosure are organized as Raid-6
- 4 x Crucial SSD C300 in the server case are organized as Raid-0
Partitioning
- Two system hard-drives are connected to internal SATA controller and mirrored as Raid-1 using Linux Software Raid.
- Devices: /dev/sda, /dev/sdb
- Partitions: /boot (2GB ext2), / (256GB ext4), /home (ext4)
- The RAID is split into the 2 partitions (GPT partition table): the fast and normal.
- Device: /dev/sdc
- The fast partition will be used to stream the data from the camera and should be able to stand throughput of 850 MB/s. The data should be moved out as soon as possible. Only a single application is allowed to write to the disk.
- Size: first 6TB of disk array
- File system: non-journaled ext4
- Mount point: /mnt/fast
- Standard partition is for short term data storage (before offloading to LSDF)
- Size: 22 TB
- File system: ext4
- Mount point: /mnt/raid
- The SSD cache
- Device: /dev/sdd
- Size: 1 TB
- File system: ext4
- Mount point: /mnt/ssd
- Partition table (/dev/sda & /dev/sdc)
Disk /dev/sda: 2000.4 GB, 2000398934016 bytes 255 heads, 63 sectors/track, 243201 cylinders, total 3907029168 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disk identifier: 0x0003874d Device Boot Start End Blocks Id System /dev/sda1 * 2048 4192255 2095104 fd Linux raid autodetect /dev/sda2 4192256 541069311 268438528 fd Linux raid autodetect /dev/sda3 541069312 3907028991 1682979840 fd Linux raid autodetect
- Partition table (/dev/sdc)
Disk /dev/sdc: 28.0TB Sector size (logical/physical): 512B/512B Partition Table: gpt_sync_mbr Number Start End Size File system Name Flags 1 1049kB 6597GB 6597GB primary 2 6597GB 28.0TB 21.4TB xfs primary
- Partition table (/dev/sdd)
Disk /dev/sdd: 1000.0 GB, 999998619648 bytes 255 heads, 63 sectors/track, 121576 cylinders, total 1953122304 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disk identifier: 0x000b6bc3 Device Boot Start End Blocks Id System /dev/sdd1 2048 1953122303 976560128 83 Linux
- Raid table:
Personalities : [raid1] [raid0] [raid10] [raid6] [raid5] [raid4] md2 : active raid1 sdb3[1] sda3[0] 1682979704 blocks super 1.0 [2/2] [UU] bitmap: 3/13 pages [12KB], 65536KB chunk md0 : active raid1 sdb1[1] sda1[0] 2095092 blocks super 1.0 [2/2] [UU] bitmap: 0/1 pages [0KB], 65536KB chunk md1 : active raid1 sda2[0] sdb2[1] 268438392 blocks super 1.0 [2/2] [UU] bitmap: 0/3 pages [0KB], 65536KB chunk
- mdadm.conf
DEVICE containers partitions ARRAY /dev/md0 UUID=7c032686:e8861a19:9ccb43c3:8f25011e ARRAY /dev/md1 UUID=4a18bb5c:4b4b4490:929fdc08:99b65f2f ARRAY /dev/md2 UUID=8e7c863e:3a75af81:321862ae:d679602e
- fstab
/dev/disk/by-id/md-uuid-4a18bb5c:4b4b4490:929fdc08:99b65f2f / ext4 acl,user_xattr 1 1 /dev/disk/by-id/md-uuid-7c032686:e8861a19:9ccb43c3:8f25011e /boot ext2 acl,user_xattr 1 2 /dev/disk/by-id/md-uuid-8e7c863e:3a75af81:321862ae:d679602e /home ext4 acl,user_xattr 1 2 /dev/disk/by-id/scsi-2001b4d2003077811-part2 /mnt/raid xfs defaults 1 2 /dev/disk/by-id/scsi-2001b4d2064473251-part1 /mnt/ssd ext4 acl,user_xattr 1 2 proc /proc proc defaults 0 0 sysfs /sys sysfs noauto 0 0 debugfs /sys/kernel/debug debugfs noauto 0 0 usbfs /proc/bus/usb usbfs noauto 0 0 devpts /dev/pts devpts mode=0620,gid=5 0 0
Software
Base System
- openSUSE 12.1
- Development: Kernel, GNOME, Python
- Default compiler: gcc-4.3
- Important packages:
zypper install bzr zypper install cmake
CUDA
- Downgrade to gcc 4.3 (gcc-4.6 is not compatible with the latest CUDA toolkit)
zypper ar http://download.opensuse.org/repositories/devel:/gcc/openSUSE_12.1/devel:gcc.repo zypper install gcc43 gcc43-c++ gcc43-locale for name in `rpm -ql gcc43 gcc43-c++ | grep "/usr/bin"`; do ln -sf $name /usr/bin/`basename $name -4.3`; done
- Currently installed
- Driver: 285.05.32
- Toolkit: 4.1 (installed into /opt/cuda)
- SDK: 4.1 (installed into /opt/cuda/sdk)
- SDK must be compiled
cd /opt/cuda/sdk make
- Allow reseting failed GPUs. Add following line into the /etc/sudoers:
ALL ALL=(ALL) NOPASSWD: /opt/cuda/sdk/C/bin/linux/release/deviceQuery
UFO Framework
- Necessary build packages
zypper install libtiff-devel
- Install vogelgesang/ufo-core and vogelgesang/ufo-filters according to the manual.
PyHST
- Necessary build packages:
zypper install freeglut-devel zypper install openmpi-devel zypper install fftw3-devel zypper install python-imaging zypper install python-numpy-devel
- Currently installed in
/opt/pyhst
- To build from pyhst/pyhst
cd /opt bzr branch http://ufo.kit.edu/sources/csa/pyhst/ cd pyhst cmake . make
- Start script
/opt/PyHST
#!/bin/bash PACKAGE_HOME=/opt/pyhst PACKAGE_SOURCE=${PACKAGE_HOME} CUDA_DIR=/opt/cuda.41 export PATH=${INSTALLATION_HOME}/bin:$PATH export LD_LIBRARY_PATH=${PACKAGE_SOURCE}:${INSTALLATION_HOME}/lib:$LD_LIBRARY_PATH export LDFLAGS="-L ${INSTALLATION_HOME}/lib" export CPPFLAGS="-I ${INSTALLATION_HOME}/include" export LD_LIBRARY_PATH=$CUDA_DIR/lib64/:$LD_LIBRARY_PATH export LD_LIBRARY_PATH=${PACKAGE_SOURCE}/hst_cuda:$LD_LIBRARY_PATH export PYTHONPATH=${PACKAGE_SOURCE}:$PYTHONPATH PYTHON=python ${PYTHON} ${PACKAGE_SOURCE}/PyHST.py $*
- Start PyHST with
/opt/PyHST <parameter_file.par>
FreeNX
Install FreeNX server to allow remote desktop
zypper ar http://download.opensuse.org/repositories/X11:/RemoteDesktop/openSUSE_12.1/X11:RemoteDesktop.repo zypper install FreeNX nxsetup
VirtualGL
Install VirtualGL to provide OpenGL forwarding (HOW?)
Usage
- Reset failed GPU devices
sudo /opt/cuda/sdk/C/bin/linux/release/deviceQuery
- Check system status
To be provided by Tomsk students...
- PyHST
/opt/PyHST <parameter_file.par>
- Framework
Attachments (1)
- menable-ds.patch (3.0 KB) - added by 13 years ago.
Download all attachments as: .zip