[QUOTE=KBOYLE;16512]jmcg wrote:
[COLOR=blue]
I hope that my ignorance is not too obvious to you or anyone else who
views this, but the more I put into this, the better I understand my
system.[/COLOR]
That is a good thing. Detailed questions make us think and even
question what has been put into place to help us better understand and
better able to identify issues.
[COLOR=blue]
First, the physical server has a RAID 1 hardware type and a hardware
type RAID 5. Those are physical discs. The host OS is SLES 11 and is
install as a Xen server, minimal extras, on the RAID 1 .
The RAID 1 shows up in the terminal (df h) as c0d0p1 in the
/dev/cciss/ directory. That is the disk space that is 78% used up.[/COLOR]
Hardware RAID is good. 78% used (or is it 86% as stated in your other
post?) doesn’t help much without knowing a bit more. Can you provide
the output from these commands run on Dom0? It will explain a lot more
than I can get by asking specific questions.
fdisk -l
cat /etc/fstab
Note: That’s a lower case ELL on the fdisk command.
Please post your results between “code” tags (using the “#” from the
web interface) to make the output more readable.
[COLOR=red]
[CODE]
SLES-XEN:/ # fdisk -l
Disk /dev/cciss/c0d0: 146.8 GB, 146778685440 bytes
255 heads, 63 sectors/track, 17844 cylinders, total 286677120 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00075c47
Device Boot Start End Blocks Id System
/dev/cciss/c0d0p1 63 4209029 2104483+ 82 Linux swap / Solaris
/dev/cciss/c0d0p2 * 4209030 286663859 141227415 83 Linux
Disk /dev/cciss/c0d1: 899.9 GB, 899898718208 bytes
255 heads, 63 sectors/track, 109406 cylinders, total 1757614684 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x0006a452
Device Boot Start End Blocks Id System
/dev/cciss/c0d1p1 32 706876064 353438016+ 5 Extended
/dev/cciss/c0d1p2 706876065 1757607389 525365662+ 83 Linux
/dev/cciss/c0d1p5 64 195309599 97654768 83 Linux
/dev/cciss/c0d1p6 195309632 292968479 48829424 83 Linux
/dev/cciss/c0d1p7 292977468 481966064 94494298+ 83 Linux
SLES-XEN:/ # cat /etc/fstab
/dev/disk/by-id/cciss-3600508b1001032373120202020200004-part2 / reiserfs acl,user_xattr 1 1
/dev/disk/by-id/cciss-3600508b1001032373120202020200004-part1 swap swap defaults 0 0
proc /proc proc defaults 0 0
sysfs /sys sysfs noauto 0 0
debugfs /sys/kernel/debug debugfs noauto 0 0
usbfs /proc/bus/usb usbfs noauto 0 0
devpts /dev/pts devpts mode=0620,gid=5 0 0
/dev/disk/by-id/cciss-3600508b1001032373120202020200005-part4 /mnt/mount ext3 rw 0 0
##/dev/cciss/c0d1p4 /mnt/mount ext3 rw 0 0
##next line edited by jmcg 21May2012 so partition not mounted
##/dev/disk/by-id/cciss-3600508b1001032373120202020200005-part3 /mnt/mount2 ext3 acl,user_xattr 1 2
/dev/disk/by-id/cciss-3600508b1001032373120202020200005-part6 /mnt/mount3 ext3 acl,user_xattr 1 2
SLES-XEN:/ #
[CODE]
[/COLOR]
[COLOR=blue]
Each DomU (guests or vms) are installed as disk files on the RAID 1.
That is, not only is the Host OS installed on the RAID 1 but also the
disk files for each of the DomUs (they are in the /var/lib/xen/images/
directory) are too, but they are technically files (disk images).[/COLOR]
So far, that’s fine. The commands I asked you to run will tell us more.
[COLOR=blue]
The RAID 5 shows up in the host directory (/dev/cciss/c0d1x) with each
of the partitions (linux) that I have created for the DomUs to use.
Each DomU has a partition of the RAID 5 attached to it. They show up
as /dev/cciss/c0d1p2 or 3 or 4, depending on the order I created
them.[/COLOR]
It is good that you separated your data storage from that which
contains your operating system.
[COLOR=blue]
Within each of the domUs, I see those partitions (RAID 5) as xvdb or
xvdd or whatever. That is what I see from the terminal of the domU
when I enter df h. I dont see them from the Dom0s terminal because
they are not attached to it, only to the DomUs.[/COLOR]
What you say makes sense but if on a particular DomU you can see
/dev/xvdd, that suggests to me you have assigned one file-backed
storage device for the operating system and three separate partitions
from your RAID 5 device. Is that correct? Why three devices for data?
(Just curious.)
[COLOR=blue]
So, it seems that what I have done is used disk files for the OS on
each DomU but am using actually using partitions on the RAID 5 for
each of the DomUs. So I am mixing and matching. Is that bad?[/COLOR]
No, that is not a bad idea depending upon what you are trying to
achieve. File-backed storage is created by default. It offers some
flexibilities and the ability to “thin provision” but is not
recommended for use with NSS data. Partitions on the RAID 5 (a block
storage device) offer better performance and are recommended for NSS.
[COLOR=blue]
Now with all that said I have one more issue. One of the DomUs that
is an OES11sp1 server that is running my primary domain for groupwise
has become sluggish. Over a period of 5 or so days it becomes less
responsive. Now I cant even get top or df h to activate from the
terminal or from an ssh session. The groupwise system is running
fine, but the server seems busy. What can I do to investigate why it
is not responsive to terminal commands? Yesterday I did a restart of
the groupwise system and it took 5 or 6 minutes to restart and it
only has 100 users, a small system. I appreciate your help.[/COLOR]
The first thing we have to determine is whether this issue lies within
your Dom0 or your DomU. If only one DomU is affected, I would suspect
the DomU (Server 2?).
Please run these commands on your DomU:
fdisk -l
cat /etc/fstab
[COLOR=red]
[CODE]
groupwise:~ # fdisk -l
Disk /dev/xvda: 21.5 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders, total 41943040 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00007bc8
Device Boot Start End Blocks Id System
/dev/xvda1 2048 4208639 2103296 82 Linux swap / Solaris
/dev/xvda2 * 4208640 41943039 18867200 83 Linux
Disk /dev/xvdb: 4189 MB, 4189161472 bytes
255 heads, 63 sectors/track, 509 cylinders, total 8181956 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk /dev/xvdb doesn’t contain a valid partition table
Disk /dev/xvdd: 96.8 GB, 96762161664 bytes
255 heads, 63 sectors/track, 11763 cylinders, total 188988597 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x0001b841
Device Boot Start End Blocks Id System
/dev/xvdd1 2048 188987391 94492672 83 Linux
groupwise:~ # cat etc/fstab
cat: etc/fstab: No such file or directory
groupwise:~ # cat /etc/fstab
/dev/xvda1 swap swap defaults 0 0
/dev/xvda2 / ext3 acl,user_xattr 1 1
proc /proc proc defaults 0 0
sysfs /sys sysfs noauto 0 0
debugfs /sys/kernel/debug debugfs noauto 0 0
devpts /dev/pts devpts mode=0620,gid=5 0 0
/dev/xvdd1 /mnt/vol ext3 acl,user_xattr 1 2
198.nnn.nnn.209:/media/nss/SVUSD1DATA/stu-domain /media/nss/SVUSD1DATA/stu-domain nfs defaults 0 0
groupwise:~ #
[CODE]
[/COLOR]
If you shutdown GroupWise, is your server more responsive?
Did you move GWIA as you planned to do in your OP?
Poor performance can be caused by IO bottlenecks. Have you checked for
excessive swapping on each server or verified that there are no disk
errors that are causing excessive retries?
One other suggestion: If this is a production server and you need to
get this resolved quickly, have you considered opening a Service
Request to get some hands-on support?
Please provide the requested information to get a better understanding
of what’s happening.
I’ll post more later today.