Weird high load on server
nodata
fedora at nodata.co.uk
Tue Jul 19 14:38:45 UTC 2005
> On 7/19/05, nodata <fedora at nodata.co.uk> wrote:
>> > Hi Guys,
>> >
>> > Hope you experts can help me out here.
>> >
>> > Basically I have server running at a very high load 2.44, although
>> > nothing is noticably high when using top. There aren't any processes
>> > running on the box except the standard linux OS tools. This box is
>> > used for backup, and only becomes active during the night.
>> >
>> > Its a compaq dl380 with a raid 5 configuration.
>> >
>> > Can anyone suggest what I can do to find out why the load is high?
>> >
>> > Thanks for your help in advance.
>> >
>> > Dan
>> >
>>
>> I bet you have hanging nfs mounts.
>> If the box is constantly at a load of around 2.44, and isn't sluggish, I
>> wouldn't worry.
>>
>> Look at iostat, sar, etc. to find out why the load is like that.
>>
>
>
> Hi
>
> I've looked at these but can't see anything. The server doesn't mount
> or export any filesystems using nfs or any other protocol. If it helps
> here are the various outputs:
>
> uptime
> 14:45:49 up 62 days, 43 min, 2 users, load average: 1.46, 1.57, 1.59
>
> sar 5 10
> Linux 2.4.21-27.0.4.ELsmp (orion.gs.moneyextra.com) 19/07/05
>
> 14:46:02 CPU %user %nice %system %idle
> 14:46:07 all 0.00 0.00 0.00 100.00
> 14:46:12 all 0.00 0.00 0.10 99.90
> 14:46:17 all 0.00 0.00 0.10 99.90
> 14:46:22 all 0.00 0.00 0.00 100.00
> 14:46:27 all 0.00 0.00 0.00 100.00
> 14:46:32 all 0.00 0.00 0.10 99.90
> 14:46:37 all 0.00 0.00 0.00 100.00
> 14:46:42 all 0.10 0.00 0.31 99.59
> 14:46:47 all 0.00 0.00 0.00 100.00
> 14:46:52 all 0.00 0.00 0.00 100.00
> Average: all 0.01 0.00 0.06 99.93
>
> vmstat -a
> procs memory swap io system
> cpu
> r b swpd free inact active si so bi bo in cs us sy
> wa id
> 0 0 0 15404 189668 202836 0 0 3 1 0 2 3 4
> 1 3
>
> free -m
> total used free shared buffers cached
> Mem: 498 483 15 0 128 301
> -/+ buffers/cache: 53 445
> Swap: 1027 0 1027
>
> iostat
> Linux 2.4.21-27.0.4.ELsmp (orion.gs.moneyextra.com) 19/07/05
>
> avg-cpu: %user %nice %sys %idle
> 3.11 0.00 3.72 93.17
>
> Device: tps Blk_read/s Blk_wrtn/s Blk_read Blk_wrtn
> /dev/ida/c0d0 19.68 427.93 279.15 2147483647 1400883506
> /dev/ida/c0d0p1
> 0.00 0.22 0.00 1087144 8986
> /dev/ida/c0d0p2
> 0.65 3.72 10.24 18680778 51401528
> /dev/ida/c0d0p3
> 0.00 0.00 0.00 248 0
> /dev/ida/c0d0p4
> 0.00 0.00 0.00 0 0
> /dev/ida/c0d0p5
> 0.74 3.90 6.88 19570498 34517568
> /dev/ida/c0d0p6
> 0.00 0.00 0.00 168 0
> /dev/ida/c0d0p7
> 0.00 0.00 0.00 168 0
> /dev/ida/c0d0p8
> 18.29 427.93 262.03 2147483647 1314955424
>
> top
> 14:47:51 up 62 days, 45 min, 2 users, load average: 1.73, 1.61, 1.59
> 61 processes: 60 sleeping, 1 running, 0 zombie, 0 stopped
> CPU states: cpu user nice system irq softirq iowait idle
> total 0.4% 0.0% 0.0% 0.0% 0.0% 0.0% 99.5%
> cpu00 0.9% 0.0% 0.0% 0.0% 0.0% 0.0% 99.0%
> cpu01 0.0% 0.0% 0.0% 0.0% 0.0% 0.0% 100.0%
> Mem: 510400k av, 495224k used, 15176k free, 0k shrd, 132000k
> buff
> 203040k actv, 182824k in_d, 6852k in_c
> Swap: 1052592k av, 0k used, 1052592k free 308668k
> cached
>
> PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME CPU COMMAND
> 13100 root 20 0 1092 1092 888 R 0.4 0.2 0:00 0 top
> 1 root 15 0 512 512 452 S 0.0 0.1 1:18 0 init
> 2 root RT 0 0 0 0 SW 0.0 0.0 0:00 0
> migration/0
> 3 root RT 0 0 0 0 SW 0.0 0.0 0:00 1
> migration/1
> 4 root 15 0 0 0 0 SW 0.0 0.0 0:00 1 keventd
> 5 root 34 19 0 0 0 SWN 0.0 0.0 0:00 0
> ksoftirqd/0
> 6 root 34 19 0 0 0 SWN 0.0 0.0 0:00 1
> ksoftirqd/1
> 9 root 15 0 0 0 0 SW 0.0 0.0 0:00 0 bdflush
> 7 root 15 0 0 0 0 SW 0.0 0.0 70:21 0 kswapd
> 8 root 15 0 0 0 0 SW 0.0 0.0 23:07 1 kscand
> 10 root 15 0 0 0 0 SW 0.0 0.0 3:30 0 kupdated
> 11 root 25 0 0 0 0 SW 0.0 0.0 0:00 0
> mdrecoveryd
> 18 root 15 0 0 0 0 SW 0.0 0.0 0:00 0 ahc_dv_0
> 19 root 25 0 0 0 0 SW 0.0 0.0 0:00 0
> scsi_eh_0
> 23 root 15 0 0 0 0 SW 0.0 0.0 2:30 1
> kjournald
> 192 root 15 0 0 0 0 SW 0.0 0.0 0:00 0
> kjournald
> 193 root 15 0 0 0 0 SW 0.0 0.0 13:57 1
> kjournald
> 194 root 15 0 0 0 0 SW 0.0 0.0 4:18 0
> kjournald
> 568 root 15 0 576 576 492 S 0.0 0.1 0:57 0 syslogd
> 572 root 15 0 472 472 408 S 0.0 0.0 0:00 1 klogd
> 582 root 15 0 452 452 388 S 0.0 0.0 5:33 1
> irqbalance
> 599 rpc 15 0 600 600 524 S 0.0 0.1 0:22 0 portmap
> 618 rpcuser 25 0 720 720 644 S 0.0 0.1 0:00 0
> rpc.statd
> 629 root 15 0 400 400 344 S 0.0 0.0 0:18 0 mdadm
> 712 root 15 0 3160 3160 2024 S 0.0 0.6 3:22 1 snmpd
> 713 root 25 0 3160 3160 2024 S 0.0 0.6 0:00 0 snmpd
> 722 root 15 0 1576 1576 1324 S 0.0 0.3 4:58 1 sshd
>
> Anyone have any ideas. Literally the box is sitting there not doing
> anything that has been scheduled.
>
> This happens occassionally then the load spontaneously goes down. Do
> you reckon it has something to do with the raid 5?
>
> Thanks
> Dan
>
ps auxw | grep " D "
More information about the users
mailing list