[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[cobalt-users] Conflicting info on XTR kernel upgrad debacle



I have seen two different and conflicting sets of instructions on how
to deal with fixing the problems from the latest official BlueLinQ
Kernel update.  So far I have tried neither. The two solutions I've
seen are:

1. Roll back the upgrade by manually copying back the saved kernel and
system map that are stored in /boot. I saw this on the Cobalt-users
list, Instructions on how to deal with future upgrades and get the
database for the web face and BlueLinQ back into sync not provided,
though this would be a good temporary fix. I will save the new kernel
and system map if I must use this.

2. Boot from ROM, by shutting down (if possible - otherwise just turn
it off I guess). Then turn it on holding down the 's' button which
enables the four arrow keys to be used for selection. Manipulate with
the arrow keys until the LCD says "boot from ROM" and press
enter. This will force the RAID array to rebuild. I saw this on the
Sun Cobalt forum.

What did happen was that the machine became so flaky, I couldn't even
get a prompt from the terminal I leave connected in emergencies. So I
turned it off and on (powercycled the machine, if you prefer). When it
rebooted it spent ages checking disks and then rebuilding RAID. The
"Active Monitor" displayed the yellow warning light, and then after a
few hours behavior seemed normal(ish). All I'm using is striping -
RAID0 (or is that 1?) so I have four partitions counting swap with the
four 30GB drives in the array.

[josh josh]$ df
Filesystem           1k-blocks      Used Available Use% Mounted on
/dev/md1                991896    829080    162816  84% /
/dev/md3                495944     42004    453940   9% /var
/dev/md4             107012296   3090144 103922152   3% /home


Currently there are two processes or groups of process causing me
enough concern that I expect to try to boot from ROM over the
weekend. (Right now I don't won't to bring down an operating
system). What is kupdate, mdrecoveryd, raid1d, and raidsynchd. While
the load averages don't look too bad, I've been noticing that I often
now go over 1.00 which almost never happened before and I used to be
hard pressed to get it over 0.10 because it is normally lightly used.

 12:36pm  up 1 day, 19:53,  3 users,  load average: 0.34, 0.72, 0.53
100 processes: 98 sleeping, 2 running, 0 zombie, 0 stopped
CPU states: 19.4% user, 10.1% system,  0.0% nice, 16.0% idle
Mem:  1035192K av,  506940K used,  528252K free,  243072K shrd,  314032K buff
Swap:  131448K av,       0K used,  131448K free                   97956K cached

  PID USER     PRI  NI  SIZE  RSS SHARE STAT  LIB %CPU %MEM   TIME COMMAND
 9455 httpd     10   0 11976  11M  1388 R       0 79.6  1.1   0:00 perl5.6.1
 9453 josh       2   0   868  868   648 R       0  2.8  0.0   0:00 top
 1206 root       0   0  2572 2572  1196 S       0  0.9  0.2   7:11 poprelayd
    1 root       0   0   472  472   400 S       0  0.0  0.0   0:18 init
    2 root       0   0     0    0     0 SW      0  0.0  0.0   0:00 kflushd
    3 root       0   0     0    0     0 SW      0  0.0  0.0  72:42 kupdate
    4 root       0   0     0    0     0 SW      0  0.0  0.0   9:56 kswapd
    5 root     -20 -20     0    0     0 SW<     0  0.0  0.0   0:00 mdrecoveryd
    6 root     -20 -20     0    0     0 SW<     0  0.0  0.0   0:00 raid1d
    7 root     -20 -20     0    0     0 SW<     0  0.0  0.0   1:51 raid1syncd
    8 root     -20 -20     0    0     0 SW<     0  0.0  0.0   0:00 raid1d
    9 root      19  19     0    0     0 SWN     0  0.0  0.0  15:51 raid1syncd
   10 root     -20 -20     0    0     0 SW<     0  0.0  0.0   0:00 raid1d
   11 root      19  19     0    0     0 SWN     0  0.0  0.0  29:16 raid1syncd
  137 root       0   0   524  524   424 S       0  0.0  0.0  12:07 syslogd
  146 root       0   0   760  760   384 S       0  0.0  0.0   1:50 klogd


Is my machine stable? I welcome opinions on the best course of action
to insure future stability.

-- 
Josh Kuperman                       
josh@xxxxxxxxxxxxxxxxxx