Fujitsu Notes
This covers specific setup notes for Fujitsu Primenergy servers on EL6.
Primergy RX200 S8
BIOS
- Advanced
- SATA Configuration
- SATA Mode -> Disabled
- SATA Configuration
- Server Mgmt
- Asset Tag -> (short host name)
- Temperature Monitoring -> Enabled
- Boot
- Bootup NumLock State -> On
- PXE Boot Option Retry -> Enabled
RAID Controller (D3116C)
- Configuration Wizard
- New Configuration -> Next
- Confirm clear config -> Yes
- Manual Configuration -> Next
- Press and hold <ctrl>, Click to highlight all Drives in left pane -> Add to Array -> Accept DG -> Next.
- Add to SPAN -> Next
- RAID Level; 1-8 driver == RAID 5, 9+ drives == RAID 6
- Write Policy -> Write Back with BBU
- Select Size -> Enter size in green text under right pane; R5 size for RAID 5, R6 size for RAID 6.
- Accept -> Confirm cache policy; Yes -> Next
- Accept
- Save the configuration; Yes -> Confirm existing data wipe; Yes
- Click to select Set Boot Drive -> Go -> Back
- Exit -> Confirm exit; Yes
Reboot.
Primergy RX300 S7
RAID
Install the MegaCLI tools;
Check for an updates MegaCLI from here (under "Management Software and Tools"). If there is an updated version, follow the
mkdir ~/temp
cd ~/temp
# Download the updated 8.04.07_MegaCLI.zip here
unzip 8.04.07_MegaCLI.zip
unzip CLI_Lin_8.04.07.zip
unzip MegaCliLin.zip
rpm -Uvh MegaCli-8.04.07-1.noarch.rpm Lib_Utils-1.00-09.noarch.rpm
# This makes MegaCli64 app available without the full path
ln -s /opt/MegaRAID/MegaCli/MegaCli64 /sbin/
If you want to install from the AN!Cache, you can do this;
rpm -Uvh https://alteeve.ca/files/Lib_Utils-1.00-09.noarch.rpm https://alteeve.ca/files/MegaCli-8.04.07-1.noarch.rpm
# This makes MegaCli64 app available without the full path
ln -s /opt/MegaRAID/MegaCli/MegaCli64 /sbin/
Once installed, verify that you can see your hardware:
Replacing a Failed Drive
Replacing a failed drive involves two steps;
- Identify the drive that is failing and gathering the data needed to request the RMA.
- Swapping the actual drive when the replacement arrives on client site.
Identify the Failing Drive
If the drive has failed entirely, the red front LED on the drive should be lit, making identification and RMA request simple.
However, if the drive has not yet failed, identifying the drive and confirming it's pending failure requires a little extra work.
Identify the failed drive:
MegaCli64 PDList aAll
<snip>
Enclosure Device ID: 10
Slot Number: 5
Drive's postion: DiskGroup: 0, Span: 0, Arm: 5
Enclosure position: 1
Device Id: 7
WWN: 5000C50054AE9C38
Sequence Number: 2
Media Error Count: 0
Other Error Count: 2
Predictive Failure Count: 0
Last Predictive Failure Event Seq Number: 0
PD Type: SAS
Raw Size: 279.396 GB [0x22ecb25c Sectors]
Non Coerced Size: 278.896 GB [0x22dcb25c Sectors]
Coerced Size: 278.875 GB [0x22dc0000 Sectors]
Firmware state: Online, Spun Up
Device Firmware Level: 5301
Shield Counter: 0
Successful diagnostics completion on : N/A
SAS Address(0): 0x5000c50054ae9c39
SAS Address(1): 0x0
Connected Port Number: 0(path0)
Inquiry Data: SEAGATE ST9300653SS 53016XN1EMF2 @#87980
FDE Capable: Not Capable
FDE Enable: Disable
Secured: Unsecured
Locked: Unlocked
Needs EKM Attention: No
Foreign State: None
Device Speed: 6.0Gb/s
Link Speed: 6.0Gb/s
Media Type: Hard Disk Device
Drive Temperature :29C (84.20 F)
PI Eligibility: No
Drive is formatted for PI information: No
PI: No PI
Port-0 :
Port status: Active
Port's Linkspeed: 6.0Gb/s
Port-1 :
Port status: Active
Port's Linkspeed: 6.0Gb/s
Drive has flagged a S.M.A.R.T alert : No
Notice the line:
Other Error Count: 2
This is a sign of impending failure, despite SMART showing the drive as still healthy. You can gather more details (which Fujitsu will require to verify impending failure), run;
MegaCli64 -AdpEventLog -GetEvents -f raid_events.log -aALL
When this finishes gathering data, it will create a file called raid_events.log. Send this file to your Fujitsu support rep. They will validate the pending failure and issue an RMA.
Identifying The Drive Prior to Replacement
If the drive has failed, identifying the drive is as simple as looking at the front of the node for the drive with the red error LED lit.
If the drive has not failed yet, then you can use the identify command to blink the LED. That is what we'll cover here.
In the previous section, we identified the failed drive using the MegaCli64 command. You need to note the following;
MegaCli64 PDList aAll
Enclosure Device ID: 10
Slot Number: 5
...
Other Error Count: 2
The two bits of information you need are the enclosure ID and slot number. In this case, that is 10 and 5, respectively.
With that info, you can trigger the drive's red LED using the following command;
MegaCli64 -PdLocate -start -physdrv [10:5] -aALL
Adapter: 0: Device at EnclId-10 SlotId-5 -- PD Locate Start Command was successfully sent to Firmware
Exit Code: 0x00
Once you've located the drive, you can stop the "locate" command using:
MegaCli64 -PdLocate -stop -physdrv [10:5] -aALL
Adapter: 0: Device at EnclId-10 SlotId-5 -- PD Locate Stop Command was successfully sent to Firmware
Exit Code: 0x00
Now that you know which drive has failed, you can OFFLINE it in preparation of replacing it.
MegaCli64 -PDOffline -physdrv [10:5] -aALL
Adapter: 0: EnclId-10 SlotId-5 state changed to OffLine.
Exit Code: 0x00
You can now physically remove the failed disk and insert the replacement disk.
Monitoring the Rebuild
When a replacement disk is inserted, the array should recognise it and automatically begin rebuilding the array. You can monitor this operation by calling;
MegaCli64 -PDRbld -ProgDsply -PhysDrv [10:5] -aALL
This will display the rebuild progress as a textual bar graph. A rebuild of a 300 GB 15,000rpm SAS drive in a 6-drive array took about 30 minutes. How long it takes in your case will vary depending on disk speed, array size and load.
Checking the BBU
Check the BBU;
MegaCli64 AdpBbuCmd aAll
Triggering a re-learn cycle.
MegaCli64 AdpBbuCmd BbuLearn aAll
Monitor a learn cycle:
watch "MegaCli64 AdpBbuCmd aAll | grep -i -e learn -e status -e charge -e operation"
Creating a Backup Set
ToDo: Explain this...
mkdir ~/base
cd ~/base
mkdir root
mkdir -p etc/sysconfig/network-scripts/
mkdir -p etc/udev/rules.d/
# Root user
rsync -av /root/.bashrc root/
rsync -av /root/.ssh root/
# Directories
rsync -av /etc/ssh etc/
rsync -av /etc/apcupsd etc/
rsync -av /etc/cluster etc/
rsync -av /etc/drbd.* etc/
rsync -av /etc/lvm etc/
# Specific files.
rsync -av /etc/sysconfig/network-scripts/ifcfg-{eth*,bond*,vbr*} etc/sysconfig/network-scripts/
rsync -av /etc/udev/rules.d/70-persistent-net.rules etc/udev/rules.d/
rsync -av /etc/sysconfig/network etc/sysconfig/
rsync -av /etc/hosts etc/
rsync -av /etc/ntp.conf etc/
# Save recreating user accounts.
rsync -av /etc/passwd etc/
rsync -av /etc/group etc/
rsync -av /etc/shadow etc/
rsync -av /etc/gshadow etc/
# If you have the cluster built and want to backup it's configs.
mkdir etc/cluster
mkdir etc/lvm
rsync -av /etc/cluster/cluster.conf etc/cluster/
rsync -av /etc/lvm/lvm.conf etc/lvm/
# NOTE: DRBD won't work until you've manually created the partitions.
rsync -av /etc/drbd.d etc/
# If you had to manually set the UUID in libvirtd;
mkdir etc/libvirt
rsync -av /etc/libvirt/libvirt.conf etc/libvirt/
# If you're running RHEL and want to backup your registration info;
rsync -av /etc/sysconfig/rhn etc/sysconfig/
# Pack it up
# NOTE: Change the name to suit your node.
tar -cvf base_an-c01n01.tar etc root
Specific Fujitsu Model Notes
RX200 S7
BIOS Changes for use in Anvil!s
To enter the BIOS, press F2 during POST.
- Advanced
- SATA Configuration
- SATA Mode -> Disaabled (if no optical drive)
- - OR -
- SATA Mode -> AHCI Mode (if optical drive)
- Onboard Device Configuration
- LAN 2 Oprom -> PXE
- Onboard SAS/SATA (SCU) -> Disabled (if no optical drive)
- SATA Configuration
- Server Mgmt
- Asset Asset Tag -> short host name of the node, lower case
- Temperature Monitoring -> Enabled
- Boot
- Bootup NumLock State -> On
Save and Exit
LSI RAID Controller Setup
Warning: These instructions assume NO existing data. This is a destructive process! |
After POST, press <ctrl> + H to enter the controller's WebBIOS
- Choose the controller (usually only one available), click Start
- Choose Configuration Wizard
- Choose New Configuration -> click Next
- Confirm the clear by clicking Yes
- Choose Manual Configuration -> click Next
- Click on the first drive, usually Slot: 0,..., press and hold <ctrl> button and then click to select the rest of the drives. Click Add to Array.
- Click on Accept DG to create the drive group then click on Next
- Click on Add to SPAN, then click on Next
- Configure the array;
- For 1 to 8 drives;
- RAID Level -> RAID 5
- Write Policy -> Write Back with BBU
- Look at the R5:xxxx size on the right and enter that size in the Select Size section. Be sure to match the GB or TB suffix.
- For 9 or more drives;
- RAID Level -> RAID 6
- Write Policy -> Write Back with BBU
- Look at the R6:xxxx size on the right and enter that size in the Select Size section. Be sure to match the GB or TB suffix.
- Click on Accept then click on Yes to accept the warning.
- The virtual disk will now be shown on the right. Click Next to proceed.
- For 1 to 8 drives;
- Click Accept and then click Yes to save the configuration. Click Yes to acknowledge the warning and initialize the drive.
- Click on Set Boot Drive and then click Go.
- Clock on Home.
- Click on Exit and then on Yes.
Reboot and you are done.
Any questions, feedback, advice, complaints or meanderings are welcome. | |||
Alteeve's Niche! | Enterprise Support: Alteeve Support |
Community Support | |
© Alteeve's Niche! Inc. 1997-2024 | Anvil! "Intelligent Availability®" Platform | ||
legal stuff: All info is provided "As-Is". Do not use anything here unless you are willing and able to take responsibility for your own actions. |