Fujitsu Notes

From Alteeve Wiki
Jump to navigation Jump to search

 AN!Wiki :: How To :: Fujitsu Notes

This covers specific setup notes for Fujitsu Primenergy servers on EL6.

Primergy RX300 S7

RAID

Install the MegaCLI tools;

Check for an updates MegaCLI from here (under "Management Software and Tools"). If there is an updated version, follow the

mkdir ~/temp
cd ~/temp
# Download the updated 8.04.07_MegaCLI.zip here
unzip 8.04.07_MegaCLI.zip
unzip CLI_Lin_8.04.07.zip 
unzip MegaCliLin.zip
rpm -Uvh MegaCli-8.04.07-1.noarch.rpm Lib_Utils-1.00-09.noarch.rpm

# This makes MegaCli64 app available without the full path
ln -s /opt/MegaRAID/MegaCli/MegaCli64 /sbin/

If you want to install from the AN!Cache, you can do this;

rpm -Uvh https://alteeve.ca/files/Lib_Utils-1.00-09.noarch.rpm https://alteeve.ca/files/MegaCli-8.04.07-1.noarch.rpm

# This makes MegaCli64 app available without the full path
ln -s /opt/MegaRAID/MegaCli/MegaCli64 /sbin/

Once installed, verify that you can see your hardware:

Replacing a Failed Drive

Replacing a failed drive involves two steps;

  1. Identify the drive that is failing and gathering the data needed to request the RMA.
  2. Swapping the actual drive when the replacement arrives on client site.

Identify the Failing Drive

If the drive has failed entirely, the red front LED on the drive should be lit, making identification and RMA request simple.

However, if the drive has not yet failed, identifying the drive and confirming it's pending failure requires a little extra work.

Identify the failed drive:

MegaCli64 PDList aAll
<snip>

Enclosure Device ID: 10
Slot Number: 5
Drive's postion: DiskGroup: 0, Span: 0, Arm: 5
Enclosure position: 1
Device Id: 7
WWN: 5000C50054AE9C38
Sequence Number: 2
Media Error Count: 0
Other Error Count: 2
Predictive Failure Count: 0
Last Predictive Failure Event Seq Number: 0
PD Type: SAS

Raw Size: 279.396 GB [0x22ecb25c Sectors]
Non Coerced Size: 278.896 GB [0x22dcb25c Sectors]
Coerced Size: 278.875 GB [0x22dc0000 Sectors]
Firmware state: Online, Spun Up
Device Firmware Level: 5301
Shield Counter: 0
Successful diagnostics completion on :  N/A
SAS Address(0): 0x5000c50054ae9c39
SAS Address(1): 0x0
Connected Port Number: 0(path0)
Inquiry Data: SEAGATE ST9300653SS     53016XN1EMF2    @#87980
FDE Capable: Not Capable
FDE Enable: Disable
Secured: Unsecured
Locked: Unlocked
Needs EKM Attention: No
Foreign State: None
Device Speed: 6.0Gb/s
Link Speed: 6.0Gb/s
Media Type: Hard Disk Device
Drive Temperature :29C (84.20 F)
PI Eligibility:  No
Drive is formatted for PI information:  No
PI: No PI
Port-0 :
Port status: Active
Port's Linkspeed: 6.0Gb/s
Port-1 :
Port status: Active
Port's Linkspeed: 6.0Gb/s
Drive has flagged a S.M.A.R.T alert : No

Notice the line:

Other Error Count: 2

This is a sign of impending failure, despite SMART showing the drive as still healthy. You can gather more details (which Fujitsu will require to verify impending failure), run;

MegaCli -AdpEventLog -GetEvents -f raid_events.log -aALL

When this finishes gathering data, it will create a file called raid_events.log. Send this file to your Fujitsu support rep. They will validate the pending failure and issue an RMA.

Identifying The Drive Prior to Replacement

If the drive has failed, identifying the drive is as simple as looking at the front of the node for the drive with the red error LED lit.

If the drive has not failed yet, then you can use the identify command to blink the LED. That is what we'll cover here.

In the previous section, we identified the failed drive using the MegaCli64 command. You need to note the following;

MegaCli64 PDList aAll
Enclosure Device ID: 10
Slot Number: 5
...
Other Error Count: 2

The two bits of information you need are the enclosure ID and slot number. In this case, that is 10 and 5, respectively.

With that info, you can trigger the drive's red LED using the following command;

MegaCli64 -PdLocate -start -physdrv [10:5] -aALL
Adapter: 0: Device at EnclId-10 SlotId-5  -- PD Locate Start Command was successfully sent to Firmware 

Exit Code: 0x00

Once you've located the drive, you can stop the "locate" command using:

MegaCli64 -PdLocate -stop -physdrv [10:5] -aALL
Adapter: 0: Device at EnclId-10 SlotId-5  -- PD Locate Stop Command was successfully sent to Firmware 

Exit Code: 0x00

Now that you know which drive has failed, you can OFFLINE it in preparation of replacing it.

MegaCli64 -PDOffline -physdrv [10:5] -aALL
Adapter: 0: EnclId-10 SlotId-5 state changed to OffLine.

Exit Code: 0x00

You can now physically remove the failed disk and insert the replacement disk.

Monitoring the Rebuild

When a replacement disk is inserted, the array should recognise it and automatically begin rebuilding the array. You can monitor this operation by calling;

MegaCli64 -PDRbld -ProgDsply -PhysDrv [10:5] -aALL

This will display the rebuild progress as a textual bar graph. A rebuild of a 300 GB 15,000rpm SAS drive in a 6-drive array took about 30 minutes. How long it takes in your case will vary depending on disk speed, array size and load.

Checking the BBU

Check the BBU;

MegaCli64 AdpBbuCmd aAll


Creating a Backup Set

ToDo: Explain this...

mkdir ~/base
cd ~/base
mkdir root
mkdir -p etc/sysconfig/network-scripts/
mkdir -p etc/udev/rules.d/

# Root user
rsync -av /root/.bashrc          root/
rsync -av /root/.ssh             root/

# Directories
rsync -av /etc/ssh               etc/
rsync -av /etc/apcupsd           etc/
rsync -av /etc/cluster           etc/
rsync -av /etc/drbd.*            etc/
rsync -av /etc/lvm               etc/

# Specific files.
rsync -av /etc/sysconfig/network-scripts/ifcfg-{eth*,bond*,vbr*} etc/sysconfig/network-scripts/
rsync -av /etc/udev/rules.d/70-persistent-net.rules etc/udev/rules.d/
rsync -av /etc/sysconfig/network etc/sysconfig/
rsync -av /etc/hosts             etc/
rsync -av /etc/ntp.conf          etc/

# Save recreating user accounts.
rsync -av /etc/passwd            etc/
rsync -av /etc/group             etc/
rsync -av /etc/shadow            etc/
rsync -av /etc/gshadow           etc/

# If you have the cluster built and want to backup it's configs.
mkdir etc/cluster
mkdir etc/lvm
rsync -av /etc/cluster/cluster.conf etc/cluster/
rsync -av /etc/lvm/lvm.conf etc/lvm/
# NOTE: DRBD won't work until you've manually created the partitions.
rsync -av /etc/drbd.d etc/

# If you had to manually set the UUID in libvirtd;
mkdir etc/libvirt
rsync -av /etc/libvirt/libvirt.conf etc/libvirt/

# If you're running RHEL and want to backup your registration info;
rsync -av /etc/sysconfig/rhn etc/sysconfig/

# Pack it up
# NOTE: Change the name to suit your node.
tar -cvf base_an-c01n01.tar etc root

Specific Fujitsu Model Notes

RX200 S7

BIOS Changes for use in Anvil!s

To enter the BIOS, press F2 during POST.

  • Advanced
    • SATA Configuration
      • SATA Mode -> Disaabled (if no optical drive)
      • - OR -
      • SATA Mode -> AHCI Mode (if optical drive)
    • Onboard Device Configuration
      • LAN 2 Oprom -> PXE
      • Onboard SAS/SATA (SCU) -> Disabled (if no optical drive)
  • Server Mgmt
    • Asset Asset Tag -> short host name of the node, lower case
    • Temperature Monitoring -> Enabled

Save and Exit

LSI RAID Controller Setup

Warning: These instructions assume NO existing data. This is a destructive process!

After POST, press <ctrl> + H to enter the controller's WebBIOS

  • Choose the controller (usually only one available), click Start
  • Choose Configuration Wizard
    • Choose New Confgiruation -> click Next
    • Confirm the clear by clicking Yes
    • Choose Manual Configuration -> click Next
      • Click on the first drive, usually Slot: 0,..., press and hold <ctrl> button and then click to select the rest of the drives. Click Add to Array.
      • Click on Accept DG to create the drive group then click on Next
      • Click on Add to SPAN, then click on Next
      • Configure the array;
        • RAID Level -> RAID 5
        • Write Policy -> Write Back with BBU
        • Look at the R5:xxxx size on the right and enter that size in the Select Size section. Be sure to match the GB or TB suffix.
        • Click on Accept then click on Yes to accept the warning.
        • The virtual disk will now be shown on the right. Click Next to proceed.
      • Click Accept and then click Yes to save the configuration. Click Yes to acknowledge the warning and initialize the drive.
    • Click on Set Boot Drive and then click Go.
    • Clock on Home.
  • Click on Exit and then on Yes.

Reboot and you are done.


 

Any questions, feedback, advice, complaints or meanderings are welcome.
Alteeve's Niche! Enterprise Support:
Alteeve Support
Community Support
© Alteeve's Niche! Inc. 1997-2024   Anvil! "Intelligent Availability®" Platform
legal stuff: All info is provided "As-Is". Do not use anything here unless you are willing and able to take responsibility for your own actions.