minor number mismatch in Solaris

We face the issue of minor number in Solaris servers , below is the issue how to fix the same .

 

Console Error Messages during the reboot:

==================================

VxVM vxvm-startup2 INFO V-5-2-503 VxVM general startup…

Feb  5 11:01:01 vxvm:vxconfigd: V-5-1-663 Group mdmadmin-dg: Duplicate virtual device number(s):

Feb  5 11:01:01 vxvm:vxconfigd:         Volume mdmadmin_vol remapped from 234,104000 to 234,131070

Feb  5 11:01:01 vxvm:vxconfigd:         Volume mdmadminprog_vol remapped from 234,104001 to 234,131069

The system is coming up.  Please wait.

Minor number’s once the server came up:

=======================================

  1. Unix01:root# pwd

/global/.devices/node@1/dev/vx/dsk

Unix01:root# cd ../mdmadmin-dg  ( go to the Path)

Unix01:root# ls -l

total 0

brw——-   1 root     root     234,131070 Feb  5 11:11 mdmadmin_vol———->>>> Minor number

brw——-   1 root     root     234,131069 Feb  5 11:11 mdmadminprog_vol ——>>>> Minor number

  1. vxprint -htg <<dg>> | grep default

 

==================================

dg cdmadmin-dg  default      default  43000    1166196781.30.Unix01

dg mdmadmin-dg  default      default  104000   1166196841.36.Unix01 ———->>>>Minor Number

dg sdmadmin-dg  default      default  10000    1166196567.24.Unix01

It clearly shows the the minor number for “mdmadmin-dg” is not the same, per above 2 outputs:

 

 

  1. Gave a new minor number “104010” to mdmadmin-dg

 

  1. Unix01:root# vxdg reminor mdmadmin-dg 104010

Unix01:root# ls -l

total 0

brw——-   1 root     root     234,104010 Feb  5 11:34 mdmadmin_vol

brw——-   1 root     root     234,104011 Feb  5 11:34 mdmadminprog_vol

Unix01:root#

5.Unix01:root# vxprint -htg mdmadmin-dg | grep dg | grep default

dg mdmadmin-dg  default      default  104010   1166196841.36.Unix01

Unix01:root#

 

 

Blogs : Supported By VK.S

EMC power path Removal failed

EMC power path Removal failed

Error :

* Error: PowerPath pseudo(s) still in-use, busy or mounted.

* Please stop all I/Os and make sure all volumes are

* not mounted then retry removing the package.

Steps to resolve the issues .

  1. Login to Single User mode (init s)
  2. export all file system which carry emcpower devices ( e.g  there are zpool so  export  all pools except rpool)
  3. remove emc package (pkgrm EMCpower)
  4. Init 0
  5. Boot –r

Disk Replacement in ZFS

Steps by steps Process to replace a disk in zfs

1.Appserver:/ # zpool status erpprod1
pool: erpprod1
state: ONLINE
scan: scrub repaired 0 in 3h35m with 0 errors on Sat Jul 9 12:10:41 2016
config:

NAME STATE READ WRITE CKSUM
erpprod1 ONLINE 0 0 0
raidz1-0 ONLINE 0 0 0
c0t1d0 ONLINE 0 0 0
c1t1d0 ONLINE 0 0 0
c3t1d0 ONLINE 0 0 0
errors: No known data errors

2.Appserver:/ # zpool offline erpprod1 c0t1d0
3.Appserver:/ # zpool status erpprod1
pool: erpprod1
state: DEGRADED
status: One or more devices has been taken offline by the administrator.
Sufficient replicas exist for the pool to continue functioning in a
degraded state.
action: Online the device using ‘zpool online’ or replace the device with
‘zpool replace’.
scan: scrub repaired 0 in 3h35m with 0 errors on Sat Jul 9 12:10:41 2016
config:

NAME STATE READ WRITE CKSUM
erpprod1 DEGRADED 0 0 0
raidz1-0 DEGRADED 0 0 0
c0t1d0 OFFLINE 0 0 0
c1t1d0 ONLINE 0 0 0
c3t1d0 ONLINE 0 0 0
errors: No known data errors

4.Appserver:/ # cfgadm -c unconfigure sata0/1
Unconfigure the device at: /devices/pci@0,0/pci1022,7458@1/pci11ab,11ab@1:1
This operation will suspend activity on the SATA device
Continue (yes/no)? yes
5.Appserver:/ # cfgadm -al | grep c0t1d0
6.Appserver:/ # cfgadm | grep sata0/1
sata0/1 disk connected unconfigured ok

7.cfgadm -c configure sata0/1
8.cfgadm | grep sata0/1
9.zpool online erpprod1 c0t1d0
10.zpool replace erpprod1 c0t1d0

Solaris10 patching in zfs

1. freezing resources

scswitch -n -M -j rac-framework-rs
scswitch -n -M -j crs-framework-rs
scswitch -n -M -j qfs-mds-rs
scswitch -n -M -j qfs-mds-oradatb-rs
scswitch -n -M -j qfs-mds-oralog-rs
scswitch -n -M -j qfs-mds-arch-rs
scswitch -n -M -j qfs-oradata-mp-rs
scswitch -n -M -j qfs-oradatb-mp-rs
scswitch -n -M -j qfs-oralog-mp-rs
scswitch -n -M -j qfs-arch-mp-rs
scswitch -n -M -j rac-v55x-proxy-rs
scswitch -n -M -j rac-v58x-proxy-rs
scswitch -n -M -j rac-v51x-proxy-rs
scswitch -n -M -j rac-v81x-proxy-rs
scswitch -n -M -j rac-v44x-proxy-rs
scswitch -n -M -j rac-v57x-proxy-rs
scswitch -n -M -j qfs-oradatd-rs
scswitch -n -M -j qfs-oradatc-rs
scswitch -n -M -j qfs-oraarch1-rs
scswitch -n -M -j qfs-oralog1-rs
scswitch -n -M -j qfs-oradatac-rs
scswitch -n -M -j qfs-oradatad-rs
scswitch -n -M -j scal-oradatd-rs
scswitch -n -M -j scal-oradatc-rs
scswitch -n -M -j scal-oraarch1-rs
scswitch -n -M -j scal-oralog1-rs
scswitch -n -M -j scal-oradatac-rs
scswitch -n -M -j scal-oradatad-rs
scswitch -n -M -j qfs-oradatf-rs
scswitch -n -M -j qfs-oradate-rs
scswitch -n -M -j qfs-oralog2-rs
scswitch -n -M -j qfs-oraarch2-rs
scswitch -n -M -j scal-oradatf-rs
scswitch -n -M -j scal-oradate-rs
scswitch -n -M -j scal-oralog2-rs
scswitch -n -M -j scal-oraarch2-rs

2. Bringdown the databases

3. Creating the BOOT ENVIRONMENT for OS patching

# lucreate -n newenv -p rpool

4 checking the boot environment status

# lustatus

5 navigating to the patch clusterand installing pre-requiste patchset

# cd /var/tmp/10_Recommended
#./installpatchset –apply-prereq –s10patchset

6 once everything is fine apply the patch on alternate boot dataset as follows

# ./installpatchset –s10patchset -B newenv

7 # lustatus

7 activate the new boot environment

# luactivate newenv

# lustatus

8 reboot the server using

# init 6

Creating service group in vcs

CVM is an extension of Veritas Volume Manager, the industry-standard storage virtualization platform. CVM extends the concepts of VxVM across multiple nodes. Each node recognizes the same logical volume layout, and more importantly, the same state of all volume resources.

CVM supports performance-enhancing capabilities, such as striping, mirroring, and mirror break-off (snapshot) for off-host backup. You can use standard VxVM commands from one node in the cluster to manage all storage. All other nodes immediately recognize any changes in disk group and volume configuration with no user interaction.

Process to create CVM SG in VCS is given below  and its tested as well.

Create mount point on both servers.
#mkdir –p /infapps/dae
In Second node also create the mount point.
#mkdir –p /infapps/dae

=========================================================

# haconf -makerw

# hagrp -add vrts_vea_cfs_int_cfsmountinfapps_dae
VCS NOTICE V-16-1-10136 Group added; populating SystemList and setting the Parallel attribute recommended before adding resources

# hagrp -modify vrts_vea_cfs_int_cfsmountinfapps_dae Parallel 1

# hagrp -modify vrts_vea_cfs_int_cfsmountinfapps_dae SystemList nj09mhf0373 0 nj09mhf0374 1

# hagrp -modify vrts_vea_cfs_int_cfsmountinfapps_dae AutoStartList nj09mhf0373 nj09mhf0374

# hares -add cvmvoldg28 CVMVolDg vrts_vea_cfs_int_cfsmountinfapps_dae
VCS NOTICE V-16-1-10242 Resource added. Enabled attribute must be set before agent monitors

# hagrp -resources vrts_vea_cfs_int_cfsmountinfapps_dae
cvmvoldg28
# hares -modify cvmvoldg28 CVMDiskGroup sdap_racdg

# hares -modify cvmvoldg28 CVMActivation sw

# hares -modify cvmvoldg28 CVMVolume infapps_dae

# hares -add cfsmount102 CFSMount vrts_vea_cfs_int_cfsmountinfapps_dae
VCS NOTICE V-16-1-10242 Resource added. Enabled attribute must be set before agent monitors

# hagrp -resources vrts_vea_cfs_int_cfsmountinfapps_dae
cvmvoldg28
cfsmount102

# hares -modify cfsmount102 MountPoint “/infapps/dae”

# hares -modify cfsmount102 BlockDevice “/dev/vx/dsk/sdap_racdg/infapps_dae”

# hares -link cfsmount102 cvmvoldg28

# hagrp -link vrts_vea_cfs_int_cfsmountinfapps_dae cvm online local firm

# hagrp -enableresources vrts_vea_cfs_int_cfsmountinfapps_dae

# haconf -dump -makero

# hagrp -online vrts_vea_cfs_int_cfsmountinfapps_dae -any
VCS NOTICE V-16-1-50735 Attempting to online group on system nj09mhf0373
VCS NOTICE V-16-1-50735 Attempting to online group on system nj09mhf0374

# hagrp -state vrts_vea_cfs_int_cfsmountinfapps_dae
#Group Attribute System Value
vrts_vea_cfs_int_cfsmountinfapps_dae State nj09mhf0373 |ONLINE|
vrts_vea_cfs_int_cfsmountinfapps_dae State nj09mhf0374 |ONLINE|

By: VK

Firmware upgrade activity in Solaris10

Firmware’s upgrade Procedure.

1.First check firmware Version

  1. prtdiag –v

————————————————————

Sun System Firmware 8.1.1.c 2011/09/12 10:54

2.make auto boot true

unixplanetary:# eeprom |grep -i auto

auto-boot-on-error?=false

auto-boot?=true

3.Download the firmware from oracle website .

# unzip 152475-04.zip
Archive: 152475-04.zip
creating: 152475-04/
inflating: 152475-04/metadata.xml
inflating: 152475-04/Sun_System_Firmware-8_9_8-SPARC_T4-1.pkg
inflating: 152475-04/Install.info
inflating: 152475-04/README.152475-04
inflating: 152475-04/license.txt
inflating: 152475-04/sysfwdownload
inflating: 152475-04/sysfwdownload.README
inflating: 152475-04/copyright
inflating: 152475-04/LEGAL_LICENSE.TXT
inflating: 152475-04/152475-04.html

————————————————————
Sun System Firmware 8.7.2.b 2015/05/06 15:23

====================== System PROM revisions =======================
Version
————————————————————
OBP 4.37.2 2015/04/10 08:59

Chassis Serial Number
———————
1216BDY152
# pwd
/tmp/152475-04
# ls
152475-04.html
Install.info
LEGAL_LICENSE.TXT
README.152475-04
Sun_System_Firmware-8_9_8-SPARC_T4-1.pkg
copyright
license.txt
metadata.xml
sysfwdownload
sysfwdownload.README
# ./sysfwdownload -u Sun_System_Firmware-8_9_8-SPARC_T4-1.pkg  ( This command will install the firmware and automatically reboot the system)
WARNING: Host will be powered down for automatic firmware update when download is completed.
Do you want to continue(yes/no)? yes

………. (5%)……..Aug 24 13:36:46 bkprod01 tldcd[3711]: TLD(2) [3711] robotic path /dev/sg/c0tw500308c0022c8803l1 does not exist
.. (11%)………. (17%)……..Aug 24 13:36:46 bkprod01 last message repeated 1 time
Aug 24 13:37:00 bkprod01 bpjava-msvc[5218]: pam_dial_auth: terminal-device not specifiedby login, returning Error in underlying service module.
.. (23%)..Aug 24 13:37:05 bkprod01 tldcd[3711]: TLD(2) [3711] robotic path /dev/sg/c0tw500308c0022c8803l1 does not exist
Aug 24 13:37:05 unixplanetary last message repeated 1 time
Aug 24 13:37:05 unixplanetary tldd[3646]: TLD(2) unavailable: initialization failed: Unable to open robotic path
…….. (29%)………. (34%)….Aug 24 13:37:22 bkprod01 bpjava-msvc[5350]: pam_dial_auth: terminal-device not specifiedby login, returning Error in underlying service module.
…… (40%)………. (46%)..Aug 24 13:37:35 bkprod01 tldcd[3711]: TLD(2) [3711] robotic path /dev/sg/c0tw500308c0022c8803l1 does not exist
…….. (52%)………. (58%)……Aug 24 13:37:35 bkprod01 last message repeated 1 time
…. (64%)………. (69%)………. (75%)………. (81%)………. (87%)………. (93%)..Aug 24 13:38:48 bkprod01 tldcd[3711]: TLD(2) [3711] robotic path /dev/sg/c0tw500308c0022c8803l1 does not exist
…….. (99%).. (100%)

Download completed successfully.

# Aug 24 13:38:48 unixplanetary last message repeated 1 time
Aug 24 13:39:03 bkprod01 unix: WARNING: Power-off requested, system will now shutdown.

Shutdown started. Thu Aug 24 13:39:03 MDT 2017

Changing to init state 5 – please wait
updating /platform/sun4v/boot_archive
Aug 24 13:39:07 unixplanetary tldcd[3711]: TLD(2) [3711] robotic path /dev/sg/c0tw500308c0022c8803l1 does not exist
Aug 24 13:39:07 unixplanetary last message repeated 1 time
Aug 24 13:39:07 unixplanetary tldd[3646]: TLD(2) unavailable: initialization failed: Unable to open robotic path
Broadcast Message from root (msglog) on bkprod01 Thu Aug 24 13:39:08…
THE SYSTEM unixplanetary IS BEING SHUT DOWN NOW ! ! !
Log off now or risk your files being damaged

3.unixplanetary-ilo login: root

Password:

 

Oracle(R) Integrated Lights Out Manager

 

Version 3.2.6.5 r115840

 

-> show /HOST

 

sysfw_version = Sun System Firmware 8.9.8 2017/03/22 13:01

You do not have write access in guest domain solaris

Whenever we are logging to guest domain we are getting below messages .

unixplanetary:# telnet 0 5001
Trying 0.0.0.0…
Connected to 0.

Escape character is ‘^]’.

Connecting to console “ldom1” in group “ldom1” ….

You do not have write access
You do not have write access

then login control domain

Check the telnet process and kill them

unixplanetary:# ps -ef |grep -i telnet
root 14672 11041 0 08:12:23 pts/2 0:00 telnet localhost 5001
root 27207 27148 0 08:41:43 pts/6 0:00 telnet 0 5001
root 824 776 0 08:48:09 pts/8 0:00 grep -i telnet

kill -9 14672

kill -9 27207

Once you killed you will be able to logged in  .

Failed to determine zone configuration for target boot environment

Failed to determine zone configuration for target boot environment

Follow the below steps.

1.lumount new-bootenv /a

2.  vi /a/etc/zones/index

3.# grep global /a/etc/zones/index
global:configured:/

Changed to

# grep global /a/etc/zones/index
global:installed:/:

4. luumount new-bootenv /a

5 .umount /a

 

Re-run ./installpatchset –s10patchset -B  New-bootenv

Firmware upgrade issue and Resolution

Error :
Unrecognized Chassis: This module is installed in an unknown or unsupported
chassis. You must upgrade the firmware to a newer version that supports
this chassis.


Solution :
-> cd /SP/network/test
-> set ping=152.159.44.105 -======== IP where the image is loaded.
Ping of 152.159.44.105 succeeded
Unrecognized Chassis: This module is installed in an unknown or unsupported
chassis. You must upgrade the firmware to a newer version that supports
this chassis.

load -source scp://:@//Sun_System_Firmware-8_3_40-SPARC_T3-1.pkg
-> load -force -source scp://tmp:qazwsx12!@152.159.44.105/tmp/152738-01/Sun_System_Firmware-8_3_40-SPARC_T3-1.pkg

NOTE: An upgrade takes several minutes to complete. ILOM
will enter a special mode to load new firmware. No
other tasks can be performed in ILOM until the
firmware upgrade is complete and ILOM is reset.

Are you sure you want to load the specified file (y/n)? y
Preserve existing configuration (y/n)? y
……………………………………………………………………………..
………………………………………………………………………………………………………..

Firmware update is complete.
ILOM will now be restarted with the new firmware.

Unrecognized Chassis: This module is installed in an unknown or unsupported
chassis. You must upgrade the firmware to a newer version that supports
this chassis.

-> /sbin/reboot

You do not need to do reboot ILOM. It will automatically reboot it and then you have start the server with
-> show /HOST status
-> start /SYS
-> show /HOST status
-> start /SP/console