I had such a thing also once. If you create a zone, but make a typo for the zonepath (e.q. Zpool/tralala instead of zonepool/tralala) the zone will be installed but won't run.
Kind regards, The out-side Op 14 okt. 2012 om 15:59 heeft Julien Peeters <[email protected]> het volgende geschreven: > Hi Jim, > > When I ran the commands you gave me, I had a core dump error. > > # zlogin -S websrv > [Connected to zone 'websrv' pts/4] > > @websrv:~$ id > uid=0(root) gid=0(root) > > @websrv:~$ zfs mount -a > Memory fault(coredump) > > @websrv:~$ zfs list > Memory fault(coredump) > > @websrv:~$ /usr/bin/svcs -xv > svc:/system/filesystem/local:default (local file system mounts) > State: maintenance since Sun Oct 14 04:50:53 2012 > Reason: Start method exited with $SMF_EXIT_ERR_FATAL. > See: http://illumos.org/msg/SMF-8000-KS > See: /var/svc/log/system-filesystem-local:default.log > Impact: 14 dependent services are not running: > svc:/system/filesystem/autofs:default > svc:/system/system-log:default > svc:/milestone/multi-user:default > svc:/milestone/multi-user-server:default > svc:/network/ssh:default > svc:/system/sysidtool:net > svc:/network/rpc/bind:default > svc:/network/inetd:default > svc:/system/sysidtool:system > svc:/network/inetd-upgrade:default > svc:/system/cron:default > svc:/system/boot-archive-update:default > svc:/network/shares/group:default > svc:/system/sac:default > > @websrv:~$ cat /var/svc/log/system-filesystem-local:default.log > [ Oct 14 04:50:39 Enabled. ] > [ Oct 14 04:50:52 Executing start method ("/lib/svc/method/fs-local"). ] > /lib/svc/method/fs-local: line 90: 10457: Memory fault(coredump) > WARNING: /usr/sbin/zfs mount -a failed: exit status 267 > [ Oct 14 04:50:53 Method "start" exited with status 95. ] > [ Oct 14 06:56:30 Enabled. ] > [ Oct 14 06:56:33 Executing start method ("/lib/svc/method/fs-local"). ] > /lib/svc/method/fs-local: line 90: 11333: Memory fault(coredump) > WARNING: /usr/sbin/zfs mount -a failed: exit status 267 > [ Oct 14 06:56:34 Method "start" exited with status 95. ] > > Here is all the configuration/outputs for the given zone. > > On the GZ when the zone is halted: > > # zonecfg -z websrv export > create -b > set zonepath=/storage/zones/websrv > set brand=ipkg > set autoboot=true > set ip-type=exclusive > add net > set physical=vnic0 > end > > # zfs list > NAME USED AVAIL REFER MOUNTPOINT > rpool 10.0G 4.64G 45.5K /rpool > rpool/ROOT 1.88G 4.64G 31K legacy > rpool/ROOT/openindiana 1.88G 4.64G 1.78G / > rpool/dump 3.94G 4.64G 3.94G - > rpool/export 31K 4.64G 31K /export > rpool/swap 4.19G 8.70G 133M - > storage/zones 450M 3.56T 202K /storage/zones > storage/zones/websrv 450M 3.56T 213K /storage/zones/websrv > storage/zones/websrv/ROOT 450M 3.56T 192K legacy > storage/zones/websrv/ROOT/zbe 449M 3.56T 449M legacy > > # ls -l /storage/zones/websrv/ > total 22 > drwxr-xr-x 2 root sys 2 2012-10-14 13:50 dev > drwxr-xr-x 16 root root 19 2012-10-14 15:45 root > > # mount > / on rpool/ROOT/openindiana read/write/setuid/devices/dev=2cd0002 on Thu Jan > 1 01:00:00 1970 > /devices on /devices read/write/setuid/devices/dev=83c0000 on Sun Oct 14 > 12:26:49 2012 > /dev on /dev read/write/setuid/devices/dev=8400000 on Sun Oct 14 12:26:49 2012 > /system/contract on ctfs read/write/setuid/devices/dev=84c0001 on Sun Oct 14 > 12:26:49 2012 > /proc on proc read/write/setuid/devices/dev=8440000 on Sun Oct 14 12:26:49 > 2012 > /etc/mnttab on mnttab read/write/setuid/devices/dev=8500001 on Sun Oct 14 > 12:26:49 2012 > /etc/svc/volatile on swap read/write/setuid/devices/xattr/dev=8540001 on Sun > Oct 14 12:26:49 2012 > /system/object on objfs read/write/setuid/devices/dev=8580001 on Sun Oct 14 > 12:26:49 2012 > /etc/dfs/sharetab on sharefs read/write/setuid/devices/dev=85c0001 on Sun Oct > 14 12:26:49 2012 > /lib/libc.so.1 on /usr/lib/libc/libc_hwcap2.so.1 > read/write/setuid/devices/dev=2cd0002 on Sun Oct 14 12:27:06 2012 > /dev/fd on fd read/write/setuid/devices/dev=86c0001 on Sun Oct 14 12:27:17 > 2012 > /tmp on swap read/write/setuid/devices/xattr/dev=8540002 on Sun Oct 14 > 12:27:18 2012 > /var/run on swap read/write/setuid/devices/xattr/dev=8540003 on Sun Oct 14 > 12:27:18 2012 > /rpool on rpool > read/write/setuid/devices/nonbmand/exec/xattr/atime/dev=2cd0007 on Sun Oct 14 > 12:27:30 2012 > /storage on storage > read/write/setuid/devices/nonbmand/exec/xattr/noatime/dev=2cd0008 on Sun Oct > 14 12:27:30 2012 > /export on rpool/export > read/write/setuid/devices/nonbmand/exec/xattr/atime/dev=2cd001e on Sun Oct 14 > 13:39:21 2012 > /storage/zones/websrv on storage/zones/websrv > read/write/setuid/devices/nonbmand/exec/xattr/noatime/dev=2cd0021 on Sun Oct > 14 13:46:34 2012 > /storage/zones/websrv/root on storage/zones/websrv/ROOT/zbe > read/write/setuid/devices/nonbmand/exec/xattr/noatime/dev=2cd0024 on Sun Oct > 14 15:49:22 2012 > > I don't understand where the problem comes from... Any other idea? > > Best, > Julien. > > On 14/10/12 14:13 , Jim Klimov wrote: >> 2012-10-14 16:06, Julien Peeters wrote: >>> Hi ML readers, >>> >>> I've recently installed a OpenIndiana system for a storage server. >>> >>> I would like to create zones for security purpose and when I do so I >>> encountered some problems. >>> >>> When I try to boot a zone, the following error appear: >> ... >>> svc:/system/filesystem/local:default: WARNING: /usr/sbin/zfs mount -a >>> failed: exit status 267 >>> Oct 14 04:50:53 svc.startd[9831]: svc:/system/filesystem/local:default: >>> Method "/lib/svc/method/fs-local" failed with exit status 95. >>> Oct 14 04:50:53 svc.startd[9831]: system/filesystem/local:default failed >>> fatally: transitioned to maintenance (see 'svcs -xv' for details) >> >> I believe your zone definition manifest includes delegation of >> some datasets from GZ into the local zone. These datasets have >> defined mountpoints relative to the zone root. >> >> Are the directories at these mountpoints empty? Non-emptiness >> and refusal to overlay-mount are the typical cause of such zfs >> error. To test this, you can "zlogin -S zonename" from GZ to >> get into administrative shell and try to run the "zfs mount -a" >> and "zfs list" commands, as well as "zfs mount pool/datasetname" >> to mount the dataset in question. If you have a hierarchy of >> datasets, mount each one alone and see if the deeper level one's >> mountpoint inside the parent dataset is empty. >> >> HTH, >> //Jim Klimov >> >> >> _______________________________________________ >> OpenIndiana-discuss mailing list >> [email protected] >> http://openindiana.org/mailman/listinfo/openindiana-discuss > > > _______________________________________________ > OpenIndiana-discuss mailing list > [email protected] > http://openindiana.org/mailman/listinfo/openindiana-discuss _______________________________________________ OpenIndiana-discuss mailing list [email protected] http://openindiana.org/mailman/listinfo/openindiana-discuss
