OpenSolaris ZFS Kernel Notice BAD TRAP

Hello All,

I have an Opensolaris file server running on  snv_126 i86pc i386 i86pc. I have been backing up my zpools to an e-sata pool weekly for about 5 months now. Last week, the machine rebooted while doing the backup. The only thing I can find in the logs is attached below.

This has now started happening on a daily basis. Can anyone shed any light on what this means?:
(#pf Page fault) rp=ffffff0010568eb0 addr=30 occurred in module "zfs" due to a NULL pointer dereference
 
This is always the last thing before the machine reboots. I will be happy to provide any other info needed, just point me in the right direction.

Thanks!
Rob
Apr  8 08:48:39 Stubborn2 unix: [ID 836849 kern.notice] 
Apr  8 08:48:39 Stubborn2 ^Mpanic[cpu7]/thread=ffffff0010569c60: 
Apr  8 08:48:39 Stubborn2 genunix: [ID 335743 kern.notice] BAD TRAP: type=e (#pf Page fault) rp=ffffff0010568eb0 addr=30 occurred in module "zfs" due to a NULL pointer dereference
Apr  8 08:48:39 Stubborn2 unix: [ID 100000 kern.notice] 
Apr  8 08:48:39 Stubborn2 unix: [ID 839527 kern.notice] sched: 
Apr  8 08:48:39 Stubborn2 unix: [ID 753105 kern.notice] #pf Page fault
Apr  8 08:48:39 Stubborn2 unix: [ID 532287 kern.notice] Bad kernel fault at addr=0x30
Apr  8 08:48:39 Stubborn2 unix: [ID 243837 kern.notice] pid=0, pc=0xfffffffff7932664, sp=0xffffff0010568fa8, eflags=0x10292
Apr  8 08:48:39 Stubborn2 unix: [ID 211416 kern.notice] cr0: 8005003b<pg,wp,ne,et,ts,mp,pe> cr4: 6f8<xmme,fxsr,pge,mce,pae,pse,de>
Apr  8 08:48:39 Stubborn2 unix: [ID 624947 kern.notice] cr2: 30
Apr  8 08:48:39 Stubborn2 unix: [ID 625075 kern.notice] cr3: 3c00000
Apr  8 08:48:39 Stubborn2 unix: [ID 625715 kern.notice] cr8: c
Apr  8 08:48:39 Stubborn2 unix: [ID 100000 kern.notice] 
Apr  8 08:48:39 Stubborn2 unix: [ID 592667 kern.notice]         rdi:                0 rsi:                8 rdx:                0
Apr  8 08:48:39 Stubborn2 unix: [ID 592667 kern.notice]         rcx:                1  r8:                0  r9:                0
Apr  8 08:48:39 Stubborn2 unix: [ID 592667 kern.notice]         rax:                0 rbx:                0 rbp: ffffff0010568fb0
Apr  8 08:48:39 Stubborn2 unix: [ID 592667 kern.notice]         r10: fffffffffb84f430 r11: ffffff02e2005930 r12:                0
Apr  8 08:48:39 Stubborn2 unix: [ID 592667 kern.notice]         r13:                0 r14:            6ded0 r15: ffffff02e1a772d0
Apr  8 08:48:39 Stubborn2 unix: [ID 592667 kern.notice]         fsb:                0 gsb: ffffff02d814eb00  ds:               4b
Apr  8 08:48:39 Stubborn2 unix: [ID 592667 kern.notice]          es:               4b  fs:                0  gs:              1c3
Apr  8 08:48:39 Stubborn2 unix: [ID 592667 kern.notice]         trp:                e err:                0 rip: fffffffff7932664
Apr  8 08:48:39 Stubborn2 unix: [ID 592667 kern.notice]          cs:               30 rfl:            10292 rsp: ffffff0010568fa8
Apr  8 08:48:39 Stubborn2 unix: [ID 266532 kern.notice]          ss:               38
Apr  8 08:48:39 Stubborn2 unix: [ID 100000 kern.notice]

Open in new window

LVL 1
tongsleeAsked:
Who is Participating?
I wear a lot of hats...

"The solutions and answers provided on Experts Exchange have been extremely helpful to me over the last few years. I wear a lot of hats - Developer, Database Administrator, Help Desk, etc., so I know a lot of things but not a lot about one thing. Experts Exchange gives me answers from people who do know a lot about one thing, in a easy to use platform." -Todd S.

Brian UtterbackPrinciple Software EngineerCommented:
Do you get a core dump? If not, do you have core dumps enabled? You will probably need a core dump to analyze the problem.
0
tongsleeAuthor Commented:
/var/adm/messages shows:
[ID 454863 kern.info] dump on /dev/zvol/dsk/rpool/dump size 4090 MB

When I use
mdb /dev/zvol/dsk/rpool/dump
I get the > prompt I enter
$c and I get mdb: command is not supported by current target.

What am I missing?

Thanks!
0
tongsleeAuthor Commented:
Ignore the last, savecore is not enabled... I'm working on enabling it. I'll post back
0
Powerful Yet Easy-to-Use Network Monitoring

Identify excessive bandwidth utilization or unexpected application traffic with SolarWinds Bandwidth Analyzer Pack.

tongsleeAuthor Commented:
Alrighty,

I've enable savecore. After creating the directories, and manually running savecore, mdb 0 reveals the code below.

Any thoughts?

Thanks again for all of the help!
Rob

rcherven@Stubborn2:/var/crash/Stubborn2$ mdb 0
Loading modules: [ unix genunix specfs dtrace mac cpu.generic uppc pcplusmp rootnex scsi_vhci zfs sd sockfs ip hook neti sctp arp usba uhci fctl md lofs fcip fcp cpc random crypto smbsrv nfs logindmux ptm ufs nsmb sppp ipc ]
> $C
ffffff000f4ef3b0 vdev_is_dead+0xc(0)
ffffff000f4ef3d0 vdev_readable+0x16(0)
ffffff000f4ef410 vdev_mirror_child_select+0x61(ffffff02fa41da10)
ffffff000f4ef450 vdev_mirror_io_start+0xda(ffffff02fa41da10)
ffffff000f4ef490 zio_vdev_io_start+0x1ba(ffffff02fa41da10)
ffffff000f4ef4c0 zio_execute+0xa0(ffffff02fa41da10)
ffffff000f4ef4e0 zio_nowait+0x42(ffffff02fa41da10)
ffffff000f4ef580 arc_read_nolock+0x82d(0, ffffff02d716b000, ffffff02e3fdc000, 0, 0, 6, 3, ffffff000f4ef65c, ffffff000f4ef670)
ffffff000f4ef620 arc_read+0x75(0, ffffff02d716b000, ffffff02e3fdc000, ffffff02e3a7f928, 0, 0, 6, 3, ffffff000f4ef65c, ffffff000f4ef670)
ffffff000f4ef6c0 dbuf_prefetch+0x131(ffffff02e3a80018, 20)
ffffff000f4ef710 dmu_zfetch_fetch+0xa8(ffffff02e3a80018, 20, 1)
ffffff000f4ef750 dmu_zfetch_dofetch+0xb8(ffffff02e3a80278, ffffff02f4c52868)
ffffff000f4ef7b0 dmu_zfetch_find+0x436(ffffff02e3a80278, ffffff000f4ef7c0, 1)
ffffff000f4ef870 dmu_zfetch+0xac(ffffff02e3a80278, 2b0000, 4000, 1)
ffffff000f4ef8d0 dbuf_read+0x170(ffffff02f3d8ea00, 0, 2)
ffffff000f4ef950 dnode_hold_impl+0xed(ffffff02e2a2f040, 1591, 1, ffffff02e4e71478, ffffff000f4ef998)
ffffff000f4ef980 dnode_hold+0x2b(ffffff02e2a2f040, 1591, ffffff02e4e71478, ffffff000f4ef998)
ffffff000f4ef9e0 dmu_tx_hold_object_impl+0x4a(ffffff02e4e71478, ffffff02e2a2f040, 1591, 2, 0, 0)
ffffff000f4efa00 dmu_tx_hold_bonus+0x2a(ffffff02e4e71478, 1591)
ffffff000f4efa50 zfs_inactive+0x99(ffffff030213ae80, ffffff02d4ed6d88, 0)
ffffff000f4efaa0 fop_inactive+0xaf(ffffff030213ae80, ffffff02d4ed6d88, 0)
ffffff000f4efac0 vn_rele+0x5f(ffffff030213ae80)
ffffff000f4efae0 smb_node_free+0x7d(ffffff02f098b2a0)
ffffff000f4efb10 smb_node_release+0x9a(ffffff02f098b2a0)
ffffff000f4efb30 smb_ofile_delete+0x76(ffffff03026d5d18)
ffffff000f4efb60 smb_ofile_release+0x84(ffffff03026d5d18)
ffffff000f4efb80 smb_request_free+0x23(ffffff02fa4b0058)
ffffff000f4efbb0 smb_session_worker+0x6e(ffffff02fa4b0058)
ffffff000f4efc40 taskq_d_thread+0xb1(ffffff02e51b9e90)
ffffff000f4efc50 thread_start+8()
> $q

Open in new window

0
Brian UtterbackPrinciple Software EngineerCommented:
What does the output of this command show:
zpool status -v rootpool
0
Brian UtterbackPrinciple Software EngineerCommented:
Try this:

Boot into single user mode, then do "rm /etc/zfs/zpool.cache" . Then reboot.
0
tongsleeAuthor Commented:
I run a scrub thru cron at night (which shows no errors,) but below is the output you requested.

Will rm'ing the cache cause any other issues? I had seen that mentioned on some older bug reports, but was hesitant to try it because I didn't know if it would be recreated auto-magically on boot.

Thanks for sticking with me!
Rob
rcherven@Stubborn2:~$ zpool status -v rpool
  pool: rpool
 state: ONLINE
 scrub: none requested
config:

        NAME         STATE     READ WRITE CKSUM
        rpool        ONLINE       0     0     0
          c11t0d0s0  ONLINE       0     0     0

errors: No known data errors

Open in new window

0
Brian UtterbackPrinciple Software EngineerCommented:
Shouldn't cause a problem. It should be recreated.
0
tongsleeAuthor Commented:
I also have two data pools of about 2 TB in size in addition to the rpool. Do I need to export those prior to deleting this? Or are there any other precautions to take with those pools?

Just for info, is there just one cache for all zpools?

Thanks!
0
Brian UtterbackPrinciple Software EngineerCommented:
Wait. I found something that says that rm zpool.cache is no longer safe. Let me research this.
0
tongsleeAuthor Commented:
Alrighty. The little bits I've seen reference the cache in respect to loading on boot. Mine boots fine, but at some point, when people begin access the share, it will panic and reboot. I've since tried export/re-import the pools, no joy.

If I can provide anymore dump/log info, please just let me know.

Thanks again for helping out
Rob
0
Brian UtterbackPrinciple Software EngineerCommented:
Can you get the kernel stack on more than one core file so we can tell what is common and what is just circumstance?
0
tongsleeAuthor Commented:
Here's the 4 that I have
rcherven@Stubborn2:/var/crash/Stubborn2$ mdb 0
Loading modules: [ unix genunix specfs dtrace mac cpu.generic uppc pcplusmp rootnex scsi_vhci zfs sd sockfs ip hook neti sctp arp usba uhci fctl md lofs fcip fcp cpc random crypto smbsrv nfs logindmux ptm ufs nsmb sppp ipc ]
> $c
vdev_is_dead+0xc(0)
vdev_readable+0x16(0)
vdev_mirror_child_select+0x61(ffffff02fa41da10)
vdev_mirror_io_start+0xda(ffffff02fa41da10)
zio_vdev_io_start+0x1ba(ffffff02fa41da10)
zio_execute+0xa0(ffffff02fa41da10)
zio_nowait+0x42(ffffff02fa41da10)
arc_read_nolock+0x82d(0, ffffff02d716b000, ffffff02e3fdc000, 0, 0, 6)
arc_read+0x75(0, ffffff02d716b000, ffffff02e3fdc000, ffffff02e3a7f928, 0, 0)
dbuf_prefetch+0x131(ffffff02e3a80018, 20)
dmu_zfetch_fetch+0xa8(ffffff02e3a80018, 20, 1)
dmu_zfetch_dofetch+0xb8(ffffff02e3a80278, ffffff02f4c52868)
dmu_zfetch_find+0x436(ffffff02e3a80278, ffffff000f4ef7c0, 1)
dmu_zfetch+0xac(ffffff02e3a80278, 2b0000, 4000, 1)
dbuf_read+0x170(ffffff02f3d8ea00, 0, 2)
dnode_hold_impl+0xed(ffffff02e2a2f040, 1591, 1, ffffff02e4e71478, ffffff000f4ef998)
dnode_hold+0x2b(ffffff02e2a2f040, 1591, ffffff02e4e71478, ffffff000f4ef998)
dmu_tx_hold_object_impl+0x4a(ffffff02e4e71478, ffffff02e2a2f040, 1591, 2, 0, 0)
dmu_tx_hold_bonus+0x2a(ffffff02e4e71478, 1591)
zfs_inactive+0x99(ffffff030213ae80, ffffff02d4ed6d88, 0)
fop_inactive+0xaf(ffffff030213ae80, ffffff02d4ed6d88, 0)
vn_rele+0x5f(ffffff030213ae80)
smb_node_free+0x7d(ffffff02f098b2a0)
smb_node_release+0x9a(ffffff02f098b2a0)
smb_ofile_delete+0x76(ffffff03026d5d18)
smb_ofile_release+0x84(ffffff03026d5d18)
smb_request_free+0x23(ffffff02fa4b0058)
smb_session_worker+0x6e(ffffff02fa4b0058)
taskq_d_thread+0xb1(ffffff02e51b9e90)
thread_start+8()
> $q
####################################################################
rcherven@Stubborn2:/var/crash/Stubborn2$ mdb 1
Loading modules: [ unix genunix specfs dtrace mac cpu.generic uppc pcplusmp rootnex scsi_vhci zfs sd sockfs ip hook neti sctp arp usba uhci fctl md lofs random crypto fcip fcp cpc smbsrv nfs logindmux ptm ufs nsmb sppp ipc ]
> $c
vdev_is_dead+0xc(0)
vdev_readable+0x16(0)
vdev_mirror_child_select+0x61(ffffff03345459f8)
vdev_mirror_io_start+0xda(ffffff03345459f8)
zio_vdev_io_start+0x1ba(ffffff03345459f8)
zio_execute+0xa0(ffffff03345459f8)
zio_nowait+0x42(ffffff03345459f8)
arc_read_nolock+0x82d(0, ffffff02d7180000, ffffff02e4d09000, 0, 0, 6)
arc_read+0x75(0, ffffff02d7180000, ffffff02e4d09000, ffffff02e4973738, 0, 0)
dbuf_prefetch+0x131(ffffff02e4ce8a80, 20)
dmu_prefetch+0x8c(ffffff02debca040, 410, 0, 0)
zfs_readdir+0x50f(ffffff03369fb940, ffffff001097a630, ffffff02d4ed6a88, ffffff001097a6ec, ffffffffc0015fd0, 1)
fop_readdir+0xab(ffffff03369fb940, ffffff001097a630, ffffff02d4ed6a88, ffffff001097a6ec, ffffffffc0015fd0, 1)
smb_vop_readdir+0x10e(ffffff03369fb940, 0, ffffff02f395fcd0, ffffff02f395fb94, ffffff001097a6ec, 0)
smb_odir_next_odirent+0xdd(ffffff02f395fb38, ffffff02fb3b6400)
smb_odir_read_fileinfo+0xcb(ffffff02e213e650, ffffff02f395fb38, ffffff001097a7d0, ffffff001097a9b4)
smb_trans2_find_entries+0xa2(ffffff02e213e650, ffffff0332de7e08, ffffff02f395fb38, ffffff001097a9c0, ffffff001097a9b4)
smb_com_trans2_find_first2+0x14a(ffffff02e213e650, ffffff0332de7e08)
smb_trans2_dispatch+0x195(ffffff02e213e650, ffffff0332de7e08)
smb_com_transaction2+0x17b(ffffff02e213e650)
smb_dispatch_request+0x455(ffffff02e213e650)
smb_session_worker+0x4a(ffffff02e213e650)
taskq_d_thread+0xb1(ffffff032dc54b68)
thread_start+8()
> $q

####################################################################
rcherven@Stubborn2:/var/crash/Stubborn2$ mdb 2
Loading modules: [ unix genunix specfs dtrace mac cpu.generic uppc pcplusmp rootnex scsi_vhci zfs sd sockfs ip hook neti sctp arp usba uhci fctl md lofs random crypto fcip fcp cpc smbsrv nfs logindmux ptm ufs nsmb sppp ipc ]
> $c
rootnex_coredma_sync+0x35(ffffff02d3b5bc88, ffffff02d7f81540, ffffff02db723000, 6, 79, 2)
rootnex_dma_sync+0x36(ffffff02d3b5bc88, ffffff02d7f81540, ffffff02db723000, 6, 79, 2)
ddi_dma_sync+0x7e(ffffff02db723000, 6, 79, 2)
e1000g_receive+0x156(ffffff02d8e93640, ffffff000fbb1b78, 0)
e1000g_intr_pciexpress+0x143(ffffff02d8e8f000)
av_dispatch_autovect+0x7c(4d)
dispatch_hardint+0x33(4d, 0)
switch_sp_and_call+0x13()
do_interrupt+0xa3(ffffff000fb7bae0, 0)
_interrupt+0xb8()
i86_mwait+0xd()
cpu_idle_mwait+0xf1()
idle+0x114()
thread_start+8()
> $q

####################################################################
rcherven@Stubborn2:/var/crash/Stubborn2$ mdb 3
Loading modules: [ unix genunix specfs dtrace mac cpu.generic uppc pcplusmp rootnex scsi_vhci zfs sd sockfs ip hook neti sctp arp usba uhci fctl md lofs random crypto smbsrv nfs fcip fcp cpc logindmux ptm ufs nsmb sppp ipc ]
> $c
vdev_is_dead+0xc(0)
vdev_readable+0x16(0)
vdev_mirror_child_select+0x61(ffffff02e0bb1020)
vdev_mirror_io_start+0xda(ffffff02e0bb1020)
zio_vdev_io_start+0x1ba(ffffff02e0bb1020)
zio_execute+0xa0(ffffff02e0bb1020)
zio_nowait+0x42(ffffff02e0bb1020)
arc_read_nolock+0x82d(0, ffffff02d8e1f000, ffffff02e323a000, 0, 0, 6)
arc_read+0x75(0, ffffff02d8e1f000, ffffff02e323a000, ffffff02f4b9f168, 0, 0)
dbuf_prefetch+0x131(ffffff030357c800, 20)
dmu_zfetch_fetch+0x70(ffffff030357c800, 1f, a)
dmu_zfetch_dofetch+0xb8(ffffff030357ca60, ffffff030bb623b8)
dmu_zfetch_find+0x436(ffffff030357ca60, ffffff000fdbde00, 1)
dmu_zfetch+0xac(ffffff030357ca60, 3c7c000, 4000, 1)
dbuf_read+0x170(ffffff030865a450, 0, 2)
dnode_hold_impl+0xed(ffffff02e6b19000, 1e3e3, 1, fffffffff7977cf0, ffffff000fdbdfd8)
dnode_hold+0x2b(ffffff02e6b19000, 1e3e3, fffffffff7977cf0, ffffff000fdbdfd8)
dmu_bonus_hold+0x33(ffffff02e6b19000, 1e3e3, 0, ffffff000fdbe028)
zfs_zget+0x5a(ffffff02ff846900, 1e3e3, ffffff000fdbe190)
zfs_dirent_lock+0x3fc(ffffff000fdbe198, ffffff030f1b9960, ffffff02ef3ed780, ffffff000fdbe190, 26, ffffff000fdbe534)
zfs_dirlook+0x12d(ffffff030f1b9960, ffffff02ef3ed780, ffffff000fdbe538, 80000, ffffff000fdbe534, ffffff000fdbe470)
zfs_lookup+0x25f(ffffff03075d1180, ffffff02ef3ed780, ffffff000fdbe538, 0, 80000, 0)
fop_lookup+0xed(ffffff03075d1180, ffffff02ef3ed780, ffffff000fdbe538, 0, 80000, 0)
smb_vop_lookup+0x163(ffffff03075d1180, ffffff02ef3ed780, ffffff000fdbe538, ffffff02e5392040, 2, ffffff000fdbe534)
smb_fsop_lookup+0x12f(ffffff02e0c49660, ffffff02ded14858, 0, ffffff0304318048, ffffff02e43bd118, ffffff02ef3ed780)
smb_odir_wildcard_fileinfo+0xe4(ffffff02e0c49660, ffffff02f153a998, ffffff02ef3ed780, ffffff000fdbe7d0)
smb_odir_read_fileinfo+0xfd(ffffff02e0c49660, ffffff02f153a998, ffffff000fdbe7d0, ffffff000fdbe9b4)
smb_trans2_find_entries+0xa2(ffffff02e0c49660, ffffff02ff4a3008, ffffff02f153a998, ffffff000fdbe9c0, ffffff000fdbe9b4)
smb_com_trans2_find_first2+0x14a(ffffff02e0c49660, ffffff02ff4a3008)
smb_trans2_dispatch+0x195(ffffff02e0c49660, ffffff02ff4a3008)
smb_com_transaction2+0x17b(ffffff02e0c49660)
smb_dispatch_request+0x455(ffffff02e0c49660)
smb_session_worker+0x4a(ffffff02e0c49660)
taskq_d_thread+0xb1(ffffff02d86736f8)
thread_start+8()
> $q

Open in new window

0
Brian UtterbackPrinciple Software EngineerCommented:
What I would do is try breaking the mirror and then putting it back and re-silvering.
0
tongsleeAuthor Commented:
What mirror are you referring to? Below is the output from zpool status, that shows devices involved. I have 2 raidz2's, and the root pool is actually done through a hardware raid mirror, not zfs.


rcherven@Stubborn2:/var/crash/Stubborn2$ zpool status
  pool: Pool1
 state: ONLINE
 scrub: scrub completed after 2h20m with 0 errors on Fri Apr  9 03:20:12 2010
config:

        NAME         STATE     READ WRITE CKSUM
        Pool1        ONLINE       0     0     0
          raidz2-0   ONLINE       0     0     0
            c10t0d0  ONLINE       0     0     0
            c10t1d0  ONLINE       0     0     0
            c10t2d0  ONLINE       0     0     0
            c10t3d0  ONLINE       0     0     0
            c10t4d0  ONLINE       0     0     0
            c10t5d0  ONLINE       0     0     0
            c10t6d0  ONLINE       0     0     0
            c10t7d0  ONLINE       0     0     0
        spares
          c10t8d0    AVAIL   

errors: No known data errors

  pool: Pool2
 state: ONLINE
 scrub: scrub completed after 1h22m with 0 errors on Fri Apr  9 04:43:17 2010
config:

        NAME        STATE     READ WRITE CKSUM
        Pool2       ONLINE       0     0     0
          raidz2-0  ONLINE       0     0     0
            c0t0d0  ONLINE       0     0     0
            c0t1d0  ONLINE       0     0     0
            c0t2d0  ONLINE       0     0     0
            c0t3d0  ONLINE       0     0     0
            c0t4d0  ONLINE       0     0     0
            c0t5d0  ONLINE       0     0     0
            c0t6d0  ONLINE       0     0     0
            c0t7d0  ONLINE       0     0     0
        spares
          c0t8d0    AVAIL   

errors: No known data errors

  pool: rpool
 state: ONLINE
 scrub: scrub completed after 0h11m with 0 errors on Fri Apr  9 04:54:57 2010
config:

        NAME         STATE     READ WRITE CKSUM
        rpool        ONLINE       0     0     0
          c11t0d0s0  ONLINE       0     0     0

errors: No known data errors

Open in new window

0
Brian UtterbackPrinciple Software EngineerCommented:
Sorry, I thought you had mirrors, but you have raidz.

I know you tried export and re-import, but have you tried export-reboot-import?
0
tongsleeAuthor Commented:
It panicked Saturday, but has been stable since then. I will report back in a day with the status.

Thank you!
Rob
0
tongsleeAuthor Commented:
Well... still no luck. We have exported, rebooted and imported. The unit reboot twice yesterday, and has done so 3 times in the last two hours today. I haven't been able to link it to any specific activity, or been able to replicate it. It seems as if it's more likely to do it when it's being accessed, versus overnight when there is no activity. I've got a ticket bugtacker ticket submitted, but as of yet have had no feedback.

If you have ANY other thoughts, please let me know.

Thanks!
Rob
0
Brian UtterbackPrinciple Software EngineerCommented:
Do you have a bug number?
0
tongsleeAuthor Commented:
http://bugs.opensolaris.org/view_bug.do?bug_id=6942181

It was also submitted through here:
http://defect.opensolaris.org/bz/show_bug.cgi?id=15546

But the second one was closed and everything should now point to 6942181.

0
Brian UtterbackPrinciple Software EngineerCommented:
Looks like they have the crash dumps you submitted, so there really isn't much else to do until they get an analysis done. Good luck.
0

Experts Exchange Solution brought to you by

Your issues matter to us.

Facing a tech roadblock? Get the help and guidance you need from experienced professionals who care. Ask your question anytime, anywhere, with no hassle.

Start your 7-day free trial
tongsleeAuthor Commented:
Thanks for the help with this. I still haven't had any resolution thru Sun, but I think we've got it fixed. It seemed all of our issues stemmed back to accessing one of the zfs sets. It was a 300 GB set of photos. Anytime a small portion of them were browsed from a windows client, the machine would panic and reboot. I ended up creating a new zfs set, and restored from a 2 week old backup. It has now been stable for over 24 hours, even while copying snapshots, and users accessing it.

So, thanks for the help. I still don't know exactly what it was, but it seems stable for now  
0
It's more than this solution.Get answers and train to solve all your tech problems - anytime, anywhere.Try it for free Edge Out The Competitionfor your dream job with proven skills and certifications.Get started today Stand Outas the employee with proven skills.Start learning today for free Move Your Career Forwardwith certification training in the latest technologies.Start your trial today
Unix OS

From novice to tech pro — start learning today.