ZFS bad session hang during "ls" of PATH/.zfs/snapshot. dtrace attached.

Description

SImple ordinary pool, a couple of levels deep, the sort of path I would "ls" without thinking twice on 11.3 to recover a lost file that's in a previous snapshot.

Except on 12-beta1 it hangs the session.

Having reported one bug ( https://ixsystems.atlassian.net/browse/NAS-107049#icft=NAS-107049) related to zfs that turned out to relate to some kind of bug asserting holds during a send/recv, I thought maybe traversing to a snapshot might also involve holds, so it could be the same or similar bug. So I ran the same dtrace on it in one window, and ran "ls (snapshot dir)" in another window.

RESULT:

The moment I hit return after "ls", the session with "ls" hung, and dtrace started dumping text to the screen.

The attached screenshots show the commands used and outputs in both sessions.

The dtrace log is a bit different from the output of https://ixsystems.atlassian.net/browse/NAS-107049#icft=NAS-107049 , so I don't know if it's the exact same bug. But if not, its a close cousin. The dtrace command used was from that bug report, and a sample from the start of the dtrace log is below:

dtrace -n 'zfs-dbgmsg { print(stringof(arg0)); } set-error { stack(); }' > ~/debug2.txt

7 74436 none:zfs-dbgmsg string [ 'z', 'i', 'o', '.', 'c', ':', '2', '0', '0', '0', ':', 'z', 'i', 'o', '', 'd', 'e', 'a', 'd', 'm', 'a', 'n', '', 'i', 'm', 'p', 'l', '(', ')', ':', ' ', 's', 'l', 'o', 'w', ' ', 'z', 'i', 'o', '[', '3', ']', ':', ' ', 'z', 'i', 'o', '=', '0', 'x', 'f', 'f', 'f', 'f', 'f', '8', '3', 'd', '7', '0', '7', '4', '1', '4', 'b', '8', 'x', ' ', 't', 'i', 'm', 'e', 's', 't', 'a', 'm', 'p', '=', '3', '4', '7', '6', '6', '0', '3', '0', '4', '7', '6', '8', '5', '2', ' ', 'd', 'e', 'l', 't', 'a', '=', '1', '3', '0', '7', '5', '0', '9', '3', '4', '8', '7', '2', '7', '1', ' ', 'q', 'u', 'e', 'u', 'e', 'd', '=', '0', ' ', 'i', 'o', '=', '3', '4', '7', '6', '6', '0', '3', '0', '4', '7', '8', '2', '9', '4', ' ', 'p', 'a', 't', 'h', '=', '/', 'd', 'e', 'v', '/', 'n', 'v', 'd', '1', 'p', '1', ' ', 'l', 'a', 's', 't', '=', '2', '4', '3', '6', '8', '1', '9', '4', '9', '3', '8', '3', '7', '9', ' ', 't', 'y', 'p', 'e', '=', '1', ' ', 'p', 'r', 'i', 'o', 'r', 'i', 't', 'y', '=', '0', ' ', 'f', 'l', 'a', 'g', 's', '=', '0', 'x', '1', '8', '0', '8', '8', '0', ' ', 's', 't', 'a', 'g', 'e', '=', '0', 'x', '1', '0', '0', '0', '0', '0', ' ', 'p', 'i', 'p', 'e', 'l', 'i', 'n', 'e', '=', '0', 'x', '1', 'f', '0', '0', '0', '0', '0', ' ', 'p', 'i', 'p', 'e', 'l', 'i', 'n', 'e', '-', 't', 'r' ]

7 74436 none:zfs-dbgmsg string "zio.c:2035:zio_deadman(): zio_wait waiting for hung I/O to pool 'Main_pool'"

2 74396 none:set-error
openzfs.ko`zfsdev_ioctl_common+0x470
openzfs.ko`zfsdev_ioctl+0x29e
kernel`devfs_ioctl+0xb0
kernel`VOP_IOCTL_APV+0x7b
kernel`vn_ioctl+0x16a
kernel`devfs_ioctl_f+0x1e
kernel`kern_ioctl+0x2b7
kernel`sys_ioctl+0xfa
kernel`amd64_syscall+0x387
kernel`0xffffffff80fa50e0

2 74396 none:set-error
openzfs.ko`vdev_config_generate+0x3ab
openzfs.ko`spa_config_generate+0x2a5
openzfs.ko`spa_open_common+0x214
openzfs.ko`spa_get_stats+0x4e
openzfs.ko`zfs_ioc_pool_stats+0x22
openzfs.ko`zfsdev_ioctl_common+0x470
openzfs.ko`zfsdev_ioctl+0x29e
kernel`devfs_ioctl+0xb0
kernel`VOP_IOCTL_APV+0x7b
kernel`vn_ioctl+0x16a
kernel`devfs_ioctl_f+0x1e
kernel`kern_ioctl+0x2b7
kernel`sys_ioctl+0xfa
kernel`amd64_syscall+0x387
kernel`0xffffffff80fa50e0

2 74396 none:set-error
openzfs.ko`vdev_config_generate+0x3d6
openzfs.ko`spa_config_generate+0x2a5
openzfs.ko`spa_open_common+0x214
openzfs.ko`spa_get_stats+0x4e
openzfs.ko`zfs_ioc_pool_stats+0x22
openzfs.ko`zfsdev_ioctl_common+0x470
openzfs.ko`zfsdev_ioctl+0x29e
kernel`devfs_ioctl+0xb0
kernel`VOP_IOCTL_APV+0x7b
kernel`vn_ioctl+0x16a
kernel`devfs_ioctl_f+0x1e
kernel`kern_ioctl+0x2b7
kernel`sys_ioctl+0xfa
kernel`amd64_syscall+0x387
kernel`0xffffffff80fa50e0

2 74396 none:set-error
openzfs.ko`zfs_ioc_pool_stats+0x36
openzfs.ko`zfsdev_ioctl_common+0x470
openzfs.ko`zfsdev_ioctl+0x29e
kernel`devfs_ioctl+0xb0
kernel`VOP_IOCTL_APV+0x7b
kernel`vn_ioctl+0x16a
kernel`devfs_ioctl_f+0x1e
kernel`kern_ioctl+0x2b7
kernel`sys_ioctl+0xfa
kernel`amd64_syscall+0x387
kernel`0xffffffff80fa50e0

2 74396 none:set-error
openzfs.ko`vdev_config_generate+0x3ab
openzfs.ko`spa_config_generate+0x2a5
openzfs.ko`spa_open_common+0x214
openzfs.ko`spa_get_stats+0x4e
openzfs.ko`zfs_ioc_pool_stats+0x22
openzfs.ko`zfsdev_ioctl_common+0x470
openzfs.ko`zfsdev_ioctl+0x29e
kernel`devfs_ioctl+0xb0
kernel`VOP_IOCTL_APV+0x7b
kernel`vn_ioctl+0x16a
kernel`devfs_ioctl_f+0x1e
kernel`kern_ioctl+0x2b7
kernel`sys_ioctl+0xfa
kernel`amd64_syscall+0x387
kernel`0xffffffff80fa50e0

2 74396 none:set-error
openzfs.ko`vdev_config_generate+0x3d6
openzfs.ko`spa_config_generate+0x2a5
openzfs.ko`spa_open_common+0x214
openzfs.ko`spa_get_stats+0x4e
openzfs.ko`zfs_ioc_pool_stats+0x22
openzfs.ko`zfsdev_ioctl_common+0x470
openzfs.ko`zfsdev_ioctl+0x29e
kernel`devfs_ioctl+0xb0
kernel`VOP_IOCTL_APV+0x7b
kernel`vn_ioctl+0x16a
kernel`devfs_ioctl_f+0x1e
kernel`kern_ioctl+0x2b7
kernel`sys_ioctl+0xfa
kernel`amd64_syscall+0x387
kernel`0xffffffff80fa50e0

Problem/Justification

None

Impact

None

SmartDraw Connector

Katalon Manual Tests (BETA)

Activity

Show:

Ryan Moeller September 3, 2020 at 7:09 PM

Let's have just the one ticket for this issue.

Stilez August 26, 2020 at 10:28 AM

Pathological scrub behaviour now collated and issue created:  https://ixsystems.atlassian.net/browse/NAS-107364#icft=NAS-107364

Stilez August 26, 2020 at 1:10 AM
Edited

To answer the last questions, 20k snaps now reduced to about 6k. On 11.3 I had around 80k and no issues at all (15 minute snaps, never got round to a cleanup for ages). The system just shouldn't be that busy, and almost certainly wasn't. Even busy, it should work well. For example local send/recv on 12-beta2 hammers the pools, but doesn't show any such emergent pathological behaviours.

---------------------------

Sorry for not getting back on this. I promise I will.
But I have some fresh insight on it.

---------------------------

When I reported this, I was fresh off the heels of NAS-107049, an issue where \zfs misbehaved around snapshots because (we found) the number of holds for operations such as replication, easily overran the inadequate size of nvlist that was defined. So when this issue happened, I saw the common theme of "issues round ZFS misbehaving when snapshots were in the picture" and drew a direct line connection with that matter.

Since then I've seen NAS-107036.  It's a closed issue, but worth reading, because a significant issue came to light thats unresolved on it, see the last couple of comments in that issue. In brief, it appears that in 12-beta2, just running scrub can trigger extreme CPU starvation even on a powerfully specified and otherwise idle server. When this happened in NAS-107036, the symptoms included extreme slowdown, and simple console prompts and commands taking 10 - 20 seconds to display.

My 1st comment on this issue remarks that a scrub was indeed in progress at the time of this issue.  The symptoms of hanging in the OP on the prompt and on simple commands like "ls" exactly match the scrub-induced CPU starvation symptoms I saw with NAS-107036.

So my guess is that I mistakenly linked the OP issue in this item to https://ixsystems.atlassian.net/browse/NAS-107049#icft=NAS-107049 due to a commonality around traversing snapshots, whereas the true commonality is more likely to be with NAS-107036, due to scrub causing extreme slowdown, via (presumed in this case, verified in the other) CPU starvation.

 

Please change title to reflect "during scrub or snapshot traversal" thanks?

Alexander Motin August 10, 2020 at 5:51 PM

I am curios, how many snapshots do you have there?  Can it be that it is just reasonably busy?  Have you looked what CPU and disks are doing during that?  I see the dtrace script logs many spa_deadman() timeouts, that may mean pool I/O being blocked by something or being extremely slow.

Stilez August 10, 2020 at 11:14 AM
Edited

I've also had it hang in a separate session, simply by typing "cd .zfs/snapshot/PRESS TAB KEY", as it traverses to find matches for "tab".

For info, a scrub was in progress on that pool at the time, which may have used some holds as well? If that's any kind of relevance to the picture. Tried to get a debug file as well, its not producing one. Possibly 12-beta1 client-server middleware issue fixed in beta2, but doesant help us right now.

Noting implication for samba, file history, or time machine, or anything else that auto-accesses and traverses snap contents.

Duplicate

Details

Assignee

Reporter

Labels

Fix versions

Affects versions

Priority

More fields

Katalon Platform

Created August 10, 2020 at 11:10 AM
Updated July 1, 2022 at 4:56 PM
Resolved September 3, 2020 at 7:09 PM

Flag notifications