So, I'm doing some failure testing, this is the first time I thought of just pulling out a drive to see what happens with the HAST/ZFS configuration.
There's something quite unfortunate: when a drive goes offline (like by pulling it out) HAST doesn't do anything, the HAST device stays up, so the zpool doesn't know it's offline.
Anybody have experience with this? A solution even? I am script capable, but not sure what I should do to handle it.
And one more thing that I just found out the FreeBSD implementation of ZFS doesn't seem to include the agent responsible for checking failures. As a test, I changed one of the HAST devs to INIT, so that the zpool could not see it. Sure enough it listed it as OFFLINE, but the hot spare that I had configured, did not engage.
The whole idea of a hot spare is to automate the replacement of a failed drive. That isn't happening, and I waited a while too.
...btw: autoreplace=ON in my zpool settings, witch is a raidz2.
thanks!
There's something quite unfortunate: when a drive goes offline (like by pulling it out) HAST doesn't do anything, the HAST device stays up, so the zpool doesn't know it's offline.
Anybody have experience with this? A solution even? I am script capable, but not sure what I should do to handle it.
And one more thing that I just found out the FreeBSD implementation of ZFS doesn't seem to include the agent responsible for checking failures. As a test, I changed one of the HAST devs to INIT, so that the zpool could not see it. Sure enough it listed it as OFFLINE, but the hot spare that I had configured, did not engage.
The whole idea of a hot spare is to automate the replacement of a failed drive. That isn't happening, and I waited a while too.
...btw: autoreplace=ON in my zpool settings, witch is a raidz2.
thanks!