Goals:
* Track each node's `STONITH_WATCHDOG_TIMEOUT`
** Most likely in the controller
** A `node_state` attribute would be good, so it persists across reboots
** Cluster nodes could advertise it as part of the join process
** Pacemaker Remote nodes already have a timeout verification when it could be done
** The local `STONITH_WATCHDOG_TIMEOUT` would be the default before a target node's value is known (watchdog fencing is unreliable before a target is first seen anyway, and a never-seen node shouldn't be running resources)
* The `stonith-watchdog-timeout` cluster option should be deprecated and replaced with one or more new options
** The new options should use standard types/validators hen practical
** The new option names should use some variant of "fencing" rather than "stonith"
** Something like `watchdog-fencing-duration` as a duration (nonnegative interval spec) to set a specific wait time
** Something like `watchdog-fencing-auto=true/false/increase/decrease` where `true` = use twice the target-specific value, `false` = use specified duration exactly, `increase` = use the higher of the specified duration or twice the target-specific value, `decrease` = use the lower of the specified duration or twice the target-specific value
** The new syntax should be designed so it is straightforward to XSL-transform the old syntax to it when we eventually drop it