1) yeah, you can guess the kernel, apply the filter, reguess, reapply, reguess... that's called blind deconvolution and "should" converge at least in some cases.
2) the entropy is the same for blurred and reconstructed image. If I compare source with blurred image, the entropy decreases (naturally). For 16 bit image data it only decreases marginally, though, that's why you can reconstruct the image. I'll try and find numbers for it.
3) rapidly. If I change one (!) pixel in the most blurred image by, say, 10%, the reconstructed image has enormous ringing artifacts. That said, for small blur radii such as in the last image, it's not really noticeable.