No, because it gives it as arguments to "rm" (which only deletes
files by defintion) and not to "rm -rf" (which also deletes
directories).
Not excluding directories from the arguments to "rm" is the BUG.
All my scripts always do that, just the kernel Makefile doesn't.
And it might even be faster, because we don't have to do all the
other tests, if it's not an directory ;-)
> N.B.: X/KDE will not run on a ramfs, because it reports as a size-zero
> filesystem in "df". Switching to tmpfs solved that for me.
This is partially a ramfs BUG, because filesystems are supposed
to do accounting ;-)
OTOH, querying the available space before transfers says NOTHING
about the sucess of an operation. kfm is not the only application
running and the fs might do delayed allocation.
Applications should not check for free space, they should try to
allocate it (and maybe touch it, if we are not IO bound) and
watch for ENOSPC instead.
Then we truncate all written stuff to the extend, where it made
sense, delete the whole file if it has become size 0 and emit an
error telling the user, that we had not enough disk space at this
time to succeed.
We don't check for available memory either, so why do it on disk?
It made sense in DOS times, but doesn't do anymore. It's only
relevant to the administrator to tell him, that we need a bigger
disk ;-)
Regards
Ingo Oeser
-- 10.+11.03.2001 - 3. Chemnitzer LinuxTag <http://www.tu-chemnitz.de/linux/tag> <<<<<<<<<<<< been there and had much fun >>>>>>>>>>>> - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/