> Is someone working on a way to overcome the 2GB file limitation in Linux? I
> currently backup several servers using a dedicated hard drive for the
> backups. Recently I saw one backup die saying the the file size had been
> exceeded. I've never had good luck with tape backups, yes they backup, but
> whenever I really need a file, it can't be retrieved.
The problem has been fixed in the file system, the kernel and
the supporting libraries for ages. Applications may need to be updated
to take advantage of 64 bit file offsets and file operations, though.
You didn't say how you were backing up those other servers. Is it a
particular backup program? If so, that program may need to be rebuilt
or updated. Are you backing up over NFS or some other file transport
where the Linux box is just acting as a file server or is the backup
program running on that Linux box?
Oh... And what version of Linux? If it's old enough, I
suppose you might still have a problem there. Recent (RedHat 7.x
and contemporaries) distributions shouldn't be a problem.
> TIA,
> Bobby
Mike
-- Michael H. Warfield | (770) 985-6132 | mhw@WittsEnd.com (The Mad Wizard) | (678) 463-0932 | http://www.wittsend.com/mhw/ NIC whois: MHW9 | An optimist believes we live in the best of all PGP Key: 0xDF1DD471 | possible worlds. A pessimist is sure of it!- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/