|
Authored by: Anonymous on Sunday, June 23 2013 @ 10:25 PM EDT |
Something to add to that. I have been having a lot of power issues here with the
machine losing power suddenly, so the file system needs to be resistant to that.
No, a UPS won't help as it is the UPS doing a scram due to line issues.
Tufty
[ Reply to This | Parent | # ]
|
|
Authored by: artp on Sunday, June 23 2013 @ 11:25 PM EDT |
I've been sticking with ext3 for now.
As for copying, modern "tar"s do a decent job - preserves
timestamps and all. If you want more security, use scp or
rdist. Both make sure that the file was copied correctly.
I think I used to use a command something like:
tar cf . - | (cd /otherplace; tar xvf -)
Options to change cd to rsh or more secure remote login are
possible. scp lets you specify remote host and directory to
start from. So does rdist.
Have fun. And remember, professionals don't need to do
backups first, just amateurs like us! ;-)
---
Userfriendly on WGA server outage:
When you're chained to an oar you don't think you should go down when the galley
sinks ?[ Reply to This | Parent | # ]
|
|
Authored by: Wol on Monday, June 24 2013 @ 07:00 PM EDT |
I've had problems with my machine locking up and stuff (suspect I've got
powerlines too near sensitive components). I use ext4, and it's given me no
trouble.
I'm also planning to mirror my system with two drives. And I've just upgraded it
from a 500MB drive to two 1TB drives. I used "cp -a" to copy stuff. I
make extensive use of hard links to share/protect files (mostly photographs) so
if I did an ordinary copy that didn't preserve links, copying /home to a drive
twice the size would probably fall over with a "disk full"! Why bother
with tar if you're not copying over a network? Yes it's great if you are because
you can compress as you copy (and I've done it, it pays dividends!), but it's
unnecessary system-hammering for a local copy.
Cheers,
Wol[ Reply to This | Parent | # ]
|
|
Authored by: Anonymous on Wednesday, June 26 2013 @ 04:22 PM EDT |
I would take a look at ZFS for Linux. In my opinion ZFS is the most resilient
file system available at this time. It has support for software mirroring,
RAID5 and RAID6. You can create a disk mirror with a single command.
I've been running a NAS system using OpenSolaris and ZFS for over 3 years and
have survived both disk failures and a controller failure with no data loss.
I would recommend using cpio to copy data, it can deal with soft links and
special files.[ Reply to This | Parent | # ]
|
|
|
|
|