Faster rsync of huge directory which was not changedFilesystem large number of files in a single directoryRsync daemon: is it really useful?Do large folder sizes slow down IO performance?rsync two large storage serversRsync takes 8+ hours to backup 15M+ filesCopying a large directory tree locally? cp or rsync?Using rsync to quickly upload a file that is similar to another fileReliable file copy (move) process - mostly Unix/LinuxPerformance problem with QNAP TS-410 NASCygwin's RSYNC for large data transferrsync hanging on local copyrsync many small files with long filenames takes a lot of bandwidthBackup with local manifestMultiple servers Rsync into one NAS simultaneausLive file syncronization across multiple Linux servers with millions of files and directories
Who were the members of the jury in the Game of Thrones finale?
Why is unzipped directory exactly 4.0K (much smaller than zipped file)?
Why was this character made Grand Maester?
Status of proof by contradiction and excluded middle throughout the history of mathematics?
Are PMR446 walkie-talkies legal in Switzerland?
Fill area of x^2+y^2>1 and x^2+y^2>4 using patterns and tikzpicture
What is Orcus doing with Mind Flayers in the art on the last page of Volo's Guide to Monsters?
The disk image is 497GB smaller than the target device
Python script to extract text from PDF with images
Is it safe to redirect stdout and stderr to the same file without file descriptor copies?
Determine direction of mass transfer
What is the limit to a Glyph of Warding's trigger?
How does the Earth's center produce heat?
Why is this integration method not valid?
ifconfig shows UP while ip link shows DOWN
Why Emacs (dired+) asks me twice to delete file?
Why does Bran want to find Drogon?
How to escape dependency hell?
Merge pdfs sequentially
Flatten not working
Alexandrov's generalization of Cauchy's rigidity theorem
What is the use case for non-breathable waterproof pants?
Why is std::ssize() introduced in C++20
Is there an idiom that means that you are in a very strong negotiation position in a negotiation?
Faster rsync of huge directory which was not changed
Filesystem large number of files in a single directoryRsync daemon: is it really useful?Do large folder sizes slow down IO performance?rsync two large storage serversRsync takes 8+ hours to backup 15M+ filesCopying a large directory tree locally? cp or rsync?Using rsync to quickly upload a file that is similar to another fileReliable file copy (move) process - mostly Unix/LinuxPerformance problem with QNAP TS-410 NASCygwin's RSYNC for large data transferrsync hanging on local copyrsync many small files with long filenames takes a lot of bandwidthBackup with local manifestMultiple servers Rsync into one NAS simultaneausLive file syncronization across multiple Linux servers with millions of files and directories
.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty height:90px;width:728px;box-sizing:border-box;
We use rsync to backup servers.
Unfortunately the network to some servers is slow.
It takes up to five minutes for rsync to detect, that nothing has changed in huge directories. These huge directory trees contain a lot of small files (about 80k files).
I guess that the rsync clients sends data for each of the 80k files.
Since the network is slow I would like to avoid to send 80k times information about each file.
Is there a way to tell rsync to make a hash-sum of a sub directory tree?
This way the rsync client would send only a few bytes for a huge directory tree.
Update
Up to now my strategy is to use rsync
. But if a different tools fits better here, I am able to switch. Both (server and client) are under my control.
Update2
There are 80k files in one directory tree. Each single directory does not have more than 2k files or sub-directories
Update3
Details on the slowness of the network:
time ssh einswp 'cd attachments/200 && ls -lLR' >/tmp/list
real 0m2.645s
Size of tmp/list file: 2MByte
time scp einswp:/tmp/list tmp/
real 0m2.821s
Conclusion: scp has the same speed (no surprise)
time scp einswp:tmp/100MB tmp/
real 1m24.049s
Speed: 1.2MB/s
rsync synchronization
add a comment |
We use rsync to backup servers.
Unfortunately the network to some servers is slow.
It takes up to five minutes for rsync to detect, that nothing has changed in huge directories. These huge directory trees contain a lot of small files (about 80k files).
I guess that the rsync clients sends data for each of the 80k files.
Since the network is slow I would like to avoid to send 80k times information about each file.
Is there a way to tell rsync to make a hash-sum of a sub directory tree?
This way the rsync client would send only a few bytes for a huge directory tree.
Update
Up to now my strategy is to use rsync
. But if a different tools fits better here, I am able to switch. Both (server and client) are under my control.
Update2
There are 80k files in one directory tree. Each single directory does not have more than 2k files or sub-directories
Update3
Details on the slowness of the network:
time ssh einswp 'cd attachments/200 && ls -lLR' >/tmp/list
real 0m2.645s
Size of tmp/list file: 2MByte
time scp einswp:/tmp/list tmp/
real 0m2.821s
Conclusion: scp has the same speed (no surprise)
time scp einswp:tmp/100MB tmp/
real 1m24.049s
Speed: 1.2MB/s
rsync synchronization
1
You might read up on zsync. I have not used it myself, but from what I read, it pre-renders the metadata on the server side and might just speed up transfers in your case. It might be worth testing anyway. Beyond that, the only other solution I am aware of is real time block level syncronization that comes with some san/nas solutions.
– Aaron
Jan 5 '16 at 4:47
add a comment |
We use rsync to backup servers.
Unfortunately the network to some servers is slow.
It takes up to five minutes for rsync to detect, that nothing has changed in huge directories. These huge directory trees contain a lot of small files (about 80k files).
I guess that the rsync clients sends data for each of the 80k files.
Since the network is slow I would like to avoid to send 80k times information about each file.
Is there a way to tell rsync to make a hash-sum of a sub directory tree?
This way the rsync client would send only a few bytes for a huge directory tree.
Update
Up to now my strategy is to use rsync
. But if a different tools fits better here, I am able to switch. Both (server and client) are under my control.
Update2
There are 80k files in one directory tree. Each single directory does not have more than 2k files or sub-directories
Update3
Details on the slowness of the network:
time ssh einswp 'cd attachments/200 && ls -lLR' >/tmp/list
real 0m2.645s
Size of tmp/list file: 2MByte
time scp einswp:/tmp/list tmp/
real 0m2.821s
Conclusion: scp has the same speed (no surprise)
time scp einswp:tmp/100MB tmp/
real 1m24.049s
Speed: 1.2MB/s
rsync synchronization
We use rsync to backup servers.
Unfortunately the network to some servers is slow.
It takes up to five minutes for rsync to detect, that nothing has changed in huge directories. These huge directory trees contain a lot of small files (about 80k files).
I guess that the rsync clients sends data for each of the 80k files.
Since the network is slow I would like to avoid to send 80k times information about each file.
Is there a way to tell rsync to make a hash-sum of a sub directory tree?
This way the rsync client would send only a few bytes for a huge directory tree.
Update
Up to now my strategy is to use rsync
. But if a different tools fits better here, I am able to switch. Both (server and client) are under my control.
Update2
There are 80k files in one directory tree. Each single directory does not have more than 2k files or sub-directories
Update3
Details on the slowness of the network:
time ssh einswp 'cd attachments/200 && ls -lLR' >/tmp/list
real 0m2.645s
Size of tmp/list file: 2MByte
time scp einswp:/tmp/list tmp/
real 0m2.821s
Conclusion: scp has the same speed (no surprise)
time scp einswp:tmp/100MB tmp/
real 1m24.049s
Speed: 1.2MB/s
rsync synchronization
rsync synchronization
edited Jan 7 '16 at 15:53
guettli
asked Jan 4 '16 at 8:53
guettliguettli
44742961
44742961
1
You might read up on zsync. I have not used it myself, but from what I read, it pre-renders the metadata on the server side and might just speed up transfers in your case. It might be worth testing anyway. Beyond that, the only other solution I am aware of is real time block level syncronization that comes with some san/nas solutions.
– Aaron
Jan 5 '16 at 4:47
add a comment |
1
You might read up on zsync. I have not used it myself, but from what I read, it pre-renders the metadata on the server side and might just speed up transfers in your case. It might be worth testing anyway. Beyond that, the only other solution I am aware of is real time block level syncronization that comes with some san/nas solutions.
– Aaron
Jan 5 '16 at 4:47
1
1
You might read up on zsync. I have not used it myself, but from what I read, it pre-renders the metadata on the server side and might just speed up transfers in your case. It might be worth testing anyway. Beyond that, the only other solution I am aware of is real time block level syncronization that comes with some san/nas solutions.
– Aaron
Jan 5 '16 at 4:47
You might read up on zsync. I have not used it myself, but from what I read, it pre-renders the metadata on the server side and might just speed up transfers in your case. It might be worth testing anyway. Beyond that, the only other solution I am aware of is real time block level syncronization that comes with some san/nas solutions.
– Aaron
Jan 5 '16 at 4:47
add a comment |
4 Answers
4
active
oldest
votes
Some unrelated points:
80K is a lot of files.
80,000 files in one directory? No operating system or app handles that situation very well by default. You just happen to notice this problem with rsync.
Check your rsync version
Modern rsync handles large directories a lot better than in the past. Be sure you are using the latest version.
Even old rsync handles large directories fairly well over high latency links... but 80k files isn't large...it is huge!
That said, rsync's memory usage is directly proportional to the number of files in a tree. Large directories take a large amount of RAM. The slowness may be due to a lack of RAM on either side. Do a test run while watching memory usage. Linux uses any left-over RAM as a disk cache, so if you are running low on RAM, there is less disk caching. If you run out of RAM and the system starts using swap, performance will be really bad.
Make sure --checksum is not being used
--checksum
(or -c
) requires reading each and every block of every file. You probably can get by with the default behavior of just reading the modification times (stored in the inode).
Split the job into small batches.
There are some projects like Gigasync which will "Chop up the workload by using perl to recurse the directory tree, building smallish lists of files to transfer with rsync."
The extra directory scan is going to be a large amount of overhead, but maybe it will be a net win.
OS defaults aren't made for this situation.
If you are using Linux/FreeBSD/etc with all the defaults, performance will be terrible for all your applications. The defaults assume smaller directories so-as not to waste RAM on oversized caches.
Tune your filesystem to better handle large directories: Do large folder sizes slow down IO performance?
Look at the "namei cache"
BSD-like operating systems have a cache that accelerates looking up a name to the inode (the "namei" cache"). There is a namei cache for each directory. If it is too small, it is a hindrance more than an optimization. Since rsync is doing a lstat() on each file, the inode is being accessed for every one of the 80k files. That might be blowing your cache. Research how to tune file directory performance on your system.
Consider a different file system
XFS was designed to handle larger directories. See Filesystem large number of files in a single directory
Maybe 5 minutes is the best you can do.
Consider calculating how many disk blocks are being read, and calculate how fast you should expect the hardware to be able to read that many blocks.
Maybe your expectations are too high. Consider how many disk blocks must be read to do an rsync with no changed files: each server will need to read the directory and read one inode per file. Let's assume nothing is cached because, well, 80k files has probably blown your cache. Let's say that it is 80k blocks to keep the math simple. That's about 40M of data, which should be readable in a few seconds. However if there needs to be a disk seek between each block, that could take much longer.
So you are going to need to read about 80,000 disk blocks. How fast can your hard drive do that? Considering that this is random I/O, not a long linear read, 5 minutes might be pretty excellent. That's 1 / (80000 / 600), or a disk read every 7.5ms. Is that fast or slow for your hard drive? It depends on the model.
Benchmark against something similar
Another way to think about it is this. If no files have changed, ls -Llr
does the same amount of disk activity but never reads any file data (just metadata). The time ls -Llr
takes to run is your upper bound.
Is rsync (with no files changed) significantly slower than
ls -Llr
? Then the options you are using for rsync can be improved. Maybe-c
is enabled or some other flag that reads more than just directories and metadata (inode data).Is rsync (with no files changed) nearly as fast as
ls -Llr
? Then you've tuned rsync as best as you can. You have to tune the OS, add RAM, get faster drives, change filesystems, etc.
Talk to your devs
80k files is just bad design. Very few file systems and system tools handle such large directories very well. If the filenames are abcdefg.txt, consider storing them in abdc/abcdefg.txt (note the repetition). This breaks the directories up into smaller ones, but doesn't require a huge change to the code.
Also.... consider using a database. If you have 80k files in a directory, maybe your developers are working around the fact that what they really want is a database. MariaDB or MySQL or PostgreSQL would be a much better option for storing large amounts of data.
Hey, what's wrong with 5 minutes?
Lastly, is 5 minutes really so bad? If you run this backup once a day, 5 minutes is not a lot of time. Yes, I love speed. However if 5 minutes is "good enough" for your customers, then it is good enough for you. If you don't have a written SLA, how about an informal discussion with your users to find out how fast they expect the backups to take.
I assume you didn't ask this question if there wasn't a need to improve the performance. However, if your customers are happy with 5 minutes, declare victory and move on to other projects that need your efforts.
Update: After some discussion we determined that the bottleneck is the network. I'm going to recommend 2 things before I give up :-).
- Try to squeeze more bandwidth out of the pipe with compression. However compression requires more CPU, so if your CPU is overloaded, it might make performance worse. Try rsync with and without
-z
, and configure your ssh with and without compression. Time all 4 combinations to see if any of them perform significantly better than others. - Watch network traffic to see if there are any pauses. If there are pauses, you can find what is causing them and optimize there. If rsync is always sending, then you really are at your limit. Your choices are:
- a faster network
- something other than rsync
- move the source and destination closer together. If you can't do that, can you rsync to a local machine then rsync to the real destination? There may be benefits to doing this if the system has to be down during the initial rsync.
80K is a lot of files.: There are 80k files in one directory tree. Each single directory does not have more than 2k files/subdirectories.
– guettli
Jan 7 '16 at 8:50
Check your rsync version: done, Make sure --checksum is not being used: done. Split the job into small batches: Thank you I will have a look at gigasync. OS defaults aren't made for this situation: done (the bottleneck is network not OS). Look at the "namei cache": done (it is net, not OS). Consider a different file system: again net, not OS. Maybe 5 minutes is the best you can do.: I think it could be much faster. Talk to your devs (use DB): This would be a giant change. Maybe an filesystem with better backup support would solve it.
– guettli
Jan 7 '16 at 8:58
2k files per directory is a lot better. thank you for the update. You hadn't mentioned that the network was slow. Is it low bandwidth, high latency, or both? rsync usually performs well on high latency links (it was developed by someone working on his PhD from Australia while dealing with computers in the U.S.). Try doing that "ls -lLR" over ssh and time how long it takes to transmit the result. "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list". Make sure the /tmp/list gets created on the local host.
– TomOnTime
Jan 7 '16 at 13:40
yes the network is slow. It is a pitty.
– guettli
Jan 7 '16 at 15:40
How slow? If you use "scp" to copy a 100M file, how long does it take? Also, what is the output of "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list"?
– TomOnTime
Jan 7 '16 at 15:41
|
show 8 more comments
No, that's not possible with rsync and it would be quite inefficient in another regard:
Normally, rsync
only compares file modification dates and file sizes. Your approach would force it to read and checksum the content of all files twice (on the local and remote system) to find changed directories.
1
AFAIK rsync checks mtime and size. If both matches, the file is not transferred again (at least in the default settings). It would be enough to send the hash of the tuples (filename, size, mtime). There is no need to checksum the content.
– guettli
Jan 4 '16 at 10:59
Yes, you are correct, but anyway,rsync
doesn't do this.
– Sven♦
Jan 4 '16 at 10:59
add a comment |
For synchronisation of large numbers of files (where little has changed), it is also worth setting noatime
on the source and destination partitions. This saves writing access times to the disk for each unchanged file.
Yes, the noatime option makes sense. We use it since several years. I guess an alternative to rsync is needed.
– guettli
Aug 28 '17 at 7:07
add a comment |
Use rsync in daemon mode at the server end to speed up the listing/checksum process:
- Rsync daemon: is it really useful?
- http://giantdorks.org/alain/achieve-faster-file-transfer-times-by-running-rsync-as-a-daemon/
Note it isn't encrypted, but may be able to be tunneled without losing the listing performance improvement.
Also having rsync do compression rather than ssh should improve performance.
add a comment |
Your Answer
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "2"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fserverfault.com%2fquestions%2f746551%2ffaster-rsync-of-huge-directory-which-was-not-changed%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
4 Answers
4
active
oldest
votes
4 Answers
4
active
oldest
votes
active
oldest
votes
active
oldest
votes
Some unrelated points:
80K is a lot of files.
80,000 files in one directory? No operating system or app handles that situation very well by default. You just happen to notice this problem with rsync.
Check your rsync version
Modern rsync handles large directories a lot better than in the past. Be sure you are using the latest version.
Even old rsync handles large directories fairly well over high latency links... but 80k files isn't large...it is huge!
That said, rsync's memory usage is directly proportional to the number of files in a tree. Large directories take a large amount of RAM. The slowness may be due to a lack of RAM on either side. Do a test run while watching memory usage. Linux uses any left-over RAM as a disk cache, so if you are running low on RAM, there is less disk caching. If you run out of RAM and the system starts using swap, performance will be really bad.
Make sure --checksum is not being used
--checksum
(or -c
) requires reading each and every block of every file. You probably can get by with the default behavior of just reading the modification times (stored in the inode).
Split the job into small batches.
There are some projects like Gigasync which will "Chop up the workload by using perl to recurse the directory tree, building smallish lists of files to transfer with rsync."
The extra directory scan is going to be a large amount of overhead, but maybe it will be a net win.
OS defaults aren't made for this situation.
If you are using Linux/FreeBSD/etc with all the defaults, performance will be terrible for all your applications. The defaults assume smaller directories so-as not to waste RAM on oversized caches.
Tune your filesystem to better handle large directories: Do large folder sizes slow down IO performance?
Look at the "namei cache"
BSD-like operating systems have a cache that accelerates looking up a name to the inode (the "namei" cache"). There is a namei cache for each directory. If it is too small, it is a hindrance more than an optimization. Since rsync is doing a lstat() on each file, the inode is being accessed for every one of the 80k files. That might be blowing your cache. Research how to tune file directory performance on your system.
Consider a different file system
XFS was designed to handle larger directories. See Filesystem large number of files in a single directory
Maybe 5 minutes is the best you can do.
Consider calculating how many disk blocks are being read, and calculate how fast you should expect the hardware to be able to read that many blocks.
Maybe your expectations are too high. Consider how many disk blocks must be read to do an rsync with no changed files: each server will need to read the directory and read one inode per file. Let's assume nothing is cached because, well, 80k files has probably blown your cache. Let's say that it is 80k blocks to keep the math simple. That's about 40M of data, which should be readable in a few seconds. However if there needs to be a disk seek between each block, that could take much longer.
So you are going to need to read about 80,000 disk blocks. How fast can your hard drive do that? Considering that this is random I/O, not a long linear read, 5 minutes might be pretty excellent. That's 1 / (80000 / 600), or a disk read every 7.5ms. Is that fast or slow for your hard drive? It depends on the model.
Benchmark against something similar
Another way to think about it is this. If no files have changed, ls -Llr
does the same amount of disk activity but never reads any file data (just metadata). The time ls -Llr
takes to run is your upper bound.
Is rsync (with no files changed) significantly slower than
ls -Llr
? Then the options you are using for rsync can be improved. Maybe-c
is enabled or some other flag that reads more than just directories and metadata (inode data).Is rsync (with no files changed) nearly as fast as
ls -Llr
? Then you've tuned rsync as best as you can. You have to tune the OS, add RAM, get faster drives, change filesystems, etc.
Talk to your devs
80k files is just bad design. Very few file systems and system tools handle such large directories very well. If the filenames are abcdefg.txt, consider storing them in abdc/abcdefg.txt (note the repetition). This breaks the directories up into smaller ones, but doesn't require a huge change to the code.
Also.... consider using a database. If you have 80k files in a directory, maybe your developers are working around the fact that what they really want is a database. MariaDB or MySQL or PostgreSQL would be a much better option for storing large amounts of data.
Hey, what's wrong with 5 minutes?
Lastly, is 5 minutes really so bad? If you run this backup once a day, 5 minutes is not a lot of time. Yes, I love speed. However if 5 minutes is "good enough" for your customers, then it is good enough for you. If you don't have a written SLA, how about an informal discussion with your users to find out how fast they expect the backups to take.
I assume you didn't ask this question if there wasn't a need to improve the performance. However, if your customers are happy with 5 minutes, declare victory and move on to other projects that need your efforts.
Update: After some discussion we determined that the bottleneck is the network. I'm going to recommend 2 things before I give up :-).
- Try to squeeze more bandwidth out of the pipe with compression. However compression requires more CPU, so if your CPU is overloaded, it might make performance worse. Try rsync with and without
-z
, and configure your ssh with and without compression. Time all 4 combinations to see if any of them perform significantly better than others. - Watch network traffic to see if there are any pauses. If there are pauses, you can find what is causing them and optimize there. If rsync is always sending, then you really are at your limit. Your choices are:
- a faster network
- something other than rsync
- move the source and destination closer together. If you can't do that, can you rsync to a local machine then rsync to the real destination? There may be benefits to doing this if the system has to be down during the initial rsync.
80K is a lot of files.: There are 80k files in one directory tree. Each single directory does not have more than 2k files/subdirectories.
– guettli
Jan 7 '16 at 8:50
Check your rsync version: done, Make sure --checksum is not being used: done. Split the job into small batches: Thank you I will have a look at gigasync. OS defaults aren't made for this situation: done (the bottleneck is network not OS). Look at the "namei cache": done (it is net, not OS). Consider a different file system: again net, not OS. Maybe 5 minutes is the best you can do.: I think it could be much faster. Talk to your devs (use DB): This would be a giant change. Maybe an filesystem with better backup support would solve it.
– guettli
Jan 7 '16 at 8:58
2k files per directory is a lot better. thank you for the update. You hadn't mentioned that the network was slow. Is it low bandwidth, high latency, or both? rsync usually performs well on high latency links (it was developed by someone working on his PhD from Australia while dealing with computers in the U.S.). Try doing that "ls -lLR" over ssh and time how long it takes to transmit the result. "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list". Make sure the /tmp/list gets created on the local host.
– TomOnTime
Jan 7 '16 at 13:40
yes the network is slow. It is a pitty.
– guettli
Jan 7 '16 at 15:40
How slow? If you use "scp" to copy a 100M file, how long does it take? Also, what is the output of "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list"?
– TomOnTime
Jan 7 '16 at 15:41
|
show 8 more comments
Some unrelated points:
80K is a lot of files.
80,000 files in one directory? No operating system or app handles that situation very well by default. You just happen to notice this problem with rsync.
Check your rsync version
Modern rsync handles large directories a lot better than in the past. Be sure you are using the latest version.
Even old rsync handles large directories fairly well over high latency links... but 80k files isn't large...it is huge!
That said, rsync's memory usage is directly proportional to the number of files in a tree. Large directories take a large amount of RAM. The slowness may be due to a lack of RAM on either side. Do a test run while watching memory usage. Linux uses any left-over RAM as a disk cache, so if you are running low on RAM, there is less disk caching. If you run out of RAM and the system starts using swap, performance will be really bad.
Make sure --checksum is not being used
--checksum
(or -c
) requires reading each and every block of every file. You probably can get by with the default behavior of just reading the modification times (stored in the inode).
Split the job into small batches.
There are some projects like Gigasync which will "Chop up the workload by using perl to recurse the directory tree, building smallish lists of files to transfer with rsync."
The extra directory scan is going to be a large amount of overhead, but maybe it will be a net win.
OS defaults aren't made for this situation.
If you are using Linux/FreeBSD/etc with all the defaults, performance will be terrible for all your applications. The defaults assume smaller directories so-as not to waste RAM on oversized caches.
Tune your filesystem to better handle large directories: Do large folder sizes slow down IO performance?
Look at the "namei cache"
BSD-like operating systems have a cache that accelerates looking up a name to the inode (the "namei" cache"). There is a namei cache for each directory. If it is too small, it is a hindrance more than an optimization. Since rsync is doing a lstat() on each file, the inode is being accessed for every one of the 80k files. That might be blowing your cache. Research how to tune file directory performance on your system.
Consider a different file system
XFS was designed to handle larger directories. See Filesystem large number of files in a single directory
Maybe 5 minutes is the best you can do.
Consider calculating how many disk blocks are being read, and calculate how fast you should expect the hardware to be able to read that many blocks.
Maybe your expectations are too high. Consider how many disk blocks must be read to do an rsync with no changed files: each server will need to read the directory and read one inode per file. Let's assume nothing is cached because, well, 80k files has probably blown your cache. Let's say that it is 80k blocks to keep the math simple. That's about 40M of data, which should be readable in a few seconds. However if there needs to be a disk seek between each block, that could take much longer.
So you are going to need to read about 80,000 disk blocks. How fast can your hard drive do that? Considering that this is random I/O, not a long linear read, 5 minutes might be pretty excellent. That's 1 / (80000 / 600), or a disk read every 7.5ms. Is that fast or slow for your hard drive? It depends on the model.
Benchmark against something similar
Another way to think about it is this. If no files have changed, ls -Llr
does the same amount of disk activity but never reads any file data (just metadata). The time ls -Llr
takes to run is your upper bound.
Is rsync (with no files changed) significantly slower than
ls -Llr
? Then the options you are using for rsync can be improved. Maybe-c
is enabled or some other flag that reads more than just directories and metadata (inode data).Is rsync (with no files changed) nearly as fast as
ls -Llr
? Then you've tuned rsync as best as you can. You have to tune the OS, add RAM, get faster drives, change filesystems, etc.
Talk to your devs
80k files is just bad design. Very few file systems and system tools handle such large directories very well. If the filenames are abcdefg.txt, consider storing them in abdc/abcdefg.txt (note the repetition). This breaks the directories up into smaller ones, but doesn't require a huge change to the code.
Also.... consider using a database. If you have 80k files in a directory, maybe your developers are working around the fact that what they really want is a database. MariaDB or MySQL or PostgreSQL would be a much better option for storing large amounts of data.
Hey, what's wrong with 5 minutes?
Lastly, is 5 minutes really so bad? If you run this backup once a day, 5 minutes is not a lot of time. Yes, I love speed. However if 5 minutes is "good enough" for your customers, then it is good enough for you. If you don't have a written SLA, how about an informal discussion with your users to find out how fast they expect the backups to take.
I assume you didn't ask this question if there wasn't a need to improve the performance. However, if your customers are happy with 5 minutes, declare victory and move on to other projects that need your efforts.
Update: After some discussion we determined that the bottleneck is the network. I'm going to recommend 2 things before I give up :-).
- Try to squeeze more bandwidth out of the pipe with compression. However compression requires more CPU, so if your CPU is overloaded, it might make performance worse. Try rsync with and without
-z
, and configure your ssh with and without compression. Time all 4 combinations to see if any of them perform significantly better than others. - Watch network traffic to see if there are any pauses. If there are pauses, you can find what is causing them and optimize there. If rsync is always sending, then you really are at your limit. Your choices are:
- a faster network
- something other than rsync
- move the source and destination closer together. If you can't do that, can you rsync to a local machine then rsync to the real destination? There may be benefits to doing this if the system has to be down during the initial rsync.
80K is a lot of files.: There are 80k files in one directory tree. Each single directory does not have more than 2k files/subdirectories.
– guettli
Jan 7 '16 at 8:50
Check your rsync version: done, Make sure --checksum is not being used: done. Split the job into small batches: Thank you I will have a look at gigasync. OS defaults aren't made for this situation: done (the bottleneck is network not OS). Look at the "namei cache": done (it is net, not OS). Consider a different file system: again net, not OS. Maybe 5 minutes is the best you can do.: I think it could be much faster. Talk to your devs (use DB): This would be a giant change. Maybe an filesystem with better backup support would solve it.
– guettli
Jan 7 '16 at 8:58
2k files per directory is a lot better. thank you for the update. You hadn't mentioned that the network was slow. Is it low bandwidth, high latency, or both? rsync usually performs well on high latency links (it was developed by someone working on his PhD from Australia while dealing with computers in the U.S.). Try doing that "ls -lLR" over ssh and time how long it takes to transmit the result. "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list". Make sure the /tmp/list gets created on the local host.
– TomOnTime
Jan 7 '16 at 13:40
yes the network is slow. It is a pitty.
– guettli
Jan 7 '16 at 15:40
How slow? If you use "scp" to copy a 100M file, how long does it take? Also, what is the output of "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list"?
– TomOnTime
Jan 7 '16 at 15:41
|
show 8 more comments
Some unrelated points:
80K is a lot of files.
80,000 files in one directory? No operating system or app handles that situation very well by default. You just happen to notice this problem with rsync.
Check your rsync version
Modern rsync handles large directories a lot better than in the past. Be sure you are using the latest version.
Even old rsync handles large directories fairly well over high latency links... but 80k files isn't large...it is huge!
That said, rsync's memory usage is directly proportional to the number of files in a tree. Large directories take a large amount of RAM. The slowness may be due to a lack of RAM on either side. Do a test run while watching memory usage. Linux uses any left-over RAM as a disk cache, so if you are running low on RAM, there is less disk caching. If you run out of RAM and the system starts using swap, performance will be really bad.
Make sure --checksum is not being used
--checksum
(or -c
) requires reading each and every block of every file. You probably can get by with the default behavior of just reading the modification times (stored in the inode).
Split the job into small batches.
There are some projects like Gigasync which will "Chop up the workload by using perl to recurse the directory tree, building smallish lists of files to transfer with rsync."
The extra directory scan is going to be a large amount of overhead, but maybe it will be a net win.
OS defaults aren't made for this situation.
If you are using Linux/FreeBSD/etc with all the defaults, performance will be terrible for all your applications. The defaults assume smaller directories so-as not to waste RAM on oversized caches.
Tune your filesystem to better handle large directories: Do large folder sizes slow down IO performance?
Look at the "namei cache"
BSD-like operating systems have a cache that accelerates looking up a name to the inode (the "namei" cache"). There is a namei cache for each directory. If it is too small, it is a hindrance more than an optimization. Since rsync is doing a lstat() on each file, the inode is being accessed for every one of the 80k files. That might be blowing your cache. Research how to tune file directory performance on your system.
Consider a different file system
XFS was designed to handle larger directories. See Filesystem large number of files in a single directory
Maybe 5 minutes is the best you can do.
Consider calculating how many disk blocks are being read, and calculate how fast you should expect the hardware to be able to read that many blocks.
Maybe your expectations are too high. Consider how many disk blocks must be read to do an rsync with no changed files: each server will need to read the directory and read one inode per file. Let's assume nothing is cached because, well, 80k files has probably blown your cache. Let's say that it is 80k blocks to keep the math simple. That's about 40M of data, which should be readable in a few seconds. However if there needs to be a disk seek between each block, that could take much longer.
So you are going to need to read about 80,000 disk blocks. How fast can your hard drive do that? Considering that this is random I/O, not a long linear read, 5 minutes might be pretty excellent. That's 1 / (80000 / 600), or a disk read every 7.5ms. Is that fast or slow for your hard drive? It depends on the model.
Benchmark against something similar
Another way to think about it is this. If no files have changed, ls -Llr
does the same amount of disk activity but never reads any file data (just metadata). The time ls -Llr
takes to run is your upper bound.
Is rsync (with no files changed) significantly slower than
ls -Llr
? Then the options you are using for rsync can be improved. Maybe-c
is enabled or some other flag that reads more than just directories and metadata (inode data).Is rsync (with no files changed) nearly as fast as
ls -Llr
? Then you've tuned rsync as best as you can. You have to tune the OS, add RAM, get faster drives, change filesystems, etc.
Talk to your devs
80k files is just bad design. Very few file systems and system tools handle such large directories very well. If the filenames are abcdefg.txt, consider storing them in abdc/abcdefg.txt (note the repetition). This breaks the directories up into smaller ones, but doesn't require a huge change to the code.
Also.... consider using a database. If you have 80k files in a directory, maybe your developers are working around the fact that what they really want is a database. MariaDB or MySQL or PostgreSQL would be a much better option for storing large amounts of data.
Hey, what's wrong with 5 minutes?
Lastly, is 5 minutes really so bad? If you run this backup once a day, 5 minutes is not a lot of time. Yes, I love speed. However if 5 minutes is "good enough" for your customers, then it is good enough for you. If you don't have a written SLA, how about an informal discussion with your users to find out how fast they expect the backups to take.
I assume you didn't ask this question if there wasn't a need to improve the performance. However, if your customers are happy with 5 minutes, declare victory and move on to other projects that need your efforts.
Update: After some discussion we determined that the bottleneck is the network. I'm going to recommend 2 things before I give up :-).
- Try to squeeze more bandwidth out of the pipe with compression. However compression requires more CPU, so if your CPU is overloaded, it might make performance worse. Try rsync with and without
-z
, and configure your ssh with and without compression. Time all 4 combinations to see if any of them perform significantly better than others. - Watch network traffic to see if there are any pauses. If there are pauses, you can find what is causing them and optimize there. If rsync is always sending, then you really are at your limit. Your choices are:
- a faster network
- something other than rsync
- move the source and destination closer together. If you can't do that, can you rsync to a local machine then rsync to the real destination? There may be benefits to doing this if the system has to be down during the initial rsync.
Some unrelated points:
80K is a lot of files.
80,000 files in one directory? No operating system or app handles that situation very well by default. You just happen to notice this problem with rsync.
Check your rsync version
Modern rsync handles large directories a lot better than in the past. Be sure you are using the latest version.
Even old rsync handles large directories fairly well over high latency links... but 80k files isn't large...it is huge!
That said, rsync's memory usage is directly proportional to the number of files in a tree. Large directories take a large amount of RAM. The slowness may be due to a lack of RAM on either side. Do a test run while watching memory usage. Linux uses any left-over RAM as a disk cache, so if you are running low on RAM, there is less disk caching. If you run out of RAM and the system starts using swap, performance will be really bad.
Make sure --checksum is not being used
--checksum
(or -c
) requires reading each and every block of every file. You probably can get by with the default behavior of just reading the modification times (stored in the inode).
Split the job into small batches.
There are some projects like Gigasync which will "Chop up the workload by using perl to recurse the directory tree, building smallish lists of files to transfer with rsync."
The extra directory scan is going to be a large amount of overhead, but maybe it will be a net win.
OS defaults aren't made for this situation.
If you are using Linux/FreeBSD/etc with all the defaults, performance will be terrible for all your applications. The defaults assume smaller directories so-as not to waste RAM on oversized caches.
Tune your filesystem to better handle large directories: Do large folder sizes slow down IO performance?
Look at the "namei cache"
BSD-like operating systems have a cache that accelerates looking up a name to the inode (the "namei" cache"). There is a namei cache for each directory. If it is too small, it is a hindrance more than an optimization. Since rsync is doing a lstat() on each file, the inode is being accessed for every one of the 80k files. That might be blowing your cache. Research how to tune file directory performance on your system.
Consider a different file system
XFS was designed to handle larger directories. See Filesystem large number of files in a single directory
Maybe 5 minutes is the best you can do.
Consider calculating how many disk blocks are being read, and calculate how fast you should expect the hardware to be able to read that many blocks.
Maybe your expectations are too high. Consider how many disk blocks must be read to do an rsync with no changed files: each server will need to read the directory and read one inode per file. Let's assume nothing is cached because, well, 80k files has probably blown your cache. Let's say that it is 80k blocks to keep the math simple. That's about 40M of data, which should be readable in a few seconds. However if there needs to be a disk seek between each block, that could take much longer.
So you are going to need to read about 80,000 disk blocks. How fast can your hard drive do that? Considering that this is random I/O, not a long linear read, 5 minutes might be pretty excellent. That's 1 / (80000 / 600), or a disk read every 7.5ms. Is that fast or slow for your hard drive? It depends on the model.
Benchmark against something similar
Another way to think about it is this. If no files have changed, ls -Llr
does the same amount of disk activity but never reads any file data (just metadata). The time ls -Llr
takes to run is your upper bound.
Is rsync (with no files changed) significantly slower than
ls -Llr
? Then the options you are using for rsync can be improved. Maybe-c
is enabled or some other flag that reads more than just directories and metadata (inode data).Is rsync (with no files changed) nearly as fast as
ls -Llr
? Then you've tuned rsync as best as you can. You have to tune the OS, add RAM, get faster drives, change filesystems, etc.
Talk to your devs
80k files is just bad design. Very few file systems and system tools handle such large directories very well. If the filenames are abcdefg.txt, consider storing them in abdc/abcdefg.txt (note the repetition). This breaks the directories up into smaller ones, but doesn't require a huge change to the code.
Also.... consider using a database. If you have 80k files in a directory, maybe your developers are working around the fact that what they really want is a database. MariaDB or MySQL or PostgreSQL would be a much better option for storing large amounts of data.
Hey, what's wrong with 5 minutes?
Lastly, is 5 minutes really so bad? If you run this backup once a day, 5 minutes is not a lot of time. Yes, I love speed. However if 5 minutes is "good enough" for your customers, then it is good enough for you. If you don't have a written SLA, how about an informal discussion with your users to find out how fast they expect the backups to take.
I assume you didn't ask this question if there wasn't a need to improve the performance. However, if your customers are happy with 5 minutes, declare victory and move on to other projects that need your efforts.
Update: After some discussion we determined that the bottleneck is the network. I'm going to recommend 2 things before I give up :-).
- Try to squeeze more bandwidth out of the pipe with compression. However compression requires more CPU, so if your CPU is overloaded, it might make performance worse. Try rsync with and without
-z
, and configure your ssh with and without compression. Time all 4 combinations to see if any of them perform significantly better than others. - Watch network traffic to see if there are any pauses. If there are pauses, you can find what is causing them and optimize there. If rsync is always sending, then you really are at your limit. Your choices are:
- a faster network
- something other than rsync
- move the source and destination closer together. If you can't do that, can you rsync to a local machine then rsync to the real destination? There may be benefits to doing this if the system has to be down during the initial rsync.
edited Nov 13 '18 at 11:28
mkucharski
32
32
answered Jan 5 '16 at 14:50
TomOnTimeTomOnTime
5,16532249
5,16532249
80K is a lot of files.: There are 80k files in one directory tree. Each single directory does not have more than 2k files/subdirectories.
– guettli
Jan 7 '16 at 8:50
Check your rsync version: done, Make sure --checksum is not being used: done. Split the job into small batches: Thank you I will have a look at gigasync. OS defaults aren't made for this situation: done (the bottleneck is network not OS). Look at the "namei cache": done (it is net, not OS). Consider a different file system: again net, not OS. Maybe 5 minutes is the best you can do.: I think it could be much faster. Talk to your devs (use DB): This would be a giant change. Maybe an filesystem with better backup support would solve it.
– guettli
Jan 7 '16 at 8:58
2k files per directory is a lot better. thank you for the update. You hadn't mentioned that the network was slow. Is it low bandwidth, high latency, or both? rsync usually performs well on high latency links (it was developed by someone working on his PhD from Australia while dealing with computers in the U.S.). Try doing that "ls -lLR" over ssh and time how long it takes to transmit the result. "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list". Make sure the /tmp/list gets created on the local host.
– TomOnTime
Jan 7 '16 at 13:40
yes the network is slow. It is a pitty.
– guettli
Jan 7 '16 at 15:40
How slow? If you use "scp" to copy a 100M file, how long does it take? Also, what is the output of "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list"?
– TomOnTime
Jan 7 '16 at 15:41
|
show 8 more comments
80K is a lot of files.: There are 80k files in one directory tree. Each single directory does not have more than 2k files/subdirectories.
– guettli
Jan 7 '16 at 8:50
Check your rsync version: done, Make sure --checksum is not being used: done. Split the job into small batches: Thank you I will have a look at gigasync. OS defaults aren't made for this situation: done (the bottleneck is network not OS). Look at the "namei cache": done (it is net, not OS). Consider a different file system: again net, not OS. Maybe 5 minutes is the best you can do.: I think it could be much faster. Talk to your devs (use DB): This would be a giant change. Maybe an filesystem with better backup support would solve it.
– guettli
Jan 7 '16 at 8:58
2k files per directory is a lot better. thank you for the update. You hadn't mentioned that the network was slow. Is it low bandwidth, high latency, or both? rsync usually performs well on high latency links (it was developed by someone working on his PhD from Australia while dealing with computers in the U.S.). Try doing that "ls -lLR" over ssh and time how long it takes to transmit the result. "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list". Make sure the /tmp/list gets created on the local host.
– TomOnTime
Jan 7 '16 at 13:40
yes the network is slow. It is a pitty.
– guettli
Jan 7 '16 at 15:40
How slow? If you use "scp" to copy a 100M file, how long does it take? Also, what is the output of "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list"?
– TomOnTime
Jan 7 '16 at 15:41
80K is a lot of files.: There are 80k files in one directory tree. Each single directory does not have more than 2k files/subdirectories.
– guettli
Jan 7 '16 at 8:50
80K is a lot of files.: There are 80k files in one directory tree. Each single directory does not have more than 2k files/subdirectories.
– guettli
Jan 7 '16 at 8:50
Check your rsync version: done, Make sure --checksum is not being used: done. Split the job into small batches: Thank you I will have a look at gigasync. OS defaults aren't made for this situation: done (the bottleneck is network not OS). Look at the "namei cache": done (it is net, not OS). Consider a different file system: again net, not OS. Maybe 5 minutes is the best you can do.: I think it could be much faster. Talk to your devs (use DB): This would be a giant change. Maybe an filesystem with better backup support would solve it.
– guettli
Jan 7 '16 at 8:58
Check your rsync version: done, Make sure --checksum is not being used: done. Split the job into small batches: Thank you I will have a look at gigasync. OS defaults aren't made for this situation: done (the bottleneck is network not OS). Look at the "namei cache": done (it is net, not OS). Consider a different file system: again net, not OS. Maybe 5 minutes is the best you can do.: I think it could be much faster. Talk to your devs (use DB): This would be a giant change. Maybe an filesystem with better backup support would solve it.
– guettli
Jan 7 '16 at 8:58
2k files per directory is a lot better. thank you for the update. You hadn't mentioned that the network was slow. Is it low bandwidth, high latency, or both? rsync usually performs well on high latency links (it was developed by someone working on his PhD from Australia while dealing with computers in the U.S.). Try doing that "ls -lLR" over ssh and time how long it takes to transmit the result. "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list". Make sure the /tmp/list gets created on the local host.
– TomOnTime
Jan 7 '16 at 13:40
2k files per directory is a lot better. thank you for the update. You hadn't mentioned that the network was slow. Is it low bandwidth, high latency, or both? rsync usually performs well on high latency links (it was developed by someone working on his PhD from Australia while dealing with computers in the U.S.). Try doing that "ls -lLR" over ssh and time how long it takes to transmit the result. "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list". Make sure the /tmp/list gets created on the local host.
– TomOnTime
Jan 7 '16 at 13:40
yes the network is slow. It is a pitty.
– guettli
Jan 7 '16 at 15:40
yes the network is slow. It is a pitty.
– guettli
Jan 7 '16 at 15:40
How slow? If you use "scp" to copy a 100M file, how long does it take? Also, what is the output of "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list"?
– TomOnTime
Jan 7 '16 at 15:41
How slow? If you use "scp" to copy a 100M file, how long does it take? Also, what is the output of "time ssh remotehost 'cd /dest && ls -lLR' >/tmp/list"?
– TomOnTime
Jan 7 '16 at 15:41
|
show 8 more comments
No, that's not possible with rsync and it would be quite inefficient in another regard:
Normally, rsync
only compares file modification dates and file sizes. Your approach would force it to read and checksum the content of all files twice (on the local and remote system) to find changed directories.
1
AFAIK rsync checks mtime and size. If both matches, the file is not transferred again (at least in the default settings). It would be enough to send the hash of the tuples (filename, size, mtime). There is no need to checksum the content.
– guettli
Jan 4 '16 at 10:59
Yes, you are correct, but anyway,rsync
doesn't do this.
– Sven♦
Jan 4 '16 at 10:59
add a comment |
No, that's not possible with rsync and it would be quite inefficient in another regard:
Normally, rsync
only compares file modification dates and file sizes. Your approach would force it to read and checksum the content of all files twice (on the local and remote system) to find changed directories.
1
AFAIK rsync checks mtime and size. If both matches, the file is not transferred again (at least in the default settings). It would be enough to send the hash of the tuples (filename, size, mtime). There is no need to checksum the content.
– guettli
Jan 4 '16 at 10:59
Yes, you are correct, but anyway,rsync
doesn't do this.
– Sven♦
Jan 4 '16 at 10:59
add a comment |
No, that's not possible with rsync and it would be quite inefficient in another regard:
Normally, rsync
only compares file modification dates and file sizes. Your approach would force it to read and checksum the content of all files twice (on the local and remote system) to find changed directories.
No, that's not possible with rsync and it would be quite inefficient in another regard:
Normally, rsync
only compares file modification dates and file sizes. Your approach would force it to read and checksum the content of all files twice (on the local and remote system) to find changed directories.
edited Jan 4 '16 at 10:55
answered Jan 4 '16 at 10:49
Sven♦Sven
88.3k10150202
88.3k10150202
1
AFAIK rsync checks mtime and size. If both matches, the file is not transferred again (at least in the default settings). It would be enough to send the hash of the tuples (filename, size, mtime). There is no need to checksum the content.
– guettli
Jan 4 '16 at 10:59
Yes, you are correct, but anyway,rsync
doesn't do this.
– Sven♦
Jan 4 '16 at 10:59
add a comment |
1
AFAIK rsync checks mtime and size. If both matches, the file is not transferred again (at least in the default settings). It would be enough to send the hash of the tuples (filename, size, mtime). There is no need to checksum the content.
– guettli
Jan 4 '16 at 10:59
Yes, you are correct, but anyway,rsync
doesn't do this.
– Sven♦
Jan 4 '16 at 10:59
1
1
AFAIK rsync checks mtime and size. If both matches, the file is not transferred again (at least in the default settings). It would be enough to send the hash of the tuples (filename, size, mtime). There is no need to checksum the content.
– guettli
Jan 4 '16 at 10:59
AFAIK rsync checks mtime and size. If both matches, the file is not transferred again (at least in the default settings). It would be enough to send the hash of the tuples (filename, size, mtime). There is no need to checksum the content.
– guettli
Jan 4 '16 at 10:59
Yes, you are correct, but anyway,
rsync
doesn't do this.– Sven♦
Jan 4 '16 at 10:59
Yes, you are correct, but anyway,
rsync
doesn't do this.– Sven♦
Jan 4 '16 at 10:59
add a comment |
For synchronisation of large numbers of files (where little has changed), it is also worth setting noatime
on the source and destination partitions. This saves writing access times to the disk for each unchanged file.
Yes, the noatime option makes sense. We use it since several years. I guess an alternative to rsync is needed.
– guettli
Aug 28 '17 at 7:07
add a comment |
For synchronisation of large numbers of files (where little has changed), it is also worth setting noatime
on the source and destination partitions. This saves writing access times to the disk for each unchanged file.
Yes, the noatime option makes sense. We use it since several years. I guess an alternative to rsync is needed.
– guettli
Aug 28 '17 at 7:07
add a comment |
For synchronisation of large numbers of files (where little has changed), it is also worth setting noatime
on the source and destination partitions. This saves writing access times to the disk for each unchanged file.
For synchronisation of large numbers of files (where little has changed), it is also worth setting noatime
on the source and destination partitions. This saves writing access times to the disk for each unchanged file.
answered Aug 26 '17 at 11:08
Andy BeverleyAndy Beverley
1413
1413
Yes, the noatime option makes sense. We use it since several years. I guess an alternative to rsync is needed.
– guettli
Aug 28 '17 at 7:07
add a comment |
Yes, the noatime option makes sense. We use it since several years. I guess an alternative to rsync is needed.
– guettli
Aug 28 '17 at 7:07
Yes, the noatime option makes sense. We use it since several years. I guess an alternative to rsync is needed.
– guettli
Aug 28 '17 at 7:07
Yes, the noatime option makes sense. We use it since several years. I guess an alternative to rsync is needed.
– guettli
Aug 28 '17 at 7:07
add a comment |
Use rsync in daemon mode at the server end to speed up the listing/checksum process:
- Rsync daemon: is it really useful?
- http://giantdorks.org/alain/achieve-faster-file-transfer-times-by-running-rsync-as-a-daemon/
Note it isn't encrypted, but may be able to be tunneled without losing the listing performance improvement.
Also having rsync do compression rather than ssh should improve performance.
add a comment |
Use rsync in daemon mode at the server end to speed up the listing/checksum process:
- Rsync daemon: is it really useful?
- http://giantdorks.org/alain/achieve-faster-file-transfer-times-by-running-rsync-as-a-daemon/
Note it isn't encrypted, but may be able to be tunneled without losing the listing performance improvement.
Also having rsync do compression rather than ssh should improve performance.
add a comment |
Use rsync in daemon mode at the server end to speed up the listing/checksum process:
- Rsync daemon: is it really useful?
- http://giantdorks.org/alain/achieve-faster-file-transfer-times-by-running-rsync-as-a-daemon/
Note it isn't encrypted, but may be able to be tunneled without losing the listing performance improvement.
Also having rsync do compression rather than ssh should improve performance.
Use rsync in daemon mode at the server end to speed up the listing/checksum process:
- Rsync daemon: is it really useful?
- http://giantdorks.org/alain/achieve-faster-file-transfer-times-by-running-rsync-as-a-daemon/
Note it isn't encrypted, but may be able to be tunneled without losing the listing performance improvement.
Also having rsync do compression rather than ssh should improve performance.
edited Mar 22 at 21:41
answered Mar 22 at 21:29
Gringo SuaveGringo Suave
314312
314312
add a comment |
add a comment |
Thanks for contributing an answer to Server Fault!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fserverfault.com%2fquestions%2f746551%2ffaster-rsync-of-huge-directory-which-was-not-changed%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
1
You might read up on zsync. I have not used it myself, but from what I read, it pre-renders the metadata on the server side and might just speed up transfers in your case. It might be worth testing anyway. Beyond that, the only other solution I am aware of is real time block level syncronization that comes with some san/nas solutions.
– Aaron
Jan 5 '16 at 4:47