aboutsummaryrefslogtreecommitdiff
path: root/doc/forum/__34__git_annex_sync__34___synced_after_8_hours.mdwn
blob: 2804828f59f5d5c0e7d8e257e0de7021a0c9cdab (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
Hi,

The git annex seem has problem with many files.

For synchronize, the operation lasts 8 hours. Here the sample for synchronizing to my local remote server (sbackup)

start at **20:12** / end at **04:13** / total time = ~ **8 hours**

    git annex sync sbackup
    
    [2015-04-13 20:12:26 CEST] call: git ["--git-dir=.git","--work-tree=.","push","sbackup","+git-annex:synced/git-annex","master:synced/master"]
    Counting objects: 792155, done.
    Delta compression using up to 4 threads.
    Compressing objects: 100% (789727/789727), done.
    Writing objects: 100% (792155/792155), 75.73 MiB | 2.35 MiB/s, done.
    Total 792155 (delta 449604), reused 1 (delta 0)
    To partage@192.168.253.53:/data/samba/git-annex/docshare
      ae182f0..fad3aca  git-annex -> synced/git-annex
      e0e67fe..5226a6f  master -> synced/master
    [2015-04-14 04:13:05 CEST] read: git ["--git-dir=.git","--work-tree=.","push","sbackup","git-annex","master"]
    ok

Another problem, I do not know exactly how many files I own (i use **find . | wc -l** )

.git = 1250633

documents = 61124

medias = 199504

it seem i own ~250000 files, but in the .git **1.2 millions files**.

The following command also very slow

    git annex info


What the best pratices for use git annex with many files > 500 000 or maintenance & reduction/cleaning method

Thanks