| Commit message (Collapse) | Author | Age |
|\ |
|
| | |
|
| | |
|
|/
|
|
|
|
|
|
|
|
|
| |
However, I don't yet have a reliable way to deal with files being modified
while they're being transferred. I have code that detects it on the sending
side, but the receiver is still free to move the wrong content into its
annex, and record that it has the content. So that's not acceptable, and
I'll need to work on it some more.
However, at this point I can use a direct mode repository as a remote and
transfer files from and to it.
|
|
|
|
| |
livedrive.com. Needs DAV version 0.3.
|
| |
|
|
|
|
| |
currently-supported AWS regions.
|
| |
|
| |
|
| |
|
|
|
|
| |
one-click enabling of the repository.
|
|
|
|
| |
remotes.
|
|
|
|
| |
The meter code does that too.
|
| |
|
| |
|
|
|
|
|
|
| |
Higher than any other remote, this is mostly due to the long retrieval
time, so it'd make sense to get a file from nearly any other remote.
(Unless it's behind a very slow connection.)
|
| |
|
| |
|
|
|
|
| |
to the repository, by setting embedcreds=yes|no when running initremote.
|
| |
|
|
|
|
|
|
|
|
|
|
| |
too small chunksize was configured.
Ensure that each file has something written to it, even if the bytestring
chunk size is greater than the configured chunksize.
This means we may write a bit larger than the configured value, but only
when the configured value is very small; ie, < 8 kb.
|
|
|
|
|
|
|
|
|
| |
Files are now written to a tmp directory in the remote, and once all
chunks are written, etc, it's moved into the final place atomically.
For now, checkpresent still checks every single chunk of a file, because
the old method could leave partially transferred files with some chunks
present and others not.
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
|
| |
Both the directory and webdav special remotes used to have to buffer
the whole file contents before it could be decrypted, as they read
from chunks. Now the chunks are streamed through gpg with no buffering.
|
| |
|
|
|
|
|
| |
The Element data type changed to use a map of attributes. Rather than
ifdef, I'm avoiding directly using that data type.
|
| |
|
|
|
|
|
| |
Note that receiving encrypted chunked content currently involves buffering.
(So does doing so with the directory special remote.)
|
|
|
|
|
|
|
|
| |
This allows deleting all chunks for a file with a single http command,
so it's a win after all.
However, does not look in the mixed case hash directories, which were
in the past used by the directory, etc remotes.
|
|
|
|
| |
Not yet getting it though.
|
|
|
|
|
| |
However, directory still uses its optimzed chunked file writer, as it uses
less memory than the generic one in the helper.
|
| |
|
|
|
|
|
|
|
|
|
|
| |
The benefit of using a compatable directory structure does not outweigh the
cost in complexity of handling the multiple locations content can be stored
in directory special remotes. And this also allows doing away with the parent
directories, which can't be made unwritable in DAV, so have no benefit
there. This will save 2 http calls per file store.
But, kept the directory hashing, just in case.
|
| |
|
| |
|
|
|
|
|
| |
Doesn't actually store anything yet, but initremote works and tests the
server.
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
| |
a key's presence.
Also, use the new withQuietOutput function to avoid running the shell to
/dev/null stderr in two other places.
|
|
|
|
| |
which doesn't work with LDAP or NIS.
|
| |
|
|
|
|
|
|
|
| |
bup 0.25 does not accept that; and bup split reads from stdin by
default if no file is given. I'm not sure what version of bup changed this.
This only affected bup special remotes that were encrypted.
|
| |
|
|\
| |
| |
| |
| |
| | |
Conflicts:
debian/changelog
git-annex.cabal
|