1
0
mirror of https://github.com/rclone/rclone.git synced 2025-12-30 15:13:55 +00:00

Compare commits

...

51 Commits

Author SHA1 Message Date
Nick Craig-Wood
a2b292b31d Version v1.71.2 2025-10-20 13:59:45 +01:00
Nick Craig-Wood
4013f006b8 docs: update sponsors 2025-10-20 12:49:19 +01:00
Nick Craig-Wood
371fc61191 docs: update sponsor images 2025-10-20 12:49:19 +01:00
Nick Craig-Wood
0d9a95125c docs: update privacy policy with a section on user data 2025-10-20 12:49:19 +01:00
Dulani Woods
386d3c5b4d gcs: add region us-east5 - fixes #8863 2025-10-20 12:49:19 +01:00
dougal
3dceb72e9b index: add missing providers 2025-10-20 12:49:19 +01:00
dougal
30b016d8c0 docs: add missing ` 2025-10-20 12:49:19 +01:00
Nick Craig-Wood
b8e6d45482 mega: fix 402 payment required errors - fixes #8758
The underlying library now supports hashcash which should fix this
problem.
2025-10-20 12:49:19 +01:00
iTrooz
dfdc69a3e1 docs: fix variants of --user-from-header 2025-10-20 12:41:15 +01:00
albertony
0908c346bb docs: add remote setup page to main docs dropdown 2025-10-20 12:41:15 +01:00
albertony
f186b7554e docs: update remote setup page 2025-10-20 12:41:15 +01:00
albertony
4bbab927fa docs: add link from authorize command docs to remote setup docs 2025-10-20 12:41:15 +01:00
albertony
1c119f9330 docs: lowercase internet and web browser instead of Internet browser 2025-10-20 12:41:15 +01:00
albertony
089c4b1a69 docs: use the term backend name instead of fs name for authorize command 2025-10-20 12:41:15 +01:00
Nick Craig-Wood
b5ba737e14 bisync: fix race when CaptureOutput is used concurrently #8815
Before this change CaptureOutput could trip the race detector when
used concurrently. In particular if go routines using the logging are
outlasting the return from `fun()`.

This fixes the problem with a mutex.
2025-10-20 12:41:15 +01:00
Vikas Bhansali
15f727f0e3 azurefiles: Fix server side copy not waiting for completion - fixes #8848 2025-10-20 12:41:15 +01:00
Youfu Zhang
01e1f90216 pikpak: fix unnecessary retries by using URL expire parameter - fixes #8601
Before this change, rclone would unnecessarily retry downloads when
the `Link.Expire` field was unreliable but the download URL contained
a valid expire query parameter. This primarily affects cases where
media links are unavailable or when `no_media_link` is enabled.

The `Link.Valid()` method now primarily checks the URL's expire query
parameter (as Unix timestamp) and falls back to the Expire field
only when URL parsing fails. This eliminates the `error no link`
retry loops while maintaining backward compatibility.

Signed-off-by: Youfu Zhang <zhangyoufu@gmail.com>
2025-10-20 12:41:15 +01:00
dougal
997c73b586 serve http: fix: logging url on start 2025-10-20 12:41:15 +01:00
Matt LaPaglia
ccd01f1a68 docs: fix typo 2025-10-20 12:41:15 +01:00
dougal
63c0dc773c b2: fix 1TB+ uploads
Before this change the minimum chunk size would default to 96M which
would allow a maximum size of just below 1TB file to be uploaded, due to
the 10000 part rule for b2.

Now the calculated chunk size is used so the chunk size can be 5GB
making a max file size of 50TB.

Fixes #8460
2025-10-20 12:41:15 +01:00
Nick Craig-Wood
f1e6bffc14 Start v1.71.2-DEV development 2025-09-24 17:33:54 +01:00
Nick Craig-Wood
1a98c36a73 Version v1.71.1 2025-09-24 16:24:23 +01:00
Nick Craig-Wood
931ab7c4db pacer: fix deadlock with --max-connections
If the pacer was used recursively and --max-connections was in use
then it could deadlock if all the connections were in use at the time
of recursive call (likely).

This affected the azureblob backend because when it receives an
InvalidBlockOrBlob error it attempts to clear the condition before
retrying. This in turn involves recursively calling the pacer.

This fixes the problem by skipping the --max-connections check if the
pacer is called recursively.

The recursive detection is done by stack inspection which isn't ideal,
but the alternative would be to add ctx to all >1,000 pacer calls. The
benchmark reveals stack inspection takes about 55nS per stack level so
it is relatively cheap.
2025-09-22 17:40:22 +01:00
Nick Craig-Wood
2b531ada34 Revert "azureblob: fix deadlock with --max-connections with InvalidBlockOrBlob errors"
This reverts commit 0c1902cc6037d81eaf95e931172879517a25d529.

This turns out not to be sufficient so we need a better approach
2025-09-22 17:40:22 +01:00
Nick Craig-Wood
45f45c987c march: fix deadlock when using --fast-list on syncs - fixes #8811
Before this change, it was possible to have a deadlock when using
--fast-list for a sync if both the source and destination supported
ListR.

This fixes the problem by shortening the locking window.
2025-09-22 17:31:08 +01:00
dougal
d2351e60b6 docs: HDFS: erasure coding limitation #8808 2025-09-22 17:31:08 +01:00
nielash
f8de6b48f1 local: fix rmdir "Access is denied" on windows - fixes #8363
Before this change, Rmdir (and other commands that rely on Rmdir) would fail
with "Access is denied" on Windows, if the directory had
FILE_ATTRIBUTE_READONLY. This could happen if, for example, an empty folder had
a custom icon added via Windows Explorer's interface (Properties => Customize =>
Change Icon...).

However, Microsoft docs indicate that "This attribute is not honored on
directories."
https://learn.microsoft.com/en-us/windows/win32/fileio/file-attribute-constants#file_attribute_readonly
Accordingly, this created an odd situation where such directories were removable
(by their owner) via File Explorer and the rd command, but not via rclone.

An upstream issue has been open since 2018, but has not yet resulted in a fix.
https://github.com/golang/go/issues/26295

This change gets around the issue by doing os.Chmod on the dir and then retrying
os.Remove. If the dir is not empty, this will still fail with "The directory is
not empty."

A bisync user confirmed that it fixed their issue in
https://forum.rclone.org/t/bisync-leaving-empty-directories-on-unc-path-1-or-local-filesystem-path-2-on-directory-renames/52456/4?u=nielash

It is likely also a fix for #8019, although @ncw is correct that Purge would be
a more efficient solution in that particular scenario.
2025-09-22 17:31:08 +01:00
nielash
616a280aac bisync: fix error handling for renamed conflicts
Before this change, rclone could crash during modifyListing if a rename's
srcNewName is known but not found in the srcList
(srcNewName != "" && new == nil).
This scenario should not happen, but if it does, we should print an error
instead of crashing.

On #8458 there is a report of this possibly happening on v1.68.2. It is unknown
what the underlying issue was, and whether it still exists in the latest
version, but if it does, the user will now see an error and debug info instead
of a crash.
2025-09-22 17:31:08 +01:00
Jean-Christophe Cura
e1833f4090 docs: pcloud: update root_folder_id instructions 2025-09-22 17:31:08 +01:00
Nick Craig-Wood
135d89d0f9 operations: fix partial name collisions for non --inplace copies
In this commit:

c63f1865f3 operations: copy: generate stable partial suffix

We made the partial suffix for non inplace copies stable. This was a
hash based off the file fingerprint.

However, given a directory of files which have the same fingerprint
the partial suffix collides. On some backends (eg the local backend)
the fingerprint is just the size and modification time so files with
different contents can collide.

The effect of collisions was hash failures on copy when using
--transfers > 1. These copies invariably retried successfully which
probably explains why this bug hasn't been reported.

This fixes the problem by adding the file name to the hash.

It also makes sure the hash is always represented as 8 hex bytes for
consistency.
2025-09-22 17:31:08 +01:00
Ed Craig-Wood
ea54bddbd5 drive: docs: update making your own client ID instructions
update instructions with the most recent changes to google cloud console
2025-09-22 17:31:08 +01:00
Nick Craig-Wood
baf6167930 internetarchive: fix server side copy files with spaces
In this commit we broke server side copy for files with spaces

4c5764204d internetarchive: fix server side copy files with &

This fixes the problem by using rest.URLPathEscapeAll which escapes
everything possible.

Fixes #8754
2025-09-22 17:31:08 +01:00
Nick Craig-Wood
1d91618d9e lib/rest: add URLPathEscapeAll to URL escape as many chars as possible 2025-09-22 17:31:08 +01:00
Nick Craig-Wood
4a7e62b79c docs: add link to MEGA S4 from MEGA page 2025-09-22 17:31:08 +01:00
anon-pradip
dcca477f39 docs: clarify subcommand description in rclone usage 2025-09-22 17:31:08 +01:00
albertony
7304ecaf18 docs: fix description of regex syntax of name transform 2025-09-22 17:31:08 +01:00
albertony
c3932ecde1 docs: add some more details about supported regex syntax 2025-09-22 17:31:08 +01:00
nielash
c9df7b1cd7 makefile: fix lib/transform docs not getting updated
As of
4280ec75cc
the lib/transform docs are generated with //go:generate and embedded with
//go:embed.

Before this change, however, they were not getting automatically updated with
subsequent changes (like
fe62a2bb4e)
because `go generate ./lib/transform` was not being run as part of the release
making process.

This change fixes that by running it in `make commanddocs`.
2025-09-22 17:31:08 +01:00
dougal
3985496e5d vfs: fix SIGHUP killing serve instead of flushing directory caches
Before, rclone serve would crash when sent a SIGHUP which contradicts
the documentation - saying it should flush the directory caches.

Moved signal handling from the mount into the vfs layer, which now
handles SIGHUP on all uses of the VFS including mount and serve.

Fixes #8607
2025-09-22 17:31:08 +01:00
albertony
0d2ef2eb20 docs: remove broken links from rc to commands 2025-09-22 17:31:08 +01:00
Claudius Ellsel
836e19243d docs: add example of how to add date as suffix 2025-09-22 17:31:08 +01:00
Nick Craig-Wood
0a6cce1bc1 box: fix about after change in API return - fixes #8776 2025-09-22 17:31:08 +01:00
albertony
cffb6732a4 docs: fix incorrectly escaped windows path separators 2025-09-22 17:31:08 +01:00
albertony
236f247c59 build: restore error handling in gendocs 2025-09-22 17:31:08 +01:00
skbeh
3b07f9d34d combine: propagate SlowHash feature 2025-09-22 17:31:08 +01:00
albertony
bad77c642f docs/oracleobjectstorage: add introduction before external links and remove broken link 2025-09-22 17:31:07 +01:00
albertony
41eef6608b docs: fix markdown lint issues in backend docs 2025-09-22 17:31:07 +01:00
albertony
fc6bd9ff79 docs: fix markdown lint issues in command docs 2025-09-22 17:31:07 +01:00
albertony
ee83cd214c docs: update markdown code block json indent size 2 2025-09-22 17:31:07 +01:00
Tilman Vogel
2c2642a927 mount: do not log successful unmount as an error - fixes #8766 2025-09-22 17:31:07 +01:00
Nick Craig-Wood
32eed8dd36 Start v1.71.1-DEV development 2025-09-22 17:15:11 +01:00
291 changed files with 19363 additions and 12695 deletions

View File

@@ -628,7 +628,7 @@ You'll need to modify the following files
- `backend/s3/s3.go` - `backend/s3/s3.go`
- Add the provider to `providerOption` at the top of the file - Add the provider to `providerOption` at the top of the file
- Add endpoints and other config for your provider gated on the provider in `fs.RegInfo`. - Add endpoints and other config for your provider gated on the provider in `fs.RegInfo`.
- Exclude your provider from generic config questions (eg `region` and `endpoint). - Exclude your provider from generic config questions (eg `region` and `endpoint`).
- Add the provider to the `setQuirks` function - see the documentation there. - Add the provider to the `setQuirks` function - see the documentation there.
- `docs/content/s3.md` - `docs/content/s3.md`
- Add the provider at the top of the page. - Add the provider at the top of the page.

8410
MANUAL.html generated

File diff suppressed because it is too large Load Diff

8128
MANUAL.md generated

File diff suppressed because it is too large Load Diff

2248
MANUAL.txt generated

File diff suppressed because it is too large Load Diff

View File

@@ -144,6 +144,7 @@ MANUAL.txt: MANUAL.md
pandoc -s --from markdown-smart --to plain MANUAL.md -o MANUAL.txt pandoc -s --from markdown-smart --to plain MANUAL.md -o MANUAL.txt
commanddocs: rclone commanddocs: rclone
go generate ./lib/transform
-@rmdir -p '$$HOME/.config/rclone' -@rmdir -p '$$HOME/.config/rclone'
XDG_CACHE_HOME="" XDG_CONFIG_HOME="" HOME="\$$HOME" USER="\$$USER" rclone gendocs --config=/notfound docs/content/ XDG_CACHE_HOME="" XDG_CONFIG_HOME="" HOME="\$$HOME" USER="\$$USER" rclone gendocs --config=/notfound docs/content/
@[ ! -e '$$HOME' ] || (echo 'Error: created unwanted directory named $$HOME' && exit 1) @[ ! -e '$$HOME' ] || (echo 'Error: created unwanted directory named $$HOME' && exit 1)

View File

@@ -1 +1 @@
v1.71.0 v1.71.2

View File

@@ -2765,8 +2765,6 @@ func (o *Object) clearUncommittedBlocks(ctx context.Context) (err error) {
blockList blockblob.GetBlockListResponse blockList blockblob.GetBlockListResponse
properties *blob.GetPropertiesResponse properties *blob.GetPropertiesResponse
options *blockblob.CommitBlockListOptions options *blockblob.CommitBlockListOptions
// Use temporary pacer as this can be called recursively which can cause a deadlock with --max-connections
pacer = fs.NewPacer(ctx, pacer.NewS3(pacer.MinSleep(minSleep), pacer.MaxSleep(maxSleep), pacer.DecayConstant(decayConstant)))
) )
properties, err = o.readMetaDataAlways(ctx) properties, err = o.readMetaDataAlways(ctx)
@@ -2778,7 +2776,7 @@ func (o *Object) clearUncommittedBlocks(ctx context.Context) (err error) {
if objectExists { if objectExists {
// Get the committed block list // Get the committed block list
err = pacer.Call(func() (bool, error) { err = o.fs.pacer.Call(func() (bool, error) {
blockList, err = blockBlobSVC.GetBlockList(ctx, blockblob.BlockListTypeAll, nil) blockList, err = blockBlobSVC.GetBlockList(ctx, blockblob.BlockListTypeAll, nil)
return o.fs.shouldRetry(ctx, err) return o.fs.shouldRetry(ctx, err)
}) })
@@ -2820,7 +2818,7 @@ func (o *Object) clearUncommittedBlocks(ctx context.Context) (err error) {
// Commit only the committed blocks // Commit only the committed blocks
fs.Debugf(o, "Committing %d blocks to remove uncommitted blocks", len(blockIDs)) fs.Debugf(o, "Committing %d blocks to remove uncommitted blocks", len(blockIDs))
err = pacer.Call(func() (bool, error) { err = o.fs.pacer.Call(func() (bool, error) {
_, err := blockBlobSVC.CommitBlockList(ctx, blockIDs, options) _, err := blockBlobSVC.CommitBlockList(ctx, blockIDs, options)
return o.fs.shouldRetry(ctx, err) return o.fs.shouldRetry(ctx, err)
}) })

View File

@@ -1313,10 +1313,29 @@ func (f *Fs) Copy(ctx context.Context, src fs.Object, remote string) (fs.Object,
} }
srcURL := srcObj.fileClient().URL() srcURL := srcObj.fileClient().URL()
fc := f.fileClient(remote) fc := f.fileClient(remote)
_, err = fc.StartCopyFromURL(ctx, srcURL, &opt) startCopy, err := fc.StartCopyFromURL(ctx, srcURL, &opt)
if err != nil { if err != nil {
return nil, fmt.Errorf("Copy failed: %w", err) return nil, fmt.Errorf("Copy failed: %w", err)
} }
// Poll for completion if necessary
//
// The for loop is never executed for same storage account copies.
copyStatus := startCopy.CopyStatus
var properties file.GetPropertiesResponse
pollTime := 100 * time.Millisecond
for copyStatus != nil && string(*copyStatus) == string(file.CopyStatusTypePending) {
time.Sleep(pollTime)
properties, err = fc.GetProperties(ctx, &file.GetPropertiesOptions{})
if err != nil {
return nil, err
}
copyStatus = properties.CopyStatus
pollTime = min(2*pollTime, time.Second)
}
dstObj, err := f.NewObject(ctx, remote) dstObj, err := f.NewObject(ctx, remote)
if err != nil { if err != nil {
return nil, fmt.Errorf("Copy: NewObject failed: %w", err) return nil, fmt.Errorf("Copy: NewObject failed: %w", err)

View File

@@ -2192,13 +2192,17 @@ func (f *Fs) OpenChunkWriter(ctx context.Context, remote string, src fs.ObjectIn
return info, nil, err return info, nil, err
} }
up, err := f.newLargeUpload(ctx, o, nil, src, f.opt.ChunkSize, false, nil, options...)
if err != nil {
return info, nil, err
}
info = fs.ChunkWriterInfo{ info = fs.ChunkWriterInfo{
ChunkSize: int64(f.opt.ChunkSize), ChunkSize: up.chunkSize,
Concurrency: o.fs.opt.UploadConcurrency, Concurrency: o.fs.opt.UploadConcurrency,
//LeavePartsOnError: o.fs.opt.LeavePartsOnError, //LeavePartsOnError: o.fs.opt.LeavePartsOnError,
} }
up, err := f.newLargeUpload(ctx, o, nil, src, f.opt.ChunkSize, false, nil, options...) return info, up, nil
return info, up, err
} }
// Remove an object // Remove an object

View File

@@ -125,10 +125,21 @@ type FolderItems struct {
Offset int `json:"offset"` Offset int `json:"offset"`
Limit int `json:"limit"` Limit int `json:"limit"`
NextMarker *string `json:"next_marker,omitempty"` NextMarker *string `json:"next_marker,omitempty"`
Order []struct { // There is some confusion about how this is actually
By string `json:"by"` // returned. The []struct has worked for many years, but in
Direction string `json:"direction"` // https://github.com/rclone/rclone/issues/8776 box was
} `json:"order"` // returning it returned not as a list. We don't actually use
// this so comment it out.
//
// Order struct {
// By string `json:"by"`
// Direction string `json:"direction"`
// } `json:"order"`
//
// Order []struct {
// By string `json:"by"`
// Direction string `json:"direction"`
// } `json:"order"`
} }
// Parent defined the ID of the parent directory // Parent defined the ID of the parent directory

View File

@@ -241,18 +241,22 @@ func NewFs(ctx context.Context, name, root string, m configmap.Mapper) (outFs fs
DirModTimeUpdatesOnWrite: true, DirModTimeUpdatesOnWrite: true,
PartialUploads: true, PartialUploads: true,
}).Fill(ctx, f) }).Fill(ctx, f)
canMove := true canMove, slowHash := true, false
for _, u := range f.upstreams { for _, u := range f.upstreams {
features = features.Mask(ctx, u.f) // Mask all upstream fs features = features.Mask(ctx, u.f) // Mask all upstream fs
if !operations.CanServerSideMove(u.f) { if !operations.CanServerSideMove(u.f) {
canMove = false canMove = false
} }
slowHash = slowHash || u.f.Features().SlowHash
} }
// We can move if all remotes support Move or Copy // We can move if all remotes support Move or Copy
if canMove { if canMove {
features.Move = f.Move features.Move = f.Move
} }
// If any of upstreams are SlowHash, propagate it
features.SlowHash = slowHash
// Enable ListR when upstreams either support ListR or is local // Enable ListR when upstreams either support ListR or is local
// But not when all upstreams are local // But not when all upstreams are local
if features.ListR == nil { if features.ListR == nil {

View File

@@ -252,6 +252,9 @@ Docs: https://cloud.google.com/storage/docs/bucket-policy-only
}, { }, {
Value: "us-east4", Value: "us-east4",
Help: "Northern Virginia", Help: "Northern Virginia",
}, {
Value: "us-east5",
Help: "Ohio",
}, { }, {
Value: "us-west1", Value: "us-west1",
Help: "Oregon", Help: "Oregon",

View File

@@ -590,7 +590,7 @@ func (f *Fs) PublicLink(ctx context.Context, remote string, expire fs.Duration,
return "", err return "", err
} }
bucket, bucketPath := f.split(remote) bucket, bucketPath := f.split(remote)
return path.Join(f.opt.FrontEndpoint, "/download/", bucket, quotePath(bucketPath)), nil return path.Join(f.opt.FrontEndpoint, "/download/", bucket, rest.URLPathEscapeAll(bucketPath)), nil
} }
// Copy src to this remote using server-side copy operations. // Copy src to this remote using server-side copy operations.
@@ -622,7 +622,7 @@ func (f *Fs) Copy(ctx context.Context, src fs.Object, remote string) (_ fs.Objec
"x-archive-auto-make-bucket": "1", "x-archive-auto-make-bucket": "1",
"x-archive-queue-derive": "0", "x-archive-queue-derive": "0",
"x-archive-keep-old-version": "0", "x-archive-keep-old-version": "0",
"x-amz-copy-source": quotePath(path.Join("/", srcBucket, srcPath)), "x-amz-copy-source": rest.URLPathEscapeAll(path.Join("/", srcBucket, srcPath)),
"x-amz-metadata-directive": "COPY", "x-amz-metadata-directive": "COPY",
"x-archive-filemeta-sha1": srcObj.sha1, "x-archive-filemeta-sha1": srcObj.sha1,
"x-archive-filemeta-md5": srcObj.md5, "x-archive-filemeta-md5": srcObj.md5,
@@ -778,7 +778,7 @@ func (o *Object) Open(ctx context.Context, options ...fs.OpenOption) (in io.Read
// make a GET request to (frontend)/download/:item/:path // make a GET request to (frontend)/download/:item/:path
opts := rest.Opts{ opts := rest.Opts{
Method: "GET", Method: "GET",
Path: path.Join("/download/", o.fs.root, quotePath(o.fs.opt.Enc.FromStandardPath(o.remote))), Path: path.Join("/download/", o.fs.root, rest.URLPathEscapeAll(o.fs.opt.Enc.FromStandardPath(o.remote))),
Options: optionsFixed, Options: optionsFixed,
} }
err = o.fs.pacer.Call(func() (bool, error) { err = o.fs.pacer.Call(func() (bool, error) {
@@ -1334,16 +1334,6 @@ func trimPathPrefix(s, prefix string, enc encoder.MultiEncoder) string {
return enc.ToStandardPath(strings.TrimPrefix(s, prefix+"/")) return enc.ToStandardPath(strings.TrimPrefix(s, prefix+"/"))
} }
// mimics urllib.parse.quote() on Python; exclude / from url.PathEscape
func quotePath(s string) string {
seg := strings.Split(s, "/")
newValues := []string{}
for _, v := range seg {
newValues = append(newValues, url.QueryEscape(v))
}
return strings.Join(newValues, "/")
}
var ( var (
_ fs.Fs = &Fs{} _ fs.Fs = &Fs{}
_ fs.Copier = &Fs{} _ fs.Copier = &Fs{}

View File

@@ -7,6 +7,7 @@ import (
"errors" "errors"
"fmt" "fmt"
"io" "io"
iofs "io/fs"
"os" "os"
"path" "path"
"path/filepath" "path/filepath"
@@ -841,7 +842,13 @@ func (f *Fs) Rmdir(ctx context.Context, dir string) error {
} else if !fi.IsDir() { } else if !fi.IsDir() {
return fs.ErrorIsFile return fs.ErrorIsFile
} }
return os.Remove(localPath) err := os.Remove(localPath)
if runtime.GOOS == "windows" && errors.Is(err, iofs.ErrPermission) { // https://github.com/golang/go/issues/26295
if os.Chmod(localPath, 0o600) == nil {
err = os.Remove(localPath)
}
}
return err
} }
// Precision of the file system // Precision of the file system

View File

@@ -0,0 +1,40 @@
//go:build windows
package local
import (
"context"
"path/filepath"
"runtime"
"syscall"
"testing"
"github.com/rclone/rclone/fs/operations"
"github.com/rclone/rclone/fstest"
"github.com/stretchr/testify/assert"
"github.com/stretchr/testify/require"
)
// TestRmdirWindows tests that FILE_ATTRIBUTE_READONLY does not block Rmdir on windows.
// Microsoft docs indicate that "This attribute is not honored on directories."
// See https://learn.microsoft.com/en-us/windows/win32/fileio/file-attribute-constants#file_attribute_readonly
// and https://github.com/golang/go/issues/26295
func TestRmdirWindows(t *testing.T) {
if runtime.GOOS != "windows" {
t.Skipf("windows only")
}
r := fstest.NewRun(t)
defer r.Finalise()
err := operations.Mkdir(context.Background(), r.Flocal, "testdir")
require.NoError(t, err)
ptr, err := syscall.UTF16PtrFromString(filepath.Join(r.Flocal.Root(), "testdir"))
require.NoError(t, err)
err = syscall.SetFileAttributes(ptr, uint32(syscall.FILE_ATTRIBUTE_DIRECTORY+syscall.FILE_ATTRIBUTE_READONLY))
require.NoError(t, err)
err = operations.Rmdir(context.Background(), r.Flocal, "testdir")
assert.NoError(t, err)
}

View File

@@ -5,6 +5,7 @@ package api
import ( import (
"fmt" "fmt"
"net/url"
"reflect" "reflect"
"strconv" "strconv"
"time" "time"
@@ -136,8 +137,25 @@ type Link struct {
} }
// Valid reports whether l is non-nil, has an URL, and is not expired. // Valid reports whether l is non-nil, has an URL, and is not expired.
// It primarily checks the URL's expire query parameter, falling back to the Expire field.
func (l *Link) Valid() bool { func (l *Link) Valid() bool {
return l != nil && l.URL != "" && time.Now().Add(10*time.Second).Before(time.Time(l.Expire)) if l == nil || l.URL == "" {
return false
}
// Primary validation: check URL's expire query parameter
if u, err := url.Parse(l.URL); err == nil {
if expireStr := u.Query().Get("expire"); expireStr != "" {
// Try parsing as Unix timestamp (seconds)
if expireInt, err := strconv.ParseInt(expireStr, 10, 64); err == nil {
expireTime := time.Unix(expireInt, 0)
return time.Now().Add(10 * time.Second).Before(expireTime)
}
}
}
// Fallback validation: use the Expire field if URL parsing didn't work
return time.Now().Add(10 * time.Second).Before(time.Time(l.Expire))
} }
// URL is a basic form of URL // URL is a basic form of URL

View File

@@ -0,0 +1,99 @@
package api
import (
"fmt"
"testing"
"time"
)
// TestLinkValid tests the Link.Valid method for various scenarios
func TestLinkValid(t *testing.T) {
tests := []struct {
name string
link *Link
expected bool
desc string
}{
{
name: "nil link",
link: nil,
expected: false,
desc: "nil link should be invalid",
},
{
name: "empty URL",
link: &Link{URL: ""},
expected: false,
desc: "empty URL should be invalid",
},
{
name: "valid URL with future expire parameter",
link: &Link{
URL: fmt.Sprintf("https://example.com/file?expire=%d", time.Now().Add(time.Hour).Unix()),
},
expected: true,
desc: "URL with future expire parameter should be valid",
},
{
name: "expired URL with past expire parameter",
link: &Link{
URL: fmt.Sprintf("https://example.com/file?expire=%d", time.Now().Add(-time.Hour).Unix()),
},
expected: false,
desc: "URL with past expire parameter should be invalid",
},
{
name: "URL expire parameter takes precedence over Expire field",
link: &Link{
URL: fmt.Sprintf("https://example.com/file?expire=%d", time.Now().Add(time.Hour).Unix()),
Expire: Time(time.Now().Add(-time.Hour)), // Fallback is expired
},
expected: true,
desc: "URL expire parameter should take precedence over Expire field",
},
{
name: "URL expire parameter within 10 second buffer should be invalid",
link: &Link{
URL: fmt.Sprintf("https://example.com/file?expire=%d", time.Now().Add(5*time.Second).Unix()),
},
expected: false,
desc: "URL expire parameter within 10 second buffer should be invalid",
},
{
name: "fallback to Expire field when no URL expire parameter",
link: &Link{
URL: "https://example.com/file",
Expire: Time(time.Now().Add(time.Hour)),
},
expected: true,
desc: "should fallback to Expire field when URL has no expire parameter",
},
{
name: "fallback to Expire field when URL expire parameter is invalid",
link: &Link{
URL: "https://example.com/file?expire=invalid",
Expire: Time(time.Now().Add(time.Hour)),
},
expected: true,
desc: "should fallback to Expire field when URL expire parameter is unparseable",
},
{
name: "invalid when both URL expire and Expire field are expired",
link: &Link{
URL: fmt.Sprintf("https://example.com/file?expire=%d", time.Now().Add(-time.Hour).Unix()),
Expire: Time(time.Now().Add(-time.Hour)),
},
expected: false,
desc: "should be invalid when both URL expire and Expire field are expired",
},
}
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
result := tt.link.Valid()
if result != tt.expected {
t.Errorf("Link.Valid() = %v, expected %v. %s", result, tt.expected, tt.desc)
}
})
}
}

View File

@@ -51,47 +51,52 @@ output. The output is typically used, free, quota and trash contents.
E.g. Typical output from ` + "`rclone about remote:`" + ` is: E.g. Typical output from ` + "`rclone about remote:`" + ` is:
Total: 17 GiB ` + "```text" + `
Used: 7.444 GiB Total: 17 GiB
Free: 1.315 GiB Used: 7.444 GiB
Trashed: 100.000 MiB Free: 1.315 GiB
Other: 8.241 GiB Trashed: 100.000 MiB
Other: 8.241 GiB
` + "```" + `
Where the fields are: Where the fields are:
* Total: Total size available. - Total: Total size available.
* Used: Total size used. - Used: Total size used.
* Free: Total space available to this user. - Free: Total space available to this user.
* Trashed: Total space used by trash. - Trashed: Total space used by trash.
* Other: Total amount in other storage (e.g. Gmail, Google Photos). - Other: Total amount in other storage (e.g. Gmail, Google Photos).
* Objects: Total number of objects in the storage. - Objects: Total number of objects in the storage.
All sizes are in number of bytes. All sizes are in number of bytes.
Applying a ` + "`--full`" + ` flag to the command prints the bytes in full, e.g. Applying a ` + "`--full`" + ` flag to the command prints the bytes in full, e.g.
Total: 18253611008 ` + "```text" + `
Used: 7993453766 Total: 18253611008
Free: 1411001220 Used: 7993453766
Trashed: 104857602 Free: 1411001220
Other: 8849156022 Trashed: 104857602
Other: 8849156022
` + "```" + `
A ` + "`--json`" + ` flag generates conveniently machine-readable output, e.g. A ` + "`--json`" + ` flag generates conveniently machine-readable output, e.g.
{ ` + "```json" + `
"total": 18253611008, {
"used": 7993453766, "total": 18253611008,
"trashed": 104857602, "used": 7993453766,
"other": 8849156022, "trashed": 104857602,
"free": 1411001220 "other": 8849156022,
} "free": 1411001220
}
` + "```" + `
Not all backends print all fields. Information is not included if it is not Not all backends print all fields. Information is not included if it is not
provided by a backend. Where the value is unlimited it is omitted. provided by a backend. Where the value is unlimited it is omitted.
Some backends does not support the ` + "`rclone about`" + ` command at all, Some backends does not support the ` + "`rclone about`" + ` command at all,
see complete list in [documentation](https://rclone.org/overview/#optional-features). see complete list in [documentation](https://rclone.org/overview/#optional-features).`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.41", "versionIntroduced": "v1.41",
// "groups": "", // "groups": "",

View File

@@ -23,21 +23,23 @@ func init() {
} }
var commandDefinition = &cobra.Command{ var commandDefinition = &cobra.Command{
Use: "authorize <fs name> [base64_json_blob | client_id client_secret]", Use: "authorize <backendname> [base64_json_blob | client_id client_secret]",
Short: `Remote authorization.`, Short: `Remote authorization.`,
Long: `Remote authorization. Used to authorize a remote or headless Long: `Remote authorization. Used to authorize a remote or headless
rclone from a machine with a browser - use as instructed by rclone from a machine with a browser. Use as instructed by rclone config.
rclone config. See also the [remote setup documentation](/remote_setup).
The command requires 1-3 arguments: The command requires 1-3 arguments:
- fs name (e.g., "drive", "s3", etc.)
- Either a base64 encoded JSON blob obtained from a previous rclone config session - Name of a backend (e.g. "drive", "s3")
- Or a client_id and client_secret pair obtained from the remote service - Either a base64 encoded JSON blob obtained from a previous rclone config session
- Or a client_id and client_secret pair obtained from the remote service
Use --auth-no-open-browser to prevent rclone to open auth Use --auth-no-open-browser to prevent rclone to open auth
link in default browser automatically. link in default browser automatically.
Use --template to generate HTML output via a custom Go template. If a blank string is provided as an argument to this flag, the default template is used.`, Use --template to generate HTML output via a custom Go template. If a blank
string is provided as an argument to this flag, the default template is used.`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.27", "versionIntroduced": "v1.27",
}, },

View File

@@ -10,7 +10,7 @@ import (
func TestAuthorizeCommand(t *testing.T) { func TestAuthorizeCommand(t *testing.T) {
// Test that the Use string is correctly formatted // Test that the Use string is correctly formatted
if commandDefinition.Use != "authorize <fs name> [base64_json_blob | client_id client_secret]" { if commandDefinition.Use != "authorize <backendname> [base64_json_blob | client_id client_secret]" {
t.Errorf("Command Use string doesn't match expected format: %s", commandDefinition.Use) t.Errorf("Command Use string doesn't match expected format: %s", commandDefinition.Use)
} }
@@ -26,7 +26,7 @@ func TestAuthorizeCommand(t *testing.T) {
} }
helpOutput := buf.String() helpOutput := buf.String()
if !strings.Contains(helpOutput, "authorize <fs name>") { if !strings.Contains(helpOutput, "authorize <backendname>") {
t.Errorf("Help output doesn't contain correct usage information") t.Errorf("Help output doesn't contain correct usage information")
} }
} }

View File

@@ -37,26 +37,33 @@ see the backend docs for definitions.
You can discover what commands a backend implements by using You can discover what commands a backend implements by using
rclone backend help remote: ` + "```sh" + `
rclone backend help <backendname> rclone backend help remote:
rclone backend help <backendname>
` + "```" + `
You can also discover information about the backend using (see You can also discover information about the backend using (see
[operations/fsinfo](/rc/#operations-fsinfo) in the remote control docs [operations/fsinfo](/rc/#operations-fsinfo) in the remote control docs
for more info). for more info).
rclone backend features remote: ` + "```sh" + `
rclone backend features remote:
` + "```" + `
Pass options to the backend command with -o. This should be key=value or key, e.g.: Pass options to the backend command with -o. This should be key=value or key, e.g.:
rclone backend stats remote:path stats -o format=json -o long ` + "```sh" + `
rclone backend stats remote:path stats -o format=json -o long
` + "```" + `
Pass arguments to the backend by placing them on the end of the line Pass arguments to the backend by placing them on the end of the line
rclone backend cleanup remote:path file1 file2 file3 ` + "```sh" + `
rclone backend cleanup remote:path file1 file2 file3
` + "```" + `
Note to run these commands on a running backend then see Note to run these commands on a running backend then see
[backend/command](/rc/#backend-command) in the rc docs. [backend/command](/rc/#backend-command) in the rc docs.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.52", "versionIntroduced": "v1.52",
"groups": "Important", "groups": "Important",

View File

@@ -4,15 +4,19 @@ package bilib
import ( import (
"bytes" "bytes"
"log/slog" "log/slog"
"sync"
"github.com/rclone/rclone/fs/log" "github.com/rclone/rclone/fs/log"
) )
// CaptureOutput runs a function capturing its output at log level INFO. // CaptureOutput runs a function capturing its output at log level INFO.
func CaptureOutput(fun func()) []byte { func CaptureOutput(fun func()) []byte {
var mu sync.Mutex
buf := &bytes.Buffer{} buf := &bytes.Buffer{}
oldLevel := log.Handler.SetLevel(slog.LevelInfo) oldLevel := log.Handler.SetLevel(slog.LevelInfo)
log.Handler.SetOutput(func(level slog.Level, text string) { log.Handler.SetOutput(func(level slog.Level, text string) {
mu.Lock()
defer mu.Unlock()
buf.WriteString(text) buf.WriteString(text)
}) })
defer func() { defer func() {
@@ -20,5 +24,7 @@ func CaptureOutput(fun func()) []byte {
log.Handler.SetLevel(oldLevel) log.Handler.SetLevel(oldLevel)
}() }()
fun() fun()
mu.Lock()
defer mu.Unlock()
return buf.Bytes() return buf.Bytes()
} }

View File

@@ -51,14 +51,15 @@ var longHelp = shortHelp + makeHelp(`
bidirectional cloud sync solution in rclone. bidirectional cloud sync solution in rclone.
It retains the Path1 and Path2 filesystem listings from the prior run. It retains the Path1 and Path2 filesystem listings from the prior run.
On each successive run it will: On each successive run it will:
- list files on Path1 and Path2, and check for changes on each side. - list files on Path1 and Path2, and check for changes on each side.
Changes include |New|, |Newer|, |Older|, and |Deleted| files. Changes include |New|, |Newer|, |Older|, and |Deleted| files.
- Propagate changes on Path1 to Path2, and vice-versa. - Propagate changes on Path1 to Path2, and vice-versa.
Bisync is considered an **advanced command**, so use with care. Bisync is considered an **advanced command**, so use with care.
Make sure you have read and understood the entire [manual](https://rclone.org/bisync) Make sure you have read and understood the entire [manual](https://rclone.org/bisync)
(especially the [Limitations](https://rclone.org/bisync/#limitations) section) before using, (especially the [Limitations](https://rclone.org/bisync/#limitations) section)
or data loss can result. Questions can be asked in the [Rclone Forum](https://forum.rclone.org/). before using, or data loss can result. Questions can be asked in the
[Rclone Forum](https://forum.rclone.org/).
See [full bisync description](https://rclone.org/bisync/) for details. See [full bisync description](https://rclone.org/bisync/) for details.`)
`)

View File

@@ -434,7 +434,6 @@ func (b *bisyncRun) listDirsOnly(listingNum int) (*fileList, error) {
} }
fulllisting, err = b.loadListingNum(listingNum) fulllisting, err = b.loadListingNum(listingNum)
if err != nil { if err != nil {
b.critical = true b.critical = true
b.retryable = true b.retryable = true
@@ -610,6 +609,11 @@ func (b *bisyncRun) modifyListing(ctx context.Context, src fs.Fs, dst fs.Fs, res
} }
} }
if srcNewName != "" { // if it was renamed and not deleted if srcNewName != "" { // if it was renamed and not deleted
if new == nil { // should not happen. log error and debug info
b.handleErr(b.renames, "internal error", fmt.Errorf("missing info for %q. Please report a bug at https://github.com/rclone/rclone/issues", srcNewName), true, true)
fs.PrettyPrint(srcList, "srcList for debugging", fs.LogLevelNotice)
continue
}
srcList.put(srcNewName, new.size, new.time, new.hash, new.id, new.flags) srcList.put(srcNewName, new.size, new.time, new.hash, new.id, new.flags)
dstList.put(srcNewName, new.size, new.time, new.hash, new.id, new.flags) dstList.put(srcNewName, new.size, new.time, new.hash, new.id, new.flags)
} }

View File

@@ -43,15 +43,21 @@ var commandDefinition = &cobra.Command{
You can use it like this to output a single file You can use it like this to output a single file
rclone cat remote:path/to/file |||sh
rclone cat remote:path/to/file
|||
Or like this to output any file in dir or its subdirectories. Or like this to output any file in dir or its subdirectories.
rclone cat remote:path/to/dir |||sh
rclone cat remote:path/to/dir
|||
Or like this to output any .txt files in dir or its subdirectories. Or like this to output any .txt files in dir or its subdirectories.
rclone --include "*.txt" cat remote:path/to/dir |||sh
rclone --include "*.txt" cat remote:path/to/dir
|||
Use the |--head| flag to print characters only at the start, |--tail| for Use the |--head| flag to print characters only at the start, |--tail| for
the end and |--offset| and |--count| to print a section in the middle. the end and |--offset| and |--count| to print a section in the middle.
@@ -62,14 +68,17 @@ Use the |--separator| flag to print a separator value between files. Be sure to
shell-escape special characters. For example, to print a newline between shell-escape special characters. For example, to print a newline between
files, use: files, use:
* bash: - bash:
rclone --include "*.txt" --separator $'\n' cat remote:path/to/dir |||sh
rclone --include "*.txt" --separator $'\n' cat remote:path/to/dir
|||
* powershell: - powershell:
rclone --include "*.txt" --separator "|n" cat remote:path/to/dir |||powershell
`, "|", "`"), rclone --include "*.txt" --separator "|n" cat remote:path/to/dir
|||`, "|", "`"),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.33", "versionIntroduced": "v1.33",
"groups": "Filter,Listing", "groups": "Filter,Listing",

View File

@@ -74,8 +74,7 @@ you what happened to it. These are reminiscent of diff files.
- |! path| means there was an error reading or hashing the source or dest. - |! path| means there was an error reading or hashing the source or dest.
The default number of parallel checks is 8. See the [--checkers](/docs/#checkers-int) The default number of parallel checks is 8. See the [--checkers](/docs/#checkers-int)
option for more information. option for more information.`, "|", "`")
`, "|", "`")
// GetCheckOpt gets the options corresponding to the check flags // GetCheckOpt gets the options corresponding to the check flags
func GetCheckOpt(fsrc, fdst fs.Fs) (opt *operations.CheckOpt, close func(), err error) { func GetCheckOpt(fsrc, fdst fs.Fs) (opt *operations.CheckOpt, close func(), err error) {

View File

@@ -17,8 +17,7 @@ var commandDefinition = &cobra.Command{
Use: "cleanup remote:path", Use: "cleanup remote:path",
Short: `Clean up the remote if possible.`, Short: `Clean up the remote if possible.`,
Long: `Clean up the remote if possible. Empty the trash or delete old file Long: `Clean up the remote if possible. Empty the trash or delete old file
versions. Not supported by all remotes. versions. Not supported by all remotes.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.31", "versionIntroduced": "v1.31",
"groups": "Important", "groups": "Important",

View File

@@ -44,8 +44,7 @@ var configCommand = &cobra.Command{
Short: `Enter an interactive configuration session.`, Short: `Enter an interactive configuration session.`,
Long: `Enter an interactive configuration session where you can setup new Long: `Enter an interactive configuration session where you can setup new
remotes and manage existing ones. You may also set or remove a remotes and manage existing ones. You may also set or remove a
password to protect your configuration. password to protect your configuration.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.39", "versionIntroduced": "v1.39",
}, },
@@ -134,9 +133,7 @@ sensitive info with XXX.
This makes the config file suitable for posting online for support. This makes the config file suitable for posting online for support.
It should be double checked before posting as the redaction may not be perfect. It should be double checked before posting as the redaction may not be perfect.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.64", "versionIntroduced": "v1.64",
}, },
@@ -178,8 +175,8 @@ var configProvidersCommand = &cobra.Command{
var updateRemoteOpt config.UpdateRemoteOpt var updateRemoteOpt config.UpdateRemoteOpt
var configPasswordHelp = strings.ReplaceAll(` var configPasswordHelp = strings.ReplaceAll(
Note that if the config process would normally ask a question the `Note that if the config process would normally ask a question the
default is taken (unless |--non-interactive| is used). Each time default is taken (unless |--non-interactive| is used). Each time
that happens rclone will print or DEBUG a message saying how to that happens rclone will print or DEBUG a message saying how to
affect the value taken. affect the value taken.
@@ -205,29 +202,29 @@ it.
This will look something like (some irrelevant detail removed): This will look something like (some irrelevant detail removed):
||| |||json
{ {
"State": "*oauth-islocal,teamdrive,,", "State": "*oauth-islocal,teamdrive,,",
"Option": { "Option": {
"Name": "config_is_local", "Name": "config_is_local",
"Help": "Use web browser to automatically authenticate rclone with remote?\n * Say Y if the machine running rclone has a web browser you can use\n * Say N if running rclone on a (remote) machine without web browser access\nIf not sure try Y. If Y failed, try N.\n", "Help": "Use web browser to automatically authenticate rclone with remote?\n * Say Y if the machine running rclone has a web browser you can use\n * Say N if running rclone on a (remote) machine without web browser access\nIf not sure try Y. If Y failed, try N.\n",
"Default": true, "Default": true,
"Examples": [ "Examples": [
{ {
"Value": "true", "Value": "true",
"Help": "Yes" "Help": "Yes"
}, },
{ {
"Value": "false", "Value": "false",
"Help": "No" "Help": "No"
} }
], ],
"Required": false, "Required": false,
"IsPassword": false, "IsPassword": false,
"Type": "bool", "Type": "bool",
"Exclusive": true, "Exclusive": true,
}, },
"Error": "", "Error": "",
} }
||| |||
@@ -250,7 +247,9 @@ The keys of |Option| are used as follows:
If |Error| is set then it should be shown to the user at the same If |Error| is set then it should be shown to the user at the same
time as the question. time as the question.
rclone config update name --continue --state "*oauth-islocal,teamdrive,," --result "true" |||sh
rclone config update name --continue --state "*oauth-islocal,teamdrive,," --result "true"
|||
Note that when using |--continue| all passwords should be passed in Note that when using |--continue| all passwords should be passed in
the clear (not obscured). Any default config values should be passed the clear (not obscured). Any default config values should be passed
@@ -264,8 +263,7 @@ not just the post config questions. Any parameters are used as
defaults for questions as usual. defaults for questions as usual.
Note that |bin/config.py| in the rclone source implements this protocol Note that |bin/config.py| in the rclone source implements this protocol
as a readable demonstration. as a readable demonstration.`, "|", "`")
`, "|", "`")
var configCreateCommand = &cobra.Command{ var configCreateCommand = &cobra.Command{
Use: "create name type [key value]*", Use: "create name type [key value]*",
Short: `Create a new remote with name, type and options.`, Short: `Create a new remote with name, type and options.`,
@@ -275,13 +273,18 @@ should be passed in pairs of |key| |value| or as |key=value|.
For example, to make a swift remote of name myremote using auto config For example, to make a swift remote of name myremote using auto config
you would do: you would do:
rclone config create myremote swift env_auth true |||sh
rclone config create myremote swift env_auth=true rclone config create myremote swift env_auth true
rclone config create myremote swift env_auth=true
|||
So for example if you wanted to configure a Google Drive remote but So for example if you wanted to configure a Google Drive remote but
using remote authorization you would do this: using remote authorization you would do this:
rclone config create mydrive drive config_is_local=false |||sh
rclone config create mydrive drive config_is_local=false
|||
`, "|", "`") + configPasswordHelp, `, "|", "`") + configPasswordHelp,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.39", "versionIntroduced": "v1.39",
@@ -344,13 +347,18 @@ pairs of |key| |value| or as |key=value|.
For example, to update the env_auth field of a remote of name myremote For example, to update the env_auth field of a remote of name myremote
you would do: you would do:
rclone config update myremote env_auth true |||sh
rclone config update myremote env_auth=true rclone config update myremote env_auth true
rclone config update myremote env_auth=true
|||
If the remote uses OAuth the token will be updated, if you don't If the remote uses OAuth the token will be updated, if you don't
require this add an extra parameter thus: require this add an extra parameter thus:
rclone config update myremote env_auth=true config_refresh_token=false |||sh
rclone config update myremote env_auth=true config_refresh_token=false
|||
`, "|", "`") + configPasswordHelp, `, "|", "`") + configPasswordHelp,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.39", "versionIntroduced": "v1.39",
@@ -388,12 +396,13 @@ The |password| should be passed in in clear (unobscured).
For example, to set password of a remote of name myremote you would do: For example, to set password of a remote of name myremote you would do:
rclone config password myremote fieldname mypassword |||sh
rclone config password myremote fieldname=mypassword rclone config password myremote fieldname mypassword
rclone config password myremote fieldname=mypassword
|||
This command is obsolete now that "config update" and "config create" This command is obsolete now that "config update" and "config create"
both support obscuring passwords directly. both support obscuring passwords directly.`, "|", "`"),
`, "|", "`"),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.39", "versionIntroduced": "v1.39",
}, },
@@ -441,8 +450,7 @@ var configReconnectCommand = &cobra.Command{
To disconnect the remote use "rclone config disconnect". To disconnect the remote use "rclone config disconnect".
This normally means going through the interactive oauth flow again. This normally means going through the interactive oauth flow again.`,
`,
RunE: func(command *cobra.Command, args []string) error { RunE: func(command *cobra.Command, args []string) error {
ctx := context.Background() ctx := context.Background()
cmd.CheckArgs(1, 1, command, args) cmd.CheckArgs(1, 1, command, args)
@@ -461,8 +469,7 @@ var configDisconnectCommand = &cobra.Command{
This normally means revoking the oauth token. This normally means revoking the oauth token.
To reconnect use "rclone config reconnect". To reconnect use "rclone config reconnect".`,
`,
RunE: func(command *cobra.Command, args []string) error { RunE: func(command *cobra.Command, args []string) error {
cmd.CheckArgs(1, 1, command, args) cmd.CheckArgs(1, 1, command, args)
f := cmd.NewFsSrc(args) f := cmd.NewFsSrc(args)
@@ -490,8 +497,7 @@ var configUserInfoCommand = &cobra.Command{
Use: "userinfo remote:", Use: "userinfo remote:",
Short: `Prints info about logged in user of remote.`, Short: `Prints info about logged in user of remote.`,
Long: `This prints the details of the person logged in to the cloud storage Long: `This prints the details of the person logged in to the cloud storage
system. system.`,
`,
RunE: func(command *cobra.Command, args []string) error { RunE: func(command *cobra.Command, args []string) error {
cmd.CheckArgs(1, 1, command, args) cmd.CheckArgs(1, 1, command, args)
f := cmd.NewFsSrc(args) f := cmd.NewFsSrc(args)
@@ -534,8 +540,7 @@ var configEncryptionCommand = &cobra.Command{
Use: "encryption", Use: "encryption",
Short: `set, remove and check the encryption for the config file`, Short: `set, remove and check the encryption for the config file`,
Long: `This command sets, clears and checks the encryption for the config file using Long: `This command sets, clears and checks the encryption for the config file using
the subcommands below. the subcommands below.`,
`,
} }
var configEncryptionSetCommand = &cobra.Command{ var configEncryptionSetCommand = &cobra.Command{
@@ -559,8 +564,7 @@ variable to distinguish which password you must supply.
Alternatively you can remove the password first (with |rclone config Alternatively you can remove the password first (with |rclone config
encryption remove|), then set it again with this command which may be encryption remove|), then set it again with this command which may be
easier if you don't mind the unencrypted config file being on the disk easier if you don't mind the unencrypted config file being on the disk
briefly. briefly.`, "|", "`"),
`, "|", "`"),
RunE: func(command *cobra.Command, args []string) error { RunE: func(command *cobra.Command, args []string) error {
cmd.CheckArgs(0, 0, command, args) cmd.CheckArgs(0, 0, command, args)
config.LoadedData() config.LoadedData()
@@ -580,8 +584,7 @@ If |--password-command| is in use, this will be called to supply the old config
password. password.
If the config was not encrypted then no error will be returned and If the config was not encrypted then no error will be returned and
this command will do nothing. this command will do nothing.`, "|", "`"),
`, "|", "`"),
RunE: func(command *cobra.Command, args []string) error { RunE: func(command *cobra.Command, args []string) error {
cmd.CheckArgs(0, 0, command, args) cmd.CheckArgs(0, 0, command, args)
config.LoadedData() config.LoadedData()
@@ -600,8 +603,7 @@ It will attempt to decrypt the config using the password you supply.
If decryption fails it will return a non-zero exit code if using If decryption fails it will return a non-zero exit code if using
|--password-command|, otherwise it will prompt again for the password. |--password-command|, otherwise it will prompt again for the password.
If the config file is not encrypted it will return a non zero exit code. If the config file is not encrypted it will return a non zero exit code.`, "|", "`"),
`, "|", "`"),
RunE: func(command *cobra.Command, args []string) error { RunE: func(command *cobra.Command, args []string) error {
cmd.CheckArgs(0, 0, command, args) cmd.CheckArgs(0, 0, command, args)
config.LoadedData() config.LoadedData()

View File

@@ -31,18 +31,27 @@ var commandDefinition = &cobra.Command{
Use: "convmv dest:path --name-transform XXX", Use: "convmv dest:path --name-transform XXX",
Short: `Convert file and directory names in place.`, Short: `Convert file and directory names in place.`,
// Warning¡ "¡" will be replaced by backticks below // Warning¡ "¡" will be replaced by backticks below
Long: strings.ReplaceAll(` Long: strings.ReplaceAll(`convmv supports advanced path name transformations for converting and renaming
convmv supports advanced path name transformations for converting and renaming files and directories by applying prefixes, suffixes, and other alterations. files and directories by applying prefixes, suffixes, and other alterations.
`+transform.Help()+`Multiple transformations can be used in sequence, applied in the order they are specified on the command line. `+transform.Help()+`The regex command generally accepts Perl-style regular expressions, the exact
syntax is defined in the [Go regular expression reference](https://golang.org/pkg/regexp/syntax/).
The replacement string may contain capturing group variables, referencing
capturing groups using the syntax ¡$name¡ or ¡${name}¡, where the name can
refer to a named capturing group or it can simply be the index as a number.
To insert a literal $, use $$.
Multiple transformations can be used in sequence, applied
in the order they are specified on the command line.
The ¡--name-transform¡ flag is also available in ¡sync¡, ¡copy¡, and ¡move¡. The ¡--name-transform¡ flag is also available in ¡sync¡, ¡copy¡, and ¡move¡.
## Files vs Directories ### Files vs Directories
By default ¡--name-transform¡ will only apply to file names. The means only the leaf file name will be transformed. By default ¡--name-transform¡ will only apply to file names. The means only the
However some of the transforms would be better applied to the whole path or just directories. leaf file name will be transformed. However some of the transforms would be
To choose which which part of the file path is affected some tags can be added to the ¡--name-transform¡. better applied to the whole path or just directories. To choose which which
part of the file path is affected some tags can be added to the ¡--name-transform¡.
| Tag | Effect | | Tag | Effect |
|------|------| |------|------|
@@ -50,42 +59,58 @@ To choose which which part of the file path is affected some tags can be added t
| ¡dir¡ | Only transform name of directories - these may appear anywhere in the path | | ¡dir¡ | Only transform name of directories - these may appear anywhere in the path |
| ¡all¡ | Transform the entire path for files and directories | | ¡all¡ | Transform the entire path for files and directories |
This is used by adding the tag into the transform name like this: ¡--name-transform file,prefix=ABC¡ or ¡--name-transform dir,prefix=DEF¡. This is used by adding the tag into the transform name like this:
¡--name-transform file,prefix=ABC¡ or ¡--name-transform dir,prefix=DEF¡.
For some conversions using all is more likely to be useful, for example ¡--name-transform all,nfc¡. For some conversions using all is more likely to be useful, for example
¡--name-transform all,nfc¡.
Note that ¡--name-transform¡ may not add path separators ¡/¡ to the name. This will cause an error. Note that ¡--name-transform¡ may not add path separators ¡/¡ to the name.
This will cause an error.
## Ordering and Conflicts ### Ordering and Conflicts
* Transformations will be applied in the order specified by the user. - Transformations will be applied in the order specified by the user.
* If the ¡file¡ tag is in use (the default) then only the leaf name of files will be transformed. - If the ¡file¡ tag is in use (the default) then only the leaf name of files
* If the ¡dir¡ tag is in use then directories anywhere in the path will be transformed will be transformed.
* If the ¡all¡ tag is in use then directories and files anywhere in the path will be transformed - If the ¡dir¡ tag is in use then directories anywhere in the path will be
* Each transformation will be run one path segment at a time. transformed
* If a transformation adds a ¡/¡ or ends up with an empty path segment then that will be an error. - If the ¡all¡ tag is in use then directories and files anywhere in the path
* It is up to the user to put the transformations in a sensible order. will be transformed
* Conflicting transformations, such as ¡prefix¡ followed by ¡trimprefix¡ or ¡nfc¡ followed by ¡nfd¡, are possible. - Each transformation will be run one path segment at a time.
* Instead of enforcing mutual exclusivity, transformations are applied in sequence as specified by the - If a transformation adds a ¡/¡ or ends up with an empty path segment then
user, allowing for intentional use cases (e.g., trimming one prefix before adding another). that will be an error.
* Users should be aware that certain combinations may lead to unexpected results and should verify - It is up to the user to put the transformations in a sensible order.
transformations using ¡--dry-run¡ before execution. - Conflicting transformations, such as ¡prefix¡ followed by ¡trimprefix¡ or
¡nfc¡ followed by ¡nfd¡, are possible.
- Instead of enforcing mutual exclusivity, transformations are applied in
sequence as specified by the user, allowing for intentional use cases
(e.g., trimming one prefix before adding another).
- Users should be aware that certain combinations may lead to unexpected
results and should verify transformations using ¡--dry-run¡ before execution.
## Race Conditions and Non-Deterministic Behavior ### Race Conditions and Non-Deterministic Behavior
Some transformations, such as ¡replace=old:new¡, may introduce conflicts where multiple source files map to the same destination name. Some transformations, such as ¡replace=old:new¡, may introduce conflicts where
This can lead to race conditions when performing concurrent transfers. It is up to the user to anticipate these. multiple source files map to the same destination name. This can lead to race
* If two files from the source are transformed into the same name at the destination, the final state may be non-deterministic. conditions when performing concurrent transfers. It is up to the user to
* Running rclone check after a sync using such transformations may erroneously report missing or differing files due to overwritten results. anticipate these.
- If two files from the source are transformed into the same name at the
destination, the final state may be non-deterministic.
- Running rclone check after a sync using such transformations may erroneously
report missing or differing files due to overwritten results.
To minimize risks, users should: To minimize risks, users should:
* Carefully review transformations that may introduce conflicts.
* Use ¡--dry-run¡ to inspect changes before executing a sync (but keep in mind that it won't show the effect of non-deterministic transformations).
* Avoid transformations that cause multiple distinct source files to map to the same destination name.
* Consider disabling concurrency with ¡--transfers=1¡ if necessary.
* Certain transformations (e.g. ¡prefix¡) will have a multiplying effect every time they are used. Avoid these when using ¡bisync¡.
`, "¡", "`"), - Carefully review transformations that may introduce conflicts.
- Use ¡--dry-run¡ to inspect changes before executing a sync (but keep in mind
that it won't show the effect of non-deterministic transformations).
- Avoid transformations that cause multiple distinct source files to map to the
same destination name.
- Consider disabling concurrency with ¡--transfers=1¡ if necessary.
- Certain transformations (e.g. ¡prefix¡) will have a multiplying effect every
time they are used. Avoid these when using ¡bisync¡.`, "¡", "`"),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.70", "versionIntroduced": "v1.70",
"groups": "Filter,Listing,Important,Copy", "groups": "Filter,Listing,Important,Copy",

View File

@@ -50,22 +50,30 @@ go there.
For example For example
rclone copy source:sourcepath dest:destpath |||sh
rclone copy source:sourcepath dest:destpath
|||
Let's say there are two files in sourcepath Let's say there are two files in sourcepath
sourcepath/one.txt |||text
sourcepath/two.txt sourcepath/one.txt
sourcepath/two.txt
|||
This copies them to This copies them to
destpath/one.txt |||text
destpath/two.txt destpath/one.txt
destpath/two.txt
|||
Not to Not to
destpath/sourcepath/one.txt |||text
destpath/sourcepath/two.txt destpath/sourcepath/one.txt
destpath/sourcepath/two.txt
|||
If you are familiar with |rsync|, rclone always works as if you had If you are familiar with |rsync|, rclone always works as if you had
written a trailing |/| - meaning "copy the contents of this directory". written a trailing |/| - meaning "copy the contents of this directory".
@@ -81,20 +89,22 @@ For example, if you have many files in /path/to/src but only a few of
them change every day, you can copy all the files which have changed them change every day, you can copy all the files which have changed
recently very efficiently like this: recently very efficiently like this:
rclone copy --max-age 24h --no-traverse /path/to/src remote: |||sh
rclone copy --max-age 24h --no-traverse /path/to/src remote:
|||
Rclone will sync the modification times of files and directories if Rclone will sync the modification times of files and directories if
the backend supports it. If metadata syncing is required then use the the backend supports it. If metadata syncing is required then use the
|--metadata| flag. |--metadata| flag.
Note that the modification time and metadata for the root directory Note that the modification time and metadata for the root directory
will **not** be synced. See https://github.com/rclone/rclone/issues/7652 will **not** be synced. See [issue #7652](https://github.com/rclone/rclone/issues/7652)
for more info. for more info.
**Note**: Use the |-P|/|--progress| flag to view real-time transfer statistics. **Note**: Use the |-P|/|--progress| flag to view real-time transfer statistics.
**Note**: Use the |--dry-run| or the |--interactive|/|-i| flag to test without copying anything. **Note**: Use the |--dry-run| or the |--interactive|/|-i| flag to test without
copying anything.
`, "|", "`") + operationsflags.Help(), `, "|", "`") + operationsflags.Help(),
Annotations: map[string]string{ Annotations: map[string]string{

View File

@@ -35,26 +35,32 @@ name. If the source is a directory then it acts exactly like the
So So
rclone copyto src dst ` + "```sh" + `
rclone copyto src dst
` + "```" + `
where src and dst are rclone paths, either remote:path or where src and dst are rclone paths, either ` + "`remote:path`" + ` or
/path/to/local or C:\windows\path\if\on\windows. ` + "`/path/to/local`" + ` or ` + "`C:\\windows\\path\\if\\on\\windows`" + `.
This will: This will:
if src is file ` + "```text" + `
copy it to dst, overwriting an existing file if it exists if src is file
if src is directory copy it to dst, overwriting an existing file if it exists
copy it to dst, overwriting existing files if they exist if src is directory
see copy command for full details copy it to dst, overwriting existing files if they exist
see copy command for full details
` + "```" + `
This doesn't transfer files that are identical on src and dst, testing This doesn't transfer files that are identical on src and dst, testing
by size and modification time or MD5SUM. It doesn't delete files from by size and modification time or MD5SUM. It doesn't delete files from
the destination. the destination.
*If you are looking to copy just a byte range of a file, please see 'rclone cat --offset X --count Y'* *If you are looking to copy just a byte range of a file, please see
` + "`rclone cat --offset X --count Y`" + `.*
**Note**: Use the ` + "`-P`" + `/` + "`--progress`" + ` flag to view real-time transfer statistics **Note**: Use the ` + "`-P`" + `/` + "`--progress`" + ` flag to view
real-time transfer statistics.
` + operationsflags.Help(), ` + operationsflags.Help(),
Annotations: map[string]string{ Annotations: map[string]string{

View File

@@ -48,7 +48,7 @@ set in HTTP headers, it will be used instead of the name from the URL.
With |--print-filename| in addition, the resulting file name will be With |--print-filename| in addition, the resulting file name will be
printed. printed.
Setting |--no-clobber| will prevent overwriting file on the Setting |--no-clobber| will prevent overwriting file on the
destination if there is one with the same name. destination if there is one with the same name.
Setting |--stdout| or making the output file name |-| Setting |--stdout| or making the output file name |-|
@@ -62,9 +62,7 @@ If you can't get |rclone copyurl| to work then here are some things you can try:
- |--bind 0.0.0.0| rclone will use IPv6 if available - try disabling it - |--bind 0.0.0.0| rclone will use IPv6 if available - try disabling it
- |--bind ::0| to disable IPv4 - |--bind ::0| to disable IPv4
- |--user agent curl| - some sites have whitelists for curl's user-agent - try that - |--user agent curl| - some sites have whitelists for curl's user-agent - try that
- Make sure the site works with |curl| directly - Make sure the site works with |curl| directly`, "|", "`"),
`, "|", "`"),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.43", "versionIntroduced": "v1.43",
"groups": "Important", "groups": "Important",

View File

@@ -37,14 +37,18 @@ checksum of the file it has just encrypted.
Use it like this Use it like this
rclone cryptcheck /path/to/files encryptedremote:path ` + "```sh" + `
rclone cryptcheck /path/to/files encryptedremote:path
` + "```" + `
You can use it like this also, but that will involve downloading all You can use it like this also, but that will involve downloading all
the files in remote:path. the files in ` + "`remote:path`" + `.
rclone cryptcheck remote:path encryptedremote:path ` + "```sh" + `
rclone cryptcheck remote:path encryptedremote:path
` + "```" + `
After it has run it will log the status of the encryptedremote:. After it has run it will log the status of the ` + "`encryptedremote:`" + `.
` + check.FlagsHelp, ` + check.FlagsHelp,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.36", "versionIntroduced": "v1.36",

View File

@@ -33,13 +33,13 @@ If you supply the ` + "`--reverse`" + ` flag, it will return encrypted file name
use it like this use it like this
rclone cryptdecode encryptedremote: encryptedfilename1 encryptedfilename2 ` + "```sh" + `
rclone cryptdecode encryptedremote: encryptedfilename1 encryptedfilename2
rclone cryptdecode --reverse encryptedremote: filename1 filename2
` + "```" + `
rclone cryptdecode --reverse encryptedremote: filename1 filename2 Another way to accomplish this is by using the ` + "`rclone backend encode` (or `decode`)" + `
command. See the documentation on the [crypt](/crypt/) overlay for more info.`,
Another way to accomplish this is by using the ` + "`rclone backend encode` (or `decode`)" + ` command.
See the documentation on the [crypt](/crypt/) overlay for more info.
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.38", "versionIntroduced": "v1.38",
}, },

View File

@@ -47,15 +47,15 @@ directories have been merged.
Next, if deduping by name, for every group of duplicate file names / Next, if deduping by name, for every group of duplicate file names /
hashes, it will delete all but one identical file it finds without hashes, it will delete all but one identical file it finds without
confirmation. This means that for most duplicated files the ` + confirmation. This means that for most duplicated files the
"`dedupe`" + ` command will not be interactive. ` + "`dedupe`" + ` command will not be interactive.
` + "`dedupe`" + ` considers files to be identical if they have the ` + "`dedupe`" + ` considers files to be identical if they have the
same file path and the same hash. If the backend does not support hashes (e.g. crypt wrapping same file path and the same hash. If the backend does not support
Google Drive) then they will never be found to be identical. If you hashes (e.g. crypt wrapping Google Drive) then they will never be found
use the ` + "`--size-only`" + ` flag then files will be considered to be identical. If you use the ` + "`--size-only`" + ` flag then files
identical if they have the same size (any hash will be ignored). This will be considered identical if they have the same size (any hash will be
can be useful on crypt backends which do not support hashes. ignored). This can be useful on crypt backends which do not support hashes.
Next rclone will resolve the remaining duplicates. Exactly which Next rclone will resolve the remaining duplicates. Exactly which
action is taken depends on the dedupe mode. By default, rclone will action is taken depends on the dedupe mode. By default, rclone will
@@ -68,71 +68,82 @@ Here is an example run.
Before - with duplicates Before - with duplicates
$ rclone lsl drive:dupes ` + "```sh" + `
6048320 2016-03-05 16:23:16.798000000 one.txt $ rclone lsl drive:dupes
6048320 2016-03-05 16:23:11.775000000 one.txt 6048320 2016-03-05 16:23:16.798000000 one.txt
564374 2016-03-05 16:23:06.731000000 one.txt 6048320 2016-03-05 16:23:11.775000000 one.txt
6048320 2016-03-05 16:18:26.092000000 one.txt 564374 2016-03-05 16:23:06.731000000 one.txt
6048320 2016-03-05 16:22:46.185000000 two.txt 6048320 2016-03-05 16:18:26.092000000 one.txt
1744073 2016-03-05 16:22:38.104000000 two.txt 6048320 2016-03-05 16:22:46.185000000 two.txt
564374 2016-03-05 16:22:52.118000000 two.txt 1744073 2016-03-05 16:22:38.104000000 two.txt
564374 2016-03-05 16:22:52.118000000 two.txt
` + "```" + `
Now the ` + "`dedupe`" + ` session Now the ` + "`dedupe`" + ` session
$ rclone dedupe drive:dupes ` + "```sh" + `
2016/03/05 16:24:37 Google drive root 'dupes': Looking for duplicates using interactive mode. $ rclone dedupe drive:dupes
one.txt: Found 4 files with duplicate names 2016/03/05 16:24:37 Google drive root 'dupes': Looking for duplicates using interactive mode.
one.txt: Deleting 2/3 identical duplicates (MD5 "1eedaa9fe86fd4b8632e2ac549403b36") one.txt: Found 4 files with duplicate names
one.txt: 2 duplicates remain one.txt: Deleting 2/3 identical duplicates (MD5 "1eedaa9fe86fd4b8632e2ac549403b36")
1: 6048320 bytes, 2016-03-05 16:23:16.798000000, MD5 1eedaa9fe86fd4b8632e2ac549403b36 one.txt: 2 duplicates remain
2: 564374 bytes, 2016-03-05 16:23:06.731000000, MD5 7594e7dc9fc28f727c42ee3e0749de81 1: 6048320 bytes, 2016-03-05 16:23:16.798000000, MD5 1eedaa9fe86fd4b8632e2ac549403b36
s) Skip and do nothing 2: 564374 bytes, 2016-03-05 16:23:06.731000000, MD5 7594e7dc9fc28f727c42ee3e0749de81
k) Keep just one (choose which in next step) s) Skip and do nothing
r) Rename all to be different (by changing file.jpg to file-1.jpg) k) Keep just one (choose which in next step)
s/k/r> k r) Rename all to be different (by changing file.jpg to file-1.jpg)
Enter the number of the file to keep> 1 s/k/r> k
one.txt: Deleted 1 extra copies Enter the number of the file to keep> 1
two.txt: Found 3 files with duplicate names one.txt: Deleted 1 extra copies
two.txt: 3 duplicates remain two.txt: Found 3 files with duplicate names
1: 564374 bytes, 2016-03-05 16:22:52.118000000, MD5 7594e7dc9fc28f727c42ee3e0749de81 two.txt: 3 duplicates remain
2: 6048320 bytes, 2016-03-05 16:22:46.185000000, MD5 1eedaa9fe86fd4b8632e2ac549403b36 1: 564374 bytes, 2016-03-05 16:22:52.118000000, MD5 7594e7dc9fc28f727c42ee3e0749de81
3: 1744073 bytes, 2016-03-05 16:22:38.104000000, MD5 851957f7fb6f0bc4ce76be966d336802 2: 6048320 bytes, 2016-03-05 16:22:46.185000000, MD5 1eedaa9fe86fd4b8632e2ac549403b36
s) Skip and do nothing 3: 1744073 bytes, 2016-03-05 16:22:38.104000000, MD5 851957f7fb6f0bc4ce76be966d336802
k) Keep just one (choose which in next step) s) Skip and do nothing
r) Rename all to be different (by changing file.jpg to file-1.jpg) k) Keep just one (choose which in next step)
s/k/r> r r) Rename all to be different (by changing file.jpg to file-1.jpg)
two-1.txt: renamed from: two.txt s/k/r> r
two-2.txt: renamed from: two.txt two-1.txt: renamed from: two.txt
two-3.txt: renamed from: two.txt two-2.txt: renamed from: two.txt
two-3.txt: renamed from: two.txt
` + "```" + `
The result being The result being
$ rclone lsl drive:dupes ` + "```sh" + `
6048320 2016-03-05 16:23:16.798000000 one.txt $ rclone lsl drive:dupes
564374 2016-03-05 16:22:52.118000000 two-1.txt 6048320 2016-03-05 16:23:16.798000000 one.txt
6048320 2016-03-05 16:22:46.185000000 two-2.txt 564374 2016-03-05 16:22:52.118000000 two-1.txt
1744073 2016-03-05 16:22:38.104000000 two-3.txt 6048320 2016-03-05 16:22:46.185000000 two-2.txt
1744073 2016-03-05 16:22:38.104000000 two-3.txt
` + "```" + `
Dedupe can be run non interactively using the ` + "`" + `--dedupe-mode` + "`" + ` flag or by using an extra parameter with the same value Dedupe can be run non interactively using the ` + "`" + `--dedupe-mode` + "`" + ` flag
or by using an extra parameter with the same value
* ` + "`" + `--dedupe-mode interactive` + "`" + ` - interactive as above. - ` + "`" + `--dedupe-mode interactive` + "`" + ` - interactive as above.
* ` + "`" + `--dedupe-mode skip` + "`" + ` - removes identical files then skips anything left. - ` + "`" + `--dedupe-mode skip` + "`" + ` - removes identical files then skips anything left.
* ` + "`" + `--dedupe-mode first` + "`" + ` - removes identical files then keeps the first one. - ` + "`" + `--dedupe-mode first` + "`" + ` - removes identical files then keeps the first one.
* ` + "`" + `--dedupe-mode newest` + "`" + ` - removes identical files then keeps the newest one. - ` + "`" + `--dedupe-mode newest` + "`" + ` - removes identical files then keeps the newest one.
* ` + "`" + `--dedupe-mode oldest` + "`" + ` - removes identical files then keeps the oldest one. - ` + "`" + `--dedupe-mode oldest` + "`" + ` - removes identical files then keeps the oldest one.
* ` + "`" + `--dedupe-mode largest` + "`" + ` - removes identical files then keeps the largest one. - ` + "`" + `--dedupe-mode largest` + "`" + ` - removes identical files then keeps the largest one.
* ` + "`" + `--dedupe-mode smallest` + "`" + ` - removes identical files then keeps the smallest one. - ` + "`" + `--dedupe-mode smallest` + "`" + ` - removes identical files then keeps the smallest one.
* ` + "`" + `--dedupe-mode rename` + "`" + ` - removes identical files then renames the rest to be different. - ` + "`" + `--dedupe-mode rename` + "`" + ` - removes identical files then renames the rest to be different.
* ` + "`" + `--dedupe-mode list` + "`" + ` - lists duplicate dirs and files only and changes nothing. - ` + "`" + `--dedupe-mode list` + "`" + ` - lists duplicate dirs and files only and changes nothing.
For example, to rename all the identically named photos in your Google Photos directory, do For example, to rename all the identically named photos in your Google Photos
directory, do
rclone dedupe --dedupe-mode rename "drive:Google Photos" ` + "```sh" + `
rclone dedupe --dedupe-mode rename "drive:Google Photos"
` + "```" + `
Or Or
rclone dedupe rename "drive:Google Photos" ` + "```sh" + `
`, rclone dedupe rename "drive:Google Photos"
` + "```",
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.27", "versionIntroduced": "v1.27",
"groups": "Important", "groups": "Important",

View File

@@ -32,26 +32,29 @@ obeys include/exclude filters so can be used to selectively delete files.
alone. If you want to delete a directory and all of its contents use alone. If you want to delete a directory and all of its contents use
the [purge](/commands/rclone_purge/) command. the [purge](/commands/rclone_purge/) command.
If you supply the |--rmdirs| flag, it will remove all empty directories along with it. If you supply the |--rmdirs| flag, it will remove all empty directories along
You can also use the separate command [rmdir](/commands/rclone_rmdir/) or with it. You can also use the separate command [rmdir](/commands/rclone_rmdir/)
[rmdirs](/commands/rclone_rmdirs/) to delete empty directories only. or [rmdirs](/commands/rclone_rmdirs/) to delete empty directories only.
For example, to delete all files bigger than 100 MiB, you may first want to For example, to delete all files bigger than 100 MiB, you may first want to
check what would be deleted (use either): check what would be deleted (use either):
rclone --min-size 100M lsl remote:path |||sh
rclone --dry-run --min-size 100M delete remote:path rclone --min-size 100M lsl remote:path
rclone --dry-run --min-size 100M delete remote:path
|||
Then proceed with the actual delete: Then proceed with the actual delete:
rclone --min-size 100M delete remote:path |||sh
rclone --min-size 100M delete remote:path
|||
That reads "delete everything with a minimum size of 100 MiB", hence That reads "delete everything with a minimum size of 100 MiB", hence
delete all files bigger than 100 MiB. delete all files bigger than 100 MiB.
**Important**: Since this can cause data loss, test first with the **Important**: Since this can cause data loss, test first with the
|--dry-run| or the |--interactive|/|-i| flag. |--dry-run| or the |--interactive|/|-i| flag.`, "|", "`"),
`, "|", "`"),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.27", "versionIntroduced": "v1.27",
"groups": "Important,Filter,Listing", "groups": "Important,Filter,Listing",

View File

@@ -19,9 +19,8 @@ var commandDefinition = &cobra.Command{
Use: "deletefile remote:path", Use: "deletefile remote:path",
Short: `Remove a single file from remote.`, Short: `Remove a single file from remote.`,
Long: `Remove a single file from remote. Unlike ` + "`" + `delete` + "`" + ` it cannot be used to Long: `Remove a single file from remote. Unlike ` + "`" + `delete` + "`" + ` it cannot be used to
remove a directory and it doesn't obey include/exclude filters - if the specified file exists, remove a directory and it doesn't obey include/exclude filters - if the
it will always be removed. specified file exists, it will always be removed.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.42", "versionIntroduced": "v1.42",
"groups": "Important", "groups": "Important",

View File

@@ -14,8 +14,7 @@ var completionDefinition = &cobra.Command{
Use: "completion [shell]", Use: "completion [shell]",
Short: `Output completion script for a given shell.`, Short: `Output completion script for a given shell.`,
Long: `Generates a shell completion script for rclone. Long: `Generates a shell completion script for rclone.
Run with ` + "`--help`" + ` to list the supported shells. Run with ` + "`--help`" + ` to list the supported shells.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.33", "versionIntroduced": "v1.33",
}, },

View File

@@ -18,17 +18,21 @@ var bashCommandDefinition = &cobra.Command{
Short: `Output bash completion script for rclone.`, Short: `Output bash completion script for rclone.`,
Long: `Generates a bash shell autocompletion script for rclone. Long: `Generates a bash shell autocompletion script for rclone.
By default, when run without any arguments, By default, when run without any arguments,
rclone completion bash ` + "```sh" + `
rclone completion bash
` + "```" + `
the generated script will be written to the generated script will be written to
/etc/bash_completion.d/rclone ` + "```sh" + `
/etc/bash_completion.d/rclone
` + "```" + `
and so rclone will probably need to be run as root, or with sudo. and so rclone will probably need to be run as root, or with sudo.
If you supply a path to a file as the command line argument, then If you supply a path to a file as the command line argument, then
the generated script will be written to that file, in which case the generated script will be written to that file, in which case
you should not need root privileges. you should not need root privileges.
@@ -39,11 +43,12 @@ can logout and login again to use the autocompletion script.
Alternatively, you can source the script directly Alternatively, you can source the script directly
. /path/to/my_bash_completion_scripts/rclone ` + "```sh" + `
. /path/to/my_bash_completion_scripts/rclone
` + "```" + `
and the autocompletion functionality will be added to your and the autocompletion functionality will be added to your
current shell. current shell.`,
`,
Run: func(command *cobra.Command, args []string) { Run: func(command *cobra.Command, args []string) {
cmd.CheckArgs(0, 1, command, args) cmd.CheckArgs(0, 1, command, args)
out := "/etc/bash_completion.d/rclone" out := "/etc/bash_completion.d/rclone"

View File

@@ -21,18 +21,21 @@ var fishCommandDefinition = &cobra.Command{
This writes to /etc/fish/completions/rclone.fish by default so will This writes to /etc/fish/completions/rclone.fish by default so will
probably need to be run with sudo or as root, e.g. probably need to be run with sudo or as root, e.g.
sudo rclone completion fish ` + "```sh" + `
sudo rclone completion fish
` + "```" + `
Logout and login again to use the autocompletion scripts, or source Logout and login again to use the autocompletion scripts, or source
them directly them directly
. /etc/fish/completions/rclone.fish ` + "```sh" + `
. /etc/fish/completions/rclone.fish
` + "```" + `
If you supply a command line argument the script will be written If you supply a command line argument the script will be written
there. there.
If output_file is "-", then the output will be written to stdout. If output_file is "-", then the output will be written to stdout.`,
`,
Run: func(command *cobra.Command, args []string) { Run: func(command *cobra.Command, args []string) {
cmd.CheckArgs(0, 1, command, args) cmd.CheckArgs(0, 1, command, args)
out := "/etc/fish/completions/rclone.fish" out := "/etc/fish/completions/rclone.fish"

View File

@@ -20,13 +20,14 @@ var powershellCommandDefinition = &cobra.Command{
To load completions in your current shell session: To load completions in your current shell session:
rclone completion powershell | Out-String | Invoke-Expression ` + "```sh" + `
rclone completion powershell | Out-String | Invoke-Expression
` + "```" + `
To load completions for every new session, add the output of the above command To load completions for every new session, add the output of the above command
to your powershell profile. to your powershell profile.
If output_file is "-" or missing, then the output will be written to stdout. If output_file is "-" or missing, then the output will be written to stdout.`,
`,
Run: func(command *cobra.Command, args []string) { Run: func(command *cobra.Command, args []string) {
cmd.CheckArgs(0, 1, command, args) cmd.CheckArgs(0, 1, command, args)
if len(args) == 0 || (len(args) > 0 && args[0] == "-") { if len(args) == 0 || (len(args) > 0 && args[0] == "-") {

View File

@@ -21,18 +21,21 @@ var zshCommandDefinition = &cobra.Command{
This writes to /usr/share/zsh/vendor-completions/_rclone by default so will This writes to /usr/share/zsh/vendor-completions/_rclone by default so will
probably need to be run with sudo or as root, e.g. probably need to be run with sudo or as root, e.g.
sudo rclone completion zsh ` + "```sh" + `
sudo rclone completion zsh
` + "```" + `
Logout and login again to use the autocompletion scripts, or source Logout and login again to use the autocompletion scripts, or source
them directly them directly
autoload -U compinit && compinit ` + "```sh" + `
autoload -U compinit && compinit
` + "```" + `
If you supply a command line argument the script will be written If you supply a command line argument the script will be written
there. there.
If output_file is "-", then the output will be written to stdout. If output_file is "-", then the output will be written to stdout.`,
`,
Run: func(command *cobra.Command, args []string) { Run: func(command *cobra.Command, args []string) {
cmd.CheckArgs(0, 1, command, args) cmd.CheckArgs(0, 1, command, args)
out := "/usr/share/zsh/vendor-completions/_rclone" out := "/usr/share/zsh/vendor-completions/_rclone"

View File

@@ -184,7 +184,12 @@ rclone.org website.`,
return fmt.Errorf("internal error: failed to find cut points: startCut = %d, endCut = %d", startCut, endCut) return fmt.Errorf("internal error: failed to find cut points: startCut = %d, endCut = %d", startCut, endCut)
} }
if endCut >= 0 { if endCut >= 0 {
doc = doc[:endCut] + "### See Also" + doc[endCut+12:] doc = doc[:endCut] + `### See Also
<!-- markdownlint-capture -->
<!-- markdownlint-disable ul-style line-length -->` + doc[endCut+12:] + `
<!-- markdownlint-restore -->
`
} }
} else { } else {
var out strings.Builder var out strings.Builder
@@ -196,7 +201,7 @@ rclone.org website.`,
if group.Flags.HasFlags() { if group.Flags.HasFlags() {
_, _ = fmt.Fprintf(&out, "#### %s Options\n\n", group.Name) _, _ = fmt.Fprintf(&out, "#### %s Options\n\n", group.Name)
_, _ = fmt.Fprintf(&out, "%s\n\n", group.Help) _, _ = fmt.Fprintf(&out, "%s\n\n", group.Help)
_, _ = out.WriteString("```\n") _, _ = out.WriteString("```text\n")
_, _ = out.WriteString(group.Flags.FlagUsages()) _, _ = out.WriteString(group.Flags.FlagUsages())
_, _ = out.WriteString("```\n\n") _, _ = out.WriteString("```\n\n")
} }
@@ -204,7 +209,12 @@ rclone.org website.`,
} else { } else {
_, _ = out.WriteString("See the [global flags page](/flags/) for global options not listed here.\n\n") _, _ = out.WriteString("See the [global flags page](/flags/) for global options not listed here.\n\n")
} }
doc = doc[:startCut] + out.String() + "### See Also" + doc[endCut+12:] doc = doc[:startCut] + out.String() + `### See Also
<!-- markdownlint-capture -->
<!-- markdownlint-disable ul-style line-length -->` + doc[endCut+12:] + `
<!-- markdownlint-restore -->
`
} }
// outdent all the titles by one // outdent all the titles by one

View File

@@ -539,7 +539,7 @@ var command = &cobra.Command{
Aliases: []string{uniqueCommandName}, Aliases: []string{uniqueCommandName},
Use: subcommandName, Use: subcommandName,
Short: "Speaks with git-annex over stdin/stdout.", Short: "Speaks with git-annex over stdin/stdout.",
Long: gitannexHelp, Long: strings.TrimSpace(gitannexHelp),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.67.0", "versionIntroduced": "v1.67.0",
}, },

View File

@@ -4,8 +4,7 @@ users.
[git-annex]: https://git-annex.branchable.com/ [git-annex]: https://git-annex.branchable.com/
Installation on Linux ### Installation on Linux
---------------------
1. Skip this step if your version of git-annex is [10.20240430] or newer. 1. Skip this step if your version of git-annex is [10.20240430] or newer.
Otherwise, you must create a symlink somewhere on your PATH with a particular Otherwise, you must create a symlink somewhere on your PATH with a particular

View File

@@ -103,14 +103,17 @@ as a relative path).
Run without a hash to see the list of all supported hashes, e.g. Run without a hash to see the list of all supported hashes, e.g.
$ rclone hashsum ` + "```sh" + `
` + hash.HelpString(4) + ` $ rclone hashsum
` + hash.HelpString(0) + "```" + `
Then Then
$ rclone hashsum MD5 remote:path ` + "```sh" + `
rclone hashsum MD5 remote:path
` + "```" + `
Note that hash names are case insensitive and values are output in lower case. Note that hash names are case insensitive and values are output in lower case.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.41", "versionIntroduced": "v1.41",
"groups": "Filter,Listing", "groups": "Filter,Listing",

View File

@@ -30,9 +30,7 @@ var Root = &cobra.Command{
mounting them, listing them in lots of different ways. mounting them, listing them in lots of different ways.
See the home page (https://rclone.org/) for installation, usage, See the home page (https://rclone.org/) for installation, usage,
documentation, changelog and configuration walkthroughs. documentation, changelog and configuration walkthroughs.`,
`,
PersistentPostRun: func(cmd *cobra.Command, args []string) { PersistentPostRun: func(cmd *cobra.Command, args []string) {
fs.Debugf("rclone", "Version %q finishing with parameters %q", fs.Version, os.Args) fs.Debugf("rclone", "Version %q finishing with parameters %q", fs.Version, os.Args)
atexit.Run() atexit.Run()

View File

@@ -29,10 +29,12 @@ var commandDefinition = &cobra.Command{
Short: `Generate public link to file/folder.`, Short: `Generate public link to file/folder.`,
Long: `Create, retrieve or remove a public link to the given file or folder. Long: `Create, retrieve or remove a public link to the given file or folder.
rclone link remote:path/to/file ` + "```sh" + `
rclone link remote:path/to/folder/ rclone link remote:path/to/file
rclone link --unlink remote:path/to/folder/ rclone link remote:path/to/folder/
rclone link --expire 1d remote:path/to/file rclone link --unlink remote:path/to/folder/
rclone link --expire 1d remote:path/to/file
` + "```" + `
If you supply the --expire flag, it will set the expiration time If you supply the --expire flag, it will set the expiration time
otherwise it will use the default (100 years). **Note** not all otherwise it will use the default (100 years). **Note** not all
@@ -45,9 +47,8 @@ don't will just ignore it.
If successful, the last line of the output will contain the If successful, the last line of the output will contain the
link. Exact capabilities depend on the remote, but the link will link. Exact capabilities depend on the remote, but the link will
always by default be created with the least constraints e.g. no always by default be created with the least constraints - e.g. no
expiry, no password protection, accessible without account. expiry, no password protection, accessible without account.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.41", "versionIntroduced": "v1.41",
}, },

View File

@@ -114,8 +114,7 @@ func newLess(orderBy string) (less lessFn, err error) {
var commandDefinition = &cobra.Command{ var commandDefinition = &cobra.Command{
Use: "listremotes [<filter>]", Use: "listremotes [<filter>]",
Short: `List all the remotes in the config file and defined in environment variables.`, Short: `List all the remotes in the config file and defined in environment variables.`,
Long: ` Long: `Lists all the available remotes from the config file, or the remotes matching
Lists all the available remotes from the config file, or the remotes matching
an optional filter. an optional filter.
Prints the result in human-readable format by default, and as a simple list of Prints the result in human-readable format by default, and as a simple list of
@@ -126,8 +125,7 @@ the source (file or environment).
Result can be filtered by a filter argument which applies to all attributes, Result can be filtered by a filter argument which applies to all attributes,
and/or filter flags specific for each attribute. The values must be specified and/or filter flags specific for each attribute. The values must be specified
according to regular rclone filtering pattern syntax. according to regular rclone filtering pattern syntax.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.34", "versionIntroduced": "v1.34",
}, },

View File

@@ -21,13 +21,15 @@ var commandDefinition = &cobra.Command{
Long: `Lists the objects in the source path to standard output in a human Long: `Lists the objects in the source path to standard output in a human
readable format with size and path. Recurses by default. readable format with size and path. Recurses by default.
Eg E.g.
$ rclone ls swift:bucket ` + "```sh" + `
60295 bevajer5jef $ rclone ls swift:bucket
90613 canole 60295 bevajer5jef
94467 diwogej7 90613 canole
37600 fubuwic 94467 diwogej7
37600 fubuwic
` + "```" + `
` + lshelp.Help, ` + lshelp.Help,
Annotations: map[string]string{ Annotations: map[string]string{

View File

@@ -7,16 +7,15 @@ import (
// Help describes the common help for all the list commands // Help describes the common help for all the list commands
// Warning! "|" will be replaced by backticks below // Warning! "|" will be replaced by backticks below
var Help = strings.ReplaceAll(` var Help = strings.ReplaceAll(`Any of the filtering options can be applied to this command.
Any of the filtering options can be applied to this command.
There are several related list commands There are several related list commands
* |ls| to list size and path of objects only - |ls| to list size and path of objects only
* |lsl| to list modification time, size and path of objects only - |lsl| to list modification time, size and path of objects only
* |lsd| to list directories only - |lsd| to list directories only
* |lsf| to list objects and directories in easy to parse format - |lsf| to list objects and directories in easy to parse format
* |lsjson| to list objects and directories in JSON format - |lsjson| to list objects and directories in JSON format
|ls|,|lsl|,|lsd| are designed to be human-readable. |ls|,|lsl|,|lsd| are designed to be human-readable.
|lsf| is designed to be human and machine-readable. |lsf| is designed to be human and machine-readable.
@@ -24,9 +23,9 @@ There are several related list commands
Note that |ls| and |lsl| recurse by default - use |--max-depth 1| to stop the recursion. Note that |ls| and |lsl| recurse by default - use |--max-depth 1| to stop the recursion.
The other list commands |lsd|,|lsf|,|lsjson| do not recurse by default - use |-R| to make them recurse. The other list commands |lsd|,|lsf|,|lsjson| do not recurse by default -
use |-R| to make them recurse.
Listing a nonexistent directory will produce an error except for Listing a nonexistent directory will produce an error except for
remotes which can't have empty directories (e.g. s3, swift, or gcs - remotes which can't have empty directories (e.g. s3, swift, or gcs -
the bucket-based remotes). the bucket-based remotes).`, "|", "`")
`, "|", "`")

View File

@@ -32,18 +32,22 @@ recurse by default. Use the ` + "`-R`" + ` flag to recurse.
This command lists the total size of the directory (if known, -1 if This command lists the total size of the directory (if known, -1 if
not), the modification time (if known, the current time if not), the not), the modification time (if known, the current time if not), the
number of objects in the directory (if known, -1 if not) and the name number of objects in the directory (if known, -1 if not) and the name
of the directory, Eg of the directory, E.g.
$ rclone lsd swift: ` + "```sh" + `
494000 2018-04-26 08:43:20 10000 10000files $ rclone lsd swift:
65 2018-04-26 08:43:20 1 1File 494000 2018-04-26 08:43:20 10000 10000files
65 2018-04-26 08:43:20 1 1File
` + "```" + `
Or Or
$ rclone lsd drive:test ` + "```sh" + `
-1 2016-10-17 17:41:53 -1 1000files $ rclone lsd drive:test
-1 2017-01-03 14:40:54 -1 2500files -1 2016-10-17 17:41:53 -1 1000files
-1 2017-07-08 14:39:28 -1 4000files -1 2017-01-03 14:40:54 -1 2500files
-1 2017-07-08 14:39:28 -1 4000files
` + "```" + `
If you just want the directory names use ` + "`rclone lsf --dirs-only`" + `. If you just want the directory names use ` + "`rclone lsf --dirs-only`" + `.

View File

@@ -52,41 +52,47 @@ standard output in a form which is easy to parse by scripts. By
default this will just be the names of the objects and directories, default this will just be the names of the objects and directories,
one per line. The directories will have a / suffix. one per line. The directories will have a / suffix.
Eg E.g.
$ rclone lsf swift:bucket ` + "```sh" + `
bevajer5jef $ rclone lsf swift:bucket
canole bevajer5jef
diwogej7 canole
ferejej3gux/ diwogej7
fubuwic ferejej3gux/
fubuwic
` + "```" + `
Use the ` + "`--format`" + ` option to control what gets listed. By default this Use the ` + "`--format`" + ` option to control what gets listed. By default this
is just the path, but you can use these parameters to control the is just the path, but you can use these parameters to control the
output: output:
p - path ` + "```text" + `
s - size p - path
t - modification time s - size
h - hash t - modification time
i - ID of object h - hash
o - Original ID of underlying object i - ID of object
m - MimeType of object if known o - Original ID of underlying object
e - encrypted name m - MimeType of object if known
T - tier of storage if known, e.g. "Hot" or "Cool" e - encrypted name
M - Metadata of object in JSON blob format, eg {"key":"value"} T - tier of storage if known, e.g. "Hot" or "Cool"
M - Metadata of object in JSON blob format, eg {"key":"value"}
` + "```" + `
So if you wanted the path, size and modification time, you would use So if you wanted the path, size and modification time, you would use
` + "`--format \"pst\"`, or maybe `--format \"tsp\"`" + ` to put the path last. ` + "`--format \"pst\"`, or maybe `--format \"tsp\"`" + ` to put the path last.
Eg E.g.
$ rclone lsf --format "tsp" swift:bucket ` + "```sh" + `
2016-06-25 18:55:41;60295;bevajer5jef $ rclone lsf --format "tsp" swift:bucket
2016-06-25 18:55:43;90613;canole 2016-06-25 18:55:41;60295;bevajer5jef
2016-06-25 18:55:43;94467;diwogej7 2016-06-25 18:55:43;90613;canole
2018-04-26 08:50:45;0;ferejej3gux/ 2016-06-25 18:55:43;94467;diwogej7
2016-06-25 18:55:40;37600;fubuwic 2018-04-26 08:50:45;0;ferejej3gux/
2016-06-25 18:55:40;37600;fubuwic
` + "```" + `
If you specify "h" in the format you will get the MD5 hash by default, If you specify "h" in the format you will get the MD5 hash by default,
use the ` + "`--hash`" + ` flag to change which hash you want. Note that this use the ` + "`--hash`" + ` flag to change which hash you want. Note that this
@@ -97,16 +103,20 @@ type.
For example, to emulate the md5sum command you can use For example, to emulate the md5sum command you can use
rclone lsf -R --hash MD5 --format hp --separator " " --files-only . ` + "```sh" + `
rclone lsf -R --hash MD5 --format hp --separator " " --files-only .
` + "```" + `
Eg E.g.
$ rclone lsf -R --hash MD5 --format hp --separator " " --files-only swift:bucket ` + "```sh" + `
7908e352297f0f530b84a756f188baa3 bevajer5jef $ rclone lsf -R --hash MD5 --format hp --separator " " --files-only swift:bucket
cd65ac234e6fea5925974a51cdd865cc canole 7908e352297f0f530b84a756f188baa3 bevajer5jef
03b5341b4f234b9d984d03ad076bae91 diwogej7 cd65ac234e6fea5925974a51cdd865cc canole
8fd37c3810dd660778137ac3a66cc06d fubuwic 03b5341b4f234b9d984d03ad076bae91 diwogej7
99713e14a4c4ff553acaf1930fad985b gixacuh7ku 8fd37c3810dd660778137ac3a66cc06d fubuwic
99713e14a4c4ff553acaf1930fad985b gixacuh7ku
` + "```" + `
(Though "rclone md5sum ." is an easier way of typing this.) (Though "rclone md5sum ." is an easier way of typing this.)
@@ -114,24 +124,28 @@ By default the separator is ";" this can be changed with the
` + "`--separator`" + ` flag. Note that separators aren't escaped in the path so ` + "`--separator`" + ` flag. Note that separators aren't escaped in the path so
putting it last is a good strategy. putting it last is a good strategy.
Eg E.g.
$ rclone lsf --separator "," --format "tshp" swift:bucket ` + "```sh" + `
2016-06-25 18:55:41,60295,7908e352297f0f530b84a756f188baa3,bevajer5jef $ rclone lsf --separator "," --format "tshp" swift:bucket
2016-06-25 18:55:43,90613,cd65ac234e6fea5925974a51cdd865cc,canole 2016-06-25 18:55:41,60295,7908e352297f0f530b84a756f188baa3,bevajer5jef
2016-06-25 18:55:43,94467,03b5341b4f234b9d984d03ad076bae91,diwogej7 2016-06-25 18:55:43,90613,cd65ac234e6fea5925974a51cdd865cc,canole
2018-04-26 08:52:53,0,,ferejej3gux/ 2016-06-25 18:55:43,94467,03b5341b4f234b9d984d03ad076bae91,diwogej7
2016-06-25 18:55:40,37600,8fd37c3810dd660778137ac3a66cc06d,fubuwic 2018-04-26 08:52:53,0,,ferejej3gux/
2016-06-25 18:55:40,37600,8fd37c3810dd660778137ac3a66cc06d,fubuwic
` + "```" + `
You can output in CSV standard format. This will escape things in " You can output in CSV standard format. This will escape things in "
if they contain , if they contain,
Eg E.g.
$ rclone lsf --csv --files-only --format ps remote:path ` + "```sh" + `
test.log,22355 $ rclone lsf --csv --files-only --format ps remote:path
test.sh,449 test.log,22355
"this file contains a comma, in the file name.txt",6 test.sh,449
"this file contains a comma, in the file name.txt",6
` + "```" + `
Note that the ` + "`--absolute`" + ` parameter is useful for making lists of files Note that the ` + "`--absolute`" + ` parameter is useful for making lists of files
to pass to an rclone copy with the ` + "`--files-from-raw`" + ` flag. to pass to an rclone copy with the ` + "`--files-from-raw`" + ` flag.
@@ -139,20 +153,25 @@ to pass to an rclone copy with the ` + "`--files-from-raw`" + ` flag.
For example, to find all the files modified within one day and copy For example, to find all the files modified within one day and copy
those only (without traversing the whole directory structure): those only (without traversing the whole directory structure):
rclone lsf --absolute --files-only --max-age 1d /path/to/local > new_files ` + "```sh" + `
rclone copy --files-from-raw new_files /path/to/local remote:path rclone lsf --absolute --files-only --max-age 1d /path/to/local > new_files
rclone copy --files-from-raw new_files /path/to/local remote:path
` + "```" + `
The default time format is ` + "`'2006-01-02 15:04:05'`" + `. The default time format is ` + "`'2006-01-02 15:04:05'`" + `.
[Other formats](https://pkg.go.dev/time#pkg-constants) can be specified with the ` + "`--time-format`" + ` flag. [Other formats](https://pkg.go.dev/time#pkg-constants) can be specified with
Examples: the ` + "`--time-format`" + ` flag. Examples:
rclone lsf remote:path --format pt --time-format 'Jan 2, 2006 at 3:04pm (MST)' ` + "```sh" + `
rclone lsf remote:path --format pt --time-format '2006-01-02 15:04:05.000000000' rclone lsf remote:path --format pt --time-format 'Jan 2, 2006 at 3:04pm (MST)'
rclone lsf remote:path --format pt --time-format '2006-01-02T15:04:05.999999999Z07:00' rclone lsf remote:path --format pt --time-format '2006-01-02 15:04:05.000000000'
rclone lsf remote:path --format pt --time-format RFC3339 rclone lsf remote:path --format pt --time-format '2006-01-02T15:04:05.999999999Z07:00'
rclone lsf remote:path --format pt --time-format DateOnly rclone lsf remote:path --format pt --time-format RFC3339
rclone lsf remote:path --format pt --time-format max rclone lsf remote:path --format pt --time-format DateOnly
` + "`--time-format max`" + ` will automatically truncate ` + "'`2006-01-02 15:04:05.000000000`'" + ` rclone lsf remote:path --format pt --time-format max
` + "```" + `
` + "`--time-format max`" + ` will automatically truncate ` + "`2006-01-02 15:04:05.000000000`" + `
to the maximum precision supported by the remote. to the maximum precision supported by the remote.
` + lshelp.Help, ` + lshelp.Help,

View File

@@ -43,25 +43,27 @@ var commandDefinition = &cobra.Command{
The output is an array of Items, where each Item looks like this: The output is an array of Items, where each Item looks like this:
{ ` + "```json" + `
"Hashes" : { {
"SHA-1" : "f572d396fae9206628714fb2ce00f72e94f2258f", "Hashes" : {
"MD5" : "b1946ac92492d2347c6235b4d2611184", "SHA-1" : "f572d396fae9206628714fb2ce00f72e94f2258f",
"DropboxHash" : "ecb65bb98f9d905b70458986c39fcbad7715e5f2fcc3b1f07767d7c83e2438cc" "MD5" : "b1946ac92492d2347c6235b4d2611184",
}, "DropboxHash" : "ecb65bb98f9d905b70458986c39fcbad7715e5f2fcc3b1f07767d7c83e2438cc"
"ID": "y2djkhiujf83u33", },
"OrigID": "UYOJVTUW00Q1RzTDA", "ID": "y2djkhiujf83u33",
"IsBucket" : false, "OrigID": "UYOJVTUW00Q1RzTDA",
"IsDir" : false, "IsBucket" : false,
"MimeType" : "application/octet-stream", "IsDir" : false,
"ModTime" : "2017-05-31T16:15:57.034468261+01:00", "MimeType" : "application/octet-stream",
"Name" : "file.txt", "ModTime" : "2017-05-31T16:15:57.034468261+01:00",
"Encrypted" : "v0qpsdq8anpci8n929v3uu9338", "Name" : "file.txt",
"EncryptedPath" : "kja9098349023498/v0qpsdq8anpci8n929v3uu9338", "Encrypted" : "v0qpsdq8anpci8n929v3uu9338",
"Path" : "full/path/goes/here/file.txt", "EncryptedPath" : "kja9098349023498/v0qpsdq8anpci8n929v3uu9338",
"Size" : 6, "Path" : "full/path/goes/here/file.txt",
"Tier" : "hot", "Size" : 6,
} "Tier" : "hot",
}
` + "```" + `
The exact set of properties included depends on the backend: The exact set of properties included depends on the backend:
@@ -118,6 +120,7 @@ will be shown ("2017-05-31T16:15:57+01:00").
The whole output can be processed as a JSON blob, or alternatively it The whole output can be processed as a JSON blob, or alternatively it
can be processed line by line as each item is written on individual lines can be processed line by line as each item is written on individual lines
(except with ` + "`--stat`" + `). (except with ` + "`--stat`" + `).
` + lshelp.Help, ` + lshelp.Help,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.37", "versionIntroduced": "v1.37",

View File

@@ -21,13 +21,15 @@ var commandDefinition = &cobra.Command{
Long: `Lists the objects in the source path to standard output in a human Long: `Lists the objects in the source path to standard output in a human
readable format with modification time, size and path. Recurses by default. readable format with modification time, size and path. Recurses by default.
Eg E.g.
$ rclone lsl swift:bucket ` + "```sh" + `
60295 2016-06-25 18:55:41.062626927 bevajer5jef $ rclone lsl swift:bucket
90613 2016-06-25 18:55:43.302607074 canole 60295 2016-06-25 18:55:41.062626927 bevajer5jef
94467 2016-06-25 18:55:43.046609333 diwogej7 90613 2016-06-25 18:55:43.302607074 canole
37600 2016-06-25 18:55:40.814629136 fubuwic 94467 2016-06-25 18:55:43.046609333 diwogej7
37600 2016-06-25 18:55:40.814629136 fubuwic
` + "```" + `
` + lshelp.Help, ` + lshelp.Help,
Annotations: map[string]string{ Annotations: map[string]string{

View File

@@ -35,8 +35,7 @@ to running ` + "`rclone hashsum MD5 remote:path`" + `.
This command can also hash data received on standard input (stdin), This command can also hash data received on standard input (stdin),
by not passing a remote:path, or by passing a hyphen as remote:path by not passing a remote:path, or by passing a hyphen as remote:path
when there is data to read (if not, the hyphen will be treated literally, when there is data to read (if not, the hyphen will be treated literally,
as a relative path). as a relative path).`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.02", "versionIntroduced": "v1.02",
"groups": "Filter,Listing", "groups": "Filter,Listing",

View File

@@ -273,7 +273,7 @@ func NewMountCommand(commandName string, hidden bool, mount MountFn) *cobra.Comm
Use: commandName + " remote:path /path/to/mountpoint", Use: commandName + " remote:path /path/to/mountpoint",
Hidden: hidden, Hidden: hidden,
Short: `Mount the remote as file system on a mountpoint.`, Short: `Mount the remote as file system on a mountpoint.`,
Long: help(commandName) + vfs.Help(), Long: help(commandName) + strings.TrimSpace(vfs.Help()),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.33", "versionIntroduced": "v1.33",
"groups": "Filter", "groups": "Filter",
@@ -396,34 +396,14 @@ func (m *MountPoint) Wait() error {
if err := m.Unmount(); err != nil { if err := m.Unmount(); err != nil {
fs.Errorf(m.MountPoint, "Failed to unmount: %v", err) fs.Errorf(m.MountPoint, "Failed to unmount: %v", err)
} else { } else {
fs.Errorf(m.MountPoint, "Unmounted rclone mount") fs.Logf(m.MountPoint, "Unmounted rclone mount")
} }
}) })
} }
fnHandle := atexit.Register(finalise) fnHandle := atexit.Register(finalise)
defer atexit.Unregister(fnHandle) defer atexit.Unregister(fnHandle)
// Reload VFS cache on SIGHUP err := <-m.ErrChan
sigHup := make(chan os.Signal, 1)
NotifyOnSigHup(sigHup)
var err error
waiting := true
for waiting {
select {
// umount triggered outside the app
case err = <-m.ErrChan:
waiting = false
// user sent SIGHUP to clear the cache
case <-sigHup:
root, err := m.VFS.Root()
if err != nil {
fs.Errorf(m.VFS.Fs(), "Error reading root: %v", err)
} else {
root.ForgetAll()
}
}
}
finalise() finalise()

View File

@@ -1,7 +1,7 @@
Rclone @ allows Linux, FreeBSD, macOS and Windows to Rclone @ allows Linux, FreeBSD, macOS and Windows to
mount any of Rclone's cloud storage systems as a file system with FUSE. mount any of Rclone's cloud storage systems as a file system with FUSE.
First set up your remote using `rclone config`. Check it works with `rclone ls` etc. First set up your remote using `rclone config`. Check it works with `rclone ls` etc.
On Linux and macOS, you can run mount in either foreground or background (aka On Linux and macOS, you can run mount in either foreground or background (aka
daemon) mode. Mount runs in foreground mode by default. Use the `--daemon` flag daemon) mode. Mount runs in foreground mode by default. Use the `--daemon` flag
@@ -16,7 +16,9 @@ mount, waits until success or timeout and exits with appropriate code
On Linux/macOS/FreeBSD start the mount like this, where `/path/to/local/mount` On Linux/macOS/FreeBSD start the mount like this, where `/path/to/local/mount`
is an **empty** **existing** directory: is an **empty** **existing** directory:
rclone @ remote:path/to/files /path/to/local/mount ```sh
rclone @ remote:path/to/files /path/to/local/mount
```
On Windows you can start a mount in different ways. See [below](#mounting-modes-on-windows) On Windows you can start a mount in different ways. See [below](#mounting-modes-on-windows)
for details. If foreground mount is used interactively from a console window, for details. If foreground mount is used interactively from a console window,
@@ -26,26 +28,30 @@ used to work with the mount until rclone is interrupted e.g. by pressing Ctrl-C.
The following examples will mount to an automatically assigned drive, The following examples will mount to an automatically assigned drive,
to specific drive letter `X:`, to path `C:\path\parent\mount` to specific drive letter `X:`, to path `C:\path\parent\mount`
(where parent directory or drive must exist, and mount must **not** exist, (where parent directory or drive must exist, and mount must **not** exist,
and is not supported when [mounting as a network drive](#mounting-modes-on-windows)), and and is not supported when [mounting as a network drive](#mounting-modes-on-windows)),
the last example will mount as network share `\\cloud\remote` and map it to an and the last example will mount as network share `\\cloud\remote` and map it to an
automatically assigned drive: automatically assigned drive:
rclone @ remote:path/to/files * ```sh
rclone @ remote:path/to/files X: rclone @ remote:path/to/files *
rclone @ remote:path/to/files C:\path\parent\mount rclone @ remote:path/to/files X:
rclone @ remote:path/to/files \\cloud\remote rclone @ remote:path/to/files C:\path\parent\mount
rclone @ remote:path/to/files \\cloud\remote
```
When the program ends while in foreground mode, either via Ctrl+C or receiving When the program ends while in foreground mode, either via Ctrl+C or receiving
a SIGINT or SIGTERM signal, the mount should be automatically stopped. a SIGINT or SIGTERM signal, the mount should be automatically stopped.
When running in background mode the user will have to stop the mount manually: When running in background mode the user will have to stop the mount manually:
# Linux ```sh
fusermount -u /path/to/local/mount # Linux
#... or on some systems fusermount -u /path/to/local/mount
fusermount3 -u /path/to/local/mount #... or on some systems
# OS X or Linux when using nfsmount fusermount3 -u /path/to/local/mount
umount /path/to/local/mount # OS X or Linux when using nfsmount
umount /path/to/local/mount
```
The umount operation can fail, for example when the mountpoint is busy. The umount operation can fail, for example when the mountpoint is busy.
When that happens, it is the user's responsibility to stop the mount manually. When that happens, it is the user's responsibility to stop the mount manually.
@@ -80,20 +86,22 @@ thumbnails for image and video files on network drives.
In most cases, rclone will mount the remote as a normal, fixed disk drive by default. In most cases, rclone will mount the remote as a normal, fixed disk drive by default.
However, you can also choose to mount it as a remote network drive, often described However, you can also choose to mount it as a remote network drive, often described
as a network share. If you mount an rclone remote using the default, fixed drive mode as a network share. If you mount an rclone remote using the default, fixed drive
and experience unexpected program errors, freezes or other issues, consider mounting mode and experience unexpected program errors, freezes or other issues, consider
as a network drive instead. mounting as a network drive instead.
When mounting as a fixed disk drive you can either mount to an unused drive letter, When mounting as a fixed disk drive you can either mount to an unused drive letter,
or to a path representing a **nonexistent** subdirectory of an **existing** parent or to a path representing a **nonexistent** subdirectory of an **existing** parent
directory or drive. Using the special value `*` will tell rclone to directory or drive. Using the special value `*` will tell rclone to
automatically assign the next available drive letter, starting with Z: and moving backward. automatically assign the next available drive letter, starting with Z: and moving
Examples: backward. Examples:
rclone @ remote:path/to/files * ```sh
rclone @ remote:path/to/files X: rclone @ remote:path/to/files *
rclone @ remote:path/to/files C:\path\parent\mount rclone @ remote:path/to/files X:
rclone @ remote:path/to/files X: rclone @ remote:path/to/files C:\path\parent\mount
rclone @ remote:path/to/files X:
```
Option `--volname` can be used to set a custom volume name for the mounted Option `--volname` can be used to set a custom volume name for the mounted
file system. The default is to use the remote name and path. file system. The default is to use the remote name and path.
@@ -103,24 +111,28 @@ to your @ command. Mounting to a directory path is not supported in
this mode, it is a limitation Windows imposes on junctions, so the remote must always this mode, it is a limitation Windows imposes on junctions, so the remote must always
be mounted to a drive letter. be mounted to a drive letter.
rclone @ remote:path/to/files X: --network-mode ```sh
rclone @ remote:path/to/files X: --network-mode
```
A volume name specified with `--volname` will be used to create the network share path. A volume name specified with `--volname` will be used to create the network share
A complete UNC path, such as `\\cloud\remote`, optionally with path path. A complete UNC path, such as `\\cloud\remote`, optionally with path
`\\cloud\remote\madeup\path`, will be used as is. Any other `\\cloud\remote\madeup\path`, will be used as is. Any other
string will be used as the share part, after a default prefix `\\server\`. string will be used as the share part, after a default prefix `\\server\`.
If no volume name is specified then `\\server\share` will be used. If no volume name is specified then `\\server\share` will be used.
You must make sure the volume name is unique when you are mounting more than one drive, You must make sure the volume name is unique when you are mounting more than one
or else the mount command will fail. The share name will treated as the volume label for drive, or else the mount command will fail. The share name will treated as the
the mapped drive, shown in Windows Explorer etc, while the complete volume label for the mapped drive, shown in Windows Explorer etc, while the complete
`\\server\share` will be reported as the remote UNC path by `\\server\share` will be reported as the remote UNC path by
`net use` etc, just like a normal network drive mapping. `net use` etc, just like a normal network drive mapping.
If you specify a full network share UNC path with `--volname`, this will implicitly If you specify a full network share UNC path with `--volname`, this will implicitly
set the `--network-mode` option, so the following two examples have same result: set the `--network-mode` option, so the following two examples have same result:
rclone @ remote:path/to/files X: --network-mode ```sh
rclone @ remote:path/to/files X: --volname \\server\share rclone @ remote:path/to/files X: --network-mode
rclone @ remote:path/to/files X: --volname \\server\share
```
You may also specify the network share UNC path as the mountpoint itself. Then rclone You may also specify the network share UNC path as the mountpoint itself. Then rclone
will automatically assign a drive letter, same as with `*` and use that as will automatically assign a drive letter, same as with `*` and use that as
@@ -128,15 +140,16 @@ mountpoint, and instead use the UNC path specified as the volume name, as if it
specified with the `--volname` option. This will also implicitly set specified with the `--volname` option. This will also implicitly set
the `--network-mode` option. This means the following two examples have same result: the `--network-mode` option. This means the following two examples have same result:
rclone @ remote:path/to/files \\cloud\remote ```sh
rclone @ remote:path/to/files * --volname \\cloud\remote rclone @ remote:path/to/files \\cloud\remote
rclone @ remote:path/to/files * --volname \\cloud\remote
```
There is yet another way to enable network mode, and to set the share path, There is yet another way to enable network mode, and to set the share path,
and that is to pass the "native" libfuse/WinFsp option directly: and that is to pass the "native" libfuse/WinFsp option directly:
`--fuse-flag --VolumePrefix=\server\share`. Note that the path `--fuse-flag --VolumePrefix=\server\share`. Note that the path
must be with just a single backslash prefix in this case. must be with just a single backslash prefix in this case.
*Note:* In previous versions of rclone this was the only supported method. *Note:* In previous versions of rclone this was the only supported method.
[Read more about drive mapping](https://en.wikipedia.org/wiki/Drive_mapping) [Read more about drive mapping](https://en.wikipedia.org/wiki/Drive_mapping)
@@ -149,11 +162,11 @@ The FUSE emulation layer on Windows must convert between the POSIX-based
permission model used in FUSE, and the permission model used in Windows, permission model used in FUSE, and the permission model used in Windows,
based on access-control lists (ACL). based on access-control lists (ACL).
The mounted filesystem will normally get three entries in its access-control list (ACL), The mounted filesystem will normally get three entries in its access-control list
representing permissions for the POSIX permission scopes: Owner, group and others. (ACL), representing permissions for the POSIX permission scopes: Owner, group and
By default, the owner and group will be taken from the current user, and the built-in others. By default, the owner and group will be taken from the current user, and
group "Everyone" will be used to represent others. The user/group can be customized the built-in group "Everyone" will be used to represent others. The user/group can
with FUSE options "UserName" and "GroupName", be customized with FUSE options "UserName" and "GroupName",
e.g. `-o UserName=user123 -o GroupName="Authenticated Users"`. e.g. `-o UserName=user123 -o GroupName="Authenticated Users"`.
The permissions on each entry will be set according to [options](#options) The permissions on each entry will be set according to [options](#options)
`--dir-perms` and `--file-perms`, which takes a value in traditional Unix `--dir-perms` and `--file-perms`, which takes a value in traditional Unix
@@ -253,58 +266,63 @@ does not suffer from the same limitations.
### Mounting on macOS ### Mounting on macOS
Mounting on macOS can be done either via [built-in NFS server](/commands/rclone_serve_nfs/), [macFUSE](https://osxfuse.github.io/) Mounting on macOS can be done either via [built-in NFS server](/commands/rclone_serve_nfs/),
(also known as osxfuse) or [FUSE-T](https://www.fuse-t.org/). macFUSE is a traditional [macFUSE](https://osxfuse.github.io/) (also known as osxfuse) or
FUSE driver utilizing a macOS kernel extension (kext). FUSE-T is an alternative FUSE system [FUSE-T](https://www.fuse-t.org/).macFUSE is a traditional FUSE driver utilizing
which "mounts" via an NFSv4 local server. a macOS kernel extension (kext). FUSE-T is an alternative FUSE system which
"mounts" via an NFSv4 local server.
##### Unicode Normalization #### Unicode Normalization
It is highly recommended to keep the default of `--no-unicode-normalization=false` It is highly recommended to keep the default of `--no-unicode-normalization=false`
for all `mount` and `serve` commands on macOS. For details, see [vfs-case-sensitivity](https://rclone.org/commands/rclone_mount/#vfs-case-sensitivity). for all `mount` and `serve` commands on macOS. For details, see [vfs-case-sensitivity](https://rclone.org/commands/rclone_mount/#vfs-case-sensitivity).
#### NFS mount #### NFS mount
This method spins up an NFS server using [serve nfs](/commands/rclone_serve_nfs/) command and mounts This method spins up an NFS server using [serve nfs](/commands/rclone_serve_nfs/)
it to the specified mountpoint. If you run this in background mode using |--daemon|, you will need to command and mounts it to the specified mountpoint. If you run this in background
send SIGTERM signal to the rclone process using |kill| command to stop the mount. mode using |--daemon|, you will need to send SIGTERM signal to the rclone process
using |kill| command to stop the mount.
Note that `--nfs-cache-handle-limit` controls the maximum number of cached file handles stored by the `nfsmount` caching handler. Note that `--nfs-cache-handle-limit` controls the maximum number of cached file
This should not be set too low or you may experience errors when trying to access files. The default is 1000000, handles stored by the `nfsmount` caching handler. This should not be set too low
or you may experience errors when trying to access files. The default is 1000000,
but consider lowering this limit if the server's system resource usage causes problems. but consider lowering this limit if the server's system resource usage causes problems.
#### macFUSE Notes #### macFUSE Notes
If installing macFUSE using [dmg packages](https://github.com/osxfuse/osxfuse/releases) from If installing macFUSE using [dmg packages](https://github.com/osxfuse/osxfuse/releases)
the website, rclone will locate the macFUSE libraries without any further intervention. from the website, rclone will locate the macFUSE libraries without any further intervention.
If however, macFUSE is installed using the [macports](https://www.macports.org/) package manager, If however, macFUSE is installed using the [macports](https://www.macports.org/)
the following addition steps are required. package manager, the following addition steps are required.
sudo mkdir /usr/local/lib ```sh
cd /usr/local/lib sudo mkdir /usr/local/lib
sudo ln -s /opt/local/lib/libfuse.2.dylib cd /usr/local/lib
sudo ln -s /opt/local/lib/libfuse.2.dylib
```
#### FUSE-T Limitations, Caveats, and Notes #### FUSE-T Limitations, Caveats, and Notes
There are some limitations, caveats, and notes about how it works. These are current as There are some limitations, caveats, and notes about how it works. These are
of FUSE-T version 1.0.14. current as of FUSE-T version 1.0.14.
##### ModTime update on read ##### ModTime update on read
As per the [FUSE-T wiki](https://github.com/macos-fuse-t/fuse-t/wiki#caveats): As per the [FUSE-T wiki](https://github.com/macos-fuse-t/fuse-t/wiki#caveats):
> File access and modification times cannot be set separately as it seems to be an > File access and modification times cannot be set separately as it seems to be an
> issue with the NFS client which always modifies both. Can be reproduced with > issue with the NFS client which always modifies both. Can be reproduced with
> 'touch -m' and 'touch -a' commands > 'touch -m' and 'touch -a' commands
This means that viewing files with various tools, notably macOS Finder, will cause rlcone This means that viewing files with various tools, notably macOS Finder, will cause
to update the modification time of the file. This may make rclone upload a full new copy rlcone to update the modification time of the file. This may make rclone upload a
of the file. full new copy of the file.
##### Read Only mounts ##### Read Only mounts
When mounting with `--read-only`, attempts to write to files will fail *silently* as When mounting with `--read-only`, attempts to write to files will fail *silently*
opposed to with a clear warning as in macFUSE. as opposed to with a clear warning as in macFUSE.
### Limitations ### Limitations
@@ -405,12 +423,14 @@ helper you should symlink rclone binary to `/sbin/mount.rclone` and optionally
rclone will detect it and translate command-line arguments appropriately. rclone will detect it and translate command-line arguments appropriately.
Now you can run classic mounts like this: Now you can run classic mounts like this:
```
```sh
mount sftp1:subdir /mnt/data -t rclone -o vfs_cache_mode=writes,sftp_key_file=/path/to/pem mount sftp1:subdir /mnt/data -t rclone -o vfs_cache_mode=writes,sftp_key_file=/path/to/pem
``` ```
or create systemd mount units: or create systemd mount units:
```
```ini
# /etc/systemd/system/mnt-data.mount # /etc/systemd/system/mnt-data.mount
[Unit] [Unit]
Description=Mount for /mnt/data Description=Mount for /mnt/data
@@ -422,7 +442,8 @@ Options=rw,_netdev,allow_other,args2env,vfs-cache-mode=writes,config=/etc/rclone
``` ```
optionally accompanied by systemd automount unit optionally accompanied by systemd automount unit
```
```ini
# /etc/systemd/system/mnt-data.automount # /etc/systemd/system/mnt-data.automount
[Unit] [Unit]
Description=AutoMount for /mnt/data Description=AutoMount for /mnt/data
@@ -434,7 +455,8 @@ WantedBy=multi-user.target
``` ```
or add in `/etc/fstab` a line like or add in `/etc/fstab` a line like
```
```sh
sftp1:subdir /mnt/data rclone rw,noauto,nofail,_netdev,x-systemd.automount,args2env,vfs_cache_mode=writes,config=/etc/rclone.conf,cache_dir=/var/cache/rclone 0 0 sftp1:subdir /mnt/data rclone rw,noauto,nofail,_netdev,x-systemd.automount,args2env,vfs_cache_mode=writes,config=/etc/rclone.conf,cache_dir=/var/cache/rclone 0 0
``` ```

View File

@@ -65,14 +65,18 @@ This takes the following parameters:
Example: Example:
rclone rc mount/mount fs=mydrive: mountPoint=/home/<user>/mountPoint ` + "```sh" + `
rclone rc mount/mount fs=mydrive: mountPoint=/home/<user>/mountPoint mountType=mount rclone rc mount/mount fs=mydrive: mountPoint=/home/<user>/mountPoint
rclone rc mount/mount fs=TestDrive: mountPoint=/mnt/tmp vfsOpt='{"CacheMode": 2}' mountOpt='{"AllowOther": true}' rclone rc mount/mount fs=mydrive: mountPoint=/home/<user>/mountPoint mountType=mount
rclone rc mount/mount fs=TestDrive: mountPoint=/mnt/tmp vfsOpt='{"CacheMode": 2}' mountOpt='{"AllowOther": true}'
` + "```" + `
The vfsOpt are as described in options/get and can be seen in the the The vfsOpt are as described in options/get and can be seen in the the
"vfs" section when running and the mountOpt can be seen in the "mount" section: "vfs" section when running and the mountOpt can be seen in the "mount" section:
rclone rc options/get ` + "```sh" + `
rclone rc options/get
` + "```" + `
`, `,
}) })
} }

View File

@@ -64,7 +64,7 @@ the backend supports it. If metadata syncing is required then use the
|--metadata| flag. |--metadata| flag.
Note that the modification time and metadata for the root directory Note that the modification time and metadata for the root directory
will **not** be synced. See https://github.com/rclone/rclone/issues/7652 will **not** be synced. See <https://github.com/rclone/rclone/issues/7652>
for more info. for more info.
**Important**: Since this can cause data loss, test first with the **Important**: Since this can cause data loss, test first with the

View File

@@ -35,18 +35,22 @@ like the [move](/commands/rclone_move/) command.
So So
rclone moveto src dst ` + "```sh" + `
rclone moveto src dst
` + "```" + `
where src and dst are rclone paths, either remote:path or where src and dst are rclone paths, either remote:path or
/path/to/local or C:\windows\path\if\on\windows. /path/to/local or C:\windows\path\if\on\windows.
This will: This will:
if src is file ` + "```text" + `
move it to dst, overwriting an existing file if it exists if src is file
if src is directory move it to dst, overwriting an existing file if it exists
move it to dst, overwriting existing files if they exist if src is directory
see move command for full details move it to dst, overwriting existing files if they exist
see move command for full details
` + "```" + `
This doesn't transfer files that are identical on src and dst, testing This doesn't transfer files that are identical on src and dst, testing
by size and modification time or MD5SUM. src will be deleted on by size and modification time or MD5SUM. src will be deleted on

View File

@@ -47,22 +47,26 @@ structure as it goes along.
You can interact with the user interface using key presses, You can interact with the user interface using key presses,
press '?' to toggle the help on and off. The supported keys are: press '?' to toggle the help on and off. The supported keys are:
` + strings.Join(helpText()[1:], "\n ") + ` ` + "```text" + `
` + strings.Join(helpText()[1:], "\n") + `
` + "```" + `
Listed files/directories may be prefixed by a one-character flag, Listed files/directories may be prefixed by a one-character flag,
some of them combined with a description in brackets at end of line. some of them combined with a description in brackets at end of line.
These flags have the following meaning: These flags have the following meaning:
e means this is an empty directory, i.e. contains no files (but ` + "```text" + `
may contain empty subdirectories) e means this is an empty directory, i.e. contains no files (but
~ means this is a directory where some of the files (possibly in may contain empty subdirectories)
subdirectories) have unknown size, and therefore the directory ~ means this is a directory where some of the files (possibly in
size may be underestimated (and average size inaccurate, as it subdirectories) have unknown size, and therefore the directory
is average of the files with known sizes). size may be underestimated (and average size inaccurate, as it
. means an error occurred while reading a subdirectory, and is average of the files with known sizes).
therefore the directory size may be underestimated (and average . means an error occurred while reading a subdirectory, and
size inaccurate) therefore the directory size may be underestimated (and average
! means an error occurred while reading this directory size inaccurate)
! means an error occurred while reading this directory
` + "```" + `
This an homage to the [ncdu tool](https://dev.yorhel.nl/ncdu) but for This an homage to the [ncdu tool](https://dev.yorhel.nl/ncdu) but for
rclone remotes. It is missing lots of features at the moment rclone remotes. It is missing lots of features at the moment
@@ -73,8 +77,7 @@ UI won't respond in the meantime since the deletion is done synchronously.
For a non-interactive listing of the remote, see the For a non-interactive listing of the remote, see the
[tree](/commands/rclone_tree/) command. To just get the total size of [tree](/commands/rclone_tree/) command. To just get the total size of
the remote you can also use the [size](/commands/rclone_size/) command. the remote you can also use the [size](/commands/rclone_size/) command.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.37", "versionIntroduced": "v1.37",
"groups": "Filter,Listing", "groups": "Filter,Listing",

View File

@@ -22,9 +22,8 @@ var commandDefinition = &cobra.Command{
Long: `In the rclone config file, human-readable passwords are Long: `In the rclone config file, human-readable passwords are
obscured. Obscuring them is done by encrypting them and writing them obscured. Obscuring them is done by encrypting them and writing them
out in base64. This is **not** a secure way of encrypting these out in base64. This is **not** a secure way of encrypting these
passwords as rclone can decrypt them - it is to prevent "eyedropping" passwords as rclone can decrypt them - it is to prevent "eyedropping" -
- namely someone seeing a password in the rclone config file by namely someone seeing a password in the rclone config file by accident.
accident.
Many equally important things (like access tokens) are not obscured in Many equally important things (like access tokens) are not obscured in
the config file. However it is very hard to shoulder surf a 64 the config file. However it is very hard to shoulder surf a 64
@@ -34,7 +33,9 @@ This command can also accept a password through STDIN instead of an
argument by passing a hyphen as an argument. This will use the first argument by passing a hyphen as an argument. This will use the first
line of STDIN as the password not including the trailing newline. line of STDIN as the password not including the trailing newline.
echo "secretpassword" | rclone obscure - ` + "```sh" + `
echo "secretpassword" | rclone obscure -
` + "```" + `
If there is no data on STDIN to read, rclone obscure will default to If there is no data on STDIN to read, rclone obscure will default to
obfuscating the hyphen itself. obfuscating the hyphen itself.

View File

@@ -24,12 +24,12 @@ include/exclude filters - everything will be removed. Use the
delete files. To delete empty directories only, use command delete files. To delete empty directories only, use command
[rmdir](/commands/rclone_rmdir/) or [rmdirs](/commands/rclone_rmdirs/). [rmdir](/commands/rclone_rmdir/) or [rmdirs](/commands/rclone_rmdirs/).
The concurrency of this operation is controlled by the ` + "`--checkers`" + ` global flag. However, some backends will The concurrency of this operation is controlled by the ` + "`--checkers`" + ` global flag.
implement this command directly, in which case ` + "`--checkers`" + ` will be ignored. However, some backends will implement this command directly, in which
case ` + "`--checkers`" + ` will be ignored.
**Important**: Since this can cause data loss, test first with the **Important**: Since this can cause data loss, test first with the
` + "`--dry-run` or the `--interactive`/`-i`" + ` flag. ` + "`--dry-run` or the `--interactive`/`-i`" + ` flag.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"groups": "Important", "groups": "Important",
}, },

View File

@@ -53,8 +53,8 @@ var commandDefinition = &cobra.Command{
Short: `Run a command against a running rclone.`, Short: `Run a command against a running rclone.`,
Long: strings.ReplaceAll(`This runs a command against a running rclone. Use the |--url| flag to Long: strings.ReplaceAll(`This runs a command against a running rclone. Use the |--url| flag to
specify an non default URL to connect on. This can be either a specify an non default URL to connect on. This can be either a
":port" which is taken to mean "http://localhost:port" or a ":port" which is taken to mean <http://localhost:port> or a
"host:port" which is taken to mean "http://host:port" "host:port" which is taken to mean <http://host:port>.
A username and password can be passed in with |--user| and |--pass|. A username and password can be passed in with |--user| and |--pass|.
@@ -63,10 +63,12 @@ Note that |--rc-addr|, |--rc-user|, |--rc-pass| will be read also for
The |--unix-socket| flag can be used to connect over a unix socket like this The |--unix-socket| flag can be used to connect over a unix socket like this
# start server on /tmp/my.socket |||sh
rclone rcd --rc-addr unix:///tmp/my.socket # start server on /tmp/my.socket
# Connect to it rclone rcd --rc-addr unix:///tmp/my.socket
rclone rc --unix-socket /tmp/my.socket core/stats # Connect to it
rclone rc --unix-socket /tmp/my.socket core/stats
|||
Arguments should be passed in as parameter=value. Arguments should be passed in as parameter=value.
@@ -81,29 +83,38 @@ options in the form |-o key=value| or |-o key|. It can be repeated as
many times as required. This is useful for rc commands which take the many times as required. This is useful for rc commands which take the
"opt" parameter which by convention is a dictionary of strings. "opt" parameter which by convention is a dictionary of strings.
-o key=value -o key2 |||text
-o key=value -o key2
|||
Will place this in the "opt" value Will place this in the "opt" value
{"key":"value", "key2","") |||json
{"key":"value", "key2","")
|||
The |-a|/|--arg| option can be used to set strings in the "arg" value. It The |-a|/|--arg| option can be used to set strings in the "arg" value. It
can be repeated as many times as required. This is useful for rc can be repeated as many times as required. This is useful for rc
commands which take the "arg" parameter which by convention is a list commands which take the "arg" parameter which by convention is a list
of strings. of strings.
-a value -a value2 |||text
-a value -a value2
|||
Will place this in the "arg" value Will place this in the "arg" value
["value", "value2"] |||json
["value", "value2"]
|||
Use |--loopback| to connect to the rclone instance running |rclone rc|. Use |--loopback| to connect to the rclone instance running |rclone rc|.
This is very useful for testing commands without having to run an This is very useful for testing commands without having to run an
rclone rc server, e.g.: rclone rc server, e.g.:
rclone rc --loopback operations/about fs=/ |||sh
rclone rc --loopback operations/about fs=/
|||
Use |rclone rc| to see a list of all possible commands.`, "|", "`"), Use |rclone rc| to see a list of all possible commands.`, "|", "`"),
Annotations: map[string]string{ Annotations: map[string]string{

View File

@@ -28,8 +28,10 @@ var commandDefinition = &cobra.Command{
Short: `Copies standard input to file on remote.`, Short: `Copies standard input to file on remote.`,
Long: `Reads from standard input (stdin) and copies it to a single remote file. Long: `Reads from standard input (stdin) and copies it to a single remote file.
echo "hello world" | rclone rcat remote:path/to/file ` + "```sh" + `
ffmpeg - | rclone rcat remote:path/to/file echo "hello world" | rclone rcat remote:path/to/file
ffmpeg - | rclone rcat remote:path/to/file
` + "```" + `
If the remote file already exists, it will be overwritten. If the remote file already exists, it will be overwritten.

View File

@@ -3,6 +3,7 @@ package rcd
import ( import (
"context" "context"
"strings"
"github.com/rclone/rclone/cmd" "github.com/rclone/rclone/cmd"
"github.com/rclone/rclone/fs" "github.com/rclone/rclone/fs"
@@ -31,7 +32,7 @@ the browser when rclone is run.
See the [rc documentation](/rc/) for more info on the rc flags. See the [rc documentation](/rc/) for more info on the rc flags.
` + libhttp.Help(rcflags.FlagPrefix) + libhttp.TemplateHelp(rcflags.FlagPrefix) + libhttp.AuthHelp(rcflags.FlagPrefix), ` + strings.TrimSpace(libhttp.Help(rcflags.FlagPrefix)+libhttp.TemplateHelp(rcflags.FlagPrefix)+libhttp.AuthHelp(rcflags.FlagPrefix)),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.45", "versionIntroduced": "v1.45",
"groups": "RC", "groups": "RC",

View File

@@ -21,8 +21,7 @@ has any objects in it, not even empty subdirectories. Use
command [rmdirs](/commands/rclone_rmdirs/) (or [delete](/commands/rclone_delete/) command [rmdirs](/commands/rclone_rmdirs/) (or [delete](/commands/rclone_delete/)
with option ` + "`--rmdirs`" + `) to do that. with option ` + "`--rmdirs`" + `) to do that.
To delete a path and any objects in it, use [purge](/commands/rclone_purge/) command. To delete a path and any objects in it, use [purge](/commands/rclone_purge/) command.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"groups": "Important", "groups": "Important",
}, },

View File

@@ -38,8 +38,7 @@ This will delete ` + "`--checkers`" + ` directories concurrently so
if you have thousands of empty directories consider increasing this number. if you have thousands of empty directories consider increasing this number.
To delete a path and any objects in it, use the [purge](/commands/rclone_purge/) To delete a path and any objects in it, use the [purge](/commands/rclone_purge/)
command. command.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.35", "versionIntroduced": "v1.35",
"groups": "Important", "groups": "Important",

View File

@@ -65,7 +65,7 @@ var cmdSelfUpdate = &cobra.Command{
Use: "selfupdate", Use: "selfupdate",
Aliases: []string{"self-update"}, Aliases: []string{"self-update"},
Short: `Update the rclone binary.`, Short: `Update the rclone binary.`,
Long: selfUpdateHelp, Long: strings.TrimSpace(selfUpdateHelp),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.55", "versionIntroduced": "v1.55",
}, },

View File

@@ -43,5 +43,5 @@ command will rename the old executable to 'rclone.old.exe' upon success.
Please note that this command was not available before rclone version 1.55. Please note that this command was not available before rclone version 1.55.
If it fails for you with the message `unknown command "selfupdate"` then If it fails for you with the message `unknown command "selfupdate"` then
you will need to update manually following the install instructions located you will need to update manually following the
at https://rclone.org/install/ [install documentation](https://rclone.org/install/).

View File

@@ -123,7 +123,7 @@ default "rclone (hostname)".
Use ` + "`--log-trace` in conjunction with `-vv`" + ` to enable additional debug Use ` + "`--log-trace` in conjunction with `-vv`" + ` to enable additional debug
logging of all UPNP traffic. logging of all UPNP traffic.
` + vfs.Help(), ` + strings.TrimSpace(vfs.Help()),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.46", "versionIntroduced": "v1.46",
"groups": "Filter", "groups": "Filter",

View File

@@ -59,7 +59,7 @@ func init() {
var Command = &cobra.Command{ var Command = &cobra.Command{
Use: "docker", Use: "docker",
Short: `Serve any remote on docker's volume plugin API.`, Short: `Serve any remote on docker's volume plugin API.`,
Long: help() + vfs.Help(), Long: help() + strings.TrimSpace(vfs.Help()),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.56", "versionIntroduced": "v1.56",
"groups": "Filter", "groups": "Filter",

View File

@@ -8,7 +8,8 @@ docker daemon and runs the corresponding code when necessary.
Docker plugins can run as a managed plugin under control of the docker daemon Docker plugins can run as a managed plugin under control of the docker daemon
or as an independent native service. For testing, you can just run it directly or as an independent native service. For testing, you can just run it directly
from the command line, for example: from the command line, for example:
```
```sh
sudo rclone serve docker --base-dir /tmp/rclone-volumes --socket-addr localhost:8787 -vv sudo rclone serve docker --base-dir /tmp/rclone-volumes --socket-addr localhost:8787 -vv
``` ```

View File

@@ -76,7 +76,6 @@ func NewDriver(ctx context.Context, root string, mntOpt *mountlib.Options, vfsOp
// start mount monitoring // start mount monitoring
drv.hupChan = make(chan os.Signal, 1) drv.hupChan = make(chan os.Signal, 1)
drv.monChan = make(chan bool, 1) drv.monChan = make(chan bool, 1)
mountlib.NotifyOnSigHup(drv.hupChan)
go drv.monitor() go drv.monitor()
// unmount all volumes on exit // unmount all volumes on exit

View File

@@ -14,6 +14,7 @@ import (
"os/user" "os/user"
"regexp" "regexp"
"strconv" "strconv"
"strings"
"sync" "sync"
"time" "time"
@@ -140,7 +141,7 @@ By default this will serve files without needing a login.
You can set a single username and password with the --user and --pass flags. You can set a single username and password with the --user and --pass flags.
` + vfs.Help() + proxy.Help, ` + strings.TrimSpace(vfs.Help()+proxy.Help),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.44", "versionIntroduced": "v1.44",
"groups": "Filter", "groups": "Filter",

View File

@@ -110,7 +110,7 @@ The server will log errors. Use ` + "`-v`" + ` to see access logs.
` + "`--bwlimit`" + ` will be respected for file transfers. Use ` + "`--stats`" + ` to ` + "`--bwlimit`" + ` will be respected for file transfers. Use ` + "`--stats`" + ` to
control the stats printing. control the stats printing.
` + libhttp.Help(flagPrefix) + libhttp.TemplateHelp(flagPrefix) + libhttp.AuthHelp(flagPrefix) + vfs.Help() + proxy.Help, ` + strings.TrimSpace(libhttp.Help(flagPrefix)+libhttp.TemplateHelp(flagPrefix)+libhttp.AuthHelp(flagPrefix)+vfs.Help()+proxy.Help),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.39", "versionIntroduced": "v1.39",
"groups": "Filter", "groups": "Filter",
@@ -208,6 +208,7 @@ func newServer(ctx context.Context, f fs.Fs, opt *Options, vfsOpt *vfscommon.Opt
// Serve HTTP until the server is shutdown // Serve HTTP until the server is shutdown
func (s *HTTP) Serve() error { func (s *HTTP) Serve() error {
s.server.Serve() s.server.Serve()
fs.Logf(s.f, "HTTP Server started on %s", s.server.URLs())
s.server.Wait() s.server.Wait()
return nil return nil
} }

View File

@@ -125,7 +125,7 @@ var Command = &cobra.Command{
Use: "nfs remote:path", Use: "nfs remote:path",
Short: `Serve the remote as an NFS mount`, Short: `Serve the remote as an NFS mount`,
Long: strings.ReplaceAll(`Create an NFS server that serves the given remote over the network. Long: strings.ReplaceAll(`Create an NFS server that serves the given remote over the network.
This implements an NFSv3 server to serve any rclone remote via NFS. This implements an NFSv3 server to serve any rclone remote via NFS.
The primary purpose for this command is to enable the [mount The primary purpose for this command is to enable the [mount
@@ -179,12 +179,16 @@ cache.
To serve NFS over the network use following command: To serve NFS over the network use following command:
rclone serve nfs remote: --addr 0.0.0.0:$PORT --vfs-cache-mode=full |||sh
rclone serve nfs remote: --addr 0.0.0.0:$PORT --vfs-cache-mode=full
|||
This specifies a port that can be used in the mount command. To mount This specifies a port that can be used in the mount command. To mount
the server under Linux/macOS, use the following command: the server under Linux/macOS, use the following command:
mount -t nfs -o port=$PORT,mountport=$PORT,tcp $HOSTNAME:/ path/to/mountpoint |||sh
mount -t nfs -o port=$PORT,mountport=$PORT,tcp $HOSTNAME:/ path/to/mountpoint
|||
Where |$PORT| is the same port number used in the |serve nfs| command Where |$PORT| is the same port number used in the |serve nfs| command
and |$HOSTNAME| is the network address of the machine that |serve nfs| and |$HOSTNAME| is the network address of the machine that |serve nfs|
@@ -198,7 +202,7 @@ is desired.
This command is only available on Unix platforms. This command is only available on Unix platforms.
`, "|", "`") + vfs.Help(), `, "|", "`") + strings.TrimSpace(vfs.Help()),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.65", "versionIntroduced": "v1.65",
"groups": "Filter", "groups": "Filter",

View File

@@ -46,41 +46,43 @@ options - it is the job of the proxy program to make a complete
config. config.
This config generated must have this extra parameter This config generated must have this extra parameter
- |_root| - root to use for the backend - |_root| - root to use for the backend
And it may have this parameter And it may have this parameter
- |_obscure| - comma separated strings for parameters to obscure - |_obscure| - comma separated strings for parameters to obscure
If password authentication was used by the client, input to the proxy If password authentication was used by the client, input to the proxy
process (on STDIN) would look similar to this: process (on STDIN) would look similar to this:
||| |||json
{ {
"user": "me", "user": "me",
"pass": "mypassword" "pass": "mypassword"
} }
||| |||
If public-key authentication was used by the client, input to the If public-key authentication was used by the client, input to the
proxy process (on STDIN) would look similar to this: proxy process (on STDIN) would look similar to this:
||| |||json
{ {
"user": "me", "user": "me",
"public_key": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDuwESFdAe14hVS6omeyX7edc...JQdf" "public_key": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDuwESFdAe14hVS6omeyX7edc...JQdf"
} }
||| |||
And as an example return this on STDOUT And as an example return this on STDOUT
||| |||json
{ {
"type": "sftp", "type": "sftp",
"_root": "", "_root": "",
"_obscure": "pass", "_obscure": "pass",
"user": "me", "user": "me",
"pass": "mypassword", "pass": "mypassword",
"host": "sftp.example.com" "host": "sftp.example.com"
} }
||| |||
@@ -102,7 +104,7 @@ password or public-key is changed the cache will need to expire (which takes 5 m
before it takes effect. before it takes effect.
This can be used to build general purpose proxies to any kind of This can be used to build general purpose proxies to any kind of
backend that rclone supports. backend that rclone supports.
`, "|", "`") `, "|", "`")

View File

@@ -108,7 +108,7 @@ The server will log errors. Use -v to see access logs.
` + "`--bwlimit`" + ` will be respected for file transfers. ` + "`--bwlimit`" + ` will be respected for file transfers.
Use ` + "`--stats`" + ` to control the stats printing. Use ` + "`--stats`" + ` to control the stats printing.
### Setting up rclone for use by restic ### ### Setting up rclone for use by restic
First [set up a remote for your chosen cloud provider](/docs/#configure). First [set up a remote for your chosen cloud provider](/docs/#configure).
@@ -119,7 +119,9 @@ following instructions.
Now start the rclone restic server Now start the rclone restic server
rclone serve restic -v remote:backup ` + "```sh" + `
rclone serve restic -v remote:backup
` + "```" + `
Where you can replace "backup" in the above by whatever path in the Where you can replace "backup" in the above by whatever path in the
remote you wish to use. remote you wish to use.
@@ -133,7 +135,7 @@ Adding ` + "`--cache-objects=false`" + ` will cause rclone to stop caching objec
returned from the List call. Caching is normally desirable as it speeds returned from the List call. Caching is normally desirable as it speeds
up downloading objects, saves transactions and uses very little memory. up downloading objects, saves transactions and uses very little memory.
### Setting up restic to use rclone ### ### Setting up restic to use rclone
Now you can [follow the restic Now you can [follow the restic
instructions](http://restic.readthedocs.io/en/latest/030_preparing_a_new_repo.html#rest-server) instructions](http://restic.readthedocs.io/en/latest/030_preparing_a_new_repo.html#rest-server)
@@ -147,38 +149,43 @@ the URL for the REST server.
For example: For example:
$ export RESTIC_REPOSITORY=rest:http://localhost:8080/ ` + "```sh" + `
$ export RESTIC_PASSWORD=yourpassword $ export RESTIC_REPOSITORY=rest:http://localhost:8080/
$ restic init $ export RESTIC_PASSWORD=yourpassword
created restic backend 8b1a4b56ae at rest:http://localhost:8080/ $ restic init
created restic backend 8b1a4b56ae at rest:http://localhost:8080/
Please note that knowledge of your password is required to access Please note that knowledge of your password is required to access
the repository. Losing your password means that your data is the repository. Losing your password means that your data is
irrecoverably lost. irrecoverably lost.
$ restic backup /path/to/files/to/backup $ restic backup /path/to/files/to/backup
scan [/path/to/files/to/backup] scan [/path/to/files/to/backup]
scanned 189 directories, 312 files in 0:00 scanned 189 directories, 312 files in 0:00
[0:00] 100.00% 38.128 MiB / 38.128 MiB 501 / 501 items 0 errors ETA 0:00 [0:00] 100.00% 38.128 MiB / 38.128 MiB 501 / 501 items 0 errors ETA 0:00
duration: 0:00 duration: 0:00
snapshot 45c8fdd8 saved snapshot 45c8fdd8 saved
#### Multiple repositories #### ` + "```" + `
#### Multiple repositories
Note that you can use the endpoint to host multiple repositories. Do Note that you can use the endpoint to host multiple repositories. Do
this by adding a directory name or path after the URL. Note that this by adding a directory name or path after the URL. Note that
these **must** end with /. Eg these **must** end with /. Eg
$ export RESTIC_REPOSITORY=rest:http://localhost:8080/user1repo/ ` + "```sh" + `
# backup user1 stuff $ export RESTIC_REPOSITORY=rest:http://localhost:8080/user1repo/
$ export RESTIC_REPOSITORY=rest:http://localhost:8080/user2repo/ # backup user1 stuff
# backup user2 stuff $ export RESTIC_REPOSITORY=rest:http://localhost:8080/user2repo/
# backup user2 stuff
` + "```" + `
#### Private repositories #### #### Private repositories
The` + "`--private-repos`" + ` flag can be used to limit users to repositories starting The` + "`--private-repos`" + ` flag can be used to limit users to repositories starting
with a path of ` + "`/<username>/`" + `. with a path of ` + "`/<username>/`" + `.
` + libhttp.Help(flagPrefix) + libhttp.AuthHelp(flagPrefix), ` + strings.TrimSpace(libhttp.Help(flagPrefix)+libhttp.AuthHelp(flagPrefix)),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.40", "versionIntroduced": "v1.40",
}, },

View File

@@ -105,7 +105,7 @@ var Command = &cobra.Command{
}, },
Use: "s3 remote:path", Use: "s3 remote:path",
Short: `Serve remote:path over s3.`, Short: `Serve remote:path over s3.`,
Long: help() + httplib.AuthHelp(flagPrefix) + httplib.Help(flagPrefix) + vfs.Help(), Long: help() + strings.TrimSpace(httplib.AuthHelp(flagPrefix)+httplib.Help(flagPrefix)+vfs.Help()),
RunE: func(command *cobra.Command, args []string) error { RunE: func(command *cobra.Command, args []string) error {
var f fs.Fs var f fs.Fs
if proxy.Opt.AuthProxy == "" { if proxy.Opt.AuthProxy == "" {

View File

@@ -33,20 +33,20 @@ cause problems for S3 clients which rely on the Etag being the MD5.
For a simple set up, to serve `remote:path` over s3, run the server For a simple set up, to serve `remote:path` over s3, run the server
like this: like this:
``` ```sh
rclone serve s3 --auth-key ACCESS_KEY_ID,SECRET_ACCESS_KEY remote:path rclone serve s3 --auth-key ACCESS_KEY_ID,SECRET_ACCESS_KEY remote:path
``` ```
For example, to use a simple folder in the filesystem, run the server For example, to use a simple folder in the filesystem, run the server
with a command like this: with a command like this:
``` ```sh
rclone serve s3 --auth-key ACCESS_KEY_ID,SECRET_ACCESS_KEY local:/path/to/folder rclone serve s3 --auth-key ACCESS_KEY_ID,SECRET_ACCESS_KEY local:/path/to/folder
``` ```
The `rclone.conf` for the server could look like this: The `rclone.conf` for the server could look like this:
``` ```ini
[local] [local]
type = local type = local
``` ```
@@ -59,7 +59,7 @@ will be visible as a warning in the logs. But it will run nonetheless.
This will be compatible with an rclone (client) remote configuration which This will be compatible with an rclone (client) remote configuration which
is defined like this: is defined like this:
``` ```ini
[serves3] [serves3]
type = s3 type = s3
provider = Rclone provider = Rclone
@@ -116,20 +116,20 @@ metadata which will be set as the modification time of the file.
`serve s3` currently supports the following operations. `serve s3` currently supports the following operations.
- Bucket - Bucket
- `ListBuckets` - `ListBuckets`
- `CreateBucket` - `CreateBucket`
- `DeleteBucket` - `DeleteBucket`
- Object - Object
- `HeadObject` - `HeadObject`
- `ListObjects` - `ListObjects`
- `GetObject` - `GetObject`
- `PutObject` - `PutObject`
- `DeleteObject` - `DeleteObject`
- `DeleteObjects` - `DeleteObjects`
- `CreateMultipartUpload` - `CreateMultipartUpload`
- `CompleteMultipartUpload` - `CompleteMultipartUpload`
- `AbortMultipartUpload` - `AbortMultipartUpload`
- `CopyObject` - `CopyObject`
- `UploadPart` - `UploadPart`
Other operations will return error `Unimplemented`. Other operations will return error `Unimplemented`.

View File

@@ -19,10 +19,11 @@ var Command = &cobra.Command{
Long: `Serve a remote over a given protocol. Requires the use of a Long: `Serve a remote over a given protocol. Requires the use of a
subcommand to specify the protocol, e.g. subcommand to specify the protocol, e.g.
rclone serve http remote: ` + "```sh" + `
rclone serve http remote:
` + "```" + `
Each subcommand has its own options which you can see in their help. Each subcommand has its own options which you can see in their help.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.39", "versionIntroduced": "v1.39",
}, },

View File

@@ -6,6 +6,7 @@ package sftp
import ( import (
"context" "context"
"fmt" "fmt"
"strings"
"github.com/rclone/rclone/cmd" "github.com/rclone/rclone/cmd"
"github.com/rclone/rclone/cmd/serve" "github.com/rclone/rclone/cmd/serve"
@@ -146,11 +147,13 @@ reachable externally then supply ` + "`--addr :2022`" + ` for example.
This also supports being run with socket activation, in which case it will This also supports being run with socket activation, in which case it will
listen on the first passed FD. listen on the first passed FD.
It can be configured with .socket and .service unit files as described in It can be configured with .socket and .service unit files as described in
https://www.freedesktop.org/software/systemd/man/latest/systemd.socket.html <https://www.freedesktop.org/software/systemd/man/latest/systemd.socket.html>.
Socket activation can be tested ad-hoc with the ` + "`systemd-socket-activate`" + `command: Socket activation can be tested ad-hoc with the ` + "`systemd-socket-activate`" + `command:
systemd-socket-activate -l 2222 -- rclone serve sftp :local:vfs/ ` + "```sh" + `
systemd-socket-activate -l 2222 -- rclone serve sftp :local:vfs/
` + "```" + `
This will socket-activate rclone on the first connection to port 2222 over TCP. This will socket-activate rclone on the first connection to port 2222 over TCP.
@@ -160,7 +163,9 @@ sftp backend, but it may not be with other SFTP clients.
If ` + "`--stdio`" + ` is specified, rclone will serve SFTP over stdio, which can If ` + "`--stdio`" + ` is specified, rclone will serve SFTP over stdio, which can
be used with sshd via ~/.ssh/authorized_keys, for example: be used with sshd via ~/.ssh/authorized_keys, for example:
restrict,command="rclone serve sftp --stdio ./photos" ssh-rsa ... ` + "```text" + `
restrict,command="rclone serve sftp --stdio ./photos" ssh-rsa ...
` + "```" + `
On the client you need to set ` + "`--transfers 1`" + ` when using ` + "`--stdio`" + `. On the client you need to set ` + "`--transfers 1`" + ` when using ` + "`--stdio`" + `.
Otherwise multiple instances of the rclone server are started by OpenSSH Otherwise multiple instances of the rclone server are started by OpenSSH
@@ -173,7 +178,7 @@ used. Omitting "restrict" and using ` + "`--sftp-path-override`" + ` to enable
checksumming is possible but less secure and you could use the SFTP server checksumming is possible but less secure and you could use the SFTP server
provided by OpenSSH in this case. provided by OpenSSH in this case.
` + vfs.Help() + proxy.Help, ` + strings.TrimSpace(vfs.Help()+proxy.Help),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.48", "versionIntroduced": "v1.48",
"groups": "Filter", "groups": "Filter",

View File

@@ -107,7 +107,7 @@ browser, or you can make a remote of type WebDAV to read and write it.
### WebDAV options ### WebDAV options
#### --etag-hash #### --etag-hash
This controls the ETag header. Without this flag the ETag will be This controls the ETag header. Without this flag the ETag will be
based on the ModTime and Size of the object. based on the ModTime and Size of the object.
@@ -119,44 +119,58 @@ to see the full list.
### Access WebDAV on Windows ### Access WebDAV on Windows
WebDAV shared folder can be mapped as a drive on Windows, however the default settings prevent it. WebDAV shared folder can be mapped as a drive on Windows, however the default
Windows will fail to connect to the server using insecure Basic authentication. settings prevent it. Windows will fail to connect to the server using insecure
It will not even display any login dialog. Windows requires SSL / HTTPS connection to be used with Basic. Basic authentication. It will not even display any login dialog. Windows
If you try to connect via Add Network Location Wizard you will get the following error: requires SSL / HTTPS connection to be used with Basic. If you try to connect
via Add Network Location Wizard you will get the following error:
"The folder you entered does not appear to be valid. Please choose another". "The folder you entered does not appear to be valid. Please choose another".
However, you still can connect if you set the following registry key on a client machine: However, you still can connect if you set the following registry key on a
HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Services\WebClient\Parameters\BasicAuthLevel to 2. client machine:
The BasicAuthLevel can be set to the following values: ` + "`HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\WebClient\\Parameters\\BasicAuthLevel`" + `
0 - Basic authentication disabled to 2. The BasicAuthLevel can be set to the following values:
1 - Basic authentication enabled for SSL connections only
2 - Basic authentication enabled for SSL connections and for non-SSL connections ` + "```text" + `
0 - Basic authentication disabled
1 - Basic authentication enabled for SSL connections only
2 - Basic authentication enabled for SSL connections and for non-SSL connections
` + "```" + `
If required, increase the FileSizeLimitInBytes to a higher value. If required, increase the FileSizeLimitInBytes to a higher value.
Navigate to the Services interface, then restart the WebClient service. Navigate to the Services interface, then restart the WebClient service.
### Access Office applications on WebDAV ### Access Office applications on WebDAV
Navigate to following registry HKEY_CURRENT_USER\Software\Microsoft\Office\[14.0/15.0/16.0]\Common\Internet Navigate to following registry
` + "`HKEY_CURRENT_USER\\Software\\Microsoft\\Office\\[14.0/15.0/16.0]\\Common\\Internet`" + `
Create a new DWORD BasicAuthLevel with value 2. Create a new DWORD BasicAuthLevel with value 2.
0 - Basic authentication disabled
1 - Basic authentication enabled for SSL connections only
2 - Basic authentication enabled for SSL and for non-SSL connections
https://learn.microsoft.com/en-us/office/troubleshoot/powerpoint/office-opens-blank-from-sharepoint ` + "```text" + `
0 - Basic authentication disabled
1 - Basic authentication enabled for SSL connections only
2 - Basic authentication enabled for SSL and for non-SSL connections
` + "```" + `
<https://learn.microsoft.com/en-us/office/troubleshoot/powerpoint/office-opens-blank-from-sharepoint>
### Serving over a unix socket ### Serving over a unix socket
You can serve the webdav on a unix socket like this: You can serve the webdav on a unix socket like this:
rclone serve webdav --addr unix:///tmp/my.socket remote:path ` + "```sh" + `
rclone serve webdav --addr unix:///tmp/my.socket remote:path
` + "```" + `
and connect to it like this using rclone and the webdav backend: and connect to it like this using rclone and the webdav backend:
rclone --webdav-unix-socket /tmp/my.socket --webdav-url http://localhost lsf :webdav: ` + "```sh" + `
rclone --webdav-unix-socket /tmp/my.socket --webdav-url http://localhost lsf :webdav:
` + "```" + `
Note that there is no authentication on http protocol - this is expected to be Note that there is no authentication on http protocol - this is expected to be
done by the permissions on the socket. done by the permissions on the socket.
` + libhttp.Help(flagPrefix) + libhttp.TemplateHelp(flagPrefix) + libhttp.AuthHelp(flagPrefix) + vfs.Help() + proxy.Help, ` + strings.TrimSpace(libhttp.Help(flagPrefix)+libhttp.TemplateHelp(flagPrefix)+libhttp.AuthHelp(flagPrefix)+vfs.Help()+proxy.Help),
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.39", "versionIntroduced": "v1.39",
"groups": "Filter", "groups": "Filter",

View File

@@ -29,16 +29,21 @@ inaccessible.true
You can use it to tier single object You can use it to tier single object
rclone settier Cool remote:path/file ` + "```sh" + `
rclone settier Cool remote:path/file
` + "```" + `
Or use rclone filters to set tier on only specific files Or use rclone filters to set tier on only specific files
rclone --include "*.txt" settier Hot remote:path/dir ` + "```sh" + `
rclone --include "*.txt" settier Hot remote:path/dir
` + "```" + `
Or just provide remote directory and all files in directory will be tiered Or just provide remote directory and all files in directory will be tiered
rclone settier tier remote:path/dir ` + "```sh" + `
`, rclone settier tier remote:path/dir
` + "```",
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.44", "versionIntroduced": "v1.44",
}, },

View File

@@ -38,8 +38,7 @@ when there is data to read (if not, the hyphen will be treated literally,
as a relative path). as a relative path).
This command can also hash data received on STDIN, if not passing This command can also hash data received on STDIN, if not passing
a remote:path. a remote:path.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.27", "versionIntroduced": "v1.27",
"groups": "Filter,Listing", "groups": "Filter,Listing",

View File

@@ -41,8 +41,7 @@ Some backends do not always provide file sizes, see for example
[Google Docs](/drive/#limitations-of-google-docs). [Google Docs](/drive/#limitations-of-google-docs).
Rclone will then show a notice in the log indicating how many such Rclone will then show a notice in the log indicating how many such
files were encountered, and count them in as empty files in the output files were encountered, and count them in as empty files in the output
of the size command. of the size command.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.23", "versionIntroduced": "v1.23",
"groups": "Filter,Listing", "groups": "Filter,Listing",

View File

@@ -42,7 +42,9 @@ want to delete files from destination, use the
**Important**: Since this can cause data loss, test first with the **Important**: Since this can cause data loss, test first with the
|--dry-run| or the |--interactive|/|i| flag. |--dry-run| or the |--interactive|/|i| flag.
rclone sync --interactive SOURCE remote:DESTINATION |||sh
rclone sync --interactive SOURCE remote:DESTINATION
|||
Files in the destination won't be deleted if there were any errors at any Files in the destination won't be deleted if there were any errors at any
point. Duplicate objects (files with the same name, on those providers that point. Duplicate objects (files with the same name, on those providers that
@@ -59,7 +61,7 @@ If dest:path doesn't exist, it is created and the source:path contents
go there. go there.
It is not possible to sync overlapping remotes. However, you may exclude It is not possible to sync overlapping remotes. However, you may exclude
the destination from the sync with a filter rule or by putting an the destination from the sync with a filter rule or by putting an
exclude-if-present file inside the destination directory and sync to a exclude-if-present file inside the destination directory and sync to a
destination that is inside the source directory. destination that is inside the source directory.
@@ -68,13 +70,15 @@ the backend supports it. If metadata syncing is required then use the
|--metadata| flag. |--metadata| flag.
Note that the modification time and metadata for the root directory Note that the modification time and metadata for the root directory
will **not** be synced. See https://github.com/rclone/rclone/issues/7652 will **not** be synced. See <https://github.com/rclone/rclone/issues/7652>
for more info. for more info.
**Note**: Use the |-P|/|--progress| flag to view real-time transfer statistics **Note**: Use the |-P|/|--progress| flag to view real-time transfer statistics
**Note**: Use the |rclone dedupe| command to deal with "Duplicate object/directory found in source/destination - ignoring" errors. **Note**: Use the |rclone dedupe| command to deal with "Duplicate
See [this forum post](https://forum.rclone.org/t/sync-not-clearing-duplicates/14372) for more info. object/directory found in source/destination - ignoring" errors.
See [this forum post](https://forum.rclone.org/t/sync-not-clearing-duplicates/14372)
for more info.
`, "|", "`") + operationsflags.Help(), `, "|", "`") + operationsflags.Help(),
Annotations: map[string]string{ Annotations: map[string]string{

View File

@@ -26,8 +26,7 @@ var commandDefinition = &cobra.Command{
in filenames in the remote:path specified. in filenames in the remote:path specified.
The data doesn't contain any identifying information but is useful for The data doesn't contain any identifying information but is useful for
the rclone developers when developing filename compression. the rclone developers when developing filename compression.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.55", "versionIntroduced": "v1.55",
}, },

View File

@@ -68,8 +68,7 @@ paths passed in and how long they can be. It can take some time. It will
write test files into the remote:path passed in. It outputs a bit of go write test files into the remote:path passed in. It outputs a bit of go
code for each one. code for each one.
**NB** this can create undeletable files and other hazards - use with care **NB** this can create undeletable files and other hazards - use with care!`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.55", "versionIntroduced": "v1.55",
}, },

View File

@@ -18,13 +18,14 @@ var Command = &cobra.Command{
Select which test command you want with the subcommand, eg Select which test command you want with the subcommand, eg
rclone test memory remote: ` + "```sh" + `
rclone test memory remote:
` + "```" + `
Each subcommand has its own options which you can see in their help. Each subcommand has its own options which you can see in their help.
**NB** Be careful running these commands, they may do strange things **NB** Be careful running these commands, they may do strange things
so reading their documentation first is recommended. so reading their documentation first is recommended.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.55", "versionIntroduced": "v1.55",
}, },

View File

@@ -61,8 +61,7 @@ time instead of the current time. Times may be specified as one of:
- 'YYYY-MM-DDTHH:MM:SS.SSS' - e.g. 2006-01-02T15:04:05.123456789 - 'YYYY-MM-DDTHH:MM:SS.SSS' - e.g. 2006-01-02T15:04:05.123456789
Note that value of ` + "`--timestamp`" + ` is in UTC. If you want local time Note that value of ` + "`--timestamp`" + ` is in UTC. If you want local time
then add the ` + "`--localtime`" + ` flag. then add the ` + "`--localtime`" + ` flag.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.39", "versionIntroduced": "v1.39",
"groups": "Filter,Listing,Important", "groups": "Filter,Listing,Important",

View File

@@ -73,16 +73,18 @@ var commandDefinition = &cobra.Command{
For example For example
$ rclone tree remote:path ` + "```text" + `
/ $ rclone tree remote:path
├── file1 /
├── file2 ├── file1
├── file3 ├── file2
└── subdir ├── file3
├── file4 └── subdir
── file5 ── file4
└── file5
1 directories, 5 files 1 directories, 5 files
` + "```" + `
You can use any of the filtering options with the tree command (e.g. You can use any of the filtering options with the tree command (e.g.
` + "`--include` and `--exclude`" + `. You can also use ` + "`--fast-list`" + `. ` + "`--include` and `--exclude`" + `. You can also use ` + "`--fast-list`" + `.
@@ -93,8 +95,7 @@ sizes with ` + "`--size`" + `. Note that not all of them have
short options as they conflict with rclone's short options. short options as they conflict with rclone's short options.
For a more interactive navigation of the remote see the For a more interactive navigation of the remote see the
[ncdu](/commands/rclone_ncdu/) command. [ncdu](/commands/rclone_ncdu/) command.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.38", "versionIntroduced": "v1.38",
"groups": "Filter,Listing", "groups": "Filter,Listing",

View File

@@ -42,15 +42,17 @@ build tags and the type of executable (static or dynamic).
For example: For example:
$ rclone version ` + "```sh" + `
rclone v1.55.0 $ rclone version
- os/version: ubuntu 18.04 (64 bit) rclone v1.55.0
- os/kernel: 4.15.0-136-generic (x86_64) - os/version: ubuntu 18.04 (64 bit)
- os/type: linux - os/kernel: 4.15.0-136-generic (x86_64)
- os/arch: amd64 - os/type: linux
- go/version: go1.16 - os/arch: amd64
- go/linking: static - go/version: go1.16
- go/tags: none - go/linking: static
- go/tags: none
` + "```" + `
Note: before rclone version 1.55 the os/type and os/arch lines were merged, Note: before rclone version 1.55 the os/type and os/arch lines were merged,
and the "go/version" line was tagged as "go version". and the "go/version" line was tagged as "go version".
@@ -58,24 +60,27 @@ Note: before rclone version 1.55 the os/type and os/arch lines were merged,
If you supply the --check flag, then it will do an online check to If you supply the --check flag, then it will do an online check to
compare your version with the latest release and the latest beta. compare your version with the latest release and the latest beta.
$ rclone version --check ` + "```sh" + `
yours: 1.42.0.6 $ rclone version --check
latest: 1.42 (released 2018-06-16) yours: 1.42.0.6
beta: 1.42.0.5 (released 2018-06-17) latest: 1.42 (released 2018-06-16)
beta: 1.42.0.5 (released 2018-06-17)
` + "```" + `
Or Or
$ rclone version --check ` + "```sh" + `
yours: 1.41 $ rclone version --check
latest: 1.42 (released 2018-06-16) yours: 1.41
upgrade: https://downloads.rclone.org/v1.42 latest: 1.42 (released 2018-06-16)
beta: 1.42.0.5 (released 2018-06-17) upgrade: https://downloads.rclone.org/v1.42
upgrade: https://beta.rclone.org/v1.42-005-g56e1e820 beta: 1.42.0.5 (released 2018-06-17)
upgrade: https://beta.rclone.org/v1.42-005-g56e1e820
` + "```" + `
If you supply the --deps flag then rclone will print a list of all the If you supply the --deps flag then rclone will print a list of all the
packages it depends on and their versions along with some other packages it depends on and their versions along with some other
information about the build. information about the build.`,
`,
Annotations: map[string]string{ Annotations: map[string]string{
"versionIntroduced": "v1.33", "versionIntroduced": "v1.33",
}, },

View File

@@ -85,11 +85,11 @@ Rclone helps you:
## Features {#features} ## Features {#features}
- Transfers - Transfers
- MD5, SHA1 hashes are checked at all times for file integrity - MD5, SHA1 hashes are checked at all times for file integrity
- Timestamps are preserved on files - Timestamps are preserved on files
- Operations can be restarted at any time - Operations can be restarted at any time
- Can be to and from network, e.g. two different cloud providers - Can be to and from network, e.g. two different cloud providers
- Can use multi-threaded downloads to local disk - Can use multi-threaded downloads to local disk
- [Copy](/commands/rclone_copy/) new or changed files to cloud storage - [Copy](/commands/rclone_copy/) new or changed files to cloud storage
- [Sync](/commands/rclone_sync/) (one way) to make a directory identical - [Sync](/commands/rclone_sync/) (one way) to make a directory identical
- [Bisync](/bisync/) (two way) to keep two directories in sync bidirectionally - [Bisync](/bisync/) (two way) to keep two directories in sync bidirectionally
@@ -136,6 +136,7 @@ WebDAV or S3, that work out of the box.)
{{< provider name="Hetzner Storage Box" home="https://www.hetzner.com/storage/storage-box" config="/sftp/#hetzner-storage-box" >}} {{< provider name="Hetzner Storage Box" home="https://www.hetzner.com/storage/storage-box" config="/sftp/#hetzner-storage-box" >}}
{{< provider name="HiDrive" home="https://www.strato.de/cloud-speicher/" config="/hidrive/" >}} {{< provider name="HiDrive" home="https://www.strato.de/cloud-speicher/" config="/hidrive/" >}}
{{< provider name="HTTP" home="https://en.wikipedia.org/wiki/Hypertext_Transfer_Protocol" config="/http/" >}} {{< provider name="HTTP" home="https://en.wikipedia.org/wiki/Hypertext_Transfer_Protocol" config="/http/" >}}
{{< provider name="Huawei OBS" home="https://www.huaweicloud.com/intl/en-us/product/obs.html" config="/s3/#huawei-obs" >}}
{{< provider name="iCloud Drive" home="https://icloud.com/" config="/iclouddrive/" >}} {{< provider name="iCloud Drive" home="https://icloud.com/" config="/iclouddrive/" >}}
{{< provider name="ImageKit" home="https://imagekit.io" config="/imagekit/" >}} {{< provider name="ImageKit" home="https://imagekit.io" config="/imagekit/" >}}
{{< provider name="Internet Archive" home="https://archive.org/" config="/internetarchive/" >}} {{< provider name="Internet Archive" home="https://archive.org/" config="/internetarchive/" >}}
@@ -178,7 +179,9 @@ WebDAV or S3, that work out of the box.)
{{< provider name="QingStor" home="https://www.qingcloud.com/products/storage" config="/qingstor/" >}} {{< provider name="QingStor" home="https://www.qingcloud.com/products/storage" config="/qingstor/" >}}
{{< provider name="Qiniu Cloud Object Storage (Kodo)" home="https://www.qiniu.com/en/products/kodo" config="/s3/#qiniu" >}} {{< provider name="Qiniu Cloud Object Storage (Kodo)" home="https://www.qiniu.com/en/products/kodo" config="/s3/#qiniu" >}}
{{< provider name="Quatrix by Maytech" home="https://www.maytech.net/products/quatrix-business" config="/quatrix/" >}} {{< provider name="Quatrix by Maytech" home="https://www.maytech.net/products/quatrix-business" config="/quatrix/" >}}
{{< provider name="RackCorp Object Storage" home="https://www.rackcorp.com/" config="/s3/#RackCorp" >}}
{{< provider name="Rackspace Cloud Files" home="https://www.rackspace.com/cloud/files" config="/swift/" >}} {{< provider name="Rackspace Cloud Files" home="https://www.rackspace.com/cloud/files" config="/swift/" >}}
{{< provider name="Rclone Serve S3" home="/commands/rclone_serve_s3/" config="/s3/#rclone" >}}
{{< provider name="rsync.net" home="https://rsync.net/products/rclone.html" config="/sftp/#rsync-net" >}} {{< provider name="rsync.net" home="https://rsync.net/products/rclone.html" config="/sftp/#rsync-net" >}}
{{< provider name="Scaleway" home="https://www.scaleway.com/object-storage/" config="/s3/#scaleway" >}} {{< provider name="Scaleway" home="https://www.scaleway.com/object-storage/" config="/s3/#scaleway" >}}
{{< provider name="Seafile" home="https://www.seafile.com/" config="/seafile/" >}} {{< provider name="Seafile" home="https://www.seafile.com/" config="/seafile/" >}}
@@ -216,10 +219,9 @@ These backends adapt or modify other storage providers:
{{< provider name="Hasher: Hash files" home="/hasher/" config="/hasher/" >}} {{< provider name="Hasher: Hash files" home="/hasher/" config="/hasher/" >}}
{{< provider name="Union: Join multiple remotes to work together" home="/union/" config="/union/" >}} {{< provider name="Union: Join multiple remotes to work together" home="/union/" config="/union/" >}}
## Links ## Links
* {{< icon "fa fa-home" >}} [Home page](https://rclone.org/) - {{< icon "fa fa-home" >}} [Home page](https://rclone.org/)
* {{< icon "fab fa-github" >}} [GitHub project page for source and bug tracker](https://github.com/rclone/rclone) - {{< icon "fab fa-github" >}} [GitHub project page for source and bug tracker](https://github.com/rclone/rclone)
* {{< icon "fa fa-comments" >}} [Rclone Forum](https://forum.rclone.org) - {{< icon "fa fa-comments" >}} [Rclone Forum](https://forum.rclone.org)
* {{< icon "fas fa-cloud-download-alt" >}}[Downloads](/downloads/) - {{< icon "fas fa-cloud-download-alt" >}}[Downloads](/downloads/)

View File

@@ -8,7 +8,7 @@ versionIntroduced: "v1.40"
The `alias` remote provides a new name for another remote. The `alias` remote provides a new name for another remote.
Paths may be as deep as required or a local path, Paths may be as deep as required or a local path,
e.g. `remote:directory/subdirectory` or `/directory/subdirectory`. e.g. `remote:directory/subdirectory` or `/directory/subdirectory`.
During the initial setup with `rclone config` you will specify the target During the initial setup with `rclone config` you will specify the target
@@ -24,9 +24,9 @@ Invoking `rclone mkdir backup:../desktop` is exactly the same as invoking
The empty path is not allowed as a remote. To alias the current directory The empty path is not allowed as a remote. To alias the current directory
use `.` instead. use `.` instead.
The target remote can also be a [connection string](/docs/#connection-strings). The target remote can also be a [connection string](/docs/#connection-strings).
This can be used to modify the config of a remote for different uses, e.g. This can be used to modify the config of a remote for different uses, e.g.
the alias `myDriveTrash` with the target remote `myDrive,trashed_only:` the alias `myDriveTrash` with the target remote `myDrive,trashed_only:`
can be used to only show the trashed files in `myDrive`. can be used to only show the trashed files in `myDrive`.
## Configuration ## Configuration
@@ -34,11 +34,13 @@ can be used to only show the trashed files in `myDrive`.
Here is an example of how to make an alias called `remote` for local folder. Here is an example of how to make an alias called `remote` for local folder.
First run: First run:
rclone config ```sh
rclone config
```
This will guide you through an interactive setup process: This will guide you through an interactive setup process:
``` ```text
No remotes found, make a new one? No remotes found, make a new one?
n) New remote n) New remote
s) Set configuration password s) Set configuration password
@@ -85,15 +87,21 @@ Once configured you can then use `rclone` like this,
List directories in top level in `/mnt/storage/backup` List directories in top level in `/mnt/storage/backup`
rclone lsd remote: ```sh
rclone lsd remote:
```
List all the files in `/mnt/storage/backup` List all the files in `/mnt/storage/backup`
rclone ls remote: ```sh
rclone ls remote:
```
Copy another local directory to the alias directory called source Copy another local directory to the alias directory called source
rclone copy /home/source remote:source ```sh
rclone copy /home/source remote:source
```
{{< rem autogenerated options start" - DO NOT EDIT - instead edit fs.RegInfo in backend/alias/alias.go then run make backenddocs" >}} {{< rem autogenerated options start" - DO NOT EDIT - instead edit fs.RegInfo in backend/alias/alias.go then run make backenddocs" >}}
### Standard options ### Standard options

View File

@@ -15,11 +15,13 @@ command.) You may put subdirectories in too, e.g.
Here is an example of making a Microsoft Azure Blob Storage Here is an example of making a Microsoft Azure Blob Storage
configuration. For a remote called `remote`. First run: configuration. For a remote called `remote`. First run:
rclone config ```sh
rclone config
```
This will guide you through an interactive setup process: This will guide you through an interactive setup process:
``` ```text
No remotes found, make a new one? No remotes found, make a new one?
n) New remote n) New remote
s) Set configuration password s) Set configuration password
@@ -55,20 +57,28 @@ y/e/d> y
See all containers See all containers
rclone lsd remote: ```sh
rclone lsd remote:
```
Make a new container Make a new container
rclone mkdir remote:container ```sh
rclone mkdir remote:container
```
List the contents of a container List the contents of a container
rclone ls remote:container ```sh
rclone ls remote:container
```
Sync `/home/local/directory` to the remote container, deleting any excess Sync `/home/local/directory` to the remote container, deleting any excess
files in the container. files in the container.
rclone sync --interactive /home/local/directory remote:container ```sh
rclone sync --interactive /home/local/directory remote:container
```
### --fast-list ### --fast-list
@@ -147,26 +157,35 @@ user with a password, depending on which environment variable are set.
It reads configuration from these variables, in the following order: It reads configuration from these variables, in the following order:
1. Service principal with client secret 1. Service principal with client secret
- `AZURE_TENANT_ID`: ID of the service principal's tenant. Also called its "directory" ID. - `AZURE_TENANT_ID`: ID of the service principal's tenant. Also called its
"directory" ID.
- `AZURE_CLIENT_ID`: the service principal's client ID - `AZURE_CLIENT_ID`: the service principal's client ID
- `AZURE_CLIENT_SECRET`: one of the service principal's client secrets - `AZURE_CLIENT_SECRET`: one of the service principal's client secrets
2. Service principal with certificate 2. Service principal with certificate
- `AZURE_TENANT_ID`: ID of the service principal's tenant. Also called its "directory" ID. - `AZURE_TENANT_ID`: ID of the service principal's tenant. Also called its
"directory" ID.
- `AZURE_CLIENT_ID`: the service principal's client ID - `AZURE_CLIENT_ID`: the service principal's client ID
- `AZURE_CLIENT_CERTIFICATE_PATH`: path to a PEM or PKCS12 certificate file including the private key. - `AZURE_CLIENT_CERTIFICATE_PATH`: path to a PEM or PKCS12 certificate file
- `AZURE_CLIENT_CERTIFICATE_PASSWORD`: (optional) password for the certificate file. including the private key.
- `AZURE_CLIENT_SEND_CERTIFICATE_CHAIN`: (optional) Specifies whether an authentication request will include an x5c header to support subject name / issuer based authentication. When set to "true" or "1", authentication requests include the x5c header. - `AZURE_CLIENT_CERTIFICATE_PASSWORD`: (optional) password for the
certificate file.
- `AZURE_CLIENT_SEND_CERTIFICATE_CHAIN`: (optional) Specifies whether an
authentication request will include an x5c header to support subject
name / issuer based authentication. When set to "true" or "1",
authentication requests include the x5c header.
3. User with username and password 3. User with username and password
- `AZURE_TENANT_ID`: (optional) tenant to authenticate in. Defaults to "organizations". - `AZURE_TENANT_ID`: (optional) tenant to authenticate in. Defaults to "organizations".
- `AZURE_CLIENT_ID`: client ID of the application the user will authenticate to - `AZURE_CLIENT_ID`: client ID of the application the user will authenticate
to
- `AZURE_USERNAME`: a username (usually an email address) - `AZURE_USERNAME`: a username (usually an email address)
- `AZURE_PASSWORD`: the user's password - `AZURE_PASSWORD`: the user's password
4. Workload Identity 4. Workload Identity
- `AZURE_TENANT_ID`: Tenant to authenticate in. - `AZURE_TENANT_ID`: Tenant to authenticate in
- `AZURE_CLIENT_ID`: Client ID of the application the user will authenticate to. - `AZURE_CLIENT_ID`: Client ID of the application the user will authenticate
- `AZURE_FEDERATED_TOKEN_FILE`: Path to projected service account token file. to
- `AZURE_AUTHORITY_HOST`: Authority of an Azure Active Directory endpoint (default: login.microsoftonline.com). - `AZURE_FEDERATED_TOKEN_FILE`: Path to projected service account token file
- `AZURE_AUTHORITY_HOST`: Authority of an Azure Active Directory endpoint
(default: login.microsoftonline.com).
##### Env Auth: 2. Managed Service Identity Credentials ##### Env Auth: 2. Managed Service Identity Credentials
@@ -193,19 +212,27 @@ Credentials created with the `az` tool can be picked up using `env_auth`.
For example if you were to login with a service principal like this: For example if you were to login with a service principal like this:
az login --service-principal -u XXX -p XXX --tenant XXX ```sh
az login --service-principal -u XXX -p XXX --tenant XXX
```
Then you could access rclone resources like this: Then you could access rclone resources like this:
rclone lsf :azureblob,env_auth,account=ACCOUNT:CONTAINER ```sh
rclone lsf :azureblob,env_auth,account=ACCOUNT:CONTAINER
```
Or Or
rclone lsf --azureblob-env-auth --azureblob-account=ACCOUNT :azureblob:CONTAINER ```sh
rclone lsf --azureblob-env-auth --azureblob-account=ACCOUNT :azureblob:CONTAINER
```
Which is analogous to using the `az` tool: Which is analogous to using the `az` tool:
az storage blob list --container-name CONTAINER --account-name ACCOUNT --auth-mode login ```sh
az storage blob list --container-name CONTAINER --account-name ACCOUNT --auth-mode login
```
#### Account and Shared Key #### Account and Shared Key
@@ -226,18 +253,24 @@ explorer in the Azure portal.
If you use a container level SAS URL, rclone operations are permitted If you use a container level SAS URL, rclone operations are permitted
only on a particular container, e.g. only on a particular container, e.g.
rclone ls azureblob:container ```sh
rclone ls azureblob:container
```
You can also list the single container from the root. This will only You can also list the single container from the root. This will only
show the container specified by the SAS URL. show the container specified by the SAS URL.
$ rclone lsd azureblob: ```sh
container/ $ rclone lsd azureblob:
container/
```
Note that you can't see or access any other containers - this will Note that you can't see or access any other containers - this will
fail fail
rclone ls azureblob:othercontainer ```sh
rclone ls azureblob:othercontainer
```
Container level SAS URLs are useful for temporarily allowing third Container level SAS URLs are useful for temporarily allowing third
parties access to a single container or putting credentials into an parties access to a single container or putting credentials into an
@@ -245,7 +278,8 @@ untrusted environment such as a CI build server.
#### Service principal with client secret #### Service principal with client secret
If these variables are set, rclone will authenticate with a service principal with a client secret. If these variables are set, rclone will authenticate with a service principal
with a client secret.
- `tenant`: ID of the service principal's tenant. Also called its "directory" ID. - `tenant`: ID of the service principal's tenant. Also called its "directory" ID.
- `client_id`: the service principal's client ID - `client_id`: the service principal's client ID
@@ -256,13 +290,18 @@ The credentials can also be placed in a file using the
#### Service principal with certificate #### Service principal with certificate
If these variables are set, rclone will authenticate with a service principal with certificate. If these variables are set, rclone will authenticate with a service principal
with certificate.
- `tenant`: ID of the service principal's tenant. Also called its "directory" ID. - `tenant`: ID of the service principal's tenant. Also called its "directory" ID.
- `client_id`: the service principal's client ID - `client_id`: the service principal's client ID
- `client_certificate_path`: path to a PEM or PKCS12 certificate file including the private key. - `client_certificate_path`: path to a PEM or PKCS12 certificate file including
the private key.
- `client_certificate_password`: (optional) password for the certificate file. - `client_certificate_password`: (optional) password for the certificate file.
- `client_send_certificate_chain`: (optional) Specifies whether an authentication request will include an x5c header to support subject name / issuer based authentication. When set to "true" or "1", authentication requests include the x5c header. - `client_send_certificate_chain`: (optional) Specifies whether an
authentication request will include an x5c header to support subject name /
issuer based authentication. When set to "true" or "1", authentication
requests include the x5c header.
**NB** `client_certificate_password` must be obscured - see [rclone obscure](/commands/rclone_obscure/). **NB** `client_certificate_password` must be obscured - see [rclone obscure](/commands/rclone_obscure/).
@@ -297,15 +336,18 @@ be explicitly specified using exactly one of the `msi_object_id`,
If none of `msi_object_id`, `msi_client_id`, or `msi_mi_res_id` is If none of `msi_object_id`, `msi_client_id`, or `msi_mi_res_id` is
set, this is is equivalent to using `env_auth`. set, this is is equivalent to using `env_auth`.
#### Fedrated Identity Credentials #### Fedrated Identity Credentials
If these variables are set, rclone will authenticate with fedrated identity. If these variables are set, rclone will authenticate with fedrated identity.
- `tenant_id`: tenant_id to authenticate in storage - `tenant_id`: tenant_id to authenticate in storage
- `client_id`: client ID of the application the user will authenticate to storage - `client_id`: client ID of the application the user will authenticate to storage
- `msi_client_id`: managed identity client ID of the application the user will authenticate to - `msi_client_id`: managed identity client ID of the application the user will
authenticate to
By default "api://AzureADTokenExchange" is used as scope for token retrieval over MSI. This token is then exchanged for actual storage token using 'tenant_id' and 'client_id'. By default "api://AzureADTokenExchange" is used as scope for token retrieval
over MSI. This token is then exchanged for actual storage token using
'tenant_id' and 'client_id'.
#### Azure CLI tool `az` {#use_az} #### Azure CLI tool `az` {#use_az}
@@ -322,7 +364,9 @@ Don't set `env_auth` at the same time.
If you want to access resources with public anonymous access then set If you want to access resources with public anonymous access then set
`account` only. You can do this without making an rclone config: `account` only. You can do this without making an rclone config:
rclone lsf :azureblob,account=ACCOUNT:CONTAINER ```sh
rclone lsf :azureblob,account=ACCOUNT:CONTAINER
```
{{< rem autogenerated options start" - DO NOT EDIT - instead edit fs.RegInfo in backend/azureblob/azureblob.go then run make backenddocs" >}} {{< rem autogenerated options start" - DO NOT EDIT - instead edit fs.RegInfo in backend/azureblob/azureblob.go then run make backenddocs" >}}
### Standard options ### Standard options

View File

@@ -14,11 +14,13 @@ e.g. `remote:path/to/dir`.
Here is an example of making a Microsoft Azure Files Storage Here is an example of making a Microsoft Azure Files Storage
configuration. For a remote called `remote`. First run: configuration. For a remote called `remote`. First run:
rclone config ```sh
rclone config
```
This will guide you through an interactive setup process: This will guide you through an interactive setup process:
``` ```text
No remotes found, make a new one? No remotes found, make a new one?
n) New remote n) New remote
s) Set configuration password s) Set configuration password
@@ -88,20 +90,28 @@ Once configured you can use rclone.
See all files in the top level: See all files in the top level:
rclone lsf remote: ```sh
rclone lsf remote:
```
Make a new directory in the root: Make a new directory in the root:
rclone mkdir remote:dir ```sh
rclone mkdir remote:dir
```
Recursively List the contents: Recursively List the contents:
rclone ls remote: ```sh
rclone ls remote:
```
Sync `/home/local/directory` to the remote directory, deleting any Sync `/home/local/directory` to the remote directory, deleting any
excess files in the directory. excess files in the directory.
rclone sync --interactive /home/local/directory remote:dir ```sh
rclone sync --interactive /home/local/directory remote:dir
```
### Modified time ### Modified time
@@ -173,26 +183,35 @@ user with a password, depending on which environment variable are set.
It reads configuration from these variables, in the following order: It reads configuration from these variables, in the following order:
1. Service principal with client secret 1. Service principal with client secret
- `AZURE_TENANT_ID`: ID of the service principal's tenant. Also called its "directory" ID. - `AZURE_TENANT_ID`: ID of the service principal's tenant. Also called its
"directory" ID.
- `AZURE_CLIENT_ID`: the service principal's client ID - `AZURE_CLIENT_ID`: the service principal's client ID
- `AZURE_CLIENT_SECRET`: one of the service principal's client secrets - `AZURE_CLIENT_SECRET`: one of the service principal's client secrets
2. Service principal with certificate 2. Service principal with certificate
- `AZURE_TENANT_ID`: ID of the service principal's tenant. Also called its "directory" ID. - `AZURE_TENANT_ID`: ID of the service principal's tenant. Also called its
"directory" ID.
- `AZURE_CLIENT_ID`: the service principal's client ID - `AZURE_CLIENT_ID`: the service principal's client ID
- `AZURE_CLIENT_CERTIFICATE_PATH`: path to a PEM or PKCS12 certificate file including the private key. - `AZURE_CLIENT_CERTIFICATE_PATH`: path to a PEM or PKCS12 certificate file
- `AZURE_CLIENT_CERTIFICATE_PASSWORD`: (optional) password for the certificate file. including the private key.
- `AZURE_CLIENT_SEND_CERTIFICATE_CHAIN`: (optional) Specifies whether an authentication request will include an x5c header to support subject name / issuer based authentication. When set to "true" or "1", authentication requests include the x5c header. - `AZURE_CLIENT_CERTIFICATE_PASSWORD`: (optional) password for the
certificate file.
- `AZURE_CLIENT_SEND_CERTIFICATE_CHAIN`: (optional) Specifies whether an
authentication request will include an x5c header to support subject
name / issuer based authentication. When set to "true" or "1",
authentication requests include the x5c header.
3. User with username and password 3. User with username and password
- `AZURE_TENANT_ID`: (optional) tenant to authenticate in. Defaults to "organizations". - `AZURE_TENANT_ID`: (optional) tenant to authenticate in. Defaults to "organizations".
- `AZURE_CLIENT_ID`: client ID of the application the user will authenticate to - `AZURE_CLIENT_ID`: client ID of the application the user will authenticate
to
- `AZURE_USERNAME`: a username (usually an email address) - `AZURE_USERNAME`: a username (usually an email address)
- `AZURE_PASSWORD`: the user's password - `AZURE_PASSWORD`: the user's password
4. Workload Identity 4. Workload Identity
- `AZURE_TENANT_ID`: Tenant to authenticate in. - `AZURE_TENANT_ID`: Tenant to authenticate in
- `AZURE_CLIENT_ID`: Client ID of the application the user will authenticate to. - `AZURE_CLIENT_ID`: Client ID of the application the user will authenticate
- `AZURE_FEDERATED_TOKEN_FILE`: Path to projected service account token file. to
- `AZURE_AUTHORITY_HOST`: Authority of an Azure Active Directory endpoint (default: login.microsoftonline.com). - `AZURE_FEDERATED_TOKEN_FILE`: Path to projected service account token file
- `AZURE_AUTHORITY_HOST`: Authority of an Azure Active Directory endpoint
(default: login.microsoftonline.com).
##### Env Auth: 2. Managed Service Identity Credentials ##### Env Auth: 2. Managed Service Identity Credentials
@@ -219,15 +238,21 @@ Credentials created with the `az` tool can be picked up using `env_auth`.
For example if you were to login with a service principal like this: For example if you were to login with a service principal like this:
az login --service-principal -u XXX -p XXX --tenant XXX ```sh
az login --service-principal -u XXX -p XXX --tenant XXX
```
Then you could access rclone resources like this: Then you could access rclone resources like this:
rclone lsf :azurefiles,env_auth,account=ACCOUNT: ```sh
rclone lsf :azurefiles,env_auth,account=ACCOUNT:
```
Or Or
rclone lsf --azurefiles-env-auth --azurefiles-account=ACCOUNT :azurefiles: ```sh
rclone lsf --azurefiles-env-auth --azurefiles-account=ACCOUNT :azurefiles:
```
#### Account and Shared Key #### Account and Shared Key
@@ -244,7 +269,8 @@ To use it leave `account`, `key` and "sas_url" blank and fill in `connection_str
#### Service principal with client secret #### Service principal with client secret
If these variables are set, rclone will authenticate with a service principal with a client secret. If these variables are set, rclone will authenticate with a service principal
with a client secret.
- `tenant`: ID of the service principal's tenant. Also called its "directory" ID. - `tenant`: ID of the service principal's tenant. Also called its "directory" ID.
- `client_id`: the service principal's client ID - `client_id`: the service principal's client ID
@@ -255,13 +281,18 @@ The credentials can also be placed in a file using the
#### Service principal with certificate #### Service principal with certificate
If these variables are set, rclone will authenticate with a service principal with certificate. If these variables are set, rclone will authenticate with a service principal
with certificate.
- `tenant`: ID of the service principal's tenant. Also called its "directory" ID. - `tenant`: ID of the service principal's tenant. Also called its "directory" ID.
- `client_id`: the service principal's client ID - `client_id`: the service principal's client ID
- `client_certificate_path`: path to a PEM or PKCS12 certificate file including the private key. - `client_certificate_path`: path to a PEM or PKCS12 certificate file including
the private key.
- `client_certificate_password`: (optional) password for the certificate file. - `client_certificate_password`: (optional) password for the certificate file.
- `client_send_certificate_chain`: (optional) Specifies whether an authentication request will include an x5c header to support subject name / issuer based authentication. When set to "true" or "1", authentication requests include the x5c header. - `client_send_certificate_chain`: (optional) Specifies whether an authentication
request will include an x5c header to support subject name / issuer based
authentication. When set to "true" or "1", authentication requests include
the x5c header.
**NB** `client_certificate_password` must be obscured - see [rclone obscure](/commands/rclone_obscure/). **NB** `client_certificate_password` must be obscured - see [rclone obscure](/commands/rclone_obscure/).
@@ -296,17 +327,21 @@ be explicitly specified using exactly one of the `msi_object_id`,
If none of `msi_object_id`, `msi_client_id`, or `msi_mi_res_id` is If none of `msi_object_id`, `msi_client_id`, or `msi_mi_res_id` is
set, this is is equivalent to using `env_auth`. set, this is is equivalent to using `env_auth`.
#### Fedrated Identity Credentials #### Fedrated Identity Credentials
If these variables are set, rclone will authenticate with fedrated identity. If these variables are set, rclone will authenticate with fedrated identity.
- `tenant_id`: tenant_id to authenticate in storage - `tenant_id`: tenant_id to authenticate in storage
- `client_id`: client ID of the application the user will authenticate to storage - `client_id`: client ID of the application the user will authenticate to storage
- `msi_client_id`: managed identity client ID of the application the user will authenticate to - `msi_client_id`: managed identity client ID of the application the user will
authenticate to
By default "api://AzureADTokenExchange" is used as scope for token retrieval
over MSI. This token is then exchanged for actual storage token using 'tenant_id'
and 'client_id'.
By default "api://AzureADTokenExchange" is used as scope for token retrieval over MSI. This token is then exchanged for actual storage token using 'tenant_id' and 'client_id'.
#### Azure CLI tool `az` {#use_az} #### Azure CLI tool `az` {#use_az}
Set to use the [Azure CLI tool `az`](https://learn.microsoft.com/en-us/cli/azure/) Set to use the [Azure CLI tool `az`](https://learn.microsoft.com/en-us/cli/azure/)
as the sole means of authentication. as the sole means of authentication.
Setting this can be useful if you wish to use the `az` CLI on a host with Setting this can be useful if you wish to use the `az` CLI on a host with

View File

@@ -15,7 +15,9 @@ command.) You may put subdirectories in too, e.g. `remote:bucket/path/to/dir`.
Here is an example of making a b2 configuration. First run Here is an example of making a b2 configuration. First run
rclone config ```sh
rclone config
```
This will guide you through an interactive setup process. To authenticate This will guide you through an interactive setup process. To authenticate
you will either need your Account ID (a short hex number) and Master you will either need your Account ID (a short hex number) and Master
@@ -23,8 +25,8 @@ Application Key (a long hex number) OR an Application Key, which is the
recommended method. See below for further details on generating and using recommended method. See below for further details on generating and using
an Application Key. an Application Key.
``` ```text
No remotes found, make a new one? No remotes found, make a new one\?
n) New remote n) New remote
q) Quit config q) Quit config
n/q> n n/q> n
@@ -60,20 +62,29 @@ This remote is called `remote` and can now be used like this
See all buckets See all buckets
rclone lsd remote: ```sh
rclone lsd remote:
```
Create a new bucket Create a new bucket
rclone mkdir remote:bucket ```sh
rclone mkdir remote:bucket
```
List the contents of a bucket List the contents of a bucket
rclone ls remote:bucket ```sh
rclone ls remote:bucket
```
Sync `/home/local/directory` to the remote bucket, deleting any Sync `/home/local/directory` to the remote bucket, deleting any
excess files in the bucket. excess files in the bucket.
rclone sync --interactive /home/local/directory remote:bucket ```sh
rclone sync --interactive /home/local/directory remote:bucket
```
### Application Keys ### Application Keys
@@ -219,7 +230,7 @@ version followed by a `cleanup` of the old versions.
Show current version and all the versions with `--b2-versions` flag. Show current version and all the versions with `--b2-versions` flag.
``` ```sh
$ rclone -q ls b2:cleanup-test $ rclone -q ls b2:cleanup-test
9 one.txt 9 one.txt
@@ -232,7 +243,7 @@ $ rclone -q --b2-versions ls b2:cleanup-test
Retrieve an old version Retrieve an old version
``` ```sh
$ rclone -q --b2-versions copy b2:cleanup-test/one-v2016-07-04-141003-000.txt /tmp $ rclone -q --b2-versions copy b2:cleanup-test/one-v2016-07-04-141003-000.txt /tmp
$ ls -l /tmp/one-v2016-07-04-141003-000.txt $ ls -l /tmp/one-v2016-07-04-141003-000.txt
@@ -241,7 +252,7 @@ $ ls -l /tmp/one-v2016-07-04-141003-000.txt
Clean up all the old versions and show that they've gone. Clean up all the old versions and show that they've gone.
``` ```sh
$ rclone -q cleanup b2:cleanup-test $ rclone -q cleanup b2:cleanup-test
$ rclone -q ls b2:cleanup-test $ rclone -q ls b2:cleanup-test
@@ -256,11 +267,13 @@ $ rclone -q --b2-versions ls b2:cleanup-test
When using `--b2-versions` flag rclone is relying on the file name When using `--b2-versions` flag rclone is relying on the file name
to work out whether the objects are versions or not. Versions' names to work out whether the objects are versions or not. Versions' names
are created by inserting timestamp between file name and its extension. are created by inserting timestamp between file name and its extension.
```
```sh
9 file.txt 9 file.txt
8 file-v2023-07-17-161032-000.txt 8 file-v2023-07-17-161032-000.txt
16 file-v2023-06-15-141003-000.txt 16 file-v2023-06-15-141003-000.txt
``` ```
If there are real files present with the same names as versions, then If there are real files present with the same names as versions, then
behaviour of `--b2-versions` can be unpredictable. behaviour of `--b2-versions` can be unpredictable.
@@ -270,7 +283,7 @@ It is useful to know how many requests are sent to the server in different scena
All copy commands send the following 4 requests: All copy commands send the following 4 requests:
``` ```text
/b2api/v1/b2_authorize_account /b2api/v1/b2_authorize_account
/b2api/v1/b2_create_bucket /b2api/v1/b2_create_bucket
/b2api/v1/b2_list_buckets /b2api/v1/b2_list_buckets
@@ -287,7 +300,7 @@ require any files to be uploaded, no more requests will be sent.
Uploading files that do not require chunking, will send 2 requests per Uploading files that do not require chunking, will send 2 requests per
file upload: file upload:
``` ```text
/b2api/v1/b2_get_upload_url /b2api/v1/b2_get_upload_url
/b2api/v1/b2_upload_file/ /b2api/v1/b2_upload_file/
``` ```
@@ -295,7 +308,7 @@ file upload:
Uploading files requiring chunking, will send 2 requests (one each to Uploading files requiring chunking, will send 2 requests (one each to
start and finish the upload) and another 2 requests for each chunk: start and finish the upload) and another 2 requests for each chunk:
``` ```text
/b2api/v1/b2_start_large_file /b2api/v1/b2_start_large_file
/b2api/v1/b2_get_upload_part_url /b2api/v1/b2_get_upload_part_url
/b2api/v1/b2_upload_part/ /b2api/v1/b2_upload_part/
@@ -309,14 +322,14 @@ rclone will show and act on older versions of files. For example
Listing without `--b2-versions` Listing without `--b2-versions`
``` ```sh
$ rclone -q ls b2:cleanup-test $ rclone -q ls b2:cleanup-test
9 one.txt 9 one.txt
``` ```
And with And with
``` ```sh
$ rclone -q --b2-versions ls b2:cleanup-test $ rclone -q --b2-versions ls b2:cleanup-test
9 one.txt 9 one.txt
8 one-v2016-07-04-141032-000.txt 8 one-v2016-07-04-141032-000.txt
@@ -336,7 +349,7 @@ permitted, so you can't upload files or delete them.
Rclone supports generating file share links for private B2 buckets. Rclone supports generating file share links for private B2 buckets.
They can either be for a file for example: They can either be for a file for example:
``` ```sh
./rclone link B2:bucket/path/to/file.txt ./rclone link B2:bucket/path/to/file.txt
https://f002.backblazeb2.com/file/bucket/path/to/file.txt?Authorization=xxxxxxxx https://f002.backblazeb2.com/file/bucket/path/to/file.txt?Authorization=xxxxxxxx
@@ -344,7 +357,7 @@ https://f002.backblazeb2.com/file/bucket/path/to/file.txt?Authorization=xxxxxxxx
or if run on a directory you will get: or if run on a directory you will get:
``` ```sh
./rclone link B2:bucket/path ./rclone link B2:bucket/path
https://f002.backblazeb2.com/file/bucket/path?Authorization=xxxxxxxx https://f002.backblazeb2.com/file/bucket/path?Authorization=xxxxxxxx
``` ```
@@ -352,7 +365,7 @@ https://f002.backblazeb2.com/file/bucket/path?Authorization=xxxxxxxx
you can then use the authorization token (the part of the url from the you can then use the authorization token (the part of the url from the
`?Authorization=` on) on any file path under that directory. For example: `?Authorization=` on) on any file path under that directory. For example:
``` ```text
https://f002.backblazeb2.com/file/bucket/path/to/file1?Authorization=xxxxxxxx https://f002.backblazeb2.com/file/bucket/path/to/file1?Authorization=xxxxxxxx
https://f002.backblazeb2.com/file/bucket/path/file2?Authorization=xxxxxxxx https://f002.backblazeb2.com/file/bucket/path/file2?Authorization=xxxxxxxx
https://f002.backblazeb2.com/file/bucket/path/folder/file3?Authorization=xxxxxxxx https://f002.backblazeb2.com/file/bucket/path/folder/file3?Authorization=xxxxxxxx

Some files were not shown because too many files have changed in this diff Show More