From 790e0083ed69e2a7c5f3fcfbb65295a1dd9ea483 Mon Sep 17 00:00:00 2001 From: Joey Hess Date: Fri, 23 Jun 2023 12:54:55 -0400 Subject: [PATCH] comment --- ...3_b53f2e881af1426adb0f45e2b11e4be6._comment | 18 ++++++++++++++++++ 1 file changed, 18 insertions(+) create mode 100644 doc/todo/Speed_up___39__import_--clean-duplicates__39__/comment_3_b53f2e881af1426adb0f45e2b11e4be6._comment diff --git a/doc/todo/Speed_up___39__import_--clean-duplicates__39__/comment_3_b53f2e881af1426adb0f45e2b11e4be6._comment b/doc/todo/Speed_up___39__import_--clean-duplicates__39__/comment_3_b53f2e881af1426adb0f45e2b11e4be6._comment new file mode 100644 index 0000000000..a3849da2ee --- /dev/null +++ b/doc/todo/Speed_up___39__import_--clean-duplicates__39__/comment_3_b53f2e881af1426adb0f45e2b11e4be6._comment @@ -0,0 +1,18 @@ +[[!comment format=mdwn + username="joey" + subject="""comment 3""" + date="2023-06-23T16:47:29Z" + content=""" +`git-annex import --from remote` has recently been sped up a lot, +and the plan is to [[todo/remove_legacy_import_directory_interface]] +in favor of it. + +I think this would work as a faster alternative to --clean-duplicates, +using a directory special remote: + + git-annex import --from remote --fast + git-annex move --from remote --copies 2 + +When there's a huge file in the directory remote, it will hash it, but only +once. On subsequent runs it will recognise the file it has seen before. +"""]]