From aa66648d2fb2082374d8692df96e1c35820ca8f1 Mon Sep 17 00:00:00 2001 From: strmd Date: Wed, 8 May 2019 17:23:24 +0000 Subject: [PATCH 1/5] Added a comment --- .../comment_4_12ae880e559bde5f75cb0c451b56b322._comment | 8 ++++++++ 1 file changed, 8 insertions(+) create mode 100644 doc/forum/Recommended_way_to_work_with_semi-permanent_filename_changes/comment_4_12ae880e559bde5f75cb0c451b56b322._comment diff --git a/doc/forum/Recommended_way_to_work_with_semi-permanent_filename_changes/comment_4_12ae880e559bde5f75cb0c451b56b322._comment b/doc/forum/Recommended_way_to_work_with_semi-permanent_filename_changes/comment_4_12ae880e559bde5f75cb0c451b56b322._comment new file mode 100644 index 0000000000..f1598d157b --- /dev/null +++ b/doc/forum/Recommended_way_to_work_with_semi-permanent_filename_changes/comment_4_12ae880e559bde5f75cb0c451b56b322._comment @@ -0,0 +1,8 @@ +[[!comment format=mdwn + username="strmd" + avatar="http://cdn.libravatar.org/avatar/035707b9756129bbdea6b36a7f7b38d3" + subject="comment 4" + date="2019-05-08T17:23:24Z" + content=""" +Thanks for answering, Joey. I was basically arriving at this. With the number of files and folders I'm dealing with, keeping the branches separated does seem to make the most sense. For less involved scenarios, having imported into \"orig\", a simple `git checkout orig /path/to/file` from master does the trick nicely. +"""]] From 63f6389e64a00664acee2ae92c4a2b8822c86dee Mon Sep 17 00:00:00 2001 From: Ilya_Shlyakhter Date: Thu, 9 May 2019 03:00:34 +0000 Subject: [PATCH 2/5] problem with multi-threaded copying to an S3 remote. --- doc/bugs/parallel_copy_to_S3_fails.mdwn | 1 + 1 file changed, 1 insertion(+) create mode 100644 doc/bugs/parallel_copy_to_S3_fails.mdwn diff --git a/doc/bugs/parallel_copy_to_S3_fails.mdwn b/doc/bugs/parallel_copy_to_S3_fails.mdwn new file mode 100644 index 0000000000..a97f4b1512 --- /dev/null +++ b/doc/bugs/parallel_copy_to_S3_fails.mdwn @@ -0,0 +1 @@ +I'm seeing various failures with the latest git-annex when doing multi-threaded copy --to S3 . Using -J1 fixes the problem. From a022f42ab28f72f14a3993acdfafaa5691621ec7 Mon Sep 17 00:00:00 2001 From: yarikoptic Date: Thu, 9 May 2019 20:42:34 +0000 Subject: [PATCH 3/5] new issue on inability to access ftp urls forwarded to from http URL --- ...ecting_to_ftp_are_no_longer_supported.mdwn | 68 +++++++++++++++++++ 1 file changed, 68 insertions(+) create mode 100644 doc/bugs/regression__58___http_downloads_redirecting_to_ftp_are_no_longer_supported.mdwn diff --git a/doc/bugs/regression__58___http_downloads_redirecting_to_ftp_are_no_longer_supported.mdwn b/doc/bugs/regression__58___http_downloads_redirecting_to_ftp_are_no_longer_supported.mdwn new file mode 100644 index 0000000000..ff2d364bad --- /dev/null +++ b/doc/bugs/regression__58___http_downloads_redirecting_to_ftp_are_no_longer_supported.mdwn @@ -0,0 +1,68 @@ +### Please describe the problem. + +Redoing example from elderly and fixed [issue](https://git-annex.branchable.com/bugs/annex_drop_fails_to_determine_availability_on_a_http_url_redirecting_to_ftp/) which now fails even one step ahead (at `addurl` not `drop` where I guess it would fail too): + +[[!format sh """ +$> git annex addurl --debug --pathdepth=-1 http://www.nitrc.org/frs/downloadlink.php/1637 +[2019-05-09 16:37:34.647058788] read: git ["--git-dir=.git","--work-tree=.","--literal-pathspecs","show-ref","git-annex"] +[2019-05-09 16:37:34.651237646] process done ExitSuccess +[2019-05-09 16:37:34.651310462] read: git ["--git-dir=.git","--work-tree=.","--literal-pathspecs","show-ref","--hash","refs/heads/git-annex"] +[2019-05-09 16:37:34.654835492] process done ExitSuccess +[2019-05-09 16:37:34.655059565] read: git ["--git-dir=.git","--work-tree=.","--literal-pathspecs","log","refs/heads/git-annex..fc38e19e73cbb843c7d4824256a40eb37cad38fc","--pretty=%H","-n1"] +[2019-05-09 16:37:34.667342071] process done ExitSuccess +[2019-05-09 16:37:34.667770788] chat: git ["--git-dir=.git","--work-tree=.","--literal-pathspecs","cat-file","--batch"] +[2019-05-09 16:37:34.668256112] chat: git ["--git-dir=.git","--work-tree=.","--literal-pathspecs","cat-file","--batch-check=%(objectname) %(objecttype) %(objectsize)"] +addurl http://www.nitrc.org/frs/downloadlink.php/1637 [2019-05-09 16:37:34.696706993] Request { + host = "www.nitrc.org" + port = 80 + secure = False + requestHeaders = [("Accept-Encoding",""),("User-Agent","git-annex/7.20190322+git133-g59922f1f4-1~ndall+1")] + path = "/frs/downloadlink.php/1637" + queryString = "" + method = "HEAD" + proxy = Nothing + rawBody = False + redirectCount = 10 + responseTimeout = ResponseTimeoutDefault + requestVersion = HTTP/1.1 +} + + +[2019-05-09 16:37:35.881856173] Request { + host = "www.nitrc.org" + port = 80 + secure = False + requestHeaders = [("Accept-Encoding","identity"),("User-Agent","git-annex/7.20190322+git133-g59922f1f4-1~ndall+1")] + path = "/frs/downloadlink.php/1637" + queryString = "" + method = "GET" + proxy = Nothing + rawBody = False + redirectCount = 10 + responseTimeout = ResponseTimeoutDefault + requestVersion = HTTP/1.1 +} + +download failed: Found +failed +[2019-05-09 16:37:36.064903879] process done ExitSuccess +[2019-05-09 16:37:36.065541286] process done ExitSuccess +git-annex: addurl: 1 failed +1 10323 ->1.....................................:Thu 09 May 2019 04:37:36 PM EDT:. +(git)smaug:/tmp[master] +$> git annex version +git-annex version: 7.20190322+git133-g59922f1f4-1~ndall+1 +build flags: Assistant Webapp Pairing S3(multipartupload)(storageclasses) WebDAV Inotify DBus DesktopNotify TorrentParser MagicMime Feeds Testsuite +dependency versions: aws-0.20 bloomfilter-2.0.1.0 cryptonite-0.25 DAV-1.3.3 feed-1.0.0.0 ghc-8.4.4 http-client-0.5.13.1 persistent-sqlite-2.8.2 torrent-10000.1.1 uuid-1.3.13 yesod-1.6.0 +key/value backends: SHA256E SHA256 SHA512E SHA512 SHA224E SHA224 SHA384E SHA384 SHA3_256E SHA3_256 SHA3_512E SHA3_512 SHA3_224E SHA3_224 SHA3_384E SHA3_384 SKEIN256E SKEIN256 SKEIN512E SKEIN512 BLAKE2B256E BLAKE2B256 BLAKE2B512E BLAKE2B512 BLAKE2B160E BLAKE2B160 BLAKE2B224E BLAKE2B224 BLAKE2B384E BLAKE2B384 BLAKE2S256E BLAKE2S256 BLAKE2S160E BLAKE2S160 BLAKE2S224E BLAKE2S224 BLAKE2SP256E BLAKE2SP256 BLAKE2SP224E BLAKE2SP224 SHA1E SHA1 MD5E MD5 WORM URL +remote types: git gcrypt p2p S3 bup directory rsync web bittorrent webdav adb tahoe glacier ddar hook external +operating system: linux x86_64 +supported repository versions: 5 7 +upgrade supported from repository versions: 0 1 2 3 4 5 6 +local repository version: 5 + +"""]] + +Originally [reported in DataLad #3321](https://github.com/datalad/datalad/issues/3321) with workarounds to force `curl` downloads + +[[!meta author=yoh]] From e0c73c7f2936eb13704b49e3b7a86978a2bc0f7e Mon Sep 17 00:00:00 2001 From: Ilya_Shlyakhter Date: Thu, 9 May 2019 21:07:39 +0000 Subject: [PATCH 4/5] Added a comment --- ...mment_2_3713c706c8a0689f527c4558176a7b48._comment | 12 ++++++++++++ 1 file changed, 12 insertions(+) create mode 100644 doc/todo/add_tests_under_concurrency/comment_2_3713c706c8a0689f527c4558176a7b48._comment diff --git a/doc/todo/add_tests_under_concurrency/comment_2_3713c706c8a0689f527c4558176a7b48._comment b/doc/todo/add_tests_under_concurrency/comment_2_3713c706c8a0689f527c4558176a7b48._comment new file mode 100644 index 0000000000..9f1feef624 --- /dev/null +++ b/doc/todo/add_tests_under_concurrency/comment_2_3713c706c8a0689f527c4558176a7b48._comment @@ -0,0 +1,12 @@ +[[!comment format=mdwn + username="Ilya_Shlyakhter" + avatar="http://cdn.libravatar.org/avatar/1647044369aa7747829c38b9dcc84df0" + subject="comment 2" + date="2019-05-09T21:07:38Z" + content=""" +\"Finding concurrency bugs seems to need the test repos to contain more files than the two or three\" -- it should be enough to run a concurrent operation many times on a small repo. This will test the different possible orders of operation. Most concurrency bugs only involve a conflict of two or three operations. + +Running under something like https://rr-project.org/ can be used to reliably reproduce concurrency bugs. + +\"I can think of only one concurrency bug that ever caused even theoretical data loss\" -- that's good. But beyond data loss, when a complex operation fails in the middle, it can be non-trivial to figure out which parts to redo. +"""]] From 0076e9fa037a8f0ef4a4dda42466a10aade5969e Mon Sep 17 00:00:00 2001 From: CandyAngel Date: Fri, 10 May 2019 11:29:10 +0000 Subject: [PATCH 5/5] --- ...te_fetches_but_does_not_add_all_files.mdwn | 21 +++++++++++++++++++ 1 file changed, 21 insertions(+) create mode 100644 doc/bugs/adb_special_remote_fetches_but_does_not_add_all_files.mdwn diff --git a/doc/bugs/adb_special_remote_fetches_but_does_not_add_all_files.mdwn b/doc/bugs/adb_special_remote_fetches_but_does_not_add_all_files.mdwn new file mode 100644 index 0000000000..884ca9e783 --- /dev/null +++ b/doc/bugs/adb_special_remote_fetches_but_does_not_add_all_files.mdwn @@ -0,0 +1,21 @@ +### Please describe the problem. +I added an adb special remote (to a fresh repo) with androiddirectory=/storage/self/primary/DCIM and while importing fetches all the files from the device, they are not all added to the repo. For example, all the thumbnail files seem to be stored, but of the 9 mp4 files in DCIM/Camera fetched, only 1 is stored. + +While git-annex does mention that "there was an error fetching files, rerun to resume", there is no error output anywhere I could see (e.g. with --debug). All the file fetches are shown as "OK". + +Making the special remote with androiddirectory=/storage/self/primary/DCIM/Camera instead, results in the 9 files being fetched and stored as expected. + +### What steps will reproduce the problem? + git init android + git annex init mycomputer + git annex initremote int_dcim type=adb androiddirectory=/storage/self/primary/DCIM encryption=none exporttree=yes importtree=yes + git annex import --from int_dcim master:somedir + +### What version of git-annex are you using? On what operating system? +* git-annex version: 7.20190503-g4da50456a3 +* Operating System: Arch Linux + +### Have you had any luck using git-annex before? +(in a gamedev IRC channel, discussing my use of git-annex) + + > Oh, I know. it is difficult to not notice how much you fangirl over it :p