From 2aacd366fb3bf7e55dddf14c62adda52afa98c3e Mon Sep 17 00:00:00 2001 From: Alan Orth Date: Wed, 14 Aug 2019 13:47:56 +0300 Subject: [PATCH] Update notes for 2019-08-14 --- content/posts/2019-08.md | 3 ++- docs/2019-08/index.html | 10 ++++++---- docs/sitemap.xml | 10 +++++----- 3 files changed, 13 insertions(+), 10 deletions(-) diff --git a/content/posts/2019-08.md b/content/posts/2019-08.md index 559427b8f..31bb48616 100644 --- a/content/posts/2019-08.md +++ b/content/posts/2019-08.md @@ -175,10 +175,11 @@ $ ~/dspace/bin/dspace metadata-import -f /tmp/bioversity.csv -e blah@blah.com ``` - This time it succeeded, and using VisualVM I noticed that the import process used a maximum of 620MB of RAM +- (oops, I realize that actually I forgot to delete items I had flagged as duplicates, so the total should be 1,427 items) ## 2019-08-14 -- I imported the 1429 Bioversity records into DSpace Test +- I imported the 1,427 Bioversity records into DSpace Test - To make sure we didn't have memory issues I reduced Tomcat's JVM heap by 512m, increased the import processes's heap to 512m, and split the input file into two parts with about 700 each - Then I had to create a few new temporary collections on DSpace Test that had been created on CGSpace after our last sync - After that the import succeeded: diff --git a/docs/2019-08/index.html b/docs/2019-08/index.html index 0e43d655c..557708855 100644 --- a/docs/2019-08/index.html +++ b/docs/2019-08/index.html @@ -27,7 +27,7 @@ Run system updates on DSpace Test (linode19) and reboot it - + @@ -59,9 +59,9 @@ Run system updates on DSpace Test (linode19) and reboot it "@type": "BlogPosting", "headline": "August, 2019", "url": "https:\/\/alanorth.github.io\/cgspace-notes\/2019-08\/", - "wordCount": "1409", + "wordCount": "1431", "datePublished": "2019-08-03T12:39:51\x2b03:00", - "dateModified": "2019-08-13T16:54:35\x2b03:00", + "dateModified": "2019-08-14T13:39:29\x2b03:00", "author": { "@type": "Person", "name": "Alan Orth" @@ -352,12 +352,14 @@ $ ~/dspace/bin/dspace metadata-import -f /tmp/bioversity.csv -e blah@blah.com
  • This time it succeeded, and using VisualVM I noticed that the import process used a maximum of 620MB of RAM

  • + +
  • (oops, I realize that actually I forgot to delete items I had flagged as duplicates, so the total should be 1,427 items)

  • 2019-08-14