From 1467eefde026952ff4c14bdeafbbfa1e890c253a Mon Sep 17 00:00:00 2001 From: Alan Orth Date: Mon, 4 Apr 2016 12:37:05 +0300 Subject: [PATCH] Add notes for 2016-04-04 Signed-off-by: Alan Orth --- content/2016-04.md | 45 +++++++++ public/2016-03/index.html | 2 +- public/2016-04/index.html | 176 +++++++++++++++++++++++++++++++++++ public/index.html | 28 ++++++ public/index.xml | 53 ++++++++++- public/sitemap.xml | 7 +- public/tags/index.html | 2 +- public/tags/notes/index.html | 26 ++++++ public/tags/notes/index.xml | 53 ++++++++++- 9 files changed, 387 insertions(+), 5 deletions(-) create mode 100644 content/2016-04.md create mode 100644 public/2016-04/index.html diff --git a/content/2016-04.md b/content/2016-04.md new file mode 100644 index 000000000..9f0e68867 --- /dev/null +++ b/content/2016-04.md @@ -0,0 +1,45 @@ ++++ +date = "2016-04-04T11:06:00+03:00" +author = "Alan Orth" +title = "April, 2016" +tags = ["notes"] +image = "../images/bg.jpg" + ++++ +## 2016-04-04 + +- Looking at log file use on CGSpace and notice that we need to work on our cron setup a bit +- We are backing up all logs in the log folder, including useless stuff like solr, cocoon, handle-plugin, etc +- After running DSpace for over five years I've never needed to look in any other log file than dspace.log, leave alone one from last year! +- This will save us a few gigs of backup space we're paying for on S3 +- Also, I noticed the `checker` log has some errors we should pay attention to: + +``` +Run start time: 03/06/2016 04:00:22 +Error retrieving bitstream ID 71274 from asset store. +java.io.FileNotFoundException: /home/cgspace.cgiar.org/assetstore/64/29/06/64290601546459645925328536011917633626 (Too many open files) + at java.io.FileInputStream.open(Native Method) + at java.io.FileInputStream.(FileInputStream.java:146) + at edu.sdsc.grid.io.local.LocalFileInputStream.open(LocalFileInputStream.java:171) + at edu.sdsc.grid.io.GeneralFileInputStream.(GeneralFileInputStream.java:145) + at edu.sdsc.grid.io.local.LocalFileInputStream.(LocalFileInputStream.java:139) + at edu.sdsc.grid.io.FileFactory.newFileInputStream(FileFactory.java:630) + at org.dspace.storage.bitstore.BitstreamStorageManager.retrieve(BitstreamStorageManager.java:525) + at org.dspace.checker.BitstreamDAO.getBitstream(BitstreamDAO.java:60) + at org.dspace.checker.CheckerCommand.processBitstream(CheckerCommand.java:303) + at org.dspace.checker.CheckerCommand.checkBitstream(CheckerCommand.java:171) + at org.dspace.checker.CheckerCommand.process(CheckerCommand.java:120) + at org.dspace.app.checker.ChecksumChecker.main(ChecksumChecker.java:236) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:606) + at org.dspace.app.launcher.ScriptLauncher.runOneCommand(ScriptLauncher.java:225) + at org.dspace.app.launcher.ScriptLauncher.main(ScriptLauncher.java:77) +****************************************************** +``` + +- So this would be the `tomcat7` Unix user, who seems to have a default limit of 1024 files in its shell +- For what it's worth, we have been setting the actual Tomcat 7 process' limit to 16384 for a few years (in `/etc/default/tomcat7`) +- Looks like cron will read limits from `/etc/security/limits.*` so we can do something for the tomcat7 user there +- Submit pull request for Tomcat 7 limits in Ansible dspace role ([#30](https://github.com/ilri/rmg-ansible-public/pull/30)) diff --git a/public/2016-03/index.html b/public/2016-03/index.html index 7a4897658..2f21a5f20 100644 --- a/public/2016-03/index.html +++ b/public/2016-03/index.html @@ -337,7 +337,7 @@ - + diff --git a/public/2016-04/index.html b/public/2016-04/index.html new file mode 100644 index 000000000..0a11455f4 --- /dev/null +++ b/public/2016-04/index.html @@ -0,0 +1,176 @@ + + + + + + + + + + + + + + + + + + + + + April, 2016 · CGSpace Notes + + + + + + + + + + + + + +
+ + +
+
+

April, 2016

+ +
+
+ + +

2016-04-04

+ +
    +
  • Looking at log file use on CGSpace and notice that we need to work on our cron setup a bit
  • +
  • We are backing up all logs in the log folder, including useless stuff like solr, cocoon, handle-plugin, etc
  • +
  • After running DSpace for over five years I’ve never needed to look in any other log file than dspace.log, leave alone one from last year!
  • +
  • This will save us a few gigs of backup space we’re paying for on S3
  • +
  • Also, I noticed the checker log has some errors we should pay attention to:
  • +
+ +
Run start time: 03/06/2016 04:00:22
+Error retrieving bitstream ID 71274 from asset store.
+java.io.FileNotFoundException: /home/cgspace.cgiar.org/assetstore/64/29/06/64290601546459645925328536011917633626 (Too many open files)
+        at java.io.FileInputStream.open(Native Method)
+        at java.io.FileInputStream.<init>(FileInputStream.java:146)
+        at edu.sdsc.grid.io.local.LocalFileInputStream.open(LocalFileInputStream.java:171)
+        at edu.sdsc.grid.io.GeneralFileInputStream.<init>(GeneralFileInputStream.java:145)
+        at edu.sdsc.grid.io.local.LocalFileInputStream.<init>(LocalFileInputStream.java:139)
+        at edu.sdsc.grid.io.FileFactory.newFileInputStream(FileFactory.java:630)
+        at org.dspace.storage.bitstore.BitstreamStorageManager.retrieve(BitstreamStorageManager.java:525)
+        at org.dspace.checker.BitstreamDAO.getBitstream(BitstreamDAO.java:60)
+        at org.dspace.checker.CheckerCommand.processBitstream(CheckerCommand.java:303)
+        at org.dspace.checker.CheckerCommand.checkBitstream(CheckerCommand.java:171)
+        at org.dspace.checker.CheckerCommand.process(CheckerCommand.java:120)
+        at org.dspace.app.checker.ChecksumChecker.main(ChecksumChecker.java:236)
+        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
+        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
+        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
+        at java.lang.reflect.Method.invoke(Method.java:606)
+        at org.dspace.app.launcher.ScriptLauncher.runOneCommand(ScriptLauncher.java:225)
+        at org.dspace.app.launcher.ScriptLauncher.main(ScriptLauncher.java:77)
+******************************************************
+
+ +
    +
  • So this would be the tomcat7 Unix user, who seems to have a default limit of 1024 files in its shell
  • +
  • For what it’s worth, we have been setting the actual Tomcat 7 process’ limit to 16384 for a few years (in /etc/default/tomcat7)
  • +
  • Looks like cron will read limits from /etc/security/limits.* so we can do something for the tomcat7 user there
  • +
  • Submit pull request for Tomcat 7 limits in Ansible dspace role (#30)
  • +
+ +
+ +
+ +
+
+ + + +
+ + + + + + + + diff --git a/public/index.html b/public/index.html index a2a83f67f..e6fd0d2d6 100644 --- a/public/index.html +++ b/public/index.html @@ -58,6 +58,34 @@
+
+
+

April, 2016

+ +
+
+ 2016-04-04 Looking at log file use on CGSpace and notice that we need to work on our cron setup a bit We are backing up all logs in the log folder, including useless stuff like solr, cocoon, handle-plugin, etc After running DSpace for over five years I’ve never needed to look in any other log file than dspace.log, leave alone one from last year! This will save us a few gigs of backup space we’re paying for on S3 Also, I noticed the checker log has some errors we should pay attention to: Run start time: 03/06/2016 04:00:22 Error retrieving bitstream ID 71274 from asset store. +
+ + + +
+ + + +
+

March, 2016

diff --git a/public/index.xml b/public/index.xml index 8d6df96f9..a675676da 100644 --- a/public/index.xml +++ b/public/index.xml @@ -6,9 +6,60 @@ Recent content on CGSpace Notes Hugo -- gohugo.io en-us - Wed, 02 Mar 2016 16:50:00 +0300 + Mon, 04 Apr 2016 11:06:00 +0300 + + April, 2016 + /cgspace-notes/2016-04/ + Mon, 04 Apr 2016 11:06:00 +0300 + + /cgspace-notes/2016-04/ + + +<h2 id="2016-04-04:c88be15f5b2f07c85f7742556a955e47">2016-04-04</h2> + +<ul> +<li>Looking at log file use on CGSpace and notice that we need to work on our cron setup a bit</li> +<li>We are backing up all logs in the log folder, including useless stuff like solr, cocoon, handle-plugin, etc</li> +<li>After running DSpace for over five years I&rsquo;ve never needed to look in any other log file than dspace.log, leave alone one from last year!</li> +<li>This will save us a few gigs of backup space we&rsquo;re paying for on S3</li> +<li>Also, I noticed the <code>checker</code> log has some errors we should pay attention to:</li> +</ul> + +<pre><code>Run start time: 03/06/2016 04:00:22 +Error retrieving bitstream ID 71274 from asset store. +java.io.FileNotFoundException: /home/cgspace.cgiar.org/assetstore/64/29/06/64290601546459645925328536011917633626 (Too many open files) + at java.io.FileInputStream.open(Native Method) + at java.io.FileInputStream.&lt;init&gt;(FileInputStream.java:146) + at edu.sdsc.grid.io.local.LocalFileInputStream.open(LocalFileInputStream.java:171) + at edu.sdsc.grid.io.GeneralFileInputStream.&lt;init&gt;(GeneralFileInputStream.java:145) + at edu.sdsc.grid.io.local.LocalFileInputStream.&lt;init&gt;(LocalFileInputStream.java:139) + at edu.sdsc.grid.io.FileFactory.newFileInputStream(FileFactory.java:630) + at org.dspace.storage.bitstore.BitstreamStorageManager.retrieve(BitstreamStorageManager.java:525) + at org.dspace.checker.BitstreamDAO.getBitstream(BitstreamDAO.java:60) + at org.dspace.checker.CheckerCommand.processBitstream(CheckerCommand.java:303) + at org.dspace.checker.CheckerCommand.checkBitstream(CheckerCommand.java:171) + at org.dspace.checker.CheckerCommand.process(CheckerCommand.java:120) + at org.dspace.app.checker.ChecksumChecker.main(ChecksumChecker.java:236) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:606) + at org.dspace.app.launcher.ScriptLauncher.runOneCommand(ScriptLauncher.java:225) + at org.dspace.app.launcher.ScriptLauncher.main(ScriptLauncher.java:77) +****************************************************** +</code></pre> + +<ul> +<li>So this would be the <code>tomcat7</code> Unix user, who seems to have a default limit of 1024 files in its shell</li> +<li>For what it&rsquo;s worth, we have been setting the actual Tomcat 7 process&rsquo; limit to 16384 for a few years (in <code>/etc/default/tomcat7</code>)</li> +<li>Looks like cron will read limits from <code>/etc/security/limits.*</code> so we can do something for the tomcat7 user there</li> +<li>Submit pull request for Tomcat 7 limits in Ansible dspace role (<a href="https://github.com/ilri/rmg-ansible-public/pull/30">#30</a>)</li> +</ul> + + + March, 2016 /cgspace-notes/2016-03/ diff --git a/public/sitemap.xml b/public/sitemap.xml index 2734de5ee..be0a4d03a 100644 --- a/public/sitemap.xml +++ b/public/sitemap.xml @@ -3,10 +3,15 @@ /cgspace-notes/ - 2016-03-02T16:50:00+03:00 + 2016-04-04T11:06:00+03:00 0 + + /cgspace-notes/2016-04/ + 2016-04-04T11:06:00+03:00 + + /cgspace-notes/2016-03/ 2016-03-02T16:50:00+03:00 diff --git a/public/tags/index.html b/public/tags/index.html index cdb55768b..d592f9853 100644 --- a/public/tags/index.html +++ b/public/tags/index.html @@ -64,7 +64,7 @@ diff --git a/public/tags/notes/index.html b/public/tags/notes/index.html index 8350308ff..4e1c6137f 100644 --- a/public/tags/notes/index.html +++ b/public/tags/notes/index.html @@ -61,6 +61,32 @@

Notes

+
+
+
+

April, 2016

+ +
+
+ 2016-04-04 Looking at log file use on CGSpace and notice that we need to work on our cron setup a bit We are backing up all logs in the log folder, including useless stuff like solr, cocoon, handle-plugin, etc After running DSpace for over five years I’ve never needed to look in any other log file than dspace.log, leave alone one from last year! This will save us a few gigs of backup space we’re paying for on S3 Also, I noticed the checker log has some errors we should pay attention to: Run start time: 03/06/2016 04:00:22 Error retrieving bitstream ID 71274 from asset store. +
+ + + +
+ +
diff --git a/public/tags/notes/index.xml b/public/tags/notes/index.xml index bbba85d4e..6ede670cf 100644 --- a/public/tags/notes/index.xml +++ b/public/tags/notes/index.xml @@ -6,9 +6,60 @@ Recent content in Notes on CGSpace Notes Hugo -- gohugo.io en-us - Wed, 02 Mar 2016 16:50:00 +0300 + Mon, 04 Apr 2016 11:06:00 +0300 + + April, 2016 + /cgspace-notes/2016-04/ + Mon, 04 Apr 2016 11:06:00 +0300 + + /cgspace-notes/2016-04/ + + +<h2 id="2016-04-04:c88be15f5b2f07c85f7742556a955e47">2016-04-04</h2> + +<ul> +<li>Looking at log file use on CGSpace and notice that we need to work on our cron setup a bit</li> +<li>We are backing up all logs in the log folder, including useless stuff like solr, cocoon, handle-plugin, etc</li> +<li>After running DSpace for over five years I&rsquo;ve never needed to look in any other log file than dspace.log, leave alone one from last year!</li> +<li>This will save us a few gigs of backup space we&rsquo;re paying for on S3</li> +<li>Also, I noticed the <code>checker</code> log has some errors we should pay attention to:</li> +</ul> + +<pre><code>Run start time: 03/06/2016 04:00:22 +Error retrieving bitstream ID 71274 from asset store. +java.io.FileNotFoundException: /home/cgspace.cgiar.org/assetstore/64/29/06/64290601546459645925328536011917633626 (Too many open files) + at java.io.FileInputStream.open(Native Method) + at java.io.FileInputStream.&lt;init&gt;(FileInputStream.java:146) + at edu.sdsc.grid.io.local.LocalFileInputStream.open(LocalFileInputStream.java:171) + at edu.sdsc.grid.io.GeneralFileInputStream.&lt;init&gt;(GeneralFileInputStream.java:145) + at edu.sdsc.grid.io.local.LocalFileInputStream.&lt;init&gt;(LocalFileInputStream.java:139) + at edu.sdsc.grid.io.FileFactory.newFileInputStream(FileFactory.java:630) + at org.dspace.storage.bitstore.BitstreamStorageManager.retrieve(BitstreamStorageManager.java:525) + at org.dspace.checker.BitstreamDAO.getBitstream(BitstreamDAO.java:60) + at org.dspace.checker.CheckerCommand.processBitstream(CheckerCommand.java:303) + at org.dspace.checker.CheckerCommand.checkBitstream(CheckerCommand.java:171) + at org.dspace.checker.CheckerCommand.process(CheckerCommand.java:120) + at org.dspace.app.checker.ChecksumChecker.main(ChecksumChecker.java:236) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:606) + at org.dspace.app.launcher.ScriptLauncher.runOneCommand(ScriptLauncher.java:225) + at org.dspace.app.launcher.ScriptLauncher.main(ScriptLauncher.java:77) +****************************************************** +</code></pre> + +<ul> +<li>So this would be the <code>tomcat7</code> Unix user, who seems to have a default limit of 1024 files in its shell</li> +<li>For what it&rsquo;s worth, we have been setting the actual Tomcat 7 process&rsquo; limit to 16384 for a few years (in <code>/etc/default/tomcat7</code>)</li> +<li>Looks like cron will read limits from <code>/etc/security/limits.*</code> so we can do something for the tomcat7 user there</li> +<li>Submit pull request for Tomcat 7 limits in Ansible dspace role (<a href="https://github.com/ilri/rmg-ansible-public/pull/30">#30</a>)</li> +</ul> + + + March, 2016 /cgspace-notes/2016-03/