mirror of
https://github.com/alanorth/cgspace-notes.git
synced 2025-01-27 05:49:12 +01:00
Add notes for 2019-11-28
This commit is contained in:
@ -9,13 +9,12 @@
|
||||
<meta property="og:description" content="Documenting day-to-day work on the [CGSpace](https://cgspace.cgiar.org) repository." />
|
||||
<meta property="og:type" content="website" />
|
||||
<meta property="og:url" content="https://alanorth.github.io/cgspace-notes/posts/" />
|
||||
|
||||
<meta property="og:updated_time" content="2019-11-04T12:20:30+02:00" />
|
||||
|
||||
<meta name="twitter:card" content="summary"/>
|
||||
<meta name="twitter:title" content="Posts"/>
|
||||
<meta name="twitter:description" content="Documenting day-to-day work on the [CGSpace](https://cgspace.cgiar.org) repository."/>
|
||||
<meta name="generator" content="Hugo 0.59.1" />
|
||||
<meta name="generator" content="Hugo 0.60.0" />
|
||||
|
||||
|
||||
|
||||
@ -100,40 +99,34 @@
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2019-02-01">2019-02-01</h2>
|
||||
|
||||
<h2 id="20190201">2019-02-01</h2>
|
||||
<ul>
|
||||
<li>Linode has alerted a few times since last night that the CPU usage on CGSpace (linode18) was high despite me increasing the alert threshold last week from 250% to 275%—I might need to increase it again!</li>
|
||||
|
||||
<li><p>The top IPs before, during, and after this latest alert tonight were:</p>
|
||||
|
||||
<li>The top IPs before, during, and after this latest alert tonight were:</li>
|
||||
</ul>
|
||||
<pre><code># zcat --force /var/log/nginx/*.log /var/log/nginx/*.log.1 | grep -E "01/Feb/2019:(17|18|19|20|21)" | awk '{print $1}' | sort | uniq -c | sort -n | tail -n 10
|
||||
245 207.46.13.5
|
||||
332 54.70.40.11
|
||||
385 5.143.231.38
|
||||
405 207.46.13.173
|
||||
405 207.46.13.75
|
||||
1117 66.249.66.219
|
||||
1121 35.237.175.180
|
||||
1546 5.9.6.51
|
||||
2474 45.5.186.2
|
||||
5490 85.25.237.71
|
||||
</code></pre></li>
|
||||
|
||||
<li><p><code>85.25.237.71</code> is the “Linguee Bot” that I first saw last month</p></li>
|
||||
|
||||
<li><p>The Solr statistics the past few months have been very high and I was wondering if the web server logs also showed an increase</p></li>
|
||||
|
||||
<li><p>There were just over 3 million accesses in the nginx logs last month:</p>
|
||||
|
||||
245 207.46.13.5
|
||||
332 54.70.40.11
|
||||
385 5.143.231.38
|
||||
405 207.46.13.173
|
||||
405 207.46.13.75
|
||||
1117 66.249.66.219
|
||||
1121 35.237.175.180
|
||||
1546 5.9.6.51
|
||||
2474 45.5.186.2
|
||||
5490 85.25.237.71
|
||||
</code></pre><ul>
|
||||
<li><code>85.25.237.71</code> is the “Linguee Bot” that I first saw last month</li>
|
||||
<li>The Solr statistics the past few months have been very high and I was wondering if the web server logs also showed an increase</li>
|
||||
<li>There were just over 3 million accesses in the nginx logs last month:</li>
|
||||
</ul>
|
||||
<pre><code># time zcat --force /var/log/nginx/* | grep -cE "[0-9]{1,2}/Jan/2019"
|
||||
3018243
|
||||
|
||||
real 0m19.873s
|
||||
user 0m22.203s
|
||||
sys 0m1.979s
|
||||
</code></pre></li>
|
||||
</ul>
|
||||
</code></pre>
|
||||
<a href='https://alanorth.github.io/cgspace-notes/2019-02/'>Read more →</a>
|
||||
</article>
|
||||
|
||||
@ -151,26 +144,23 @@ sys 0m1.979s
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2019-01-02">2019-01-02</h2>
|
||||
|
||||
<h2 id="20190102">2019-01-02</h2>
|
||||
<ul>
|
||||
<li>Linode alerted that CGSpace (linode18) had a higher outbound traffic rate than normal early this morning</li>
|
||||
|
||||
<li><p>I don’t see anything interesting in the web server logs around that time though:</p>
|
||||
|
||||
<pre><code># zcat --force /var/log/nginx/*.log /var/log/nginx/*.log.1 | grep -E "02/Jan/2019:0(1|2|3)" | awk '{print $1}' | sort | uniq -c | sort -n | tail -n 10
|
||||
92 40.77.167.4
|
||||
99 210.7.29.100
|
||||
120 38.126.157.45
|
||||
177 35.237.175.180
|
||||
177 40.77.167.32
|
||||
216 66.249.75.219
|
||||
225 18.203.76.93
|
||||
261 46.101.86.248
|
||||
357 207.46.13.1
|
||||
903 54.70.40.11
|
||||
</code></pre></li>
|
||||
<li>I don't see anything interesting in the web server logs around that time though:</li>
|
||||
</ul>
|
||||
<pre><code># zcat --force /var/log/nginx/*.log /var/log/nginx/*.log.1 | grep -E "02/Jan/2019:0(1|2|3)" | awk '{print $1}' | sort | uniq -c | sort -n | tail -n 10
|
||||
92 40.77.167.4
|
||||
99 210.7.29.100
|
||||
120 38.126.157.45
|
||||
177 35.237.175.180
|
||||
177 40.77.167.32
|
||||
216 66.249.75.219
|
||||
225 18.203.76.93
|
||||
261 46.101.86.248
|
||||
357 207.46.13.1
|
||||
903 54.70.40.11
|
||||
</code></pre>
|
||||
<a href='https://alanorth.github.io/cgspace-notes/2019-01/'>Read more →</a>
|
||||
</article>
|
||||
|
||||
@ -188,16 +178,13 @@ sys 0m1.979s
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2018-12-01">2018-12-01</h2>
|
||||
|
||||
<h2 id="20181201">2018-12-01</h2>
|
||||
<ul>
|
||||
<li>Switch CGSpace (linode18) to use OpenJDK instead of Oracle JDK</li>
|
||||
<li>I manually installed OpenJDK, then removed Oracle JDK, then re-ran the <a href="http://github.com/ilri/rmg-ansible-public">Ansible playbook</a> to update all configuration files, etc</li>
|
||||
<li>Then I ran all system updates and restarted the server</li>
|
||||
</ul>
|
||||
|
||||
<h2 id="2018-12-02">2018-12-02</h2>
|
||||
|
||||
<h2 id="20181202">2018-12-02</h2>
|
||||
<ul>
|
||||
<li>I noticed that there is another issue with PDF thumbnails on CGSpace, and I see there was another <a href="https://usn.ubuntu.com/3831-1/">Ghostscript vulnerability last week</a></li>
|
||||
</ul>
|
||||
@ -218,15 +205,12 @@ sys 0m1.979s
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2018-11-01">2018-11-01</h2>
|
||||
|
||||
<h2 id="20181101">2018-11-01</h2>
|
||||
<ul>
|
||||
<li>Finalize AReS Phase I and Phase II ToRs</li>
|
||||
<li>Send a note about my <a href="https://github.com/ilri/dspace-statistics-api">dspace-statistics-api</a> to the dspace-tech mailing list</li>
|
||||
</ul>
|
||||
|
||||
<h2 id="2018-11-03">2018-11-03</h2>
|
||||
|
||||
<h2 id="20181103">2018-11-03</h2>
|
||||
<ul>
|
||||
<li>Linode has been sending mails a few times a day recently that CGSpace (linode18) has had high CPU usage</li>
|
||||
<li>Today these are the top 10 IPs:</li>
|
||||
@ -248,11 +232,10 @@ sys 0m1.979s
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2018-10-01">2018-10-01</h2>
|
||||
|
||||
<h2 id="20181001">2018-10-01</h2>
|
||||
<ul>
|
||||
<li>Phil Thornton got an ORCID identifier so we need to add it to the list on CGSpace and tag his existing items</li>
|
||||
<li>I created a GitHub issue to track this <a href="https://github.com/ilri/DSpace/issues/389">#389</a>, because I’m super busy in Nairobi right now</li>
|
||||
<li>I created a GitHub issue to track this <a href="https://github.com/ilri/DSpace/issues/389">#389</a>, because I'm super busy in Nairobi right now</li>
|
||||
</ul>
|
||||
<a href='https://alanorth.github.io/cgspace-notes/2018-10/'>Read more →</a>
|
||||
</article>
|
||||
@ -271,13 +254,12 @@ sys 0m1.979s
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2018-09-02">2018-09-02</h2>
|
||||
|
||||
<h2 id="20180902">2018-09-02</h2>
|
||||
<ul>
|
||||
<li>New <a href="https://jdbc.postgresql.org/documentation/changelog.html#version_42.2.5">PostgreSQL JDBC driver version 42.2.5</a></li>
|
||||
<li>I’ll update the DSpace role in our <a href="https://github.com/ilri/rmg-ansible-public">Ansible infrastructure playbooks</a> and run the updated playbooks on CGSpace and DSpace Test</li>
|
||||
<li>Also, I’ll re-run the <code>postgresql</code> tasks because the custom PostgreSQL variables are dynamic according to the system’s RAM, and we never re-ran them after migrating to larger Linodes last month</li>
|
||||
<li>I’m testing the new DSpace 5.8 branch in my Ubuntu 18.04 environment and I’m getting those autowire errors in Tomcat 8.5.30 again:</li>
|
||||
<li>I'll update the DSpace role in our <a href="https://github.com/ilri/rmg-ansible-public">Ansible infrastructure playbooks</a> and run the updated playbooks on CGSpace and DSpace Test</li>
|
||||
<li>Also, I'll re-run the <code>postgresql</code> tasks because the custom PostgreSQL variables are dynamic according to the system's RAM, and we never re-ran them after migrating to larger Linodes last month</li>
|
||||
<li>I'm testing the new DSpace 5.8 branch in my Ubuntu 18.04 environment and I'm getting those autowire errors in Tomcat 8.5.30 again:</li>
|
||||
</ul>
|
||||
<a href='https://alanorth.github.io/cgspace-notes/2018-09/'>Read more →</a>
|
||||
</article>
|
||||
@ -296,27 +278,20 @@ sys 0m1.979s
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2018-08-01">2018-08-01</h2>
|
||||
|
||||
<h2 id="20180801">2018-08-01</h2>
|
||||
<ul>
|
||||
<li><p>DSpace Test had crashed at some point yesterday morning and I see the following in <code>dmesg</code>:</p>
|
||||
|
||||
<li>DSpace Test had crashed at some point yesterday morning and I see the following in <code>dmesg</code>:</li>
|
||||
</ul>
|
||||
<pre><code>[Tue Jul 31 00:00:41 2018] Out of memory: Kill process 1394 (java) score 668 or sacrifice child
|
||||
[Tue Jul 31 00:00:41 2018] Killed process 1394 (java) total-vm:15601860kB, anon-rss:5355528kB, file-rss:0kB, shmem-rss:0kB
|
||||
[Tue Jul 31 00:00:41 2018] oom_reaper: reaped process 1394 (java), now anon-rss:0kB, file-rss:0kB, shmem-rss:0kB
|
||||
</code></pre></li>
|
||||
|
||||
<li><p>Judging from the time of the crash it was probably related to the Discovery indexing that starts at midnight</p></li>
|
||||
|
||||
<li><p>From the DSpace log I see that eventually Solr stopped responding, so I guess the <code>java</code> process that was OOM killed above was Tomcat’s</p></li>
|
||||
|
||||
<li><p>I’m not sure why Tomcat didn’t crash with an OutOfMemoryError…</p></li>
|
||||
|
||||
<li><p>Anyways, perhaps I should increase the JVM heap from 5120m to 6144m like we did a few months ago when we tried to run the whole CGSpace Solr core</p></li>
|
||||
|
||||
<li><p>The server only has 8GB of RAM so we’ll eventually need to upgrade to a larger one because we’ll start starving the OS, PostgreSQL, and command line batch processes</p></li>
|
||||
|
||||
<li><p>I ran all system updates on DSpace Test and rebooted it</p></li>
|
||||
</code></pre><ul>
|
||||
<li>Judging from the time of the crash it was probably related to the Discovery indexing that starts at midnight</li>
|
||||
<li>From the DSpace log I see that eventually Solr stopped responding, so I guess the <code>java</code> process that was OOM killed above was Tomcat's</li>
|
||||
<li>I'm not sure why Tomcat didn't crash with an OutOfMemoryError…</li>
|
||||
<li>Anyways, perhaps I should increase the JVM heap from 5120m to 6144m like we did a few months ago when we tried to run the whole CGSpace Solr core</li>
|
||||
<li>The server only has 8GB of RAM so we'll eventually need to upgrade to a larger one because we'll start starving the OS, PostgreSQL, and command line batch processes</li>
|
||||
<li>I ran all system updates on DSpace Test and rebooted it</li>
|
||||
</ul>
|
||||
<a href='https://alanorth.github.io/cgspace-notes/2018-08/'>Read more →</a>
|
||||
</article>
|
||||
@ -335,19 +310,16 @@ sys 0m1.979s
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2018-07-01">2018-07-01</h2>
|
||||
|
||||
<h2 id="20180701">2018-07-01</h2>
|
||||
<ul>
|
||||
<li><p>I want to upgrade DSpace Test to DSpace 5.8 so I took a backup of its current database just in case:</p>
|
||||
|
||||
<pre><code>$ pg_dump -b -v -o --format=custom -U dspace -f dspace-2018-07-01.backup dspace
|
||||
</code></pre></li>
|
||||
|
||||
<li><p>During the <code>mvn package</code> stage on the 5.8 branch I kept getting issues with java running out of memory:</p>
|
||||
|
||||
<pre><code>There is insufficient memory for the Java Runtime Environment to continue.
|
||||
</code></pre></li>
|
||||
<li>I want to upgrade DSpace Test to DSpace 5.8 so I took a backup of its current database just in case:</li>
|
||||
</ul>
|
||||
<pre><code>$ pg_dump -b -v -o --format=custom -U dspace -f dspace-2018-07-01.backup dspace
|
||||
</code></pre><ul>
|
||||
<li>During the <code>mvn package</code> stage on the 5.8 branch I kept getting issues with java running out of memory:</li>
|
||||
</ul>
|
||||
<pre><code>There is insufficient memory for the Java Runtime Environment to continue.
|
||||
</code></pre>
|
||||
<a href='https://alanorth.github.io/cgspace-notes/2018-07/'>Read more →</a>
|
||||
</article>
|
||||
|
||||
@ -365,32 +337,27 @@ sys 0m1.979s
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2018-06-04">2018-06-04</h2>
|
||||
|
||||
<h2 id="20180604">2018-06-04</h2>
|
||||
<ul>
|
||||
<li>Test the <a href="https://tracker.atmire.com/tickets-cgiar-ilri/view-ticket?id=560">DSpace 5.8 module upgrades from Atmire</a> (<a href="https://github.com/ilri/DSpace/pull/378">#378</a>)
|
||||
|
||||
<ul>
|
||||
<li>There seems to be a problem with the CUA and L&R versions in <code>pom.xml</code> because they are using SNAPSHOT and it doesn’t build</li>
|
||||
</ul></li>
|
||||
<li>There seems to be a problem with the CUA and L&R versions in <code>pom.xml</code> because they are using SNAPSHOT and it doesn't build</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li>I added the new CCAFS Phase II Project Tag <code>PII-FP1_PACCA2</code> and merged it into the <code>5_x-prod</code> branch (<a href="https://github.com/ilri/DSpace/pull/379">#379</a>)</li>
|
||||
|
||||
<li><p>I proofed and tested the ILRI author corrections that Peter sent back to me this week:</p>
|
||||
|
||||
<li>I proofed and tested the ILRI author corrections that Peter sent back to me this week:</li>
|
||||
</ul>
|
||||
<pre><code>$ ./fix-metadata-values.py -i /tmp/2018-05-30-Correct-660-authors.csv -db dspace -u dspace -p 'fuuu' -f dc.contributor.author -t correct -m 3 -n
|
||||
</code></pre></li>
|
||||
|
||||
<li><p>I think a sane proofing workflow in OpenRefine is to apply the custom text facets for check/delete/remove and illegal characters that I developed in <a href="/cgspace-notes/2018-03/">March, 2018</a></p></li>
|
||||
|
||||
<li><p>Time to index ~70,000 items on CGSpace:</p>
|
||||
|
||||
</code></pre><ul>
|
||||
<li>I think a sane proofing workflow in OpenRefine is to apply the custom text facets for check/delete/remove and illegal characters that I developed in <a href="/cgspace-notes/2018-03/">March, 2018</a></li>
|
||||
<li>Time to index ~70,000 items on CGSpace:</li>
|
||||
</ul>
|
||||
<pre><code>$ time schedtool -D -e ionice -c2 -n7 nice -n19 [dspace]/bin/dspace index-discovery -b
|
||||
|
||||
real 74m42.646s
|
||||
user 8m5.056s
|
||||
sys 2m7.289s
|
||||
</code></pre></li>
|
||||
</ul>
|
||||
</code></pre>
|
||||
<a href='https://alanorth.github.io/cgspace-notes/2018-06/'>Read more →</a>
|
||||
</article>
|
||||
|
||||
@ -408,15 +375,14 @@ sys 2m7.289s
|
||||
|
||||
</p>
|
||||
</header>
|
||||
<h2 id="2018-05-01">2018-05-01</h2>
|
||||
|
||||
<h2 id="20180501">2018-05-01</h2>
|
||||
<ul>
|
||||
<li>I cleared the Solr statistics core on DSpace Test by issuing two commands directly to the Solr admin interface:
|
||||
|
||||
<ul>
|
||||
<li><a href="http://localhost:3000/solr/statistics/update?stream.body=%3Cdelete%3E%3Cquery%3E*:*%3C/query%3E%3C/delete%3E">http://localhost:3000/solr/statistics/update?stream.body=%3Cdelete%3E%3Cquery%3E*:*%3C/query%3E%3C/delete%3E</a></li>
|
||||
<li><a href="http://localhost:3000/solr/statistics/update?stream.body=%3Ccommit/%3E">http://localhost:3000/solr/statistics/update?stream.body=%3Ccommit/%3E</a></li>
|
||||
</ul></li>
|
||||
<li>http://localhost:3000/solr/statistics/update?stream.body=%3Cdelete%3E%3Cquery%3E*:*%3C/query%3E%3C/delete%3E</li>
|
||||
<li>http://localhost:3000/solr/statistics/update?stream.body=%3Ccommit/%3E</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li>Then I reduced the JVM heap size from 6144 back to 5120m</li>
|
||||
<li>Also, I switched it to use OpenJDK instead of Oracle Java, as well as re-worked the <a href="https://github.com/ilri/rmg-ansible-public">Ansible infrastructure scripts</a> to support hosts choosing which distribution they want to use</li>
|
||||
</ul>
|
||||
|
Reference in New Issue
Block a user