Update notes for 2016-09-14

This commit is contained in:
Alan Orth 2016-09-14 20:17:19 +03:00
parent 1f0435daed
commit 4df9dc4ca7
Signed by: alanorth
GPG Key ID: 0FB860CC9C45B1B9
4 changed files with 242 additions and 0 deletions

View File

@ -224,3 +224,59 @@ Exception in thread "http-bio-127.0.0.1-8081-exec-25" java.lang.OutOfMemoryError
2016-09-14 12:23:07,981 ERROR org.dspace.storage.rdbms.DatabaseManager @ SQL connection Error -
org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error Timeout waiting for idle object
```
- Looking at REST requests, it seems there is one IP hitting us nonstop:
```
# awk '{print $1}' /var/log/nginx/rest.log | sort -n | uniq -c | sort -h | tail -n 3
820 50.87.54.15
12872 70.32.99.142
25744 70.32.83.92
# awk '{print $1}' /var/log/nginx/rest.log.1 | sort -n | uniq -c | sort -h | tail -n 3
7966 181.118.144.29
54706 70.32.99.142
109412 70.32.83.92
```
- Those are the same IPs that were hitting us heavily in July, 2016 as well...
- I think the stability issues are definitely from REST
- Crashed AGAIN, errors from dspace.log:
```
2016-09-14 14:31:43,069 ERROR org.dspace.storage.rdbms.DatabaseManager @ SQL connection Error -
org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error Timeout waiting for idle object
```
- And more heap space errors:
```
# grep -rsI "OutOfMemoryError" /var/log/tomcat7/catalina.* | wc -l
19
```
- There are no more rest requests since the last crash, so maybe there are other things causing this.
- Hmm, I noticed a shitload of IPs from 180.76.0.0/16 are connecting to both CGSpace and DSpace Test (58 unique IPs concurrently!)
- They seem to be coming from Baidu, and so far during today alone account for 1/6 of every connection:
```
# grep -c ip_addr= /home/cgspace.cgiar.org/log/dspace.log.2016-09-14
29084
# grep -c ip_addr=180.76.15 /home/cgspace.cgiar.org/log/dspace.log.2016-09-14
5192
```
- Other recent days are the same... hmmm.
- From the activity control panel I can see 58 unique IPs hitting the site *concurrently*, which has GOT to hurt our stability
- A list of all 2000 unique IPs from CGSpace logs today:
```
# grep ip_addr= /home/cgspace.cgiar.org/log/dspace.log.2016-09-11 | awk -F: '{print $5}' | sort -n | uniq -c | sort -h | tail -n 100
```
- Looking at the top 20 IPs or so, most are Yahoo, MSN, Google, Baidu, TurnitIn (iParadigm), etc... do we have any real users?
- Generate a list of all Affiliations for Peter Ballantyne to go through, make corrections, and create a lookup list from:
```
dspacetest=# \copy (select text_value, count(*) from metadatavalue where resource_type_id=2 and metadata_field_id=211 group by text_value order by count desc)
to /tmp/affiliations.csv with csv;
```

View File

@ -339,6 +339,68 @@ org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error
org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error Timeout waiting for idle object
</code></pre>
<ul>
<li>Looking at REST requests, it seems there is one IP hitting us nonstop:</li>
</ul>
<pre><code># awk '{print $1}' /var/log/nginx/rest.log | sort -n | uniq -c | sort -h | tail -n 3
820 50.87.54.15
12872 70.32.99.142
25744 70.32.83.92
# awk '{print $1}' /var/log/nginx/rest.log.1 | sort -n | uniq -c | sort -h | tail -n 3
7966 181.118.144.29
54706 70.32.99.142
109412 70.32.83.92
</code></pre>
<ul>
<li>Those are the same IPs that were hitting us heavily in July, 2016 as well&hellip;</li>
<li>I think the stability issues are definitely from REST</li>
<li>Crashed AGAIN, errors from dspace.log:</li>
</ul>
<pre><code>2016-09-14 14:31:43,069 ERROR org.dspace.storage.rdbms.DatabaseManager @ SQL connection Error -
org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error Timeout waiting for idle object
</code></pre>
<ul>
<li>And more heap space errors:</li>
</ul>
<pre><code># grep -rsI &quot;OutOfMemoryError&quot; /var/log/tomcat7/catalina.* | wc -l
19
</code></pre>
<ul>
<li>There are no more rest requests since the last crash, so maybe there are other things causing this.</li>
<li>Hmm, I noticed a shitload of IPs from 180.76.0.0/16 are connecting to both CGSpace and DSpace Test (58 unique IPs concurrently!)</li>
<li>They seem to be coming from Baidu, and so far during today alone account for <sup>1</sup>&frasl;<sub>6</sub> of every connection:</li>
</ul>
<pre><code># grep -c ip_addr= /home/cgspace.cgiar.org/log/dspace.log.2016-09-14
29084
# grep -c ip_addr=180.76.15 /home/cgspace.cgiar.org/log/dspace.log.2016-09-14
5192
</code></pre>
<ul>
<li>Other recent days are the same&hellip; hmmm.</li>
<li>From the activity control panel I can see 58 unique IPs hitting the site <em>concurrently</em>, which has GOT to hurt our stability</li>
<li>A list of all 2000 unique IPs from CGSpace logs today:</li>
</ul>
<pre><code># grep ip_addr= /home/cgspace.cgiar.org/log/dspace.log.2016-09-11 | awk -F: '{print $5}' | sort -n | uniq -c | sort -h | tail -n 100
</code></pre>
<ul>
<li>Looking at the top 20 IPs or so, most are Yahoo, MSN, Google, Baidu, TurnitIn (iParadigm), etc&hellip; do we have any real users?</li>
<li>Generate a list of all Affiliations for Peter Ballantyne to go through, make corrections, and create a lookup list from:</li>
</ul>
<pre><code>dspacetest=# \copy (select text_value, count(*) from metadatavalue where resource_type_id=2 and metadata_field_id=211 group by text_value order by count desc)
to /tmp/affiliations.csv with csv;
</code></pre>
</section>
<footer>

View File

@ -276,6 +276,68 @@ org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error
&lt;pre&gt;&lt;code&gt;2016-09-14 12:23:07,981 ERROR org.dspace.storage.rdbms.DatabaseManager @ SQL connection Error -
org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error Timeout waiting for idle object
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;Looking at REST requests, it seems there is one IP hitting us nonstop:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;# awk &#39;{print $1}&#39; /var/log/nginx/rest.log | sort -n | uniq -c | sort -h | tail -n 3
820 50.87.54.15
12872 70.32.99.142
25744 70.32.83.92
# awk &#39;{print $1}&#39; /var/log/nginx/rest.log.1 | sort -n | uniq -c | sort -h | tail -n 3
7966 181.118.144.29
54706 70.32.99.142
109412 70.32.83.92
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;Those are the same IPs that were hitting us heavily in July, 2016 as well&amp;hellip;&lt;/li&gt;
&lt;li&gt;I think the stability issues are definitely from REST&lt;/li&gt;
&lt;li&gt;Crashed AGAIN, errors from dspace.log:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;2016-09-14 14:31:43,069 ERROR org.dspace.storage.rdbms.DatabaseManager @ SQL connection Error -
org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error Timeout waiting for idle object
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;And more heap space errors:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;# grep -rsI &amp;quot;OutOfMemoryError&amp;quot; /var/log/tomcat7/catalina.* | wc -l
19
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;There are no more rest requests since the last crash, so maybe there are other things causing this.&lt;/li&gt;
&lt;li&gt;Hmm, I noticed a shitload of IPs from 180.76.0.0/16 are connecting to both CGSpace and DSpace Test (58 unique IPs concurrently!)&lt;/li&gt;
&lt;li&gt;They seem to be coming from Baidu, and so far during today alone account for &lt;sup&gt;1&lt;/sup&gt;&amp;frasl;&lt;sub&gt;6&lt;/sub&gt; of every connection:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;# grep -c ip_addr= /home/cgspace.cgiar.org/log/dspace.log.2016-09-14
29084
# grep -c ip_addr=180.76.15 /home/cgspace.cgiar.org/log/dspace.log.2016-09-14
5192
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;Other recent days are the same&amp;hellip; hmmm.&lt;/li&gt;
&lt;li&gt;From the activity control panel I can see 58 unique IPs hitting the site &lt;em&gt;concurrently&lt;/em&gt;, which has GOT to hurt our stability&lt;/li&gt;
&lt;li&gt;A list of all 2000 unique IPs from CGSpace logs today:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;# grep ip_addr= /home/cgspace.cgiar.org/log/dspace.log.2016-09-11 | awk -F: &#39;{print $5}&#39; | sort -n | uniq -c | sort -h | tail -n 100
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;Looking at the top 20 IPs or so, most are Yahoo, MSN, Google, Baidu, TurnitIn (iParadigm), etc&amp;hellip; do we have any real users?&lt;/li&gt;
&lt;li&gt;Generate a list of all Affiliations for Peter Ballantyne to go through, make corrections, and create a lookup list from:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;dspacetest=# \copy (select text_value, count(*) from metadatavalue where resource_type_id=2 and metadata_field_id=211 group by text_value order by count desc)
to /tmp/affiliations.csv with csv;
&lt;/code&gt;&lt;/pre&gt;
</description>
</item>

View File

@ -276,6 +276,68 @@ org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error
&lt;pre&gt;&lt;code&gt;2016-09-14 12:23:07,981 ERROR org.dspace.storage.rdbms.DatabaseManager @ SQL connection Error -
org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error Timeout waiting for idle object
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;Looking at REST requests, it seems there is one IP hitting us nonstop:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;# awk &#39;{print $1}&#39; /var/log/nginx/rest.log | sort -n | uniq -c | sort -h | tail -n 3
820 50.87.54.15
12872 70.32.99.142
25744 70.32.83.92
# awk &#39;{print $1}&#39; /var/log/nginx/rest.log.1 | sort -n | uniq -c | sort -h | tail -n 3
7966 181.118.144.29
54706 70.32.99.142
109412 70.32.83.92
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;Those are the same IPs that were hitting us heavily in July, 2016 as well&amp;hellip;&lt;/li&gt;
&lt;li&gt;I think the stability issues are definitely from REST&lt;/li&gt;
&lt;li&gt;Crashed AGAIN, errors from dspace.log:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;2016-09-14 14:31:43,069 ERROR org.dspace.storage.rdbms.DatabaseManager @ SQL connection Error -
org.apache.commons.dbcp.SQLNestedException: Cannot get a connection, pool error Timeout waiting for idle object
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;And more heap space errors:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;# grep -rsI &amp;quot;OutOfMemoryError&amp;quot; /var/log/tomcat7/catalina.* | wc -l
19
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;There are no more rest requests since the last crash, so maybe there are other things causing this.&lt;/li&gt;
&lt;li&gt;Hmm, I noticed a shitload of IPs from 180.76.0.0/16 are connecting to both CGSpace and DSpace Test (58 unique IPs concurrently!)&lt;/li&gt;
&lt;li&gt;They seem to be coming from Baidu, and so far during today alone account for &lt;sup&gt;1&lt;/sup&gt;&amp;frasl;&lt;sub&gt;6&lt;/sub&gt; of every connection:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;# grep -c ip_addr= /home/cgspace.cgiar.org/log/dspace.log.2016-09-14
29084
# grep -c ip_addr=180.76.15 /home/cgspace.cgiar.org/log/dspace.log.2016-09-14
5192
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;Other recent days are the same&amp;hellip; hmmm.&lt;/li&gt;
&lt;li&gt;From the activity control panel I can see 58 unique IPs hitting the site &lt;em&gt;concurrently&lt;/em&gt;, which has GOT to hurt our stability&lt;/li&gt;
&lt;li&gt;A list of all 2000 unique IPs from CGSpace logs today:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;# grep ip_addr= /home/cgspace.cgiar.org/log/dspace.log.2016-09-11 | awk -F: &#39;{print $5}&#39; | sort -n | uniq -c | sort -h | tail -n 100
&lt;/code&gt;&lt;/pre&gt;
&lt;ul&gt;
&lt;li&gt;Looking at the top 20 IPs or so, most are Yahoo, MSN, Google, Baidu, TurnitIn (iParadigm), etc&amp;hellip; do we have any real users?&lt;/li&gt;
&lt;li&gt;Generate a list of all Affiliations for Peter Ballantyne to go through, make corrections, and create a lookup list from:&lt;/li&gt;
&lt;/ul&gt;
&lt;pre&gt;&lt;code&gt;dspacetest=# \copy (select text_value, count(*) from metadatavalue where resource_type_id=2 and metadata_field_id=211 group by text_value order by count desc)
to /tmp/affiliations.csv with csv;
&lt;/code&gt;&lt;/pre&gt;
</description>
</item>