mirror of
https://github.com/alanorth/cgspace-notes.git
synced 2025-01-27 05:49:12 +01:00
Add notes for 2018-01-03
This commit is contained in:
@ -79,3 +79,109 @@ dspace.log.2018-01-02:34
|
||||
- Danny wrote to ask for help renewing the wildcard ilri.org certificate and I advised that we should probably use Let's Encrypt if it's just a handful of domains
|
||||
|
||||
<!--more-->
|
||||
|
||||
## 2018-01-03
|
||||
|
||||
- I woke up to more up and down of CGSpace, this time UptimeRobot noticed a few rounds of up and down of a few minutes each and Linode also notified of high CPU load from 12 to 2 PM
|
||||
- Looks like I need to increase the database pool size again:
|
||||
|
||||
```
|
||||
$ grep -c "Timeout: Pool empty." dspace.log.2018-01-*
|
||||
dspace.log.2018-01-01:0
|
||||
dspace.log.2018-01-02:1972
|
||||
dspace.log.2018-01-03:1909
|
||||
```
|
||||
|
||||
- For some reason there were a lot of "active" connections last night:
|
||||
|
||||

|
||||
|
||||
- The active IPs in XMLUI are:
|
||||
|
||||
```
|
||||
# cat /var/log/nginx/access.log /var/log/nginx/access.log.1 /var/log/nginx/library-access.log /var/log/nginx/library-access.log.1 | grep -E "3/Jan/2018" | awk '{print $1}' | sort -n | uniq -c | sort -h | tail
|
||||
607 40.77.167.141
|
||||
611 2a00:23c3:8c94:7800:392c:a491:e796:9c50
|
||||
663 188.226.169.37
|
||||
759 157.55.39.245
|
||||
887 68.180.229.254
|
||||
1037 157.55.39.175
|
||||
1068 216.244.66.245
|
||||
1495 66.249.64.91
|
||||
1934 104.196.152.243
|
||||
2219 134.155.96.78
|
||||
```
|
||||
|
||||
- 134.155.96.78 appears to be at the University of Mannheim in Germany
|
||||
- They identify as: Mozilla/5.0 (compatible; heritrix/3.2.0 +http://ifm.uni-mannheim.de)
|
||||
- This appears to be the [Internet Archive's open source bot](https://github.com/internetarchive/heritrix3)
|
||||
- They seem to be re-using their Tomcat session so I don't need to do anything to them just yet:
|
||||
|
||||
```
|
||||
$ grep 134.155.96.78 dspace.log.2018-01-03 | grep -o -E 'session_id=[A-Z0-9]{32}' | sort -n | uniq | wc -l
|
||||
2
|
||||
```
|
||||
|
||||
- The API logs show the normal users:
|
||||
|
||||
```
|
||||
# cat /var/log/nginx/rest.log /var/log/nginx/rest.log.1 /var/log/nginx/oai.log /var/log/nginx/oai.log.1 | grep -E "3/Jan/2018" | awk '{print $1}' | sort -n | uniq -c | sort -h | tail
|
||||
32 207.46.13.182
|
||||
38 40.77.167.132
|
||||
38 68.180.229.254
|
||||
43 66.249.64.91
|
||||
46 40.77.167.141
|
||||
49 157.55.39.245
|
||||
79 157.55.39.175
|
||||
1533 50.116.102.77
|
||||
4069 70.32.83.92
|
||||
9355 45.5.184.196
|
||||
```
|
||||
|
||||
- In other related news I see a sizeable amount of requests coming from python-requests
|
||||
- For example, just in the last day there were 1700!
|
||||
|
||||
```
|
||||
# cat /var/log/nginx/access.log /var/log/nginx/access.log.1 /var/log/nginx/library-access.log /var/log/nginx/library-access.log.1 | grep -c python-requests
|
||||
1773
|
||||
```
|
||||
|
||||
- But they come from hundreds of IPs, many of which are 54.x.x.x:
|
||||
|
||||
```
|
||||
# cat /var/log/nginx/access.log /var/log/nginx/access.log.1 /var/log/nginx/library-access.log /var/log/nginx/library-access.log.1 | grep python-requests | awk '{print $1}' | sort -n | uniq -c | sort -h | tail -n 30
|
||||
9 54.144.87.92
|
||||
9 54.146.222.143
|
||||
9 54.146.249.249
|
||||
9 54.158.139.206
|
||||
9 54.161.235.224
|
||||
9 54.163.41.19
|
||||
9 54.163.4.51
|
||||
9 54.196.195.107
|
||||
9 54.198.89.134
|
||||
9 54.80.158.113
|
||||
10 54.198.171.98
|
||||
10 54.224.53.185
|
||||
10 54.226.55.207
|
||||
10 54.227.8.195
|
||||
10 54.242.234.189
|
||||
10 54.242.238.209
|
||||
10 54.80.100.66
|
||||
11 54.161.243.121
|
||||
11 54.205.154.178
|
||||
11 54.234.225.84
|
||||
11 54.87.23.173
|
||||
11 54.90.206.30
|
||||
12 54.196.127.62
|
||||
12 54.224.242.208
|
||||
12 54.226.199.163
|
||||
13 54.162.149.249
|
||||
13 54.211.182.255
|
||||
19 50.17.61.150
|
||||
21 54.211.119.107
|
||||
139 164.39.7.62
|
||||
```
|
||||
|
||||
- I have no idea what these are but they seem to be coming from Amazon...
|
||||
- I guess for now I just have to increase the database connection pool's max active
|
||||
- It's currently 75 and normally I'd just bump it by 25 but let me be a bit daring and push it by 50 to 125, because I used to see at least 121 connections in pg_stat_activity before when we were using the shitty default pooling
|
||||
|
Reference in New Issue
Block a user