August, 2019
2019-08-03
- Look at Bioversity’s latest migration CSV and now I see that Francesco has cleaned up the extra columns and the newline at the end of the file, but many of the column headers have an extra space in the name…
2019-08-04
- Deploy ORCID identifier updates requested by Bioversity to CGSpace
- Run system updates on CGSpace (linode18) and reboot it
- Before updating it I checked Solr and verified that all statistics cores were loaded properly…
- After rebooting, all statistics cores were loaded… wow, that’s lucky.
- Run system updates on DSpace Test (linode19) and reboot it
2019-08-05
- Update Tomcat to 7.0.96 in the Ansible infrastructure playbooks
- Update PostgreSQL JDBC driver to 42.2.6 in the Ansible infrastrucutre playbooks
- Deploy both on DSpace Test (linode19)
Looking at the 1429 records for Bioversity migration again
- The following items use the same exact PDF and seem to be duplicates:
- https://www.bioversityinternational.org/index.php?id=244&tx_news_pi1[news]=10191
- https://www.bioversityinternational.org/index.php?id=244&tx_news_pi1[news]=342
- The following items use the same exact PDF, but one seems to be incorrect:
- https://www.bioversityinternational.org/index.php?id=244&tx_news_pi1[news]=5347
- https://www.bioversityinternational.org/index.php?id=244&tx_news_pi1[news]=5340
- The following PDFs are used by several items incorrectly:
Report_of_a_Working_Group_on_Allium_7.pdf
Report_of_a_Working_Group_on_Allium_Fourth_meeting_1696.pdf
- I checked the SHA1 hashes of each PDF and found that some appear more than once…
- The following items use the same PDF with a different name, but seem to be duplicates (pick one?):
- https://www.bioversityinternational.org/index.php?id=244&tx_news_pi1[news]=433
- https://www.bioversityinternational.org/index.php?id=244&tx_news_pi1[news]=10189
- The following items use the same PDF with a different name, but seem to be duplicates (pick one?):
- https://www.bioversityinternational.org/index.php?id=244&tx_news_pi1[news]=332
- https://www.bioversityinternational.org/index.php?id=244&tx_news_pi1[news]=10187
- There are about thirty PDFs that have French or Spanish filenames and there seems to be an encoding issue
- I asked Francesco if he can give me a PDF URL column instead of a “filename” column so I can download the files myself
At least the ~50 filenames identified by the following GREL will have issues:
or( isNotNull(value.match(/^.*’.*$/)), isNotNull(value.match(/^.*é.*$/)), isNotNull(value.match(/^.*á.*$/)), isNotNull(value.match(/^.*è.*$/)), isNotNull(value.match(/^.*í.*$/)), isNotNull(value.match(/^.*ó.*$/)), isNotNull(value.match(/^.*ú.*$/)), isNotNull(value.match(/^.*à.*$/)), isNotNull(value.match(/^.*û.*$/)) ).toString()
I tried to extract the filenames and construct a URL to download the PDFs with my
generate-thumbnails.py
script, but there seem to be several paths for PDFs so I can’t guess it properlyI will have to wait for Francesco to respond about the PDFs, or perhaps proceed with a metadata-only upload so we can do other checks on DSpace Test
2019-08-06
- Francesca responded to address my feedback yesterday
- I made some changes to the CSV based on her feedback (remove two duplicates, change one PDF file name, change two titles)
- Then I found some items that have PDFs in multiple languages that only list one language in
dc.language.iso
so I changed them - Strangley, one item was referring to a 7zip file…
- After removing the two duplicates there are now 1427 records
- Fix one invalid ISSN: 1020-2002→1020-3362
2019-08-07
- Daniel Haile-Michael asked about using a logical OR with the DSpace OpenSearch, but I looked in the DSpace manual and it does not seem to be possible
2019-08-08
Moayad noticed that the HTTPS certificate expired on the AReS dev server (linode20)
- The first problem was that there is a Docker container listening on port 80, so it conflicts with the ACME http-01 validation
- The second problem was that we only allow access to port 80 from localhost
I adjusted the
renew-letsencrypt
systemd service so it stops/starts the Docker container and firewall:# /opt/certbot-auto renew --standalone --pre-hook "/usr/bin/docker stop angular_nginx; /bin/systemctl stop firewalld" --post-hook "/bin/systemctl start firewalld; /usr/bin/docker start angular_nginx"
It is important that the firewall starts back up before the Docker container or else Docker will complain about missing iptables chains
Also, I updated to the latest TLS Intermediate settings as appropriate for Ubuntu 18.04’s OpenSSL 1.1.0g with nginx 1.16.0
Run all system updates on AReS dev server (linode20) and reboot it
Get a list of all PDFs from the Bioversity migration that fail to download and save them so I can try again with a different path in the URL:
$ ./generate-thumbnails.py -i /tmp/2019-08-05-Bioversity-Migration.csv -w --url-field-name url -d | tee /tmp/2019-08-08-download-pdfs.txt $ grep -B1 "Download failed" /tmp/2019-08-08-download-pdfs.txt | grep "Downloading" | sed -e 's/> Downloading //' -e 's/\.\.\.//' | sed -r 's/\x1B\[([0-9]{1,2}(;[0-9]{1,2})?)?[mGK]//g' | csvcut -H -c 1,1 > /tmp/user-upload.csv $ ./generate-thumbnails.py -i /tmp/user-upload.csv -w --url-field-name url -d | tee /tmp/2019-08-08-download-pdfs2.txt $ grep -B1 "Download failed" /tmp/2019-08-08-download-pdfs2.txt | grep "Downloading" | sed -e 's/> Downloading //' -e 's/\.\.\.//' | sed -r 's/\x1B\[([0-9]{1,2}(;[0-9]{1,2})?)?[mGK]//g' | csvcut -H -c 1,1 > /tmp/user-upload2.csv $ ./generate-thumbnails.py -i /tmp/user-upload2.csv -w --url-field-name url -d | tee /tmp/2019-08-08-download-pdfs3.txt
(the weird sed regex removes color codes, because my generate-thumbnails script prints pretty colors)
Some PDFs are uploaded in different paths so I have to try a few times to get them all:
/fileadmin/_migrated/uploads/tx_news/
/fileadmin/user_upload/online_library/publications/pdfs/
/fileadmin/user_upload/
Even so, there are still 52 items with incorrect filenames, so I can’t derive their PDF URLs…
- For example,
Wild_cherry_Prunus_avium_859.pdf
is here (with double underscore): https://www.bioversityinternational.org/fileadmin/_migrated/uploads/tx_news/Wild_cherry__Prunus_avium__859.pdf
- For example,
I will proceed with a metadata-only upload first and then let them know about the missing PDFs
Troubleshoot an issue we had with proxying to the new development version of AReS from DSpace Test (linode19)
- For some reason the host header in the proxy pass is not set so nginx on DSpace Test makes a request to the upstream nginx on an IP-based virtual host
- The upstream nginx returns HTTP 444 because we configured it to not answer when a request does not send a valid hostname
The solution is to set the host header when proxy passing:
proxy_set_header Host dev.ares.codeobia.com;
Though I am really wondering why this happened now, because the configuration has been working for months…
Improve the output of the suspicious characters check in csv-metadata-quality script and tag version 0.2.0