2022-06-06 08:45:43 +02:00
---
title: "June, 2022"
date: 2022-06-06T09:01:36+03:00
author: "Alan Orth"
categories: ["Notes"]
---
## 2022-06-06
- Look at the Solr statistics on CGSpace
- I see 167,000 hits from a bunch of Microsoft IPs with reverse DNS "msnbot-" using the Solr query `dns:*msnbot* AND dns:*.msn.com`
- I purged these first so I could see the other "real" IPs in the Solr facets
- I see 47,500 hits from 80.248.237.167 on a data center ISP in Sweden, using a normal user agent
- I see 13,000 hits from 163.237.216.11 on a data center ISP in Australia, using a normal user agent
- I see 7,300 hits from 208.185.238.57 from Britanica, using a normal user agent
- There seem to be many more of these:
<!-- more -->
```console
# zcat --force /var/log/nginx/access.log* | grep 208.185.238. | awk '{print $1}' | sort | uniq -c | sort -h
2 208.185.238.1
166 208.185.238.54
1293 208.185.238.51
2587 208.185.238.59
4692 208.185.238.56
5480 208.185.238.53
6277 208.185.238.52
6400 208.185.238.58
8261 208.185.238.55
17549 208.185.238.57
```
- I see 3,000 hits from 178.208.75.33 by a Russian-owned IP in the Netherlands that is making a GET to / every one minute, using a normal user agent
- I see 3,000 hits from 134.122.124.196 on Digital Ocean to the REST API with a normal user agent
- I purged all these hits from IPs for a total of about 265,000
- Then I faceted by user agent and found
- 1,000 hits by `insomnia/2022.2.1` , which I also saw last month and submitted to COUNTER-Robots
- 265 hits by `omgili/0.5 +http://omgili.com`
- 150 hits by `Vizzit`
- 132 hits by `MetaInspector/5.7.0 (+https://github.com/jaimeiniesta/metainspector)`
- 73 hits by `Scoop.it`
- 62 hits by `bitdiscovery`
- 59 hits by `Asana/1.4.0 WebsiteMetadataRetriever`
- 32 hits by `Sprout Social (Link Attachment)`
- 29 hits by `CyotekWebCopy/1.9 CyotekHTTP/6.2`
- 20 hits by `Hootsuite-Authoring/1.0`
- I purged about 4,100 hits from these user agents
2022-06-06 15:54:08 +02:00
- Run all system updates on AReS server (linode20) and reboot
- I want to try to update some of the build dependencies of OpenRXV since Node.js 12 is no longer supported
- Upgrade linode20 to Ubuntu 22.04 and start an AReS harvest
2022-06-08 14:36:09 +02:00
- I merged the [Mirage 2 build fix ](https://github.com/DSpace/DSpace/pull/8292 ) to `dspace-6_x` for DSpace 6.4
## 2022-06-07
- I tested Node.js 14 one more time with vanilla DSpace 6.4-SNAPSHOT and with the CGSpace source and it worked well
- I made [a pull request ](https://github.com/DSpace/DSpace/pull/8331 ) to DSpace to use Node.js 14 for Mirage 2
- I even tested Node.js 16 and it works, but that is enough for now...
## 2022-06-08
- Work on AReS a bit since I wasn't able to harvest after doing the updates on the server and in the containers a few days ago
- I don't know what the problem was really, but on the server I had to enable IPv4 forwarding so the frontend container would build
- Once I downed and upped AReS with docker-compose I was able to start a new harvest
- I also did some tests to enable ES2020 target in the backend because we're on Node.js 14 there now
2022-06-06 08:45:43 +02:00
2022-06-14 07:45:07 +02:00
## 2022-06-13
- Create a user for Mohammed Salem to test MEL submission on DSpace Test:
```console
$ dspace user -a -m mel-submit@cgiar.org -g MEL -s Submit -p 'owwwwwwww'
```
- According to my notes from [2020-10 ]({{< relref "2020-10.md" >}} ) the account must be in the admin group in order to submit via the REST API
2022-06-16 15:25:51 +02:00
## 2022-06-14
- Start a harvest on AReS
## 2022-06-16
- Francesca asked us to add the CC-BY-3.0-IGO license to the submission form on CGSpace
- I remember I [had requested SPDX to add CC-BY-NC-ND-3.0-IGO ](https://github.com/spdx/license-list-XML/issues/767 ) in 2019-02, and they finally [merged it ](https://github.com/spdx/license-list-XML/pull/1068 ) in 2020-07, but I never added it to CGSpace
- I will add the full suite of CC 3.0 IGO licenses to CGSpace and then make a request to SPDX for the others:
- CC-BY-3.0-IGO
- CC-BY-SA-3.0-IGO
- CC-BY-ND-3.0-IGO
- CC-BY-NC-3.0-IGO
- CC-BY-NC-SA-3.0-IGO
- CC-BY-NC-ND-3.0-IGO
- I filed [an issue asking for SPDX to add CC-BY-3.0-IGO ](https://github.com/spdx/license-list-XML/issues/1525 )
- Meeting with Moayad from CodeObia to discuss OpenRXV
- He added the ability to use multiple indexes / dashboards, and to be able to embed them in iframes
- Add `cg.contributor.initiative` with a controlled vocabulary based on CLARISA's list to the CGSpace submission form
2022-06-16 18:51:59 +02:00
- Switch to the `linux-virtual-hwe-20.04` kernel on CGSpace (linode18), run all system updates, and reboot
2022-06-16 15:25:51 +02:00
2022-06-17 15:46:56 +02:00
## 2022-06-17
- I noticed a few ORCID identifiers missing for some scientists so I added them to the controlled vocabulary and then tagged them on CGSpace:
```console
$ cat 2022-06-17-add-orcids.csv
dc.contributor.author,cg.creator.identifier
"Tijjani, A.","Abdulfatai Tijjani: 0000-0002-0793-9059"
"Tijjani, Abdulfatai","Abdulfatai Tijjani: 0000-0002-0793-9059"
"Mrode, Raphael A.","Raphael Mrode: 0000-0003-1964-5653"
"Okeyo Mwai, Ally","Ally Okeyo Mwai: 0000-0003-2379-7801"
"Ojango, Julie M.K.","Ojango J.M.K.: 0000-0003-0224-5370"
"Prendergast, J.G.D.","James Prendergast: 0000-0001-8916-018X"
"Ekine-Dzivenu, Chinyere","Chinyere Ekine-Dzivenu: 0000-0002-8526-435X"
"Ekine, C.","Chinyere Ekine-Dzivenu: 0000-0002-8526-435X"
"Ekine-Dzivenu, C.C","Chinyere Ekine-Dzivenu: 0000-0002-8526-435X"
"Shilomboleni, Helena","Helena Shilomboleni: 0000-0002-9875-6484"
$ ./ilri/add-orcid-identifiers-csv.py -i /tmp/2022-06-17-add-orcids.csv -db dspace -u dspace -p 'fuuu' | tee /tmp/orcids.log
$ grep -c 'Adding ORCID' /tmp/orcids2.log
304
```
- Also make some changes to the Discovery facets and item view
- I reduced the number of items to show for CRP facets from 20 to 5
- I added a facet for the Initiatives
- I re-organized a few parts of the item view to add Action Areas and the list of author affiliations
2022-06-18 19:39:37 +02:00
## 2022-06-18
- I deployed the changes on CGSpace and started a full Discovery index for the new Initiatives facet
- Run `dspace cleanup -v` on CGSpace
2022-06-21 15:59:04 +02:00
## 2022-06-20
- Add missing ORCID identifier for ILRI staff to CGSpace and tag their items
## 2022-06-21
- Work on OpenRXV backend dependencies
- Update Elasticsearch and TypeScript and eslint
- Sit in on webinar about contributing terms to AGROVOC
- I agreed that I would send Sara Jani from ICARDA a list of new terms we have that don't match AGROVOC by end of June
- I need to indicate which center is using them so we can have an appropriate expert review the terms
2022-06-23 07:40:53 +02:00
## 2022-06-22
- I re-deployed AReS with the latest OpenRXV changes then started a fresh harvest
- Meeting with Salem to discuss metadata between CGSpace and MEL
- We started working through his spreadsheet and then the Internet dropped
2022-06-24 13:49:37 +02:00
## 2022-06-23
- Start looking at country names between MEL, CGSpace, and standards like UN M.49 and GeoNames
- I used `xmllint` to extract the countries from CGSpace's input forms:
```console
$ xmllint --xpath '//value-pairs[@value-pairs-name="countrylist"]/pair/stored-value/node()' dspace/config/input-forms.xml > /tmp/cgspace-countries.txt
```
- Then I wrote a Python script (`countries-to-csv.py`) to read them and save their names alongside the ISO 3166-1 Alpha2 code
- Then I joined them with the other lists:
```console
$ csvjoin --outer -c alpha2 ~/Downloads/clarisa-countries.csv ~/Downloads/UNSD\ —\ Methodology.csv ~/Downloads/geonames-countries.csv /tmp/cgspace-countries.csv /tmp/mel-countries.csv> /tmp/countries.csv
```
- This mostly worked fine, and is much easier than writing another Python script with Pandas...
## 2022-06-24
- Spent some more time working on my `countries-to-csv.py` script to fix some logic errors
- Then re-export the UN M.49 countries to a clean list because the one I did yesterday somehow has errors:
```console
2024-03-19 07:01:13 +01:00
$ csvcut -d ';' -c 'ISO-alpha2 Code,Country or Area' ~/Downloads/UNSD\ —\ Methodology.csv | sed -e '1s/ISO-alpha2 Code/alpha2/' -e '1s/Country or Area/UN M.49 Name/' > ~/Downloads/un-countries.csv
2022-06-24 13:49:37 +02:00
```
- Check the number of lines in each file:
```
$ wc -l clarisa-countries.csv un-countries.csv cgspace-countries.csv mel-countries.csv
250 clarisa-countries.csv
250 un-countries.csv
198 cgspace-countries.csv
258 mel-countries.csv
```
- I am seeing strange results with csvjoin's `--outer` join that I need to keep unmatched terms from both left and right files...
- Using `xsv join --full` is giving me better results:
```
$ xsv join --full alpha2 ~/Downloads/clarisa-countries.csv alpha2 ~/Downloads/un-countries.csv | xsv select '!alpha2[1]' > /tmp/clarisa-un-xsv-full.csv
```
- Then adding the CGSpace and MEL countries:
```console
$ xsv join --full alpha2 /tmp/clarisa-un-xsv-full.csv alpha2 /tmp/cgspace-countries.csv | xsv select '!alpha2[1]' > /tmp/clarisa-un-cgspace-xsv-full.csv
$ xsv join --full alpha2 /tmp/clarisa-un-cgspace-xsv-full.csv alpha2 /tmp/mel-countries.csv | xsv select '!alpha2[1]' > /tmp/clarisa-un-cgspace-mel-xsv-full.csv
```
2022-06-26 17:11:33 +02:00
## 2022-06-26
- Start a harvest on AReS
2022-06-30 08:41:54 +02:00
## 2022-06-28
2023-04-27 22:10:13 +02:00
- Start working on the CGSpace subject export for FAO / AGROVOC
2022-06-30 08:41:54 +02:00
- First I exported a list of all metadata in our `dcterms.subject` and other center-specific subject fields with their counts:
```console
localhost/dspacetest= ☘ \COPY (SELECT DISTINCT text_value AS "subject", count(*) FROM metadatavalue WHERE dspace_object_id in (SELECT dspace_object_id FROM item) AND metadata_field_id IN (187, 120, 210, 122, 215, 127, 208, 124, 128, 123, 125, 135, 203, 236, 238, 119) GROUP BY "subject" ORDER BY count DESC) to /tmp/2022-06-28-cgspace-subjects.csv WITH CSV HEADER;
COPY 27010
```
- Then I extracted the subjects and looked them up against AGROVOC:
```console
$ csvcut -c subject /tmp/2022-06-28-cgspace-subjects.csv | sed '1d' > /tmp/2022-06-28-cgspace-subjects.txt
$ ./ilri/agrovoc-lookup.py -i /tmp/2022-06-28-cgspace-subjects.txt -o /tmp/2022-06-28-cgspace-subjects-results.csv
```
- I keep getting timeouts after every five or ten requests, so this will not be feasible for 27,000 subjects!
- I think I will have to write some custom script to use the AGROVOC RDF file
- Using rdflib to open the 1.2GB `agrovoc_lod.rdf` file takes several minutes and doesn't seem very efficient
2023-04-27 22:10:13 +02:00
- I tried using [lightrdf ](https://github.com/ozekik/lightrdf ) and it's much quicker, but the documentation is limited and I'm not sure how to search yet
2022-06-30 08:41:54 +02:00
- I had to try in different Python versions because 3.10.x is apparently too new
- For future reference I was able to search with lightrdf:
2022-08-03 20:01:39 +02:00
```python
2022-06-30 08:41:54 +02:00
import lightrdf
parser = lightrdf.Parser()
# prints millions of lines
for triple in parser.parse("./agrovoc_lod.rdf", base_iri=None):
print(triple)
agrovoc = lightrdf.RDFDocument('agrovoc_lod.rdf');
# all results for prefix http://aims.fao.org/aos/agrovoc/c_5
for triple in agrovoc.search_triples('http://aims.fao.org/aos/agrovoc/c_5', None, None):
print(triple)
('http://aims.fao.org/aos/agrovoc/c_5', 'http://www.w3.org/2004/02/skos/core#altLabel', '"Abalone"@de')
('http://aims.fao.org/aos/agrovoc/c_5', 'http://www.w3.org/2004/02/skos/core#prefLabel', '"abalones"@en')
# all stuff for abalones in English
for triple in agrovoc.search_triples(None, None, '"abalones"@en'):
print(triple)
```
- I ran the `agrovoc-lookup.py` from a Linode server and it completed without issues... hmmm
## 2022-06-29
- Continue working on the list of non-AGROVOC subject to report to FAO
2022-06-30 15:48:03 +02:00
- I got a one liner to get the list of non-AGROVOC subjects and join them with their counts (updated to use regex in csvgrep):
2022-06-30 08:41:54 +02:00
```console
2022-06-30 15:48:03 +02:00
$ csvgrep -c 'number of matches' -r '^0$' /tmp/2022-06-28-cgspace-subjects-results.csv \
2022-06-30 08:41:54 +02:00
| csvcut -c subject \
| csvjoin -c subject /tmp/2022-06-28-cgspace-subjects.csv - \
> /tmp/2022-06-28-cgspace-non-agrovoc.csv
```
2022-06-30 15:48:03 +02:00
## 2022-06-30
- Check some AfricaRice records for potential duplicates on CGSpace for Abenet:
```console
$ csvcut -l -c dc.title,dcterms.issued,dcterms.type ~/Downloads/Africarice_2ndBatch_ay.csv | sed '1s/line_number/id/' > /tmp/africarice.csv
$ csv-metadata-quality -i /tmp/africarice.csv -o /tmp/africarice-cleaned.csv -u
$ ./ilri/check-duplicates.py -i /tmp/africarice-cleaned.csv -u dspacetest -db dspacetest -p 'dom@in34sniper' -o /tmp/africarice-duplicates.csv
```
- Looking at the non-AGROVOC subjects again, I see some in our list that are duplicated in uppercase and lowercase, so I will run it again with all lowercase:
```console
localhost/dspacetest= ☘ \COPY (SELECT DISTINCT(lower(text_value)) AS "subject", count(*) FROM metadatavalue WHERE dspace_object_id in (SELECT dspace_object_id FROM item) AND metadata_field_id IN (187, 120, 210, 122, 215, 127, 208, 124, 128, 123, 125, 135, 203, 236, 238, 119) GROUP BY "subject" ORDER BY count DESC) to /tmp/2022-06-30-cgspace-subjects.csv WITH CSV HEADER;
```
- Also, I see there might be something wrong with my csvjoin because nigeria shows up in the final list as having not matched...
- Ah, I was using `csvgrep -m 0` to find rows that didn't match, but that also matched items that had 10, 100, 50, etc...
- We need to use a regex:
```console
$ csvgrep -c 'number of matches' -r '^0$' /tmp/2022-06-30-cgspace-subjects-results.csv \
| csvcut -c subject \
| csvjoin -c subject /tmp/2022-06-30-cgspace-subjects.csv - \
> /tmp/2022-06-30-cgspace-non-agrovoc.csv
```
- Then I took all the terms with fifty or more occurences and put them on a Google Sheet
- There I started removing any term that was a variation of an existing AGROVOC term (like cowpea/cowpeas, policy/policies) or a compound concept
2022-07-04 08:25:14 +02:00
- pnbecker on DSpace Slack mentioned that they made a JSPUI deduplication step that is open source: https://github.com/the-library-code/deduplication
- It uses Levenshtein distance via PostgreSQL's fuzzystrmatch extension
2022-06-30 15:48:03 +02:00
2022-06-06 08:45:43 +02:00
<!-- vim: set sw=2 ts=2: -->