Update notes for 2017-02-28

This commit is contained in:
2017-02-28 22:58:29 +02:00
parent a3f0d88945
commit 56a24bf456
5 changed files with 59 additions and 9 deletions

View File

@ -310,4 +310,14 @@ dspace=# \copy (select resource_id, metadata_value_id from metadatavalue where r
COPY 1968
```
- And then using awk or uniq to either remove or print the lines that have a duplicate `resource_id` (meaning they belong to the same item in DSpace and are therefore duplicates), and then using the `metadata_value_id` to delete them
- And then use awk to print the duplicate lines to a separate file:
```
$ awk -F',' 'seen[$1]++' /tmp/ciat.csv > /tmp/ciat-dupes.csv
```
- From that file I can create a list of 279 deletes and put them in a batch script like:
```
delete from metadatavalue where resource_type_id=2 and metadata_field_id=3 and metadata_value_id=2742061;
```