mirror of
https://github.com/alanorth/cgspace-notes.git
synced 2024-11-22 22:55:04 +01:00
Add notes for 2016-02-15
Signed-off-by: Alan Orth <alan.orth@gmail.com>
This commit is contained in:
parent
450965091c
commit
6a4cb0aca6
@ -185,3 +185,15 @@ Processing 64195.pdf
|
|||||||
- A few items link to PDFs on IFPRI's e-Library or Research Gate
|
- A few items link to PDFs on IFPRI's e-Library or Research Gate
|
||||||
- A few items have no item
|
- A few items have no item
|
||||||
- Also, I'm not sure if we import these items, will be remove the `dc.identifier.url` field from the records?
|
- Also, I'm not sure if we import these items, will be remove the `dc.identifier.url` field from the records?
|
||||||
|
|
||||||
|
## 2016-02-12
|
||||||
|
|
||||||
|
- Looking at CIAT's records again, there are some files linking to PDFs on Slide Share, Embrapa, UEA UK, and Condesan, so I'm not sure if we can use those
|
||||||
|
- 265 items have dirty, URL-encoded filenames:
|
||||||
|
|
||||||
|
```
|
||||||
|
$ ls | grep -c -E "%"
|
||||||
|
265
|
||||||
|
```
|
||||||
|
|
||||||
|
- I suggest that we import ~850 or so of the clean ones first, then do the rest after I can find a clean/reliable way to decode the filenames
|
||||||
|
@ -283,6 +283,21 @@ Processing 64195.pdf
|
|||||||
<li>A few items link to PDFs on IFPRI’s e-Library or Research Gate</li>
|
<li>A few items link to PDFs on IFPRI’s e-Library or Research Gate</li>
|
||||||
<li>A few items have no item</li>
|
<li>A few items have no item</li>
|
||||||
<li>Also, I’m not sure if we import these items, will be remove the <code>dc.identifier.url</code> field from the records?</li>
|
<li>Also, I’m not sure if we import these items, will be remove the <code>dc.identifier.url</code> field from the records?</li>
|
||||||
|
</ul>
|
||||||
|
|
||||||
|
<h2 id="2016-02-12-1:124a59adbaa8ef13e1518d003fc03981">2016-02-12</h2>
|
||||||
|
|
||||||
|
<ul>
|
||||||
|
<li>Looking at CIAT’s records again, there are some files linking to PDFs on Slide Share, Embrapa, UEA UK, and Condesan, so I’m not sure if we can use those</li>
|
||||||
|
<li>265 items have dirty, URL-encoded filenames:</li>
|
||||||
|
</ul>
|
||||||
|
|
||||||
|
<pre><code>$ ls | grep -c -E "%"
|
||||||
|
265
|
||||||
|
</code></pre>
|
||||||
|
|
||||||
|
<ul>
|
||||||
|
<li>I suggest that we import ~850 or so of the clean ones first, then do the rest after I can find a clean/reliable way to decode the filenames</li>
|
||||||
</ul>
|
</ul>
|
||||||
|
|
||||||
</section>
|
</section>
|
||||||
|
@ -222,6 +222,21 @@ Processing 64195.pdf
|
|||||||
<li>A few items have no item</li>
|
<li>A few items have no item</li>
|
||||||
<li>Also, I&rsquo;m not sure if we import these items, will be remove the <code>dc.identifier.url</code> field from the records?</li>
|
<li>Also, I&rsquo;m not sure if we import these items, will be remove the <code>dc.identifier.url</code> field from the records?</li>
|
||||||
</ul>
|
</ul>
|
||||||
|
|
||||||
|
<h2 id="2016-02-12-1:124a59adbaa8ef13e1518d003fc03981">2016-02-12</h2>
|
||||||
|
|
||||||
|
<ul>
|
||||||
|
<li>Looking at CIAT&rsquo;s records again, there are some files linking to PDFs on Slide Share, Embrapa, UEA UK, and Condesan, so I&rsquo;m not sure if we can use those</li>
|
||||||
|
<li>265 items have dirty, URL-encoded filenames:</li>
|
||||||
|
</ul>
|
||||||
|
|
||||||
|
<pre><code>$ ls | grep -c -E &quot;%&quot;
|
||||||
|
265
|
||||||
|
</code></pre>
|
||||||
|
|
||||||
|
<ul>
|
||||||
|
<li>I suggest that we import ~850 or so of the clean ones first, then do the rest after I can find a clean/reliable way to decode the filenames</li>
|
||||||
|
</ul>
|
||||||
</description>
|
</description>
|
||||||
</item>
|
</item>
|
||||||
|
|
||||||
|
@ -222,6 +222,21 @@ Processing 64195.pdf
|
|||||||
<li>A few items have no item</li>
|
<li>A few items have no item</li>
|
||||||
<li>Also, I&rsquo;m not sure if we import these items, will be remove the <code>dc.identifier.url</code> field from the records?</li>
|
<li>Also, I&rsquo;m not sure if we import these items, will be remove the <code>dc.identifier.url</code> field from the records?</li>
|
||||||
</ul>
|
</ul>
|
||||||
|
|
||||||
|
<h2 id="2016-02-12-1:124a59adbaa8ef13e1518d003fc03981">2016-02-12</h2>
|
||||||
|
|
||||||
|
<ul>
|
||||||
|
<li>Looking at CIAT&rsquo;s records again, there are some files linking to PDFs on Slide Share, Embrapa, UEA UK, and Condesan, so I&rsquo;m not sure if we can use those</li>
|
||||||
|
<li>265 items have dirty, URL-encoded filenames:</li>
|
||||||
|
</ul>
|
||||||
|
|
||||||
|
<pre><code>$ ls | grep -c -E &quot;%&quot;
|
||||||
|
265
|
||||||
|
</code></pre>
|
||||||
|
|
||||||
|
<ul>
|
||||||
|
<li>I suggest that we import ~850 or so of the clean ones first, then do the rest after I can find a clean/reliable way to decode the filenames</li>
|
||||||
|
</ul>
|
||||||
</description>
|
</description>
|
||||||
</item>
|
</item>
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user