Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/dfreelon/news_extract
Python module to extract articles from NexisUni and Factiva.
https://github.com/dfreelon/news_extract
Last synced: 3 months ago
JSON representation
Python module to extract articles from NexisUni and Factiva.
- Host: GitHub
- URL: https://github.com/dfreelon/news_extract
- Owner: dfreelon
- License: bsd-3-clause
- Created: 2019-07-10T18:23:39.000Z (over 5 years ago)
- Default Branch: master
- Last Pushed: 2019-07-11T15:50:20.000Z (over 5 years ago)
- Last Synced: 2024-07-08T14:28:21.119Z (4 months ago)
- Language: Python
- Size: 48.8 KB
- Stars: 36
- Watchers: 1
- Forks: 9
- Open Issues: 4
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# news_extract
Python module to extract articles from NexisUni and Factiva.
## Requirements* Python 3
* [pandas](https://pandas.pydata.org/)## Installation
```pip install news_extract```
## Overview
```news_extract``` allows the output of the NexisUni and Factiva databases to be imported into Python. **Note, you must export your documents manually first! This module does not scrape the databases directly; rather, it extracts articles and associated metadata from pre-exported output files.** To use it, you must subscribe to at least one of these databases and use the following instructions to export your articles from each database:
### NexisUni export instructions
1. Make sure you are exporting full documents with no attachments, not just the results list.
2. Export in RTF format. (Note: you can export up to 100 articles at a time if you create an individual NexisUni account and change your personal settings accordingly.)
3. Save documents in a single file.
4. Uncheck all options on the "Formatting Options" tab.### Factiva export instructions
1. For Factiva, you must export your documents using the Firefox browser.
2. After conducting your search, click the "View Selected Articles" button that looks like an eye.
3. On the right, click the "Display Options" text and select "Full Article/Report plus Indexing."
4. Click the "Format for Saving" button that looks like a 3.5" floppy disk and select "Article Format."
5. On the resulting page, select "Save Page As..." from the Firefox menu.
6. In the "Save as type" dropdown, select "Text Files" and save your file.
7. [This animated gif](https://rtemis.hypotheses.org/files/2017/02/Factiva-animated-tutorial.gif) shows how to do steps 2-4 (warning: French)Once you've exported your file(s), you can do the following:
```python
import news_extract as ne
nu_file = 'results1.rtf' #file exported from NexisUni
fc_file = 'results2.txt' #file exported from Factiva
nu_data = ne.nexis_rtf_extract(nu_file)
fc_data = ne.factiva_extract(fc_file)print(nu_data[0].keys()) #view field names for NexisUni articles
print(fc_data[0].keys()) #view field names for first Factiva articlefor i in nu_data:
print(i['HEADLINE']) #show all NexisUni headlines
for i in fc_data:
print(i['HD']) #show all Factiva headlines
```## Output
Both ```nexis_rtf_extract``` and ```factiva_extract``` return lists of dicts wherein each dict corresponds to an article. The dict keys are field names, while the dict values are the metadata. One major difference between the two functions is that ```nexis_rtf_extract``` outputs the same set of metadata for all articles, while ```factiva_extract``` auto-extracts the specific field names and values attached to each article. This is due to differences in how the two types of export files are formatted.
## Combining Factiva and NexisUni output
### Converting fieldnames
You can use the function ```fix_fac_fieldnames``` to convert Factiva fieldnames to their longer and more descriptive NexisUni equivalents like so:
```python
#note that this will only convert eight common field names, leaving the rest intact
fc_converted = ne.fix_fac_fieldnames(fc_data)
```### Merging Factiva and NexisUni data into a single Pandas variable
If you want to analyze data from NexisUni and Factiva in the same project, here's how to do it:
```python
nu_plus_fc = nu_data + fc_converted
combined = ne.news_export(nu_plus_fc)
```The ```news_export``` function performs several operations, including removing duplicates (using a custom algorithm based on the Jaccard coefficient and time of publication) and resolving conflicts between articles with different metadata fields. For the latter, the function attempts to export all fields included in at least half the articles by default. This proportion can be adjusted using the ```field_threshold``` parameter, which accepts proportions between 0 and 1. 0 will attempt to include every metadata field present in at least one article, while 1 will include only those fields present in all articles.
By default, ```news_export``` returns a Pandas DataFrame containing the output data. You can save individual JSON files to disk (i.e. one article per file) by setting the ```to_pandas``` parameter to ```False```.