https://github.com/hinthornw/supreme_predictions
Princeton COS 424 final project: Predicting Supreme Court Decisions
https://github.com/hinthornw/supreme_predictions
forests law random
Last synced: 2 months ago
JSON representation
Princeton COS 424 final project: Predicting Supreme Court Decisions
- Host: GitHub
- URL: https://github.com/hinthornw/supreme_predictions
- Owner: hinthornw
- Created: 2017-04-22T19:27:10.000Z (about 8 years ago)
- Default Branch: master
- Last Pushed: 2017-05-16T21:33:52.000Z (about 8 years ago)
- Last Synced: 2025-02-07T23:47:42.454Z (4 months ago)
- Topics: forests, law, random
- Language: Jupyter Notebook
- Homepage:
- Size: 3.73 MB
- Stars: 0
- Watchers: 7
- Forks: 1
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Predicting the Decisions of Supreme Court Justices
## Authors: William Hinthorn, Maia Ezratty, Mihika Kapoor, Alice Zheng
### COS 424 Final Project
For more notes, see our [wiki](https://github.com/hinthornw/supreme_predictions/wiki).
A pdf copy of our findings is stored [here](/writeup/424_Final_Project.pdf).
Dataset:
Harold J. Spaeth, Lee Epstein, Andrew D. Martin, Jeffrey A. Segal, Theodore J. Ruger, and Sara C. Benesh. 2016 Supreme Court Database, Version 2016 Release 01. URL: http://Supremecourtdatabase.orgTo download the dataset, go to the [website](http://scdb.wustl.edu/data.php) and download the case-centered [csv file](http://scdb.wustl.edu/_brickFiles/2016_01/SCDB_2016_01_caseCentered_Citation.csv.zip).
##### The dataset contains of 8,737 cases with the following headers:
caseId,
docketId,
caseIssuesId,
voteId,
dateDecision,
decisionType,
usCite,
sctCite,
ledCite,
lexisCite,
term,
naturalCourt,
chief,
docket,
caseName,
dateArgument,
dateRearg,
petitioner,
petitionerState,
respondent,
respondentState,
jurisdiction,
adminAction,
adminActionState,
threeJudgeFdc,
caseOrigin,
caseOriginState,
caseSource,
caseSourceState,
lcDisagreement,
certReason,
lcDisposition,
lcDispositionDirection,
declarationUncon,
caseDisposition,
partyWinning,
precedentAlteration,
voteUnclear,
issue,
issueArea,
decisionDirection,
decisionDirectionDissent,
authorityDecision1,
authorityDecision2,
lawType,
lawSupp,
lawMinor,
majOpinWriter,
majOpinAssigner,
splitVote,
majVotes, and
minVotes##### The second dataset is [justice centered](http://scdb.wustl.edu/_brickFiles/2016_01/SCDB_2016_01_justiceCentered_Citation.csv.zip) (same source).
It has field names:
caseId,
docketId,
caseIssuesId,
voteId,
dateDecision,
decisionType,
usCite,
sctCite,
ledCite,
lexisCite,
term,
naturalCourt,
chief,
docket,
caseName,
dateArgument,
dateRearg,
petitioner,
petitionerState,
respondent,
respondentState,
jurisdiction,
adminAction,
adminActionState,
threeJudgeFdc,
caseOrigin,
caseOriginState,
caseSource,
caseSourceState,
lcDisagreement,
certReason,
lcDisposition,
lcDispositionDirection,
declarationUncon,
caseDisposition,
caseDispositionUnusual,
partyWinning,
precedentAlteration,
voteUnclear,
issue,
issueArea,
decisionDirection,
decisionDirectionDissent,
authorityDecision1,
authorityDecision2,
lawType,
lawSupp,
lawMinor,
majOpinWriter,
majOpinAssigner,
splitVote,
majVotes,
minVotes,
justice,
justiceName,
vote,
opinion,
direction,
majority,
firstAgreement,
secondAgreement,We used feature selection, generation, and ranom forests classifiers to build a flexible model which predicts the decisions of the court with an average accuracy of ~80-90%. We obtained similar results by building dedicated models for individual justices.
Here are some cool findings:
We found out that judges tend to stay true to their original biases throughout the length of their term.

Except for a few outliers, like Harry Blackmun, who we assume was influenced by the world wars and the New Deal programs to turn an early conservative Justice into an liberally-minded older man. Note that the moderates tend to have shorter terms (though we do not imply any health benefits of having a political bias).
These biases play a huge role on the bias of the court at large. We show the cumulative sum of justice's voting biases spread out over the cases that are included within the dataset.

We looked at correlations in voting habits of justices over the length of the term covered by the modern dataset (seen below)

And generated features based on clustering methods:

We then used textual data from [CourtListener](https://www.courtlistener.com) and correlated bags of words with political bias of the individual Justice using there $\chi^2$ values. Below are some word clouds representing our findings.
The first two show words that are strongly correlated with a justice being in the majority vote (left) or dissent (right).

The next two show words strongly correlated with a Justice voting in along a more liberal line (left) or conservative one (right)

