Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/julien-c/ios11nlp-benchmark
https://github.com/julien-c/ios11nlp-benchmark
Last synced: about 1 month ago
JSON representation
- Host: GitHub
- URL: https://github.com/julien-c/ios11nlp-benchmark
- Owner: julien-c
- Created: 2017-08-30T08:43:17.000Z (about 7 years ago)
- Default Branch: master
- Last Pushed: 2017-08-30T17:12:27.000Z (about 7 years ago)
- Last Synced: 2024-10-02T23:41:14.486Z (about 2 months ago)
- Language: Perl
- Size: 1.46 MB
- Stars: 4
- Watchers: 3
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
## iOS 11: Apple's NLP vs. server-side NLP
Medium blog post: https://medium.com/huggingface/ios-11-are-apples-new-nlp-capabilities-game-changers-5ab71f706f00
TL;DR: Apple’s on-device NLP gets average accuracy on CoNLL datasets. Out of the box, spaCy (a server-side, Python based NLP framework) consistently gets better precision and recall.
The CoNLL dataset ships with `conlleval`, a Perl script that evaluates a model’s accuracy.
### Disclaimer
Important disclaimer: It is very important to note here, that we evaluate the model on a dataset that is different from the one that trained it.
NER is a general task, and the CoNLL dataset is designed to be generic enough and close to most use cases, so a good general NER model should perform well on the dataset. However, it’s not completely fair to directly compare F1 scores with those of models that were trained on this data (and only this data). Still, it gives us directionally correct indications.