https://github.com/zouharvi/stolen-subwords
Zero-data blackbox machine translation model distillation / stealing
https://github.com/zouharvi/stolen-subwords
machine-translation model-distillation
Last synced: 3 months ago
JSON representation
Zero-data blackbox machine translation model distillation / stealing
- Host: GitHub
- URL: https://github.com/zouharvi/stolen-subwords
- Owner: zouharvi
- Created: 2022-10-06T09:44:20.000Z (about 3 years ago)
- Default Branch: main
- Last Pushed: 2025-01-30T09:32:36.000Z (9 months ago)
- Last Synced: 2025-01-30T10:29:24.418Z (9 months ago)
- Topics: machine-translation, model-distillation
- Language: Python
- Homepage:
- Size: 827 KB
- Stars: 0
- Watchers: 2
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Machine Translation Vocabulary Stealing
[](https://arxiv.org/abs/2401.16055)
Code accompanying the report [Stolen Subwords: Importance of Vocabularies for Machine Translation Model Stealing](https://arxiv.org/abs/2401.16055).
> **Abstract**: In learning-based functionality stealing, the attacker is trying to build a local model based on the victim's outputs.
> The attacker has to make choices regarding the local model's architecture, optimization method and, specifically for NLP models, subword vocabulary, such as BPE.
> On the machine translation task, we explore (1) whether the choice of the vocabulary plays a role in model stealing scenarios and (2) if it is possible to extract the victim's vocabulary.
> We find that the vocabulary itself does not have a large effect on the local model's performance.
> Given gray-box model access, it is possible to collect the victim's vocabulary by collecting the outputs (detokenized subwords on the output).
> The results of the minimum effect of vocabulary choice are important more broadly for black-box knowledge distillation.