Neural Monkey is a universal toolkit for training neural models for sequence-to-sequence tasks. The system has been successfully tested on machine translation, multimodal machine translation, or automatic post-editing. It can be used, however, for many other tasks, including image captioning, part-of-speech tagging, sequence classification, etc.
Neural Monkey's primary goal is to allow for fast prototyping and easy extension, which makes it a toolkit-of-choice for researchers who want to implement and/or modify recently published techniques.
Neural Monkey is written in Python 3 and built on the TensorFlow library. It supports training on GPUs with a minimum required effort.
If you want to start using Neural Monkey, clone it from its GitHub page. There is a bunch of tutorials and plenty of other useful information either in the package README, or in the documentation.
Neural Monkey was used in the following publications:
If you make use of Neural Monkey, please consider citing the following paper:
@article{NeuralMonkey:2017,
author = {Jind{\v{r}}ich Helcl and Jind{\v{r}}ich Libovick{\'{y}}},
title = {Neural Monkey: An Open-source Tool for Sequence Learning},
journal = {The Prague Bulletin of Mathematical Linguistics},
year = {2017},
address = {Prague, Czech Republic},
number = {107},
pages = {5--17},
issn = {0032-6585},
doi = {10.1515/pralin-2017-0001},
url = {http://ufal.mff.cuni.cz/pbml/107/art-helcl-libovicky.pdf}
}