Neural Monkey is a universal toolkit for training neural models for sequence-to-sequence tasks. The system has been successfully tested on machine translation, multimodal machine translation, or automatic post-editing. It can be used, however, for many other tasks, including image captioning, part-of-speech tagging, sequence classification, etc.
Neural Monkey's primary goal is to allow for fast prototyping and easy extension, which makes it a toolkit-of-choice for researchers who want to implement and/or modify recently published techniques.
Neural Monkey is written in Python 3 and built on the TensorFlow library. It supports training on GPUs with a minimum required effort.
If you want to start using Neural Monkey, clone it from its GitHub page. There is a bunch of tutorials and plenty of other useful information either in the package README, or in the documentation.
Neural Monkey was used in the following publications:
If you make use of Neural Monkey, please consider citing the following paper:
@article{NeuralMonkey:2017, author = {Jind{\v{r}}ich Helcl and Jind{\v{r}}ich Libovick{\'{y}}}, title = {Neural Monkey: An Open-source Tool for Sequence Learning}, journal = {The Prague Bulletin of Mathematical Linguistics}, year = {2017}, address = {Prague, Czech Republic}, number = {107}, pages = {5--17}, issn = {0032-6585}, doi = {10.1515/pralin-2017-0001}, url = {http://ufal.mff.cuni.cz/pbml/107/art-helcl-libovicky.pdf} }