M. I. Jordan, Serial order: A parallel, distributed processing approach Advances in Connectionist Theory: Speech, 1989.

J. L. Elman, Finding Structure in Time, Cognitive Science, vol.49, issue.2, pp.179-211, 1990.
DOI : 10.1207/s15516709cog1402_1

R. Collobert and J. Weston, A unified architecture for natural language processing, Proceedings of the 25th international conference on Machine learning, ICML '08, pp.160-167, 2008.
DOI : 10.1145/1390156.1390177

R. Collobert, J. Weston, L. Bottou, M. Karlen, K. Kavukcuoglu et al., Natural language processing (almost) from scratch, J. Mach. Learn. Res, pp.12-2493, 2011.

G. Mesnil, X. He, L. Deng, and Y. Bengio, Investigation of recurrent-neuralnetwork architectures and learning methods for spoken language understanding, 2013.

V. Vukotic, C. Raymond, and G. Gravier, Is it time to switch to word embedding and recurrent neural networks for spoken language understanding? In: InterSpeech, 2015.

W. Xu, M. Auli, and S. Clark, CCG Supertagging with a Recurrent Neural Network, Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pp.250-255, 2015.
DOI : 10.3115/v1/P15-2041

J. Lafferty, A. Mccallum, and F. Pereira, Conditional random fields: Probabilistic models for segmenting and labeling sequence data, Proceedings of the Eighteenth International Conference on Machine Learning (ICML), pp.282-289, 2001.

T. Mikolov, M. Karafiát, L. Burget, J. Cernock´ycernock´y, and S. Khudanpur, Recurrent neural network based language model, 11th Annual Conference of the International Speech Communication Association, pp.1045-1048, 2010.

T. Mikolov, S. Kombrink, L. Burget, J. Cernock, and S. Khudanpur, Extensions of recurrent neural network language model, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.5528-5531, 2011.
DOI : 10.1109/ICASSP.2011.5947611

O. Zennaki, N. Semmar, and L. Besacier, Unsupervised and lightly supervised partof-speech tagging using recurrent neural networks, Proceedings of the 29th Pacific Asia Conference on Language, Information and Computation, PACLIC 29, 2015.

T. Mikolov, K. Chen, G. Corrado, and J. Dean, Efficient estimation of word representations in vector space, p.3781, 2013.

T. Mikolov, W. Yih, and G. Zweig, Linguistic regularities in continuous space word representations In: Human Language Technologies: Conference of the North American Chapter, the Association of Computational Linguistics, pp.746-751, 2013.

A. Abeillé, L. Clément, and F. Toussenel, Building a Treebank for French, pp.165-188, 2003.
DOI : 10.1007/978-94-010-0201-1_10

P. Denis and B. Sagot, Coupling an annotated corpus and a lexicon for state-of-theart pos tagging, Lang. Resour. Eval, pp.46-721, 2012.

D. Mori, R. Bechet, F. Hakkani-tur, D. Mctear, M. Riccardi et al., Spoken language understanding, IEEE Signal Processing Magazine, vol.25, issue.3, pp.50-58, 2008.
DOI : 10.1109/MSP.2008.918413

URL : https://hal.archives-ouvertes.fr/hal-01314884

D. A. Dahl, M. Bates, M. Brown, W. Fisher, K. Hunicke-smith et al., Expanding the scope of the ATIS task, Proceedings of the workshop on Human Language Technology , HLT '94, pp.43-48, 1994.
DOI : 10.3115/1075812.1075823

H. Bonneau-maynard, C. Ayache, F. Bechet, A. Denis, A. Kuhn et al., Results of the french evalda-media evaluation campaign for literal understanding, pp.2054-2059, 2006.
URL : https://hal.archives-ouvertes.fr/hal-01160167

Y. Bengio, R. Ducharme, P. Vincent, and C. Jauvin, Neural Probabilistic Language Models, JOURNAL OF MACHINE LEARNING RESEARCH, vol.3, pp.1137-1155, 2003.
DOI : 10.1007/3-540-33486-6_6

URL : https://hal.archives-ouvertes.fr/hal-01434258

M. Schuster and K. Paliwal, Bidirectional recurrent neural networks, IEEE Transactions on Signal Processing, vol.45, issue.11, pp.2673-2681, 1997.
DOI : 10.1109/78.650093

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.331.9441

Y. Bengio, Practical Recommendations for Gradient-Based Training of Deep Architectures, p.5533, 2012.
DOI : 10.1162/089976602317318938

P. Werbos, Backpropagation through time: what it does and how to do it, Proceedings of IEEE, pp.1550-1560, 1990.
DOI : 10.1109/5.58337

D. Chen and C. Manning, A Fast and Accurate Dependency Parser using Neural Networks, Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp.740-750, 2014.
DOI : 10.3115/v1/D14-1082

L. Ramshaw and M. Marcus, Text Chunking Using Transformation-Based Learning, Proceedings of the 3rd Workshop on Very Large Corpora, pp.84-94, 1995.
DOI : 10.1007/978-94-017-2390-9_10

URL : http://arxiv.org/abs/cmp-lg/9505040

G. Mesnil, Y. Dauphin, K. Yao, Y. Bengio, L. Deng et al., Using Recurrent Neural Networks for Slot Filling in Spoken Language Understanding, Speech, and Language Processing, 2015.
DOI : 10.1109/TASLP.2014.2383614

N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, Dropout: A simple way to prevent neural networks from overfitting, Journal of Machine Learning Research, vol.15, pp.1929-1958, 2014.