18 | | 1. paper1 |
19 | | 1. paper2 |
20 | | 1. paper3 |
| 15 | 1. Bengio, Yoshua, et al. "A neural probabilistic language model." The Journal of Machine Learning Research 3 (2003): 1137-1155. |
| 16 | 1. Mikolov, Tomas, et al. "Efficient estimation of word representations in vector space." arXiv preprint arXiv:1301.3781 (2013). |
| 17 | 1. Mikolov, Tomas, et al. "Distributed representations of words and phrases and their compositionality." Advances in Neural Information Processing Systems. 2013. |
| 18 | 1. Chelba, Ciprian, et al. "One billion word benchmark for measuring progress in statistical language modeling." arXiv preprint arXiv:1312.3005 (2013). |
24 | | Concrete description of work assignment for students for the second one-hour part of the lecture. The work will consist of tasks connected with practical implementations of algorithms connected with the current topic (probably not the state-of-the-art algorithms mentioned in the first part) and with real data. Students can test the algorithms, evaluate them and possibly try some short adaptations for various subtasks. |
25 | | |
26 | | Students can also be required to generate some results of their work and hand them in to prove completing the tasks. |
| 22 | We will build a simple language model (skip-gram) which has very interesting properties. When trained properly, the vectors of words obey simple space arithmetics, e.g. |
| 23 | vector "king" − vector "man" + vector "woman" ~= vector of "queen". |
| 24 | We will train this model on a large Czech and English corpora and evaluate the result. |