Using Neural Networks for Modeling and Representing Natural Language

|
|
|
Using Neural Networks for Modeling and Representing Natural Language

Course Description

This course will be introduced basic concepts in natural language processing and in the area of artificial neural networks and deep learning. This course will give some basic algorithms that can be used to learn features from text and to build neural language models. Finally, this course will describe the algorithms used in fastText which can classify text extremely efficiently. This tutorial does not assume any extensive prior knowledge of neural networks and is supposed to be mainly helpful to beginners in the area of NLP.

Prerequisites

  • Basic knowledge of machine learning and math
Level of complexity of course Basics

Lecturer

Dr. Tomas Mikolov Researcher at CIIRC, Czech Technical University Tomas works on artificial intelligence research, and he is interested in building advanced models of language. His most famous results are the recurrent neural language models (RNNLM), vector representations of words (word2vec), and efficient text classification (fastText). Currently,Tomas is interested in developing less supervised models of language (and intelligence), which can develop in time through evolutionary mechanisms. Fields of interests: Deep learning, NLP, Complex Systems, Artificial Intelligence Contacts: tmikolov@gmail.com

Про факультет

Важлива інформація

Контактна інформація