Differences
This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision | ||
course:esslli2021:start [2021/08/04 19:30] schtepf [Organizational information] |
course:esslli2021:start [2022/08/11 12:26] schtepf [Hands-on Distributional Semantics (ESSLLI 2021 / 2022)] |
||
---|---|---|---|
Line 1: | Line 1: | ||
- | ====== Hands-on Distributional Semantics (ESSLLI 2021) ====== | + | ====== Hands-on Distributional Semantics (ESSLLI 2021 / 2022) ====== |
**Hands-on Distributional Semantics – From first steps to interdisciplinary applications** | **Hands-on Distributional Semantics – From first steps to interdisciplinary applications** | ||
+ | [[https:// | ||
[[https:// | [[https:// | ||
\\ | \\ | ||
// | // | ||
+ | \\ | ||
+ | |||
+ | **Hands-on Distributional Semantics for Linguistics using R** | ||
+ | \\ | ||
+ | // | ||
+ | \\ | ||
+ | |||
* [[course: | * [[course: | ||
Line 10: | Line 18: | ||
{{: | {{: | ||
- | * check this page for updates and instructions by **Friday, August 6th, 2021** (late afternoon CEST) | + | * update of all materials |
- | * please install software and download data sets on the weekend before the course | + | * //Thanks for attending our course! It's a pleasure working with you.// |
===== Course description ===== | ===== Course description ===== | ||
- | Distributional semantic models (DSM) are based on the assumption that the meaning of a word can (at least to a certain extent) be inferred from its usage, i.e. its distribution in text. Therefore, these models dynamically build semantic representations through a statistical analysis of the contexts in which words occur. DSMs are a promising technique for solving the lexical acquisition bottleneck by unsupervised learning, and their distributed representation provides a cognitively plausible, robust and flexible architecture for the organisation and processing of semantic information. | + | Distributional semantic models (DSM) – also known as “word space”, “distributional similarity”, |
- | In this introductory course we will highlight the interdisciplinary potential of DSM beyond standard semantic similarity tasks, with applications in cognitive modeling and theoretical linguistics. This course aims to equip participants with the background knowledge and skills needed to build different kinds of DSM representations and apply them to a wide range of tasks. | + | In this introductory course we will highlight the interdisciplinary potential of DSM beyond standard semantic similarity tasks, with applications in cognitive modeling and theoretical linguistics. This course aims to equip participants with the background knowledge and skills needed to build different kinds of DSM representations |
- | **Lecturers: | + | |
+ | **Lecturers: | ||
===== Organizational information ===== | ===== Organizational information ===== | ||
- | Please make sure you have up-to-date versions of **[[https:// | + | Please make sure you have up-to-date versions of **[[https:// |
+ | Additional instructions will be given in the first session on Monday. In particular, you will be asked to download and install the '' | ||
- | We will answer questions during lectures and in the afternoon via the course' | + | /* We will answer questions during lectures and in the afternoon via the course' |
- | {{: | ||
- | |||
- | \\ | ||
- | **Further information on required software packages and data sets will be provided by Saturday, August 7th, 2021.** | ||
- | \\ | ||
- | \\ | ||
===== Schedule & handouts ===== | ===== Schedule & handouts ===== | ||
- | |||
- | {{: | ||
- | |||
- | \\ | ||
- | **Handouts and other materials will be made available along with the course.** | ||
- | \\ | ||
- | \\ | ||
- | |||
- | |||
- | <!-- | ||
=== Day 1: Introduction === | === Day 1: Introduction === | ||
- | {{:course:esslli2018:dsm_tutorial_part1.slides.pdf|Presentation | + | [[http:// |
- | + | ||
- | * motivation and brief history of distributional semantics | + | |
- | * common DSM architectures & prototypical applications | + | |
- | * first practical exercises with the '' | + | |
- | + | ||
- | + | ||
- | === Day 2: The parameters of a DSM === | + | |
- | + | ||
- | {{: | + | |
+ | * motivation and geometric intuition | ||
+ | * distributional vs. semantic similarity | ||
+ | * outline of the course | ||
+ | * practice: //software setup, first steps with the '' | ||
- | * taxonomy of DSM parameters: context representation, | + | === Day 2: Building a DSM === |
- | * overview of common parameter settings & best-practice recommendations | + | |
- | * practical exercises: building DSMs and exploring their parameters | + | |
+ | [[http:// | ||
+ | [[http:// | ||
+ | R code: [[http:// | ||
+ | bonus material: [[http:// | ||
- | === Day 3: Applications | + | * formal definition of a DSM, taxonomy of parameters |
+ | * collecting co-occurrence data: what counts as a context? | ||
+ | * mathematical operations on DSM vectors | ||
+ | * computing distances/ | ||
+ | * practice: //building DSMs and exploring different parameter settings// | ||
- | {{:course: | + | === Day 3: Which aspects of meaning does a DSM capture? === |
- | * attributional and relational similarity, clustering and semantic categorization, | + | [[http:// |
- | /* * supervised & unsupervised classification based on DSM vectors */ | + | [[http://wordspace.collocations.de/ |
- | * insights from recent parameter evaluation studies | + | R code: [[http:// |
- | * practical exercises: implementation and evaluation of selected tasks | + | |
+ | * evaluation: conceptual coordinates | ||
+ | * standard evaluation tasks (multiple choice, correlation, | ||
+ | * narrowing down similarity: classifying semantic relations | ||
+ | * practice: // | ||
- | === Day 4: Elements of matrix algebra | + | === Day 4: DS beyond NLP – Linguistic theory |
- | {{:course:esslli2018:dsm_tutorial_part4.slides.pdf|Presentation | + | [[http:// |
+ | [[http:// | ||
+ | R code: [[http:// | ||
+ | bonus material: [[http:// | ||
- | * basic matrix and vector operations, orthogonal projection & dimensionality reduction | ||
- | * singular value decomposition (SVD) | ||
- | * practical exercises: roll your own DSM with matrix operations | ||
+ | * linguistic exploitation of DSM representations | ||
+ | * a textbook challenge for DSMs: polysemy | ||
+ | * success stories: semantic compositionality, | ||
+ | * issues: not all words have a distributional meaning | ||
+ | * practice: //different exercises with linguistic data sets// | ||
- | === Day 5: Making sense of DSMs === | + | === Day 5: DS beyond NLP – Cognitive modelling |
- | * mathematical properties of and relations between different types of DSM | + | [[http:// |
- | * singular value decomposition | + | [[http:// |
- | * comparison with neural vector embeddings | + | R code: [[http:// |
+ | bonus task: [[http:// | ||
+ | bonus material: [[http:// | ||
- | --> | + | * DSMs for cognitive modelling |
+ | * free association norms as a window into the mental lexicon | ||
+ | * predicting free associations with DSMs | ||
+ | * practice: //combining DSMs with first-order co-occurrence for the FAST task// | ||