Mrinmaya Sachan: Katalogdaten im Frühjahrssemester 2023 |
Name | Herr Prof. Dr. Mrinmaya Sachan |
Lehrgebiet | Maschinelles Lernen und Natürliche Sprachverarbeitung |
Adresse | Masch.Lernen und Nat.Sprachverarb. ETH Zürich, OAT Y 22.2 Andreasstrasse 5 8092 Zürich SWITZERLAND |
mrinmaya.sachan@inf.ethz.ch | |
Departement | Informatik |
Beziehung | Assistenzprofessor (Tenure Track) |
Nummer | Titel | ECTS | Umfang | Dozierende | |
---|---|---|---|---|---|
252-0945-16L | Doctoral Seminar Machine Learning (FS23) Only for Computer Science Ph.D. students. This doctoral seminar is intended for PhD students affiliated with the Institute for Machine Learning. Other PhD students who work on machine learning projects or related topics need approval by at least one of the organizers to register for the seminar. | 2 KP | 1S | N. He, V. Boeva, J. M. Buhmann, R. Cotterell, T. Hofmann, A. Krause, M. Sachan, J. Vogt, F. Yang | |
Kurzbeschreibung | An essential aspect of any research project is dissemination of the findings arising from the study. Here we focus on oral communication, which includes: appropriate selection of material, preparation of the visual aids (slides and/or posters), and presentation skills. | ||||
Lernziel | The seminar participants should learn how to prepare and deliver scientific talks as well as to deal with technical questions. Participants are also expected to actively contribute to discussions during presentations by others, thus learning and practicing critical thinking skills. | ||||
Voraussetzungen / Besonderes | This doctoral seminar of the Machine Learning Laboratory of ETH is intended for PhD students who work on a machine learning project, i.e., for the PhD students of the ML lab. | ||||
263-5000-00L | Computational Semantics for Natural Language Processing | 6 KP | 2V + 1U + 2A | M. Sachan | |
Kurzbeschreibung | This course presents an introduction to Natural language processing (NLP) with an emphasis on computational semantics i.e. the process of constructing and reasoning with meaning representations of natural language text. | ||||
Lernziel | The objective of the course is to learn about various topics in computational semantics and its importance in natural language processing methodology and research. Exercises and the project will be key parts of the course so the students will be able to gain hands-on experience with state-of-the-art techniques in the field. | ||||
Inhalt | We will take a modern view of the topic, and focus on various statistical and deep learning approaches for computation semantics. We will also overview various primary areas of research in language processing and discuss how the computational semantics view can help us make advances in NLP. | ||||
Skript | Lecture slides will be made available at the course Web site. | ||||
Literatur | No textbook is required, but there will be regularly assigned readings from research literature, linked to the course website. | ||||
Voraussetzungen / Besonderes | The student should have successfully completed a graduate level class in machine learning (252-0220-00L), deep learning (263-3210-00L) or natural language processing (252-3005-00L) before. Similar courses from other universities are acceptable too. | ||||
263-5354-00L | Large Language Models | 8 KP | 3V + 2U + 2A | R. Cotterell, M. Sachan, F. Tramèr, C. Zhang | |
Kurzbeschreibung | Large language models have become one of the most commonly deployed NLP inventions. In the past half-decade, their integration into core natural language processing tools has dramatically increased the performance of such tools, and they have entered the public discourse surrounding artificial intelligence. | ||||
Lernziel | To understand the mathematical foundations of large language models as well as how to implement them. | ||||
Inhalt | We start with the probabilistic foundations of language models, i.e., covering what constitutes a language model from a formal, theoretical perspective. We then discuss how to construct and curate training corpora, and introduce many of the neural-network architectures often used to instantiate language models at scale. The course covers aspects of systems programming, discussion of privacy and harms, as well as applications of language models in NLP and beyond. | ||||
Literatur | The lecture notes will be supplemented with various readings from the literature. |