Suchergebnis: Katalogdaten im Frühjahrssemester 2020

DAS in Data Science Information
Kernfächer
Einführungskurse
NummerTitelTypECTSUmfangDozierende
252-0220-00LIntroduction to Machine Learning Information Belegung eingeschränkt - Details anzeigen
Limited number of participants. Preference is given to students in programmes in which the course is being offered. All other students will be waitlisted. Please do not contact Prof. Krause for any questions in this regard. If necessary, please contact Link
W8 KP4V + 2U + 1AA. Krause
KurzbeschreibungThe course introduces the foundations of learning and making predictions based on data.
LernzielThe course will introduce the foundations of learning and making predictions from data. We will study basic concepts such as trading goodness of fit and model complexitiy. We will discuss important machine learning algorithms used in practice, and provide hands-on experience in a course project.
Inhalt- Linear regression (overfitting, cross-validation/bootstrap, model selection, regularization, [stochastic] gradient descent)
- Linear classification: Logistic regression (feature selection, sparsity, multi-class)
- Kernels and the kernel trick (Properties of kernels; applications to linear and logistic regression); k-nearest neighbor
- Neural networks (backpropagation, regularization, convolutional neural networks)
- Unsupervised learning (k-means, PCA, neural network autoencoders)
- The statistical perspective (regularization as prior; loss as likelihood; learning as MAP inference)
- Statistical decision theory (decision making based on statistical models and utility functions)
- Discriminative vs. generative modeling (benefits and challenges in modeling joint vy. conditional distributions)
- Bayes' classifiers (Naive Bayes, Gaussian Bayes; MLE)
- Bayesian approaches to unsupervised learning (Gaussian mixtures, EM)
LiteraturTextbook: Kevin Murphy, Machine Learning: A Probabilistic Perspective, MIT Press
Voraussetzungen / BesonderesDesigned to provide a basis for following courses:
- Advanced Machine Learning
- Deep Learning
- Probabilistic Artificial Intelligence
- Seminar "Advanced Topics in Machine Learning"
401-3632-00LComputational StatisticsW8 KP3V + 1UM. H. Maathuis
KurzbeschreibungWe discuss modern statistical methods for data analysis, including methods for data exploration, prediction and inference. We pay attention to algorithmic aspects, theoretical properties and practical considerations. The class is hands-on and methods are applied using the statistical programming language R.
LernzielThe student obtains an overview of modern statistical methods for data analysis, including their algorithmic aspects and theoretical properties. The methods are applied using the statistical programming language R.
Voraussetzungen / BesonderesAt least one semester of (basic) probability and statistics.

Programming experience is helpful but not required.
Capstone Project
NummerTitelTypECTSUmfangDozierende
266-0100-00LCapstone Project Belegung eingeschränkt - Details anzeigen
Only for DAS in Data Science.
O8 KP17AF. Perez Cruz, O. Verscheure, T. Hofmann
KurzbeschreibungThe capstone project is part of the DAS in Data Science and is an opportunity to apply the knowledge acquired in the program in an independent, real-world project.
LernzielTo apply the knowledge acquired in the program in an independent, real-world project.
InhaltThe capstone project can be done under the supervision of the Swiss Data Science Center, or of any core or adjunct faculty of Data Science.

The project has to be finished within 6 months.
Deadline for a project the following semester conducted at the SDSC is mid June/mid December.
Vertiefungen
Hardware for Machine Learning
NummerTitelTypECTSUmfangDozierende
227-0150-00LSystems-on-chip for Data Analytics and Machine Learning
Previously "Energy-Efficient Parallel Computing Systems for Data Analytics"
W6 KP4GL. Benini
KurzbeschreibungSystems-on-chip architecture and related design issues with a focus on machine learning and data analytics applications. It will cover multi-cores, many-cores, vector engines, GP-GPUs, application-specific processors and heterogeneous compute accelerators. Special emphasis given to energy-efficiency issues and hardware-software techniques for power and energy minimization.
LernzielGive in-depth understanding of the links and dependencies between architectures and their energy-efficient implementation and to get a comprehensive exposure to state-of-the-art systems-on-chip platforms for machine learning and data analytics. Practical experience will also be gained through practical exercises and mini-projects (hardware and software) assigned on specific topics.
InhaltThe course will cover advanced system-on-chip architectures, with an in-depth view on design challenges related to advanced silicon technology and state-of-the-art system integration options (nanometer silicon technology, novel storage devices, three-dimensional integration, advanced system packaging). The emphasis will be on programmable parallel architectures with application focus on machine learning and data analytics. The main SoC architectural families will be covered: namely, multi and many- cores, GPUs, vector accelerators, application-specific processors, heterogeneous platforms. The course will cover the complex design choices required to achieve scalability and energy proportionality. The course will will also delve into system design, touching on hardware-software tradeoffs and full-system analysis and optimization taking into account non-functional constraints and quality metrics, such as power consumption, thermal dissipation, reliability and variability. The application focus will be on machine learning both in the cloud and at the edges (near-sensor analytics).
SkriptSlides will be provided to accompany lectures. Pointers to scientific literature will be given. Exercise scripts and tutorials will be provided.
LiteraturJohn L. Hennessy, David A. Patterson, Computer Architecture: A Quantitative Approach (The Morgan Kaufmann Series in Computer Architecture and Design) 6th Edition, 2017.
Voraussetzungen / BesonderesKnowledge of digital design at the level of "Design of Digital Circuits SS12" is required.

Knowledge of basic VLSI design at the level of "VLSI I: Architectures of VLSI Circuits" is required
227-0155-00LMachine Learning on Microcontrollers Belegung eingeschränkt - Details anzeigen
Registration in this class requires the permission of the instructors. Class size will be limited to 30.
Preference is given to students in the MSc EEIT.
W6 KP3G + 2AM. Magno, L. Benini
KurzbeschreibungMachine Learning (ML) and artificial intelligence are pervading the digital society. Today, even low power embedded systems are incorporating ML, becoming increasingly “smart”. This lecture gives an overview of ML methods and algorithms to process and extract useful near-sensor information in end-nodes of the “internet-of-things”, using low-power microcontrollers/ processors (ARM-Cortex-M; RISC-V)
LernzielLearn how to Process data from sensors and how to extract useful information with low power microprocessors using ML techniques. We will analyze data coming from real low-power sensors (accelerometers, microphones, ExG bio-signals, cameras…). The main objective is to study in details how Machine Learning algorithms can be adapted to the performance constraints and limited resources of low-power microcontrollers.
InhaltThe final goal of the course is a deep understanding of machine learning and its practical implementation on single- and multi-core microcontrollers, coupled with performance and energy efficiency analysis and optimization. The main topics of the course include:

- Sensors and sensor data acquisition with low power embedded systems

- Machine Learning: Overview of supervised and unsupervised learning and in particular supervised learning (Bayes Decision Theory, Decision Trees, Random Forests, kNN-Methods, Support Vector Machines, Convolutional Networks and Deep Learning)

- Low-power embedded systems and their architecture. Low Power microcontrollers (ARM-Cortex M) and RISC-V-based Parallel Ultra Low Power (PULP) systems-on-chip.

- Low power smart sensor system design: hardware-software tradeoffs, analysis, and optimization. Implementation and performance evaluation of ML in battery-operated embedded systems.

The laboratory exercised will show how to address concrete design problems, like motion, gesture recognition, emotion detection, image and sound classification, using real sensors data and real MCU boards.

Presentations from Ph.D. students and the visit to the Digital Circuits and Systems Group will introduce current research topics and international research projects.
SkriptScript and exercise sheets. Books will be suggested during the course.
Voraussetzungen / BesonderesPrerequisites: Good experience in C language programming. Microprocessors and computer architecture. Basics of Digital Signal Processing. Some exposure to machine learning concepts is also desirable.
Image Analysis & Computer Vision
NummerTitelTypECTSUmfangDozierende
227-0391-00LMedical Image Analysis
Basic knowledge of computer vision would be helpful.
W3 KP2GE. Konukoglu, M. A. Reyes Aguirre
KurzbeschreibungIt is the objective of this lecture to introduce the basic concepts used
in Medical Image Analysis. In particular the lecture focuses on shape
representation schemes, segmentation techniques, machine learning based predictive models and various image registration methods commonly used in Medical Image Analysis applications.
LernzielThis lecture aims to give an overview of the basic concepts of Medical Image Analysis and its application areas.
Voraussetzungen / BesonderesPrerequisites:
Basic concepts of mathematical analysis and linear algebra.

Preferred:
Basic knowledge of computer vision and machine learning would be helpful.

The course will be held in English.
227-1034-00LComputational Vision (University of Zurich)
No enrolment to this course at ETH Zurich. Book the corresponding module directly at UZH.
UZH Module Code: INI402

Mind the enrolment deadlines at UZH:
Link
W6 KP2V + 1UD. Kiper
KurzbeschreibungThis course focuses on neural computations that underlie visual perception. We study how visual signals are processed in the retina, LGN and visual cortex. We study the morpholgy and functional architecture of cortical circuits responsible for pattern, motion, color, and three-dimensional vision.
LernzielThis course considers the operation of circuits in the process of neural computations. The evolution of neural systems will be considered to demonstrate how neural structures and mechanisms are optimised for energy capture, transduction, transmission and representation of information. Canonical brain circuits will be described as models for the analysis of sensory information. The concept of receptive fields will be introduced and their role in coding spatial and temporal information will be considered. The constraints of the bandwidth of neural channels and the mechanisms of normalization by neural circuits will be discussed.
The visual system will form the basis of case studies in the computation of form, depth, and motion. The role of multiple channels and collective computations for object recognition will
be considered. Coordinate transformations of space and time by cortical and subcortical mechanisms will be analysed. The means by which sensory and motor systems are integrated to allow for adaptive behaviour will be considered.
InhaltThis course considers the operation of circuits in the process of neural computations. The evolution of neural systems will be considered to demonstrate how neural structures and mechanisms are optimised for energy capture, transduction, transmission and representation of information. Canonical brain circuits will be described as models for the analysis of sensory information. The concept of receptive fields will be introduced and their role in coding spatial and temporal information will be considered. The constraints of the bandwidth of neural channels and the mechanisms of normalization by neural circuits will be discussed.
The visual system will form the basis of case studies in the computation of form, depth, and motion. The role of multiple channels and collective computations for object recognition will
be considered. Coordinate transformations of space and time by cortical and subcortical mechanisms will be analysed. The means by which sensory and motor systems are integrated to allow for adaptive behaviour will be considered.
LiteraturBooks: (recommended references, not required)
1. An Introduction to Natural Computation, D. Ballard (Bradford Books, MIT Press) 1997.
2. The Handbook of Brain Theorie and Neural Networks, M. Arbib (editor), (MIT Press) 1995.
252-0579-00L3D Vision Information W5 KP3G + 1AM. Pollefeys, V. Larsson
KurzbeschreibungThe course covers camera models and calibration, feature tracking and matching, camera motion estimation via simultaneous localization and mapping (SLAM) and visual odometry (VO), epipolar and mult-view geometry, structure-from-motion, (multi-view) stereo, augmented reality, and image-based (re-)localization.
LernzielAfter attending this course, students will:
1. understand the core concepts for recovering 3D shape of objects and scenes from images and video.
2. be able to implement basic systems for vision-based robotics and simple virtual/augmented reality applications.
3. have a good overview over the current state-of-the art in 3D vision.
4. be able to critically analyze and asses current research in this area.
InhaltThe goal of this course is to teach the core techniques required for robotic and augmented reality applications: How to determine the motion of a camera and how to estimate the absolute position and orientation of a camera in the real world. This course will introduce the basic concepts of 3D Vision in the form of short lectures, followed by student presentations discussing the current state-of-the-art. The main focus of this course are student projects on 3D Vision topics, with an emphasis on robotic vision and virtual and augmented reality applications.
263-3710-00LMachine Perception Information Belegung eingeschränkt - Details anzeigen
Number of participants limited to 200.
W5 KP2V + 1U + 1AO. Hilliges
KurzbeschreibungRecent developments in neural networks (aka “deep learning”) have drastically advanced the performance of machine perception systems in a variety of areas including computer vision, robotics, and intelligent UIs. This course is a deep dive into deep learning algorithms and architectures with applications to a variety of perceptual tasks.
LernzielStudents will learn about fundamental aspects of modern deep learning approaches for perception. Students will learn to implement, train and debug their own neural networks and gain a detailed understanding of cutting-edge research in learning-based computer vision, robotics and HCI. The final project assignment will involve training a complex neural network architecture and applying it on a real-world dataset of human activity.

The core competency acquired through this course is a solid foundation in deep-learning algorithms to process and interpret human input into computing systems. In particular, students should be able to develop systems that deal with the problem of recognizing people in images, detecting and describing body parts, inferring their spatial configuration, performing action/gesture recognition from still images or image sequences, also considering multi-modal data, among others.
InhaltWe will focus on teaching: how to set up the problem of machine perception, the learning algorithms, network architectures and advanced deep learning concepts in particular probabilistic deep learning models

The course covers the following main areas:
I) Foundations of deep-learning.
II) Probabilistic deep-learning for generative modelling of data (latent variable models, generative adversarial networks and auto-regressive models).
III) Deep learning in computer vision, human-computer interaction and robotics.

Specific topics include: 
I) Deep learning basics:
a) Neural Networks and training (i.e., backpropagation)
b) Feedforward Networks
c) Timeseries modelling (RNN, GRU, LSTM)
d) Convolutional Neural Networks for classification
II) Probabilistic Deep Learning:
a) Latent variable models (VAEs)
b) Generative adversarial networks (GANs)
c) Autoregressive models (PixelCNN, PixelRNN, TCNs)
III) Deep Learning techniques for machine perception:
a) Fully Convolutional architectures for dense per-pixel tasks (i.e., instance segmentation)
b) Pose estimation and other tasks involving human activity
c) Deep reinforcement learning
IV) Case studies from research in computer vision, HCI, robotics and signal processing
LiteraturDeep Learning
Book by Ian Goodfellow and Yoshua Bengio
Voraussetzungen / BesonderesThis is an advanced grad-level course that requires a background in machine learning. Students are expected to have a solid mathematical foundation, in particular in linear algebra, multivariate calculus, and probability. The course will focus on state-of-the-art research in deep-learning and will not repeat basics of machine learning

Please take note of the following conditions:
1) The number of participants is limited to 200 students (MSc and PhDs).
2) Students must have taken the exam in Machine Learning (252-0535-00) or have acquired equivalent knowledge
3) All practical exercises will require basic knowledge of Python and will use libraries such as TensorFlow, scikit-learn and scikit-image. We will provide introductions to TensorFlow and other libraries that are needed but will not provide introductions to basic programming or Python.

The following courses are strongly recommended as prerequisite:
* "Visual Computing" or "Computer Vision"

The course will be assessed by a final written examination in English. No course materials or electronic devices can be used during the examination. Note that the examination will be based on the contents of the lectures, the associated reading materials and the exercises.
Neural Information Processing
NummerTitelTypECTSUmfangDozierende
227-0395-00LNeural SystemsW6 KP2V + 1U + 1AR. Hahnloser, M. F. Yanik, B. Grewe
KurzbeschreibungThis course introduces principles of information processing in neural systems. It covers basic neuroscience for engineering students, experiment techniques used in animal research and methods for inferring neural mechanisms. Students learn about neural information processing and basic principles of natural intelligence and their impact on artificially intelligent systems.
LernzielThis course introduces
- Basic neurophysiology and mathematical descriptions of neurons
- Methods for dissecting animal behavior
- Neural recordings in intact nervous systems and information decoding principles
- Methods for manipulating the state and activity in selective neuron types
- Neuromodulatory systems and their computational roles
- Reward circuits and reinforcement learning
- Imaging methods for reconstructing the synaptic networks among neurons
- Birdsong and language
- Neurobiological principles for machine learning.
InhaltFrom active membranes to propagation of action potentials. From synaptic physiology to synaptic learning rules. From receptive fields to neural population decoding. From fluorescence imaging to connectomics. Methods for reading and manipulation neural ensembles. From classical conditioning to reinforcement learning. From the visual system to deep convolutional networks. Brain architectures for learning and memory. From birdsong to computational linguistics.
Voraussetzungen / BesonderesBefore taking this course, students are encouraged to complete "Bioelectronics and Biosensors" (227-0393-10L).

As part of the exercises for this class, students are expected to complete a programming or literature review project to be defined at the beginning of the semester.
227-0973-00LTranslational Neuromodeling Information W8 KP3V + 2U + 1AK. Stephan
KurzbeschreibungThis course provides a systematic introduction to Translational Neuromodeling (the development of mathematical models for diagnostics of brain diseases) and their application to concrete clinical questions (Computational Psychiatry/Psychosomatics). It focuses on a generative modeling strategy and teaches (hierarchical) Bayesian models of neuroimaging data and behaviour, incl. exercises.
LernzielTo obtain an understanding of the goals, concepts and methods of Translational Neuromodeling and Computational Psychiatry/Psychosomatics, particularly with regard to Bayesian models of neuroimaging (fMRI, EEG) and behavioural data.
InhaltThis course provides a systematic introduction to Translational Neuromodeling (the development of mathematical models for diagnostics of brain diseases) and their application to concrete clinical questions (Computational Psychiatry/Psychosomatics). The first part of the course will introduce disease concepts from psychiatry and psychosomatics, their history, and clinical priority problems. The second part of the course concerns computational modeling of neuronal and cognitive processes for clinical applications. A particular focus is on Bayesian methods and generative models, for example, dynamic causal models for inferring neuronal processes from neuroimaging data, and hierarchical Bayesian models for inference on cognitive processes from behavioural data. The course discusses the mathematical and statistical principles behind these models, illustrates their application to various psychiatric diseases, and outlines a general research strategy based on generative models.

Lecture topics include:
1. Introduction to Translational Neuromodeling and Computational Psychiatry/Psychosomatics
2. Psychiatric nosology
3. Pathophysiology of psychiatric disease mechanisms
4. Principles of Bayesian inference and generative modeling
5. Variational Bayes (VB)
6. Bayesian model selection
7. Markov Chain Monte Carlo techniques (MCMC)
8. Bayesian frameworks for understanding psychiatric and psychosomatic diseases
9. Generative models of fMRI data
10. Generative models of electrophysiological data
11. Generative models of behavioural data
12. Computational concepts of schizophrenia, depression and autism
13. Model-based predictions about individual patients

Practical exercises include mathematical derivations and the implementation of specific models and inference methods. In additional project work, students are required to use one of the examples discussed in the course as a basis for developing their own generative model and use it for simulations and/or inference in application to a clinical question. Group work (up to 3 students) is permitted.
LiteraturSee TNU website:
Link
Voraussetzungen / BesonderesGood knowledge of principles of statistics, good programming skills (MATLAB or Python)
227-1032-00LNeuromorphic Engineering II Information
Information für UZH Studierende:
Die Lerneinheit kann nur an der ETH belegt werden. Die Belegung des Moduls INI405 ist an der UZH nicht möglich.

Beachten Sie die Einschreibungstermine an der ETH für UZH Studierende: Link
W6 KP5GS.‑C. Liu, T. Delbrück, G. Indiveri
KurzbeschreibungThis course teaches the basics of analog chip design and layout with an emphasis on neuromorphic circuits, which are introduced in the fall semester course "Neuromorphic Engineering I".
LernzielDesign of a neuromorphic circuit for implementation with CMOS technology.
InhaltThis course teaches the basics of analog chip design and layout with an emphasis on neuromorphic circuits, which are introduced in the autumn semester course "Neuromorphic Engineering I".

The principles of CMOS processing technology are presented. Using a set of inexpensive software tools for simulation, layout and verification, suitable for neuromorphic circuits, participants learn to simulate circuits on the transistor level and to make their layouts on the mask level. Important issues in the layout of neuromorphic circuits will be explained and illustrated with examples. In the latter part of the semester students simulate and layout a neuromorphic chip. Schematics of basic building blocks will be provided. The layout will then be fabricated and will be tested by students during the following fall semester.
LiteraturS.-C. Liu et al.: Analog VLSI Circuits and Principles; software documentation.
Voraussetzungen / BesonderesPrerequisites: Neuromorphic Engineering I strongly recommended
227-1034-00LComputational Vision (University of Zurich)
No enrolment to this course at ETH Zurich. Book the corresponding module directly at UZH.
UZH Module Code: INI402

Mind the enrolment deadlines at UZH:
Link
W6 KP2V + 1UD. Kiper
KurzbeschreibungThis course focuses on neural computations that underlie visual perception. We study how visual signals are processed in the retina, LGN and visual cortex. We study the morpholgy and functional architecture of cortical circuits responsible for pattern, motion, color, and three-dimensional vision.
LernzielThis course considers the operation of circuits in the process of neural computations. The evolution of neural systems will be considered to demonstrate how neural structures and mechanisms are optimised for energy capture, transduction, transmission and representation of information. Canonical brain circuits will be described as models for the analysis of sensory information. The concept of receptive fields will be introduced and their role in coding spatial and temporal information will be considered. The constraints of the bandwidth of neural channels and the mechanisms of normalization by neural circuits will be discussed.
The visual system will form the basis of case studies in the computation of form, depth, and motion. The role of multiple channels and collective computations for object recognition will
be considered. Coordinate transformations of space and time by cortical and subcortical mechanisms will be analysed. The means by which sensory and motor systems are integrated to allow for adaptive behaviour will be considered.
InhaltThis course considers the operation of circuits in the process of neural computations. The evolution of neural systems will be considered to demonstrate how neural structures and mechanisms are optimised for energy capture, transduction, transmission and representation of information. Canonical brain circuits will be described as models for the analysis of sensory information. The concept of receptive fields will be introduced and their role in coding spatial and temporal information will be considered. The constraints of the bandwidth of neural channels and the mechanisms of normalization by neural circuits will be discussed.
The visual system will form the basis of case studies in the computation of form, depth, and motion. The role of multiple channels and collective computations for object recognition will
be considered. Coordinate transformations of space and time by cortical and subcortical mechanisms will be analysed. The means by which sensory and motor systems are integrated to allow for adaptive behaviour will be considered.
LiteraturBooks: (recommended references, not required)
1. An Introduction to Natural Computation, D. Ballard (Bradford Books, MIT Press) 1997.
2. The Handbook of Brain Theorie and Neural Networks, M. Arbib (editor), (MIT Press) 1995.
Statistics
NummerTitelTypECTSUmfangDozierende
401-0102-00LApplied Multivariate StatisticsW5 KP2V + 1UF. Sigrist
KurzbeschreibungMultivariate statistics analyzes data on several random variables simultaneously. This course introduces the basic concepts and provides an overview of classical and modern methods of multivariate statistics including visualization, dimension reduction, supervised and unsupervised learning for multivariate data. An emphasis is on applications and solving problems with the statistical software R.
LernzielAfter the course, you are able to:
- describe the various methods and the concepts behind them
- identify adequate methods for a given statistical problem
- use the statistical software R to efficiently apply these methods
- interpret the output of these methods
InhaltVisualization, multivariate outliers, the multivariate normal distribution, dimension reduction, principal component analysis, multidimensional scaling, factor analysis, cluster analysis, classification, multivariate tests and multiple testing
SkriptNone
Literatur1) "An Introduction to Applied Multivariate Analysis with R" (2011) by Everitt and Hothorn
2) "An Introduction to Statistical Learning: With Applications in R" (2013) by Gareth, Witten, Hastie and Tibshirani

Electronic versions (pdf) of both books can be downloaded for free from the ETH library.
Voraussetzungen / BesonderesThis course is targeted at students with a non-math background.

Requirements:
==========
1) Introductory course in statistics (min: t-test, regression; ideal: conditional probability, multiple regression)
2) Good understanding of R (if you don't know R, it is recommended that you study chapters 1,2,3,4, and 5 of "Introductory Statistics with R" from Peter Dalgaard, which is freely available online from the ETH library)

An alternative course with more emphasis on theory is 401-6102-00L "Multivariate Statistics" (only every second year).

401-0102-00L and 401-6102-00L are mutually exclusive. You can register for only one of these two courses.
401-4632-15LCausality Information W4 KP2GC. Heinze-Deml
KurzbeschreibungIn statistics, we are used to search for the best predictors of some random variable. In many situations, however, we are interested in predicting a system's behavior under manipulations. For such an analysis, we require knowledge about the underlying causal structure of the system. In this course, we study concepts and theory behind causal inference.
LernzielAfter this course, you should be able to
- understand the language and concepts of causal inference
- know the assumptions under which one can infer causal relations from observational and/or interventional data
- describe and apply different methods for causal structure learning
- given data and a causal structure, derive causal effects and predictions of interventional experiments
Voraussetzungen / BesonderesPrerequisites: basic knowledge of probability theory and regression
401-6102-00LMultivariate Statistics
Findet dieses Semester nicht statt.
W4 KP2Gkeine Angaben
KurzbeschreibungMultivariate Statistics deals with joint distributions of several random variables. This course introduces the basic concepts and provides an overview over classical and modern methods of multivariate statistics. We will consider the theory behind the methods as well as their applications.
LernzielAfter the course, you should be able to:
- describe the various methods and the concepts and theory behind them
- identify adequate methods for a given statistical problem
- use the statistical software "R" to efficiently apply these methods
- interpret the output of these methods
InhaltVisualization / Principal component analysis / Multidimensional scaling / The multivariate Normal distribution / Factor analysis / Supervised learning / Cluster analysis
SkriptNone
LiteraturThe course will be based on class notes and books that are available electronically via the ETH library.
Voraussetzungen / BesonderesTarget audience: This course is the more theoretical version of "Applied Multivariate Statistics" (401-0102-00L) and is targeted at students with a math background.

Prerequisite: A basic course in probability and statistics.

Note: The courses 401-0102-00L and 401-6102-00L are mutually exclusive. You may register for at most one of these two course units.
401-6624-11LApplied Time SeriesW5 KP2V + 1UM. Dettling
KurzbeschreibungThe course starts with an introduction to time series analysis (examples, goal, mathematical notation). In the following, descriptive techniques, modeling and prediction as well as advanced topics will be covered.
LernzielGetting to know the mathematical properties of time series, as well as the requirements, descriptive techniques, models, advanced methods and software that are necessary such that the student can independently run an applied time series analysis.
InhaltThe course starts with an introduction to time series analysis that comprises of examples and goals. We continue with notation and descriptive analysis of time series. A major part of the course will be dedicated to modeling and forecasting of time series using the flexible class of ARMA models. More advanced topics that will be covered in the following are time series regression, state space models and spectral analysis.
SkriptA script will be available.
Voraussetzungen / BesonderesThe course starts with an introduction to time series analysis that comprises of examples and goals. We continue with notation and descriptive analysis of time series. A major part of the course will be dedicated to modeling and forecasting of time series using the flexible class of ARMA models. More advanced topics that will be covered in the following are time series regression, state space models and spectral analysis.
Machine Learning and Artificial Intelligence
NummerTitelTypECTSUmfangDozierende
252-0526-00LStatistical Learning Theory Information W7 KP3V + 2U + 1AJ. M. Buhmann, C. Cotrini Jimenez
KurzbeschreibungThe course covers advanced methods of statistical learning:

- Variational methods and optimization.
- Deterministic annealing.
- Clustering for diverse types of data.
- Model validation by information theory.
LernzielThe course surveys recent methods of statistical learning. The fundamentals of machine learning, as presented in the courses "Introduction to Machine Learning" and "Advanced Machine Learning", are expanded from the perspective of statistical learning.
Inhalt- Variational methods and optimization. We consider optimization approaches for problems where the optimizer is a probability distribution. We will discuss concepts like maximum entropy, information bottleneck, and deterministic annealing.

- Clustering. This is the problem of sorting data into groups without using training samples. We discuss alternative notions of "similarity" between data points and adequate optimization procedures.

- Model selection and validation. This refers to the question of how complex the chosen model should be. In particular, we present an information theoretic approach for model validation.

- Statistical physics models. We discuss approaches for approximately optimizing large systems, which originate in statistical physics (free energy minimization applied to spin glasses and other models). We also study sampling methods based on these models.
SkriptA draft of a script will be provided. Lecture slides will be made available.
LiteraturHastie, Tibshirani, Friedman: The Elements of Statistical Learning, Springer, 2001.

L. Devroye, L. Gyorfi, and G. Lugosi: A probabilistic theory of pattern recognition. Springer, New York, 1996
Voraussetzungen / BesonderesKnowledge of machine learning (introduction to machine learning and/or advanced machine learning)
Basic knowledge of statistics.
252-3005-00LNatural Language Understanding Information
Findet dieses Semester nicht statt.
Findet im HS20 wieder statt.
W5 KP2V + 1U + 1ANoch nicht bekannt
KurzbeschreibungThis course presents topics in natural language processing with an emphasis on modern techniques, primarily focusing on statistical and deep learning approaches. The course provides an overview of the primary areas of research in language processing as well as a detailed exploration of the models and techniques used both in research and in commercial natural language systems.
LernzielThe objective of the course is to learn the basic concepts in the statistical processing of natural languages. The course will be project-oriented so that the students can also gain hands-on experience with state-of-the-art tools and techniques.
InhaltThis course presents an introduction to general topics and techniques used in natural language processing today, primarily focusing on statistical approaches. The course provides an overview of the primary areas of research in language processing as well as a detailed exploration of the models and techniques used both in research and in commercial natural language systems.
LiteraturLectures will make use of textbooks such as the one by Jurafsky and Martin where appropriate, but will also make use of original research and survey papers.
261-5110-00LOptimization for Data Science Information W8 KP3V + 2U + 2AB. Gärtner, D. Steurer
KurzbeschreibungThis course provides an in-depth theoretical treatment of optimization methods that are particularly relevant in data science.
LernzielUnderstanding the theoretical guarantees (and their limits) of relevant optimization methods used in data science. Learning general paradigms to deal with optimization problems arising in data science.
InhaltThis course provides an in-depth theoretical treatment of optimization methods that are particularly relevant in machine learning and data science.

In the first part of the course, we will first give a brief introduction to convex optimization, with some basic motivating examples from machine learning. Then we will analyse classical and more recent first and second order methods for convex optimization: gradient descent, projected gradient descent, subgradient descent, stochastic gradient descent, Nesterov's accelerated method, Newton's method, and Quasi-Newton methods. The emphasis will be on analysis techniques that occur repeatedly in convergence analyses for various classes of convex functions. We will also discuss some classical and recent theoretical results for nonconvex optimization.

In the second part, we discuss convex programming relaxations as a powerful and versatile paradigm for designing efficient algorithms to solve computational problems arising in data science. We will learn about this paradigm and develop a unified perspective on it through the lens of the sum-of-squares semidefinite programming hierarchy. As applications, we are discussing non-negative matrix factorization, compressed sensing and sparse linear regression, matrix completion and phase retrieval, as well as robust estimation.
Voraussetzungen / BesonderesAs background, we require material taught in the course "252-0209-00L Algorithms, Probability, and Computing". It is not necessary that participants have actually taken the course, but they should be prepared to catch up if necessary.
  •  Seite  1  von  2 Nächste Seite Letzte Seite     Alle