Search result: Catalogue data in Autumn Semester 2021

DAS in Data Science Information
Core Courses
Foundations Courses
NumberTitleTypeECTSHoursLecturers
227-0105-00LIntroduction to Estimation and Machine Learning Restricted registration - show details W6 credits4GH.‑A. Loeliger
AbstractMathematical basics of estimation and machine learning, with a view towards applications in signal processing.
ObjectiveStudents master the basic mathematical concepts and algorithms of estimation and machine learning.
ContentReview of probability theory;
basics of statistical estimation;
least squares and linear learning;
Hilbert spaces;
Gaussian random variables;
singular-value decomposition;
kernel methods, neural networks, and more
Lecture notesLecture notes will be handed out as the course progresses.
Prerequisites / Noticesolid basics in linear algebra and probability theory
Capstone Project
NumberTitleTypeECTSHoursLecturers
266-0100-00LCapstone Project Information Restricted registration - show details
Only for DAS in Data Science.
O8 credits17ASupervisors
AbstractThe capstone project is part of the DAS in Data Science and is an opportunity to apply the knowledge acquired in the program in an independent, real-world project.
ObjectiveTo apply the knowledge acquired in the program in an independent, real-world project.
ContentThe capstone project can be done under the supervision of the Swiss Data Science Center, or of any core or adjunct faculty of Data Science.
The project has to be finished within 6 months. Deadline for a project the following semester conducted at the SDSC is mid June/mid December.
Specialisation Track
Hardware for Machine Learning
Offered in the Spring Semester.
NumberTitleTypeECTSHoursLecturers
227-0155-00LMachine Learning on Microcontrollers Restricted registration - show details
Registration in this class requires the permission of the instructors. Class size will be limited to 25.
Preference is given to students in the MSc EEIT.
W6 credits3GM. Magno, L. Benini
AbstractMachine Learning (ML) and artificial intelligence are pervading the digital society. Today, even low power embedded systems are incorporating ML, becoming increasingly “smart”. This lecture gives an overview of ML methods and algorithms to process and extract useful near-sensor information in end-nodes of the “internet-of-things”, using low-power microcontrollers/ processors (ARM-Cortex-M; RISC-V)
ObjectiveLearn how to Process data from sensors and how to extract useful information with low power microprocessors using ML techniques. We will analyze data coming from real low-power sensors (accelerometers, microphones, ExG bio-signals, cameras…). The main objective is to study in details how Machine Learning algorithms can be adapted to the performance constraints and limited resources of low-power microcontrollers.
ContentThe final goal of the course is a deep understanding of machine learning and its practical implementation on single- and multi-core microcontrollers, coupled with performance and energy efficiency analysis and optimization. The main topics of the course include:

- Sensors and sensor data acquisition with low power embedded systems

- Machine Learning: Overview of supervised and unsupervised learning and in particular supervised learning (Bayes Decision Theory, Decision Trees, Random Forests, kNN-Methods, Support Vector Machines, Convolutional Networks and Deep Learning)

- Low-power embedded systems and their architecture. Low Power microcontrollers (ARM-Cortex M) and RISC-V-based Parallel Ultra Low Power (PULP) systems-on-chip.

- Low power smart sensor system design: hardware-software tradeoffs, analysis, and optimization. Implementation and performance evaluation of ML in battery-operated embedded systems.

The laboratory exercised will show how to address concrete design problems, like motion, gesture recognition, emotion detection, image and sound classification, using real sensors data and real MCU boards.

Presentations from Ph.D. students and the visit to the Digital Circuits and Systems Group will introduce current research topics and international research projects.
Lecture notesScript and exercise sheets. Books will be suggested during the course.
Prerequisites / NoticePrerequisites: C language programming. Basics of Digital Signal Processing. Basics of processor and computer architecture. Some exposure to machine learning concepts is also desirable
Image Analysis & Computer Vision
NumberTitleTypeECTSHoursLecturers
263-5902-00LComputer Vision Information W8 credits3V + 1U + 3AM. Pollefeys, S. Tang, F. Yu
AbstractThe goal of this course is to provide students with a good understanding of computer vision and image analysis techniques. The main concepts and techniques will be studied in depth and practical algorithms and approaches will be discussed and explored through the exercises.
ObjectiveThe objectives of this course are:
1. To introduce the fundamental problems of computer vision.
2. To introduce the main concepts and techniques used to solve those.
3. To enable participants to implement solutions for reasonably complex problems.
4. To enable participants to make sense of the computer vision literature.
ContentCamera models and calibration, invariant features, Multiple-view geometry, Model fitting, Stereo Matching, Segmentation, 2D Shape matching, Shape from Silhouettes, Optical flow, Structure from motion, Tracking, Object recognition, Object category recognition
Prerequisites / NoticeIt is recommended that students have taken the Visual Computing lecture or a similar course introducing basic image processing concepts before taking this course.
Neural Information Processing
NumberTitleTypeECTSHoursLecturers
227-0421-00LDeep Learning in Artificial and Biological Neuronal NetworksW4 credits3GB. Grewe
AbstractDeep-Learning (DL) a brain-inspired weak for of AI allows training of large artificial neuronal networks (ANNs) that, like humans, can learn real-world tasks such as recognizing objects in images. However, DL is far from being understood and investigating learning in biological networks might serve again as a compelling inspiration to think differently about state-of-the-art ANN training methods.
ObjectiveThe main goal of this lecture is to provide a comprehensive overview into the learning principles neuronal networks as well as to introduce a diverse skill set (e.g. simulating a spiking neuronal network) that is required to understand learning in large, hierarchical neuronal networks. To achieve this the lectures and exercises will merge ideas, concepts and methods from machine learning and neuroscience. These will include training basic ANNs, simulating spiking neuronal networks as well as being able to read and understand the main ideas presented in today’s neuroscience papers.
After this course students will be able to:
- read and understand the main ideas and methods that are presented in today’s neuroscience papers
- explain the basic ideas and concepts of plasticity in the mammalian brain
- implement alternative ANN learning algorithms to ‘error backpropagation’ in order to train deep neuronal networks.
- use a diverse set of ANN regularization methods to improve learning
- simulate spiking neuronal networks that learn simple (e.g. digit classification) tasks in a supervised manner.
ContentDeep-learning a brain-inspired weak form of AI allows training of large artificial neuronal networks (ANNs) that, like humans, can learn real-world tasks such as recognizing objects in images. The origins of deep hierarchical learning can be traced back to early neuroscience research by Hubel and Wiesel in the 1960s, who first described the neuronal processing of visual inputs in the mammalian neocortex. Similar to their neocortical counterparts ANNs seem to learn by interpreting and structuring the data provided by the external world. However, while on specific tasks such as playing (video) games deep ANNs outperform humans (Minh et al, 2015, Silver et al., 2018), ANNs are still not performing on par when it comes to recognizing actions in movie data and their ability to act as generalizable problem solvers is still far behind of what the human brain seems to achieve effortlessly. Moreover, biological neuronal networks can learn far more effectively with fewer training examples, they achieve a much higher performance in recognizing complex patterns in time series data (e.g. recognizing actions in movies), they dynamically adapt to new tasks without losing performance and they achieve unmatched performance to detect and integrate out-of-domain data examples (data they have not been trained with). In other words, many of the big challenges and unknowns that have emerged in the field of deep learning over the last years are already mastered exceptionally well by biological neuronal networks in our brain. On the other hand, many facets of typical ANN design and training algorithms seem biologically implausible, such as the non-local weight updates, discrete processing of time, and scalar communication between neurons. Recent evidence suggests that learning in biological systems is the result of the complex interplay of diverse error feedback signaling processes acting at multiple scales, ranging from single synapses to entire networks.
Lecture notesThe lecture slides will be provided as a PDF after each lecture.
Prerequisites / NoticeThis advanced level lecture requires some basic background in machine/deep learning. Thus, students are expected to have a basic mathematical foundation, including linear algebra, multivariate calculus, and probability. The course is not to be meant as an extended tutorial of how to train deep networks in PyTorch or Tensorflow, although these tools used.
The participation in the course is subject to the following conditions:

1) The number of participants is limited to 120 students (MSc and PhDs).

2) Students must have taken the exam in Deep Learning (263-3210-00L) or have acquired equivalent knowledge.
227-1033-00LNeuromorphic Engineering I Restricted registration - show details
Registration in this class requires the permission of the instructors. Class size will be limited to available lab spots.
Preference is given to students that require this class as part of their major.

Information for UZH students:
Enrolment to this course unit only possible at ETH. No enrolment to module INI404 at UZH.
Please mind the ETH enrolment deadlines for UZH students: Link
W6 credits2V + 3UT. Delbrück, G. Indiveri, S.‑C. Liu
AbstractThis course covers analog circuits with emphasis on neuromorphic engineering: MOS transistors in CMOS technology, static circuits, dynamic circuits, systems (silicon neuron, silicon retina, silicon cochlea) with an introduction to multi-chip systems. The lectures are accompanied by weekly laboratory sessions.
ObjectiveUnderstanding of the characteristics of neuromorphic circuit elements.
ContentNeuromorphic circuits are inspired by the organizing principles of biological neural circuits. Their computational primitives are based on physics of semiconductor devices. Neuromorphic architectures often rely on collective computation in parallel networks. Adaptation, learning and memory are implemented locally within the individual computational elements. Transistors are often operated in weak inversion (below threshold), where they exhibit exponential I-V characteristics and low currents. These properties lead to the feasibility of high-density, low-power implementations of functions that are computationally intensive in other paradigms. Application domains of neuromorphic circuits include silicon retinas and cochleas for machine vision and audition, real-time emulations of networks of biological neurons, and the development of autonomous robotic systems. This course covers devices in CMOS technology (MOS transistor below and above threshold, floating-gate MOS transistor, phototransducers), static circuits (differential pair, current mirror, transconductance amplifiers, etc.), dynamic circuits (linear and nonlinear filters, adaptive circuits), systems (silicon neuron, silicon retina and cochlea) and an introduction to multi-chip systems that communicate events analogous to spikes. The lectures are accompanied by weekly laboratory sessions on the characterization of neuromorphic circuits, from elementary devices to systems.
LiteratureS.-C. Liu et al.: Analog VLSI Circuits and Principles; various publications.
Prerequisites / NoticeParticular: The course is highly recommended for those who intend to take the spring semester course 'Neuromorphic Engineering II', that teaches the conception, simulation, and physical layout of such circuits with chip design tools.

Prerequisites: Background in basics of semiconductor physics helpful, but not required.
Statistics
NumberTitleTypeECTSHoursLecturers
401-0625-01LApplied Analysis of Variance and Experimental Design Information W5 credits2V + 1UL. Meier
AbstractPrinciples of experimental design, one-way analysis of variance, contrasts and multiple comparisons, multi-factor designs and analysis of variance, complete block designs, Latin square designs, random effects and mixed effects models, split-plot designs, incomplete block designs, two-series factorials and fractional designs, power.
ObjectiveParticipants will be able to plan and analyze efficient experiments in the fields of natural sciences. They will gain practical experience by using the software R.
ContentPrinciples of experimental design, one-way analysis of variance, contrasts and multiple comparisons, multi-factor designs and analysis of variance, complete block designs, Latin square designs, random effects and mixed effects models, split-plot designs, incomplete block designs, two-series factorials and fractional designs, power.
LiteratureG. Oehlert: A First Course in Design and Analysis of Experiments, W.H. Freeman and Company, New York, 2000.
Prerequisites / NoticeThe exercises, but also the classes will be based on procedures from the freely available, open-source statistical software R, for which an introduction will be held.
401-0649-00LApplied Statistical RegressionW5 credits2V + 1UM. Dettling
AbstractThis course offers a practically oriented introduction into regression modeling methods. The basic concepts and some mathematical background are included, with the emphasis lying in learning "good practice" that can be applied in every student's own projects and daily work life. A special focus will be laid in the use of the statistical software package R for regression analysis.
ObjectiveThe students acquire advanced practical skills in linear regression analysis and are also familiar with its extensions to generalized linear modeling.
ContentThe course starts with the basics of linear modeling, and then proceeds to parameter estimation, tests, confidence intervals, residual analysis, model choice, and prediction. More rarely touched but practically relevant topics that will be covered include variable transformations, multicollinearity problems and model interpretation, as well as general modeling strategies.

The last third of the course is dedicated to an introduction to generalized linear models: this includes the generalized additive model, logistic regression for binary response variables, binomial regression for grouped data and poisson regression for count data.
Lecture notesA script will be available.
LiteratureFaraway (2005): Linear Models with R
Faraway (2006): Extending the Linear Model with R
Draper & Smith (1998): Applied Regression Analysis
Fox (2008): Applied Regression Analysis and GLMs
Montgomery et al. (2006): Introduction to Linear Regression Analysis
Prerequisites / NoticeThe exercises, but also the classes will be based on procedures from the freely available, open-source statistical software package R, for which an introduction will be held.

In the Mathematics Bachelor and Master programmes, the two course units 401-0649-00L "Applied Statistical Regression" and 401-3622-00L "Statistical Modelling" are mutually exclusive. Registration for the examination of one of these two course units is only allowed if you have not registered for the examination of the other course unit.
CompetenciesCompetencies
Subject-specific CompetenciesConcepts and Theoriesassessed
Techniques and Technologiesassessed
Method-specific CompetenciesAnalytical Competenciesassessed
Decision-makingassessed
Media and Digital Technologiesassessed
Problem-solvingassessed
Project Managementfostered
Social CompetenciesCommunicationassessed
Cooperation and Teamworkfostered
Customer Orientationfostered
Leadership and Responsibilityfostered
Self-presentation and Social Influence fostered
Sensitivity to Diversityfostered
Negotiationfostered
Personal CompetenciesAdaptability and Flexibilityassessed
Creative Thinkingassessed
Critical Thinkingassessed
Integrity and Work Ethicsassessed
Self-awareness and Self-reflection fostered
Self-direction and Self-management fostered
401-3612-00LStochastic Simulation
Does not take place this semester.
W5 credits3G
AbstractThis course provides an introduction to statistical Monte Carlo methods. This includes applications of simulations in various fields (Bayesian statistics, statistical mechanics, operations research, financial mathematics), algorithms for the generation of random variables (accept-reject, importance sampling), estimating the precision, variance reduction, introduction to Markov chain Monte Carlo.
ObjectiveStochastic simulation (also called Monte Carlo method) is the experimental analysis of a stochastic model by implementing it on a computer. Probabilities and expected values can be approximated by averaging simulated values, and the central limit theorem gives an estimate of the error of this approximation. The course shows examples of the many applications of stochastic simulation and explains different algorithms used for simulation. These algorithms are illustrated with the statistical software R.
ContentExamples of simulations in different fields (computer science, statistics, statistical mechanics, operations research, financial mathematics). Generation of uniform random variables. Generation of random variables with arbitrary distributions (quantile transform, accept-reject, importance sampling), simulation of Gaussian processes and diffusions. The precision of simulations, methods for variance reduction. Introduction to Markov chains and Markov chain Monte Carlo (Metropolis-Hastings, Gibbs sampler, Hamiltonian Monte Carlo, reversible jump MCMC).
Lecture notesA script will be available in English.
LiteratureP. Glasserman, Monte Carlo Methods in Financial Engineering.
Springer 2004.

B. D. Ripley. Stochastic Simulation. Wiley, 1987.

Ch. Robert, G. Casella. Monte Carlo Statistical Methods.
Springer 2004 (2nd edition).
Prerequisites / NoticeFamiliarity with basic concepts of probability theory (random variables, joint and conditional distributions, laws of large numbers and central limit theorem) will be assumed.
401-3621-00LFundamentals of Mathematical Statistics Information W10 credits4V + 1US. van de Geer
AbstractThe course covers the basics of inferential statistics.
Objective
401-3622-00LStatistical Modelling Information W8 credits4GC. Heinze-Deml
AbstractIn regression, the dependency of a random response variable on other variables is examined. We consider the theory of linear regression with one or more covariates, high-dimensional linear models, nonlinear models and generalized linear models, robust methods, model choice and nonparametric models. Several numerical examples will illustrate the theory.
ObjectiveIntroduction into theory and practice of a broad and popular area of statistics, from a modern viewpoint.
ContentIn der Regression wird die Abhängigkeit einer beobachteten quantitativen Grösse von einer oder mehreren anderen (unter Berücksichtigung zufälliger Fehler) untersucht. Themen der Vorlesung sind: Einfache und multiple Regression, Theorie allgemeiner linearer Modelle, Hoch-dimensionale Modelle, Ausblick auf nichtlineare Modelle. Querverbindungen zur Varianzanalyse, Modellsuche, Residuenanalyse; Einblicke in Robuste Regression. Durchrechnung und Diskussion von Anwendungsbeispielen.
Prerequisites / NoticeThis is the course unit with former course title "Regression".
Credits cannot be recognised for both courses 401-3622-00L Statistical Modelling and 401-0649-00L Applied Statistical Regression in the Mathematics Bachelor and Master programmes (to be precise: one course in the Bachelor and the other course in the Master is also forbidden).
401-3628-14LBayesian StatisticsW4 credits2VF. Sigrist
AbstractIntroduction to the Bayesian approach to statistics: decision theory, prior distributions, hierarchical Bayes models, empirical Bayes, Bayesian tests and model selection, empirical Bayes, Laplace approximation, Monte Carlo and Markov chain Monte Carlo methods.
ObjectiveStudents understand the conceptual ideas behind Bayesian statistics and are familiar with common techniques used in Bayesian data analysis.
ContentTopics that we will discuss are:

Difference between the frequentist and Bayesian approach (decision theory, principles), priors (conjugate priors, noninformative priors, Jeffreys prior), tests and model selection (Bayes factors, hyper-g priors for regression),hierarchical models and empirical Bayes methods, computational methods (Laplace approximation, Monte Carlo and Markov chain Monte Carlo methods)
Lecture notesA script will be available in English.
LiteratureChristian Robert, The Bayesian Choice, 2nd edition, Springer 2007.

A. Gelman et al., Bayesian Data Analysis, 3rd edition, Chapman & Hall (2013).

Additional references will be given in the course.
Prerequisites / NoticeFamiliarity with basic concepts of frequentist statistics and with basic concepts of probability theory (random variables, joint and conditional distributions, laws of large numbers and central limit theorem) will be assumed.
401-4623-00LTime Series Analysis
Does not take place this semester.
W6 credits3GF. Balabdaoui
AbstractThe course offers an introduction into analyzing times series, that is observations which occur in time. The material will cover Stationary Models, ARMA processes, Spectral Analysis, Forecasting, Nonstationary Models, ARIMA Models and an introduction to GARCH models.
ObjectiveThe goal of the course is to have a a good overview of the different types of time series and the approaches used in their statistical analysis.
ContentThis course treats modeling and analysis of time series, that is random variables which change in time. As opposed to the i.i.d. framework, the main feature exibited by time series is the dependence between successive observations.

The key topics which will be covered as:

Stationarity
Autocorrelation
Trend estimation
Elimination of seasonality
Spectral analysis, spectral densities
Forecasting
ARMA, ARIMA, Introduction into GARCH models
LiteratureThe main reference for this course is the book "Introduction to Time Series and Forecasting", by P. J. Brockwell and R. A. Davis
Prerequisites / NoticeBasic knowledge in probability and statistics
Machine Learning and Artificial Intelligence
NumberTitleTypeECTSHoursLecturers
227-0689-00LSystem IdentificationW4 credits2V + 1UR. Smith
AbstractTheory and techniques for the identification of dynamic models from experimentally obtained system input-output data.
ObjectiveTo provide a series of practical techniques for the development of dynamical models from experimental data, with the emphasis being on the development of models suitable for feedback control design purposes. To provide sufficient theory to enable the practitioner to understand the trade-offs between model accuracy, data quality and data quantity.
ContentIntroduction to modeling: Black-box and grey-box models; Parametric and non-parametric models; ARX, ARMAX (etc.) models.

Predictive, open-loop, black-box identification methods. Time and frequency domain methods. Subspace identification methods.

Optimal experimental design, Cramer-Rao bounds, input signal design.

Parametric identification methods. On-line and batch approaches.

Closed-loop identification strategies. Trade-off between controller performance and information available for identification.
Literature"System Identification; Theory for the User" Lennart Ljung, Prentice Hall (2nd Ed), 1999.

Additional papers will be available via the course Moodle.
Prerequisites / NoticeControl systems (227-0216-00L) or equivalent.
252-0535-00LAdvanced Machine Learning Information W10 credits3V + 2U + 4AJ. M. Buhmann, C. Cotrini Jimenez
AbstractMachine learning algorithms provide analytical methods to search data sets for characteristic patterns. Typical tasks include the classification of data, function fitting and clustering, with applications in image and speech analysis, bioinformatics and exploratory data analysis. This course is accompanied by practical machine learning projects.
ObjectiveStudents will be familiarized with advanced concepts and algorithms for supervised and unsupervised learning; reinforce the statistics knowledge which is indispensible to solve modeling problems under uncertainty. Key concepts are the generalization ability of algorithms and systematic approaches to modeling and regularization. Machine learning projects will provide an opportunity to test the machine learning algorithms on real world data.
ContentThe theory of fundamental machine learning concepts is presented in the lecture, and illustrated with relevant applications. Students can deepen their understanding by solving both pen-and-paper and programming exercises, where they implement and apply famous algorithms to real-world data.

Topics covered in the lecture include:

Fundamentals:
What is data?
Bayesian Learning
Computational learning theory

Supervised learning:
Ensembles: Bagging and Boosting
Max Margin methods
Neural networks

Unsupservised learning:
Dimensionality reduction techniques
Clustering
Mixture Models
Non-parametric density estimation
Learning Dynamical Systems
Lecture notesNo lecture notes, but slides will be made available on the course webpage.
LiteratureC. Bishop. Pattern Recognition and Machine Learning. Springer 2007.

R. Duda, P. Hart, and D. Stork. Pattern Classification. John Wiley &
Sons, second edition, 2001.

T. Hastie, R. Tibshirani, and J. Friedman. The Elements of Statistical
Learning: Data Mining, Inference and Prediction. Springer, 2001.

L. Wasserman. All of Statistics: A Concise Course in Statistical
Inference. Springer, 2004.
Prerequisites / NoticeThe course requires solid basic knowledge in analysis, statistics and numerical methods for CSE as well as practical programming experience for solving assignments.
Students should have followed at least "Introduction to Machine Learning" or an equivalent course offered by another institution.

PhD students are required to obtain a passing grade in the course (4.0 or higher based on project and exam) to gain credit points.
252-3005-00LNatural Language Processing Information Restricted registration - show details
Number of participants limited to 400.
W5 credits2V + 2U + 1AR. Cotterell
AbstractThis course presents topics in natural language processing with an emphasis on modern techniques, primarily focusing on statistical and deep learning approaches. The course provides an overview of the primary areas of research in language processing as well as a detailed exploration of the models and techniques used both in research and in commercial natural language systems.
ObjectiveThe objective of the course is to learn the basic concepts in the statistical processing of natural languages. The course will be project-oriented so that the students can also gain hands-on experience with state-of-the-art tools and techniques.
ContentThis course presents an introduction to general topics and techniques used in natural language processing today, primarily focusing on statistical approaches. The course provides an overview of the primary areas of research in language processing as well as a detailed exploration of the models and techniques used both in research and in commercial natural language systems.
LiteratureLectures will make use of textbooks such as the one by Jurafsky and Martin where appropriate, but will also make use of original research and survey papers.
263-2400-00LReliable and Trustworthy Artificial Intelligence Information W6 credits2V + 2U + 1AM. Vechev
AbstractCreating reliable and explainable probabilistic models is a fundamental challenge to solving the artificial intelligence problem. This course covers some of the latest and most exciting advances that bring us closer to constructing such models.
ObjectiveThe main objective of this course is to expose students to the latest and most exciting research in the area of explainable and interpretable artificial intelligence, a topic of fundamental and increasing importance. Upon completion of the course, the students should have mastered the underlying methods and be able to apply them to a variety of problems.

To facilitate deeper understanding, an important part of the course will be a group hands-on programming project where students will build a system based on the learned material.
ContentThis comprehensive course covers some of the latest and most important research advances (over the last 3 years) underlying the creation of safe, trustworthy, and reliable AI (more information here: Link):

* Adversarial Attacks on Deep Learning (noise-based, geometry attacks, sound attacks, physical attacks, autonomous driving, out-of-distribution)
* Defenses against attacks
* Combining gradient-based optimization with logic for encoding background knowledge
* Complete Certification of deep neural networks via automated reasoning (e.g., via numerical relaxations, mixed-integer solvers).
* Probabilistic certification of deep neural networks
* Training deep neural networks to be provably robust via automated reasoning
* Fairness (different notions of fairness, certifiably fair representation learning)
* Federated Learning (introduction, security considerations)
Prerequisites / NoticeWhile not a formal requirement, the course assumes familiarity with basics of machine learning (especially linear algebra, gradient descent, and neural networks as well as basic probability theory). These topics are usually covered in “Intro to ML” classes at most institutions (e.g., “Introduction to Machine Learning” at ETH).

For solving assignments, some programming experience in Python is expected.
263-3210-00LDeep Learning Information Restricted registration - show details
Number of participants limited to 320.
W8 credits3V + 2U + 2AF. Perez Cruz, A. Lucchi
AbstractDeep learning is an area within machine learning that deals with algorithms and models that automatically induce multi-level data representations.
ObjectiveIn recent years, deep learning and deep networks have significantly improved the state-of-the-art in many application domains such as computer vision, speech recognition, and natural language processing. This class will cover the mathematical foundations of deep learning and provide insights into model design, training, and validation. The main objective is a profound understanding of why these methods work and how. There will also be a rich set of hands-on tasks and practical projects to familiarize students with this emerging technology.
Prerequisites / NoticeThis is an advanced level course that requires some basic background in machine learning. More importantly, students are expected to have a very solid mathematical foundation, including linear algebra, multivariate calculus, and probability. The course will make heavy use of mathematics and is not (!) meant to be an extended tutorial of how to train deep networks with tools like Torch or Tensorflow, although that may be a side benefit.

The participation in the course is subject to the following condition:
- Students must have taken the exam in Advanced Machine Learning (252-0535-00) or have acquired equivalent knowledge, see exhaustive list below:

Advanced Machine Learning
Link

Computational Intelligence Lab
Link

Introduction to Machine Learning
Link

Statistical Learning Theory
Link

Computational Statistics
Link

Probabilistic Artificial Intelligence
Link
263-5210-00LProbabilistic Artificial Intelligence Information Restricted registration - show details W8 credits3V + 2U + 2AA. Krause
AbstractThis course introduces core modeling techniques and algorithms from machine learning, optimization and control for reasoning and decision making under uncertainty, and study applications in areas such as robotics.
ObjectiveHow can we build systems that perform well in uncertain environments? How can we develop systems that exhibit "intelligent" behavior, without prescribing explicit rules? How can we build systems that learn from experience in order to improve their performance? We will study core modeling techniques and algorithms from statistics, optimization, planning, and control and study applications in areas such as robotics. The course is designed for graduate students.
ContentTopics covered:
- Probability
- Probabilistic inference (variational inference, MCMC)
- Bayesian learning (Gaussian processes, Bayesian deep learning)
- Probabilistic planning (MDPs, POMPDPs)
- Multi-armed bandits and Bayesian optimization
- Reinforcement learning
Prerequisites / NoticeSolid basic knowledge in statistics, algorithms and programming.
The material covered in the course "Introduction to Machine Learning" is considered as a prerequisite.
Big Data Systems
NumberTitleTypeECTSHoursLecturers
252-0834-00LInformation Systems for Engineers Information W4 credits2V + 1UG. Fourny
AbstractThis course provides the basics of relational databases from the perspective of the user.

We will discover why tables are so incredibly powerful to express relations, learn the SQL query language, and how to make the most of it. The course also covers support for data cubes (analytics).
ObjectiveThis lesson is complementary with Big Data for Engineers as they cover different time periods of database history and practices -- you can take them in any order, even though it might be more enjoyable to take this lecture first.

After visiting this course, you will be capable to:

1. Explain, in the big picture, how a relational database works and what it can do in your own words.

2. Explain the relational data model (tables, rows, attributes, primary keys, foreign keys), formally and informally, including the relational algebra operators (select, project, rename, all kinds of joins, division, cartesian product, union, intersection, etc).

3. Perform non-trivial reading SQL queries on existing relational databases, as well as insert new data, update and delete existing data.

4. Design new schemas to store data in accordance to the real world's constraints, such as relationship cardinality

5. Explain what bad design is and why it matters.

6. Adapt and improve an existing schema to make it more robust against anomalies, thanks to a very good theoretical knowledge of what is called "normal forms".

7. Understand how indices work (hash indices, B-trees), how they are implemented, and how to use them to make queries faster.

8. Access an existing relational database from a host language such as Java, using bridges such as JDBC.

9. Explain what data independence is all about and didn't age a bit since the 1970s.

10. Explain, in the big picture, how a relational database is physically implemented.

11. Know and deal with the natural syntax for relational data, CSV.

12. Explain the data cube model including slicing and dicing.

13. Store data cubes in a relational database.

14. Map cube queries to SQL.

15. Slice and dice cubes in a UI.

And of course, you will think that tables are the most wonderful object in the world.
ContentUsing a relational database
=================
1. Introduction
2. The relational model
3. Data definition with SQL
4. The relational algebra
5. Queries with SQL

Taking a relational database to the next level
=================
6. Database design theory
7. Databases and host languages
8. Databases and host languages
9. Indices and optimization
10. Database architecture and storage

Analytics on top of a relational database
=================
12. Data cubes

Outlook
=================
13. Outlook
Literature- Lecture material (slides).

- Book: "Database Systems: The Complete Book", H. Garcia-Molina, J.D. Ullman, J. Widom
(It is not required to buy the book, as the library has it)
Prerequisites / NoticeFor non-CS/DS students only, BSc and MSc
Elementary knowledge of set theory and logics
Knowledge as well as basic experience with a programming language such as Pascal, C, C++, Java, Haskell, Python
  •  Page  1  of  2 Next page Last page     All