Suchergebnis: Katalogdaten im Frühjahrssemester 2021
Cyber Security Master | ||||||
Ergänzung | ||||||
Distributed Systems | ||||||
Wahlfächer | ||||||
Nummer | Titel | Typ | ECTS | Umfang | Dozierende | |
---|---|---|---|---|---|---|
252-0312-00L | Ubiquitous Computing | W | 6 KP | 2V + 3A | C. Holz | |
Kurzbeschreibung | Ubiquitous Computing means interacting with information and with each other anywhere, mediated through miniature technology everywhere. We will investigate the technical aspects of Ubicomp, particularly sensing, processing, and sense making: input (touch & gesture), activity, monitoring cardiovascular health and neurological conditions, context & location sensing, affective computing. | |||||
Lernziel | The course will combine high-level concepts with low-level technical methods needed to sense, detect, and understand them. High-level: – input modalities for interactive systems (touch, gesture) – "activities" and "events" (exercises and other mechanical activities such as movements and resulting vibrations) – health monitoring (basic cardiovascular physiology) – location (GPS, urban simulations, smart cities and development) – affective computing (emotions, mood, personality) Low-level: – sampling (Shannon Nyquist) and filtering (FIR, IIR), time and frequency domains (Fourier transforms) – cross-modal sensor systems, signal synchronization and correlation – event detection, classification, prediction using basic signal processing as well as learning-based methods – sensor types: optical, mechanical/acoustic, electromagnetic – signals modalities and processing of: application (modalities/methods) * touch detection (resistive sensing, capacitive sensing, diffuse illumination/DI, spectral reflections, frustrated total internal reflection/FTIR, fingerprint scanning, surface-acoustic waves) * gesture recognition (inertial sensing through accelerometers, gyroscopes) * activity detection and tracking (inertial, acoustic, vibrotactile for classification, counting, vibrometry) * occupation and use (electricity monitoring, water consumption, single-point sensing) * cardiovascular (electrocardioagraphy, photoplethysmography, pulse oximetry, ballistocardiography, blood pressure, pulse transit time, bio impedance) * affective computing (heart rate variability, R-R intervals, electrodermal activity, sympathetic tone, facial expressions) * neurological (fatigue, fatigability) * location (GPS, BLE, Wifi) | |||||
Inhalt | "The most profound technologies are those that disappear. They weave themselves into the fabric of everyday life until they are indistinguishable from it" — Mark Weiser, 1991. This is the premise of Ubiquitous Computing, a vision that is slowly becoming reality as everything is a device and we can interact with information and with each other anywhere, mediated through miniature technology. Along with this change, interaction modalities have changed, too, from explicit input on keyboards and mice to implicit and passively observed input through sensors in the environment (e.g., speakers, cameras, temperature/occupancy detectors) and those we now wear on our bodies (e.g., health sensors, activity sensors, miniature computers we call smartwatches). In this course, we will look at the technical side of Ubicomp, particularly – sensing (incl. 'signals', sampling, data acquisition methods, controlled user studies, uncontrolled studies in-the-wild), – processing (incl. frequencies, feature extraction, detection), and – sense making: input sensing (touch & gesture), activity sensing (motion), monitoring cardiovascular health, affective state, neurological conditions (with basics on cardiovascular physiology + PPG, PulseOx, ECG, EDA, BCG, SCG, HRV, BioZ, IPG, PAT, PTT), context & location sensing (GPS/Wifi, motion). Lectures will be accompanied by practical sessions that focus on sensor modalities and signal processing. Here, we will work on existing data sets and devise methods to record our own data for processing and prediction purposes. A series of reading assignments, covering both well-established publications in Ubicomp as well as emerging results and methods, will bridge the fundamentals and topics taught in class to academic research and real-world problems. More information on the course site: https://teaching.siplab.org/ubiquitous_computing/2021/ | |||||
Skript | Copies of slides will be made available. Lectures will be recorded and made available online. More information on the course site: https://teaching.siplab.org/ubiquitous_computing/2021/ | |||||
Literatur | Will be provided in the lecture. To put you in the mood: Mark Weiser: The Computer for the 21st Century. Scientific American, September 1991, pp. 94-104 | |||||
252-0817-00L | Distributed Systems Laboratory | W | 10 KP | 9P | G. Alonso, T. Hoefler, A. Klimovic, A. Singla, R. Wattenhofer, C. Zhang | |
Kurzbeschreibung | Entwicklung und / oder Evaluation eines umfangreicheren praktischen Systems mit Technologien aus dem Gebiet der verteilten Systeme. Das Projekt kann aus unterschiedlichen Teilbereichen (von Web-Services bis hin zu ubiquitären Systemen) stammen; typische Technologien umfassen drahtlose Ad-hoc-Netze oder Anwendungen auf Mobiltelefonen. | |||||
Lernziel | Erwerb praktischer Kenntnisse bei Entwicklung und / oder Evaluation eines umfangreicheren praktischen Systems mit Technologien aus dem Gebiet der verteilten Systeme. | |||||
Inhalt | Entwicklung und / oder Evaluation eines umfangreicheren praktischen Systems mit Technologien aus dem Gebiet der verteilten Systeme. Das Projekt kann aus unterschiedlichen Teilbereichen (von Web-Services bis hin zu ubiquitären Systemen) stammen; typische Technologien umfassen drahtlose Ad-hoc-Netze oder Anwendungen auf Mobiltelefonen. Zu diesem Praktikum existiert keine Vorlesung. Bei Interesse bitte einen der beteiligten Professoren oder einen Assistenten der Forschungsgruppen kontaktieren. | |||||
263-3710-00L | Machine Perception Number of participants limited to 200. | W | 8 KP | 3V + 2U + 2A | O. Hilliges, S. Tang | |
Kurzbeschreibung | Recent developments in neural networks (aka “deep learning”) have drastically advanced the performance of machine perception systems in a variety of areas including computer vision, robotics, and intelligent UIs. This course is a deep dive into deep learning algorithms and architectures with applications to a variety of perceptual tasks. | |||||
Lernziel | Students will learn about fundamental aspects of modern deep learning approaches for perception. Students will learn to implement, train and debug their own neural networks and gain a detailed understanding of cutting-edge research in learning-based computer vision, robotics and HCI. The final project assignment will involve training a complex neural network architecture and applying it on a real-world dataset of human activity. The core competency acquired through this course is a solid foundation in deep-learning algorithms to process and interpret human input into computing systems. In particular, students should be able to develop systems that deal with the problem of recognizing people in images, detecting and describing body parts, inferring their spatial configuration, performing action/gesture recognition from still images or image sequences, also considering multi-modal data, among others. | |||||
Inhalt | We will focus on teaching: how to set up the problem of machine perception, the learning algorithms, network architectures and advanced deep learning concepts in particular probabilistic deep learning models The course covers the following main areas: I) Foundations of deep-learning. II) Probabilistic deep-learning for generative modelling of data (latent variable models, generative adversarial networks and auto-regressive models). III) Deep learning in computer vision, human-computer interaction and robotics. Specific topics include: I) Deep learning basics: a) Neural Networks and training (i.e., backpropagation) b) Feedforward Networks c) Timeseries modelling (RNN, GRU, LSTM) d) Convolutional Neural Networks for classification II) Probabilistic Deep Learning: a) Latent variable models (VAEs) b) Generative adversarial networks (GANs) c) Autoregressive models (PixelCNN, PixelRNN, TCNs) III) Deep Learning techniques for machine perception: a) Fully Convolutional architectures for dense per-pixel tasks (i.e., instance segmentation) b) Pose estimation and other tasks involving human activity c) Deep reinforcement learning IV) Case studies from research in computer vision, HCI, robotics and signal processing | |||||
Literatur | Deep Learning Book by Ian Goodfellow and Yoshua Bengio | |||||
Voraussetzungen / Besonderes | *** In accordance with the ETH Covid-19 master plan the lecture will be fully virtual. Details on the course website. *** This is an advanced grad-level course that requires a background in machine learning. Students are expected to have a solid mathematical foundation, in particular in linear algebra, multivariate calculus, and probability. The course will focus on state-of-the-art research in deep-learning and will not repeat basics of machine learning Please take note of the following conditions: 1) The number of participants is limited to 200 students (MSc and PhDs). 2) Students must have taken the exam in Machine Learning (252-0535-00) or have acquired equivalent knowledge 3) All practical exercises will require basic knowledge of Python and will use libraries such as Pytorch, scikit-learn and scikit-image. We will provide introductions to Pytorch and other libraries that are needed but will not provide introductions to basic programming or Python. The following courses are strongly recommended as prerequisite: * "Visual Computing" or "Computer Vision" The course will be assessed by a final written examination in English. No course materials or electronic devices can be used during the examination. Note that the examination will be based on the contents of the lectures, the associated reading materials and the exercises. |
- Seite 1 von 1