Suchergebnis: Katalogdaten im Herbstsemester 2019

Elektrotechnik und Informationstechnologie Master Information
Master-Studium (Studienreglement 2018)
Signal Processing and Machine Learning
The core courses and specialisation courses below are a selection for students who wish to specialise in the area of "Signal Processing and Machine Learning ", see Link.

The individual study plan is subject to the tutor's approval.
Vertiefungsfächer
These specialisation courses are particularly recommended for the area of "Signal Processing and Machine Learning", but you are free to choose courses from any other field in agreement with your tutor.

A minimum of 40 credits must be obtained from specialisation courses during the MSc EEIT.
NummerTitelTypECTSUmfangDozierende
227-0116-00LVLSI I: From Architectures to VLSI Circuits and FPGAs Information W6 KP5GF. K. Gürkaynak, L. Benini
KurzbeschreibungThis first course in a series that extends over three consecutive terms is concerned with tailoring algorithms and with devising high performance hardware architectures for their implementation as ASIC or with FPGAs. The focus is on front end design using HDLs and automatic synthesis for producing industrial-quality circuits.
LernzielUnderstand Very-Large-Scale Integrated Circuits (VLSI chips), Application-Specific Integrated Circuits (ASIC), and Field-Programmable Gate-Arrays (FPGA). Know their organization and be able to identify suitable application areas. Become fluent in front-end design from architectural conception to gate-level netlists. How to model digital circuits with SystemVerilog. How to ensure they behave as expected with the aid of simulation, testbenches, and assertions. How to take advantage of automatic synthesis tools to produce industrial-quality VLSI and FPGA circuits. Gain practical experience with the hardware description language SystemVerilog and with industrial Electronic Design Automation (EDA) tools.
InhaltThis course is concerned with system-level issues of VLSI design and FPGA implementations. Topics include:
- Overview on design methodologies and fabrication depths.
- Levels of abstraction for circuit modeling.
- Organization and configuration of commercial field-programmable components.
- FPGA design flows.
- Dedicated and general purpose architectures compared.
- How to obtain an architecture for a given processing algorithm.
- Meeting throughput, area, and power goals by way of architectural transformations.
- Hardware Description Languages (HDL) and the underlying concepts.
- SystemVerilog
- Register Transfer Level (RTL) synthesis and its limitations.
- Building blocks of digital VLSI circuits.
- Functional verification techniques and their limitations.
- Modular and largely reusable testbenches.
- Assertion-based verification.
- Synchronous versus asynchronous circuits.
- The case for synchronous circuits.
- Periodic events and the Anceau diagram.
- Case studies, ASICs compared to microprocessors, DSPs, and FPGAs.

During the exercises, students learn how to model FPGAs with SystemVerilog. They write testbenches for simulation purposes and synthesize gate-level netlists for FPGAs. Commercial EDA software by leading vendors is being used throughout.
SkriptTextbook and all further documents in English.
LiteraturH. Kaeslin: "Top-Down Digital VLSI Design, from Architectures to Gate-Level Circuits and FPGAs", Elsevier, 2014, ISBN 9780128007303.
Voraussetzungen / BesonderesPrerequisites:
Basics of digital circuits.

Examination:
In written form following the course semester (spring term). Problems are given in English, answers will be accepted in either English oder German.

Further details:
Link
227-0121-00LKommunikationssysteme Information W6 KP4GA. Wittneben
KurzbeschreibungInformationstheorie, Signalraumanalyse, Basisbandübertragung, Passbandübertragung, Systembeispiel und Kanal, Sicherungsschicht, MAC, Beispiele Layer 2, Layer 3, Internet
LernzielZiel der Vorlesung ist die Einführung der wichtigsten Konzepte und Verfahren, die in modernen digitalen Kommunikationssystemen Anwendung finden, sowie eine Übersicht über bestehende und zukünftige Systeme.
InhaltEs werden die untersten drei Schichten des OSI-Referenzmodells behandelt: die Bitübertragungsschicht, die Sicherungsschicht mit dem Zugriff auf das Übertragungsmedium und die Vermittlung. Die wichtigsten Begriffe der Informationstheorie werden eingeführt. Anschliessend konzentrieren sich die Betrachtungen auf die Verfahren der Punkt-zu-Punkt-Übertragung, welche sich mittels der Signalraumdarstellung elegant und kohärent behandeln lassen. Den Methoden der Fehlererkennung und –korrektur, sowie Protokollen für die erneute Übermittlung gestörter Daten wird Rechnung getragen. Auch der Vielfachzugriff bei geteiltem Übertragungsmedium wird diskutiert. Den Abschluss bilden Algorithmen für das Routing in Kommunikationsnetzen und der Flusssteuerung.

Die Anwendung der grundlegenden Verfahren wird ausführlich anhand von bestehenden und zukünftigen drahtlosen und drahtgebundenen Systemen erläutert.
SkriptVorlesungsfolien
Literatur[1] Simon Haykin, Communication Systems, 4. Auflage, John Wiley & Sons, 2001
[2] Andrew S. Tanenbaum, Computernetzwerke, 3. Auflage, Pearson Studium, 2003
[3] M. Bossert und M. Breitbach, Digitale Netze, 1. Auflage, Teubner, 1999
227-0225-00LLinear System TheoryW6 KP5GJ. Lygeros
KurzbeschreibungThe class is intended to provide a comprehensive overview of the theory of linear dynamical systems, stability analysis, and their use in control and estimation. The focus is on the mathematics behind the physical properties of these systems and on understanding and constructing proofs of properties of linear control systems.
LernzielStudents should be able to apply the fundamental results in linear system theory to analyze and control linear dynamical systems.
Inhalt- Proof techniques and practices.
- Linear spaces, normed linear spaces and Hilbert spaces.
- Ordinary differential equations, existence and uniqueness of solutions.
- Continuous and discrete-time, time-varying linear systems. Time domain solutions. Time invariant systems treated as a special case.
- Controllability and observability, duality. Time invariant systems treated as a special case.
- Stability and stabilization, observers, state and output feedback, separation principle.
SkriptAvailable on the course Moodle platform.
Voraussetzungen / BesonderesSufficient mathematical maturity, in particular in linear algebra, analysis.
227-0417-00LInformation Theory IW6 KP4GA. Lapidoth
KurzbeschreibungThis course covers the basic concepts of information theory and of communication theory. Topics covered include the entropy rate of a source, mutual information, typical sequences, the asymptotic equi-partition property, Huffman coding, channel capacity, the channel coding theorem, the source-channel separation theorem, and feedback capacity.
LernzielThe fundamentals of Information Theory including Shannon's source coding and channel coding theorems
InhaltThe entropy rate of a source, Typical sequences, the asymptotic equi-partition property, the source coding theorem, Huffman coding, Arithmetic coding, channel capacity, the channel coding theorem, the source-channel separation theorem, feedback capacity
LiteraturT.M. Cover and J. Thomas, Elements of Information Theory (second edition)
227-0421-00LLearning in Deep Artificial and Biological Neuronal NetworksW4 KP3GB. Grewe
KurzbeschreibungDeep-Learning (DL) a brain-inspired weak for of AI allows training of large artificial neuronal networks (ANNs) that, like humans, can learn real-world tasks such as recognizing objects in images. However, DL is far from being understood and investigating learning in biological networks might serve again as a compelling inspiration to think differently about state-of-the-art ANN training methods.
LernzielThe main goal of this lecture is to provide a comprehensive overview into the learning principles neuronal networks as well as to introduce a diverse skill set (e.g. simulating a spiking neuronal network) that is required to understand learning in large, hierarchical neuronal networks. To achieve this the lectures and exercises will merge ideas, concepts and methods from machine learning and neuroscience. These will include training basic ANNs, simulating spiking neuronal networks as well as being able to read and understand the main ideas presented in today’s neuroscience papers.
After this course students will be able to:
- read and understand the main ideas and methods that are presented in today’s neuroscience papers
- explain the basic ideas and concepts of plasticity in the mammalian brain
- implement alternative ANN learning algorithms to ‘error backpropagation’ in order to train deep neuronal networks.
- use a diverse set of ANN regularization methods to improve learning
- simulate spiking neuronal networks that learn simple (e.g. digit classification) tasks in a supervised manner.
InhaltDeep-learning a brain-inspired weak form of AI allows training of large artificial neuronal networks (ANNs) that, like humans, can learn real-world tasks such as recognizing objects in images. The origins of deep hierarchical learning can be traced back to early neuroscience research by Hubel and Wiesel in the 1960s, who first described the neuronal processing of visual inputs in the mammalian neocortex. Similar to their neocortical counterparts ANNs seem to learn by interpreting and structuring the data provided by the external world. However, while on specific tasks such as playing (video) games deep ANNs outperform humans (Minh et al, 2015, Silver et al., 2018), ANNs are still not performing on par when it comes to recognizing actions in movie data and their ability to act as generalizable problem solvers is still far behind of what the human brain seems to achieve effortlessly. Moreover, biological neuronal networks can learn far more effectively with fewer training examples, they achieve a much higher performance in recognizing complex patterns in time series data (e.g. recognizing actions in movies), they dynamically adapt to new tasks without losing performance and they achieve unmatched performance to detect and integrate out-of-domain data examples (data they have not been trained with). In other words, many of the big challenges and unknowns that have emerged in the field of deep learning over the last years are already mastered exceptionally well by biological neuronal networks in our brain. On the other hand, many facets of typical ANN design and training algorithms seem biologically implausible, such as the non-local weight updates, discrete processing of time, and scalar communication between neurons. Recent evidence suggests that learning in biological systems is the result of the complex interplay of diverse error feedback signaling processes acting at multiple scales, ranging from single synapses to entire networks.
SkriptThe lecture slides will be provided as a PDF after each lecture.
Voraussetzungen / BesonderesThis advanced level lecture requires some basic background in machine/deep learning. Thus, students are expected to have a basic mathematical foundation, including linear algebra, multivariate calculus, and probability. The course is not to be meant as an extended tutorial of how to train deep networks in PyTorch or Tensorflow, although these tools used.
The participation in the course is subject to the following conditions:

1) The number of participants is limited to 120 students (MSc and PhDs).

2) Students must have taken the exam in Deep Learning (263-3210-00L) or have acquired equivalent knowledge.
227-0477-00LAcoustics IW6 KP4GK. Heutschi
KurzbeschreibungIntroduction to the fundamentals of acoustics in the area of sound field calculations, measurement of acoustical events, outdoor sound propagation and room acoustics of large and small enclosures.
LernzielIntroduction to acoustics. Understanding of basic acoustical mechanisms. Survey of the technical literature. Illustration of measurement techniques in the laboratory.
InhaltFundamentals of acoustics, measuring and analyzing of acoustical events, anatomy and properties of the ear. Outdoor sound propagation, absorption and transmission of sound, room acoustics of large and small enclosures, architectural acoustics, noise and noise control, calculation of sound fields.
Skriptyes
263-5210-00LProbabilistic Artificial Intelligence Information Belegung eingeschränkt - Details anzeigen W5 KP2V + 1U + 1AA. Krause
KurzbeschreibungThis course introduces core modeling techniques and algorithms from statistics, optimization, planning, and control and study applications in areas such as sensor networks, robotics, and the Internet.
LernzielHow can we build systems that perform well in uncertain environments and unforeseen situations? How can we develop systems that exhibit "intelligent" behavior, without prescribing explicit rules? How can we build systems that learn from experience in order to improve their performance? We will study core modeling techniques and algorithms from statistics, optimization, planning, and control and study applications in areas such as sensor networks, robotics, and the Internet. The course is designed for upper-level undergraduate and graduate students.
InhaltTopics covered:
- Search (BFS, DFS, A*), constraint satisfaction and optimization
- Tutorial in logic (propositional, first-order)
- Probability
- Bayesian Networks (models, exact and approximative inference, learning) - Temporal models (Hidden Markov Models, Dynamic Bayesian Networks)
- Probabilistic palnning (MDPs, POMPDPs)
- Reinforcement learning
- Combining logic and probability
Voraussetzungen / BesonderesSolid basic knowledge in statistics, algorithms and programming
401-4619-67LAdvanced Topics in Computational Statistics
Findet dieses Semester nicht statt.
W4 KP2Vkeine Angaben
KurzbeschreibungThis lecture covers selected advanced topics in computational statistics. This year the focus will be on graphical modelling.
LernzielStudents learn the theoretical foundations of the selected methods, as well as practical skills to apply these methods and to interpret their outcomes.
InhaltThe main focus will be on graphical models in various forms:
Markov properties of undirected graphs; Belief propagation; Hidden Markov Models; Structure estimation and parameter estimation; inference for high-dimensional data; causal graphical models
Voraussetzungen / BesonderesWe assume a solid background in mathematics, an introductory lecture in probability and statistics, and at least one more advanced course in statistics.
401-3901-00LMathematical Optimization Information W11 KP4V + 2UR. Zenklusen
KurzbeschreibungMathematical treatment of diverse optimization techniques.
LernzielThe goal of this course is to get a thorough understanding of various classical mathematical optimization techniques with an emphasis on polyhedral approaches. In particular, we want students to develop a good understanding of some important problem classes in the field, of structural mathematical results linked to these problems, and of solution approaches based on this structural understanding.
InhaltKey topics include:
- Linear programming and polyhedra;
- Flows and cuts;
- Combinatorial optimization problems and techniques;
- Equivalence between optimization and separation;
- Brief introduction to Integer Programming.
Literatur- Bernhard Korte, Jens Vygen: Combinatorial Optimization. 6th edition, Springer, 2018.
- Alexander Schrijver: Combinatorial Optimization: Polyhedra and Efficiency. Springer, 2003. This work has 3 volumes.
- Ravindra K. Ahuja, Thomas L. Magnanti, James B. Orlin. Network Flows: Theory, Algorithms, and Applications. Prentice Hall, 1993.
- Alexander Schrijver: Theory of Linear and Integer Programming. John Wiley, 1986.
Voraussetzungen / BesonderesSolid background in linear algebra.
401-3621-00LFundamentals of Mathematical Statistics Information W10 KP4V + 1US. van de Geer
KurzbeschreibungThe course covers the basics of inferential statistics.
Lernziel
227-0155-00LMachine Learning on Microcontrollers Belegung eingeschränkt - Details anzeigen
Registration in this class requires the permission of the instructors. Class size will be limited to 25.
Preference is given to students in the MSc EEIT.
W3 KP2GM. Magno, L. Benini
KurzbeschreibungMachine Learning (ML) and artificial intelligence are pervading the digital society. Today, even low power embedded systems are incorporating ML, becoming increasingly “smart”. This lecture gives an overview of ML methods and algorithms to process and extract useful near-sensor information in end-nodes of the “internet-of-things”, using low-power microcontrollers/ processors (ARM-Cortex-M; RISC-V)
LernzielLearn how to Process data from sensors and how to extract useful information with low power microprocessors using ML techniques. We will analyze data coming from real low-power sensors (accelerometers, microphones, ExG bio-signals, cameras…). The main objective is to study in details how Machine Learning algorithms can be adapted to the performance constraints and limited resources of low-power microcontrollers.
InhaltThe final goal of the course is a deep understanding of machine learning and its practical implementation on single- and multi-core microcontrollers, coupled with performance and energy efficiency analysis and optimization. The main topics of the course include:

- Sensors and sensor data acquisition with low power embedded systems

- Machine Learning: Overview of supervised and unsupervised learning and in particular supervised learning (Bayes Decision Theory, Decision Trees, Random Forests, kNN-Methods, Support Vector Machines, Convolutional Networks and Deep Learning)

- Low-power embedded systems and their architecture. Low Power microcontrollers (ARM-Cortex M) and RISC-V-based Parallel Ultra Low Power (PULP) systems-on-chip.

- Low power smart sensor system design: hardware-software tradeoffs, analysis, and optimization. Implementation and performance evaluation of ML in battery-operated embedded systems.

The laboratory exercised will show how to address concrete design problems, like motion, gesture recognition, emotion detection, image and sound classification, using real sensors data and real MCU boards.

Presentations from Ph.D. students and the visit to the Digital Circuits and Systems Group will introduce current research topics and international research projects.
SkriptScript and exercise sheets. Books will be suggested during the course.
Voraussetzungen / BesonderesPrerequisites: C language programming. Basics of Digital Signal Processing. Basics of processor and computer architecture. Some exposure to machine learning concepts is also desirable
  •  Seite  1  von  1