Special programming and simulation methods in high energy physics and astrophysics

Course: Quantum field theory

Structural unit: Faculty of Physics

Title
Special programming and simulation methods in high energy physics and astrophysics
Code
ВБ 4.1
Module type
Вибіркова дисципліна для ОП
Educational cycle
Second
Year of study when the component is delivered
2021/2022
Semester/trimester when the component is delivered
3 Semester
Number of ECTS credits allocated
6
Learning outcomes
To know the theory and problems of modern machine learning models To know the main approaches to the problems of high-energy astrophysics using neural networks Be able to use Python, Wolfram Mathematica libraries for data preparation and metric classification, regression problems To be able to apply theoretical and practical approaches to multidimensional classification problems Demonstrate the ability to communicate freely in the language of instruction Use knowledge of foreign languages ​​to read technical literature on programming in preparation for seminar classes To participate in professional discussions during classroom work
Form of study
Full-time form
Prerequisites and co-requisites
1. Know the basics of mathematical analysis, probability theory, statistical physics and thermodynamics. 2. To be able to apply previously acquired knowledge from the courses of mathematical analysis, probability theory, statistical physics and thermodynamics to solve practical problems. 3. To have basic calculation skills from the theory of differential equations course.
Course content
Part 1. Basic methods 1 Basic concepts of neural networks 2 Stochastic gradient descent, the role of saddle points in the loss landscape, automatic differentiation and backpropagation 3 Basic properties of entropy, differential entropy, Bayes: motivation, Bayes formula, physical examples for updating Bayes, conditional entropy 4 Generative competitive networks, recurrent neural networks 5 natural gradient, Kullback-Leibler divergence and its properties, Fisher information, mutual information Part 2 Neural networks 6 Mutual information, renormalized mutual information, inverted neural networks 7 Variational autoencoder, generative competitive networks 8 Implicit layers: solving equations and neural differential equations. Hamiltonian and Lagrangian neural networks. Reinforcement learning 9 Big Data architecture, design of data warehouses
Recommended or required reading and other learning resources/tools
1. Pattern Recognition and Machine Learning, Bishop, Christopher (2006). 2. Linked: The New Science Of Networks Science Of Networks, Perseus Books Group (2002). 3. Deep Learning (by Goodfellow, Bengio, Courville) https://www.deeplearningbook.org/ 4. An Introduction to Statistical Learning, Gareth Games, Daniela Witten, Travor Hastie, Robert Tibshirani, Springer (2013). 5. Neural Networks A Systematic Introduction, R Rojas, Springer (1996).
Planned learning activities and teaching methods
Lectures Practical Individual work
Assessment methods and criteria
Oral survey Projects Report Exam work
Language of instruction
Ukrainian

Lecturers

This discipline is taught by the following teachers

Departments

The following departments are involved in teaching the above discipline