Back

💫 Preliminary Schedule 24/25

Subject to change

Speech2text

In the first block of the semester, we will focus on automatic speech recognition systems (ASR). In other words, systems that can map speech in form an audio signal to text.

  • W38 Waw2vec / FAIR / 2019: Unsupervised Pretraining for Speech Recognition
  • W39 Conformer / Google / 2020: Convolution-augmented Transformer for Speech Recognition
  • W40 Whisper / OpenAI / 2022: Robust Speech Recognition via Large-Scale Weak Supervision
  • W41 Industry talk TBA
  • W42 Autumn Break

Graph Neural Networks

In this block, we venture into graph neural networks and their application in football tactics.

Training our own GPT

We have now read many interesting research papers. Now, it is time to get our hands dirty. We will make use of Andrej Karpathy’s open source implementation of GPT-2 to train our small language model. We will go through the whole setup from collecting and preparing the data to logging the training results. First week, we will spend on setting up the baseline model. The following week, we will try to add some tweaks to it to see if we can improve its performance.

Industry talks