Tom (Tomáš) Pecher

CS & AI Student | Aspiring Software Developer | Passion for ML and Wolfram

Why I study AI

In my totally unbiased opinion, the study of artificial intelligence is by far the most interesting field in all of science. However, I also think that the name omits the most interesting aspect of the field. Despite the concept of AI being fairly old, experts still fundamentally disagree on what "artificial intelligence" is even supposed to mean. I interpret this field as the study of intelligence (both natural and artificial) and how simple systems can display complex behaviour. I think that people often overlook the fact that AI is a study of ourselves just as much as it is a study of computer systems. Neural networks are intelligent agents, but so are humans and all living things, even society as a whole can be considered as a single self-sustaining agent. To me, the most interesting questions are not, "can we do this or that with AI?", eventually, we will probably do most things with AI. Rather, we should ask, what does this say about us as individuals, as a species, and how will we be changed because of it? No technological advancement is ever purely technological; people change technology and technology changes us back. I believe that this change is inevitable, but we have the capacity to let AI be a change for better or for worse. And so, I study AI so that this change is a positive one, for all intelligent agents.

AI Experience

Bipedal Walking in Increasingly Treacherous Terrain (2025, RL)

As part of a group project (and our collective introduction to RL), we conducted an experiment in which we implemented a variaty of RL-based methods and set them to train on the OpenAI Gym bipedal walker environment. Specifically, we pretrained the models on the base environment (a flat surface) and then tested the best performing models on the "hardcore" environment (a surface with random bumps and holes). This is a common RL problem and is widely considered to be quite difficult. Never the less we managed to train a Soft Actor-Critic (SAC) and a SUNRISE agent to solve the hardcore environment (reach 300+ reward). More impressive however, the SUNRISE agent managed to converge to this optimal strategy five times faster than existing models we could find. This project was great fun and has made me go down an RL rabbit hole that I am still exploring in my individual project. Many thanks to my group members for their hard work and dedication that made this project possible (they are all great programmers and great people so check out their LinkedIns here: Marilyn D'Costa, Ptolemy Morris, Dhru Randeria, George Rawlinson).

Voice-based Action Selection (2024, NLP)

As part of the the VIP study "Creating immersive training experiences in VR", we intended to create a VR simulation for training users to be "effective bystanders" when witnessing sexual harassment. As lead developer, my role was to create a system that would classify the user's speech towards the perpetrator into one of a set of predetermined actions (such as distract the perpetrator etc.). Using Tensorflow, I implemented and fine-tuned an LSTM model that managed to reach 97% accuracy on test data. The system is currently in the testing phase and I hope my contribution will help the study reach its goals.

Hypercustomizable Subtitling System (2024, NLP)

As part of the "Experimental Systems Project" group module, we created a system that generates subtitles for any video in real-time. A key goal with our system was to ensure that the underlying model would be able to perform even in noisy scenarios where the audio quality was poor. We achieved this by dynamically switching between models to best adapt to the audio conditions. The system was a success, and we were able to demonstrate it to our peers and lecturers.

Current Work

Building a Robust and Scalable Traffic Control System using Reinforcement Learning (2025, RL)

As my third-year individual project, I am working on creating an RL-based traffic control system that can dynamically adapt to traffic conditions and (hopefully) outperform simple actuated systems. A decent amount of research has been dedicated to this area, however most systems struggle with robustness (exhibit unpredictible behaviour) and scalability (struggle responding to new scenarios). This project aims to marry these two ideas in such a way that would make RL-based traffic control systems not only viable but preferable over existing fixed-control and actuated systems.

Future Plans

When I am less busy, I hope to delve into some of these potential avenues:

Glossary