Questions about me?
Chat now.
Welcome!
I am Jérémie N. Mabiala. If you're from an English-speaking culture, you can call me Jeremy.
I am currently a student in Artificial Intelligence at the African Masters in Machine Intelligence (AMMI),
a pan-African master's program in Artificial Intelligence founded by Google and Meta,
hosted at AIMS Senegal. I am an AI enthusiast and a math lover.
Before that, I graduated from Stellenbosch University and AIMS South Africa
in February 2024 with a Master's degree in Mathematical Sciences. I worked on Mathematical Statistics, specifically Functional Data Analysis.
My master's thesis is titled Gaussian Processes for Multivariate Functional Data,
and it is available at the AIMS Archive.
I also hold a Bachelor's degree (equivalent to Bac +5) in Mathematics from the University of Kinshasa, the Congo's leading university. There I graduated in the top 5% in my department with "Grande Distinction" (the Congolese equivalent of Summa Cum Laude).
During my time there, I worked on Functional Analysis, which was my early math interest,
and I served as a teaching assistant and subsidiary lecturer for two years.
I am passionate about teaching, gaining knowledge, and sharing it with others.
I believe this passion comes from my father and grandfather, who were both teachers.
I am particularly interested in
Mathematics, Theoretical Aspects of Machine Learning and/or Deep Learning and their applications. My interests also extend to
Mathematical Modeling, Functional Data Analysis, Large Language Models, and Reinforcement Learning.
I am a hobbyist developer and an apprentice writer.
I am currently updating this website. Please, feel free to reach out to firstname(in french ) at aims.ac.za or firstname (in english ) at aimsammi.org for any inquiries.
This course was provided by Prof. Moustapha Cisse. The course had two parts: Statistical Learning Theory and an introduction to Deep Learning. In the first part, we explored theoretical aspects of machine learning, particularly statistical learning theory, covering both supervised and unsupervised learning. We discussed the ERM problem, bias-variance tradeoff, overfitting and underfitting, linear regression, classification, generalized linear models, and dimensionality reduction techniques such as:
This course was given by J.P. Vert from Owkin, with T.As. Romain Menegeaux and Juliette-Marie from Inria. The course covered basic concepts of machine learning in high dimensions and the importance of regularization. We also discussed Support Vector Machines (SVM). We studied in detail high-dimensional linear models regularized by the Euclidean norm or the Manhattan norm, including ridge and lasso regression, as well as ridge classification. One of the key ideas of this course was to demonstrate that some learning problems, which can be complex to solve in finite-dimensional Euclidean space, can be easily addressed in infinite-dimensional spaces, provided a good feature mapping. The kernel is simply a similarity function between data points. We were then shown how positive kernels, via the representer theorem, transform the ERM problem in Euclidean space into an ERM problem in large-dimensional space. Kernels allow for transforming linear models into non-linear models, applicable even to non-euclidean data such as strings and graphs. We also discussed Reproducing Kernel Hilbert Spaces (RKHS) and related theorems. Please find my code and the course page here.
Many problems in statistical learning consist of finding the best set of parameters or the best functions given some data. These are estimation problems. These problems, encountered almost everywhere in Machine Learning and Deep Learning, can easily be formulated as optimization problems. These formulations help to understand the performance of learning algorithms. In this course, classical convex optimization theory was presented. Classical gradient methods and their variants were discussed, as well as variance-reduced methods and accelerated gradient methods. The purpose of the course was to present the theoretical formulation of convex problems and the asymptotic behaviors of (Stochastic) Gradient methods. For instance, it was shown that although Stochastic Gradient Descent (SGD) is efficient—because it does not make use of the full dataset—it never converges to the optimal solution without restrictive assumptions. We also implemented these algorithms from scratch and worked on theoretical exercises. Please refer to the course material. The course was given by Dr. Lionel Tondji, a former student of AMMI. Find the goodbye photos here.
This course was split into two parts: the first part was an introduction to computer vision, taught by Dr. Laurens van der Maaten from Meta AI, and co-inventor of t-SNE. The second part focused on object detection and was given by Dr. Natalia Neverova from Meta AI. In the first part, we discussed CNNs, RNNs, U-Net, and implemented these architectures and related ones from scratch. See my repo for the code. The second part focused on detection and diffusion models. We covered vision models, diffusion models, distillation, image and video understanding, and more. The main tool used in this part was FAIR's Detectron2, with which we experimented using vision models like R-CNN, Mask R-CNN, Fast R-CNN, and RPN. Please check out the course the course materials, or my tutorials here on Object Detection with Detectron2.
.This is the current research project I am working on with colleagues from Tunisia, Senegal, and DR. Congo. The goal of this project is to conduct a comparative study of various Deep Reinforcement Learning (DRL) models combinded with GNN and XAI. We aim to evaluate their performance in optimizing routes and how well each model explains its decisions. The idea is to provide a transparent decision-making process for electric vehicle routing.