Kantegger

Best Personal Statements are Here


Beyond the Black Box: Advancing AI Through Bayesian Methods

The main reason for applying for the MSc in Computing (AI and ML) at a prestigious institution is that I want to study with Dr. [First Placeholder Name] and Dr. [Second Placeholder Name]. They have done great works in the field that I want to dive into: Bayesian deep learning.

My interest in Bayesian deep learning is based on my past research experience. Ever since I stepped into the world of AI research, I have been a fan of deep learning and believed these powerful and complex networks are and will be the most important tools. The success of deep learning largely relies on its scalability and in this era of big data, with inexpensive high-speed computing available, 175-billion-parameter models can be trained. However, the downside of this powerfulness is they are too complex to understand, so we have to treat them as black boxes. Such intransparency makes human unable to trust and deploy them in the real world. Therefore, as I was trying to improve deep neural networks (DNNs) for monocular depth estimation (MDE), a task of predicting scene depth from a single image, I decided to take methodologies in explainable AI. I first discovered that in well-trained DNNs for MDE, some hidden units were selective to some ranges of depth. Later, I was influenced by Professor [Third Placeholder Name], who argued that researchers should focus on make ML models more inherently interpretable instead of explainable because post-hoc explanations are always inaccurate and often misleading. Inspired by this idea, I proposed a training method making every hidden unit selective and boosted the DNNs’ interpretability.

While I am excited by getting my first-authored paper published at a top conference, I keep reflecting the broader impact of my research. What kind of interpretability or explainability do we need from AI for real-world deployment? After all, humans are not interpretable, either—we are able to explain our decisions, but when it comes to ‘System 1’ decisions (intuitions), explanations are indeed not accurate. So, The intransparency is not the problem; the problem is they are not safe enough to be trusted.

Dr. [First Placeholder Name]’s tutorial on approximate inference at NeurIPS 2020 triggered me to think about the importance of uncertainty for trustworthy intelligence. Unlike deep learning models, humans can tell when they are not sure. A responsible doctor will ask for further examines if she thinks it is hard to tell the tumour is benign or malignant, making her more reliable than DNNs. This ability to know what it doesn’t know can be quantified as a model’s uncertainty, and fortunately, Bayesian inference provides us with principled methods to represent uncertainty. In recent years, researchers tackled the limitation of deep learning from Bayesian perspectives and achieved many advances in an arising field called Bayesian deep learning (BDL), where strengths from deep learning like scalability can be combined with abilities of Bayesian methods like uncertainty reasoning. BDL is advantageous because this theoretical-grounded framework captures not only aleatoric uncertainty (irreducible data uncertainty) elegantly but also epistemic uncertainty (reducible model uncertainty), which is more useful for safety-critical applications. And, only in this framework can uncertainty be propagated forward in complicated intelligent systems.

Particularly, in my future study towards postgraduate and doctoral degrees, I plan to work on the adversarial robustness of DNNs from BDL perspectives. Studies have shown that DNNs are vulnerable to adversarial attacks and such issue could be dangerous for real-world AI systems. I also think it is an intriguing direction because research has shown that the adversarial training framework proposed by [Fourth Placeholder Name] et al. for adversarial robustness could disentangle the robust and non-robust features of DNNs. But do BNNs also learn robust and non-robust features, and can they be marginalized? Are non-robust features also transferable across different BNNs? Do adversarially robust BNNs also transfer better? Can adversarial robustness be incorporated into the Bayesian inference framework as the prior? Maybe more importantly, is learning robust posterior help learn the true posterior?

In the tutorial at NeurIPS 2020, Dr. [First Placeholder Name] and collaborator introduced approximate inference, methods that approximate difficult-to-compute probability densities, especially integrals in Bayesian inference. It is one of the core topics in BDL because when computing Bayesian inference for complex models like DNNs, the integrals are impossible to compute exactly. I am especially attracted to her work on the adversarial robustness of deep generative classifiers, motivated by the cause of DNNs’ vulnerability by its discriminative nature and the poor performance of non-deep Bayesian generative classifiers on image classification. They proposed a method to improve classical Naive Bayes classifier and achieved better robustness because it not only leaded to more successful detection for adversarial examples but also perform well under attacks that detectors failed to detect. Besides, Dr. [Second Placeholder Name] is another researcher that I dream to work with. Dr. [Second Placeholder Name] defended his PhD thesis on Gaussian process three years ago. In the more recent research, he and his colleagues connect GP and DNN and achieve a more flexible and accurate deep GP. He also contributes to invariance learning in neural networks. Within a Bayesian context, data augmentation can be replaced by optimising the marginal likelihood to learn invariance (which can be seen as more general robustness than adversarial robustness). I am curious if I can adapt this scheme to the adversarial training framework to solve the inner maximisation problem more efficiently.

In the taught programme, the only compulsory module Mathematics for Machine Learning will be taught by Dr. [Second Placeholder Name]. I will be armed with a solid mathematical background and skills, and with his instruction, I believe I will be able to connect these fundamental ideas with cutting-edge research. He will also be teaching Probabilistic Inference in the Spring term, discussing topics in Bayesian machine learning like Gaussian processes and approximate inference. In addition, Dr. [First Placeholder Name] is the instructor of Deep Learning. Thanks to her rich research experience in BDL, the course taught at the university emphasizes more on the mathematical principles of deep learning models, especially for those which are optimised by Bayesian approaches like VAE. More importantly, I will get the opportunity to conduct a research project with their help, in which I will explore the adversarial robustness of deep learning from a Bayesian view. Moreover, course including Natural Language Processing, Robot Learning and Control, and Logic-Based Learning will supplement me the knowledge of other AI application fields and methods besides computer vision and make me fully understand the potential of AI. In sum, I believe the programme is my perfect next step and I am ready for it.



One response to “Beyond the Black Box: Advancing AI Through Bayesian Methods”

  1. Comment from GPT-4:

    Strengths:

    Clear Research Focus and Relevance: The applicant demonstrates a strong, clear focus on Bayesian deep learning (BDL), articulating both the academic and practical motivations behind their interest. The discussion on the importance of interpretability, explainability, and the unique advantages of BDL in addressing these issues is particularly compelling.

    Engagement with Current Research: The statement showcases a deep engagement with current research in the field, including specific references to works and methodologies by named researchers. This not only indicates a solid grounding in the subject matter but also highlights the applicant’s ability to engage critically with ongoing debates and developments.

    Research Experience and Contributions: The mention of past research experience, particularly the development of a training method that enhances DNN interpretability, and the publication of a first-authored paper, strongly supports the applicant’s claim to both technical ability and innovative thinking.

    Future Research Goals: The applicant articulates a coherent and ambitious set of research goals, focusing on the adversarial robustness of DNNs from a BDL perspective. This suggests a forward-thinking approach and a commitment to contributing to the field’s development.

    Program-Specific Motivations: The detailed discussion of specific courses, instructors, and how these align with the applicant’s academic goals demonstrates a thoughtful consideration of how the program fits their research ambitions. It shows that the applicant has not only researched the program thoroughly but also envisages how it will benefit their future career.

    Areas for Improvement:

    Broader Impact Reflection: While the applicant reflects on the broader impact of their research, further elaboration on how their work could contribute to societal challenges or ethical considerations in AI might strengthen the statement.

    Personal Journey and Motivation: The statement is heavily focused on academic and research achievements. Including more personal insights into what initially sparked their interest in AI and BDL, or challenges they’ve overcome in their research journey, could add a more compelling, human element to the narrative.

    Diversity of Interests: Although the focus on BDL is a strength, mentioning interests or openness to exploring adjacent areas within AI could demonstrate adaptability and a broader intellectual curiosity.

    Conclusion and Decision:

    Based on the strengths of the application, particularly the applicant’s clear research focus, engagement with current research, demonstrated research ability, and thoughtful consideration of how the program aligns with their academic and professional goals, I would be inclined to offer a place to this applicant. Their statement indicates that they would not only benefit from the program but also contribute significantly to the academic community. The potential for future research contributions, especially in the critical area of adversarial robustness in BDL, suggests that the applicant could become a leader in the field. Therefore, I recommend offering a place to this applicant, anticipating that they will excel in the program and make meaningful contributions to the broader AI research community.

Leave a Reply

Your email address will not be published. Required fields are marked *

About Me

As a dedicated Personal Statement Mentor, I specialise in guiding students through the process of crafting personal statements that truly reflect their individuality and experiences. My role involves helping students highlight their unique journeys and stand out in competitive admissions processes.

“Tell me, what is it you plan to do with your one wild and precious life?”

– Mary Oliver