Monash University
Browse

Explorations of Neural Network Architectures and Training Paradigms

Download (8.59 MB)
thesis
posted on 2025-10-30, 10:13 authored by Lachlan Samuel O'Neill
This thesis explores three methods for improving neural networks, respectively making them more interpretable, efficient, and effective. ‘Regression Networks’ generalize existing methods to offer explainable machine learning, but require careful use due to subtle complexities while interpreting their predictions. ‘Iterative Permanent Dropout’ shrinks neural networks while preserving accuracy. Finally, an analysis of the neural network architecture behind systems like ChatGPT motivates a new design more suitable to advanced intelligent systems, named 'Architecture Designed for Artificial General Intelligence' (or 'ADAGI'), while a discussion of broader ethical and practical safety concerns leads to some proposed regulations designed for safer AGI development.

History

Campus location

Australia

Principal supervisor

David L Dowe

Additional supervisor 1

Nader Chmait

Year of Award

2025

Department, School or Centre

Data Science & Artificial Intelligence

Course

Doctor of Philosophy

Degree Type

DOCTORATE

Faculty

Faculty of Information Technology

Rights Statement

The author retains copyright of this thesis. It must only be used for personal non-commercial research, education and study. It must not be used for any other purposes and may not be transmitted or shared with others without prior permission. For further terms use the In Copyright link under the License field.