posted on 2025-10-30, 10:13authored byLachlan Samuel O'Neill
This thesis explores three methods for improving neural networks, respectively making them more interpretable, efficient, and effective. ‘Regression Networks’ generalize existing methods to offer explainable machine learning, but require careful use due to subtle complexities while interpreting their predictions. ‘Iterative Permanent Dropout’ shrinks neural networks while preserving accuracy. Finally, an analysis of the neural network architecture behind systems like ChatGPT motivates a new design more suitable to advanced intelligent systems, named 'Architecture Designed for Artificial General Intelligence' (or 'ADAGI'), while a discussion of broader ethical and practical safety concerns leads to some proposed regulations designed for safer AGI development.
History
Campus location
Australia
Principal supervisor
David L Dowe
Additional supervisor 1
Nader Chmait
Year of Award
2025
Department, School or Centre
Data Science & Artificial Intelligence
Course
Doctor of Philosophy
Degree Type
DOCTORATE
Faculty
Faculty of Information Technology
Rights Statement
The author retains copyright of this thesis. It must only be used for personal non-commercial research, education and study. It must not be used for any other purposes and may not be transmitted or shared with others without prior permission. For further terms use the In Copyright link under the License field.