Skip to main content

Honours and Masters project

Displaying 11 - 20 of 255 honours projects.


Haptic Devices in Multi-User VR

Multi-user virtual reality (VR) systems are becoming more common as these head-mounted displays (e.g., Oculus Quest, HTC Vive) become less expensive and thus more widely available. It is now feasible to have multiple users sharing a co-located VR space, and objects in the space. Traditionally, VR systems tend not to support the sense of touch, but recent advances in haptics (devices simulating physical objects) now make this possible.

Cybersickness Amelioration in Virtual Reality

Cybersickness (nausea, disorientation) due to exposure to virtual environments has long been a problem in virtual reality (VR) and has been shown to reduce the effectiveness of VR environments. It usually occurs due to mismatches between visual and vestibular (motion) cues, for example, when moving through a 3D environment using a joystick, which does not yield correct motion cues. There are several approaches to reducing cybersickness in VR, most notably, reducing the field of view ("tunneling") during motion, or discrete motion ("snapping" movement and rotation).

Empirical study in software systems

In this project, we will conduct an empirical study to understand certain problems in software systems. 

Software testing and debugging with/without AI/LLMs.

In this project, students and me will work together to develop a new technique for software testing and debugging. 

 

The subject under test may be AI/LLM. The technique may involve AI/LLM as well. 

Sketched networks: how do we assess their quality?

Network visualisation (or 'graph drawing') algorithms allow us to see connections and patterns in a network clearly. There are a large number of such algorithms that depict ('lay out') networks 'nicely' - according to a set of well-established criteria. These criteria apply to neat diagrams - with clear straight (nor neatly bent or curved) lines ('edges').

This project will investigate how networks that have been sketched (with curved/ wavy lines) can be adapted so that they, too, can be assessed by typical graph layout criteria.

Temporal Analytics

Time series are an ever growing form of data, generated by numerous types of sensors and automated processes. However, machine learning and deep learning methods for analysing time series are much less advanced than for other forms of data.

Our research is revolutionising the analysis of time series data. But it is early days, and many more impactful challenges are yet to be overcome.

This project is funded by the Australian Research Council and will be conducted as part of a large world-leading research team.

Bayesian Poisson regression with global-local shrinkage priors

Poisson regression is a fundamental tool for modeling count data, appearing ubiquitously in applications ranging from epidemiology (disease counts) and ecology (species abundance) to economics (patent counts) and social sciences (event frequencies). The classical generalized linear model (GLM) framework treats count outcomes as Poisson-distributed with a log-linear relationship to covariates.

MML decision trees for survival analysis

Decision trees are powerful, interpretable models for prediction and classification that recursively partition the feature space into regions with homogeneous outcomes. Traditional decision tree algorithms like CART and C4.5 rely on heuristic splitting criteria and require ad-hoc pruning methods to prevent overfitting. In contrast, the Minimum Message Length (MML) framework provides a principled, information-theoretic approach to tree induction that naturally balances model complexity against data fit without requiring separate pruning phases.

Inductive inference with Minimum Message Length

Minimum Message Length (MML) is an elegant information-theoretic framework for statistical inference and model selection developed by Chris Wallace and colleagues. The fundamental insight of MML is that both parameter estimation and model selection can be interpreted as problems of data compression. The principle is simple: if we can compress data, we have learned something about its underlying structure.

Using AI and machine learning to improve polygenic risk prediction of disease

We are interested in understanding genetic variation among individuals and how it relates to disease. To do this, we study genomic markers or variants called single nucleotide polymorphisms, or SNPs for short. A SNP is a single base position in DNA that varies among human individuals. The Human Genome Project has found that these single letter changes occur are all over the human genomes; each person has about 5M of them!  While most SNPs have no effect, some can influence traits or increase the risk of certain diseases.