Auckland University of Technology
·
Large language models (LLMs) hold significant potential for mental health support, capable of generating empathetic responses and simulating therapeutic conversations. However, existing LLM-based approaches often lack the clinical grounding necessary for real-world psychological counseling, particularly in explicit diagnostic reasoning aligned with standards like the DSM/ICD and incorporating diverse therapeutic modalities beyond basic empathy or single strategies. To address these critical limitations, we propose PsyLLM, the first large language model designed to systematically integrate both diagnostic and therapeutic reasoning for mental health counseling. To develop PsyLLM, we design a novel automated data synthesis pipeline that processes real-world mental health posts collected from Reddit, where users frequently share psychological distress and seek community support. This pipeline processes real-world mental health posts, generates multi-turn dialogue structures, and leverages LLMs guided by international diagnostic standards (e.g., DSM/ICD) and multiple therapeutic frameworks (e.g., CBT, ACT, psychodynamic) to simulate detailed clinical reasoning processes. Rigorous multi-dimensional filtering ensures the generation of high-quality, clinically aligned dialogue data. In addition, we introduce a new benchmark and evaluation protocol, assessing counseling quality across four key dimensions. Our experiments demonstrate that PsyLLM significantly outperforms state-of-the-art baseline models on this benchmark. The model weights and dataset have been publicly released at this https URL.
·
University of Macau, East China Normal University, and Auckland University of Technology researchers develop LongBench-T2I, the first comprehensive benchmark featuring 500 intricate multi-sentence instructions across nine visual dimensions for evaluating complex text-to-image generation, alongside Plan2Gen, a training-free agent framework that uses LLM-driven scene decomposition into background-midground-foreground layers with iterative refinement to achieve state-of-the-art performance (3.73 average score vs 3.70 for GPT-4o on Gemini-2.0-Flash evaluation), demonstrating that structured planning and validation can enhance instruction-following capabilities without requiring additional model training while revealing a counter-intuitive positive correlation between instruction perplexity and visual quality in smaller models that diminishes with scale.
19
·
Researchers from East China Normal University, University of Macau, and Auckland University of Technology introduce ComplexBench-Edit, a new benchmark for evaluating image editing models on complex, multi-step instructions with compositional dependencies. They also present a Chain-of-Thought (CoT) based approach that significantly boosts existing models' instruction-following capabilities on these challenging tasks.
19
·
EmoBench-M introduces a comprehensive, psychologically grounded benchmark for evaluating multimodal large language models' emotional intelligence across dynamic video, audio, and text. It reveals a substantial performance gap between state-of-the-art MLLMs and human emotional understanding, particularly in conversational contexts.
A hybrid attention-based deep learning architecture, SETransformer, is developed for robust Human Activity Recognition (HAR) using wearable sensor data. The model achieves 84.68% accuracy and 84.64% macro F1-score on the WISDM dataset, improving upon existing CNN and RNN baselines by over 13 percentage points.
We describe our ongoing research that centres on the application of natural language processing (NLP) to software engineering and systems development activities. In particular, this paper addresses the use of NLP in the requirements analysis and systems design processes. We have developed a prototype toolset that can assist the systems analyst or software engineer to select and verify terms relevant to a project. In this paper we describe the processes employed by the system to extract and classify objects of interest from requirements documents. These processes are illustrated using a small example.
AI-based music generation has made significant progress in recent years. However, generating symbolic music that is both long-structured and expressive remains a significant challenge. In this paper, we propose PerceiverS (Segmentation and Scale), a novel architecture designed to address this issue by leveraging both Effective Segmentation and Multi-Scale attention mechanisms. Our approach enhances symbolic music generation by simultaneously learning long-term structural dependencies and short-term expressive details. By combining cross-attention and self-attention in a Multi-Scale setting, PerceiverS captures long-range musical structure while preserving performance nuances. The proposed model has been evaluated using the Maestro dataset and has demonstrated improvements in generating coherent and diverse music, characterized by both structural consistency and expressive variation. The project demos and the generated music samples can be accessed through the link: this https URL.
Recently, diffusion models have been proven to perform remarkably well in text-to-image synthesis tasks in a number of studies, immediately presenting new study opportunities for image generation. Google's Imagen follows this research trend and outperforms DALLE2 as the best model for text-to-image generation. However, Imagen merely uses a T5 language model for text processing, which cannot ensure learning the semantic information of the text. Furthermore, the Efficient UNet leveraged by Imagen is not the best choice in image processing. To address these issues, we propose the Swinv2-Imagen, a novel text-to-image diffusion model based on a Hierarchical Visual Transformer and a Scene Graph incorporating a semantic layout. In the proposed model, the feature vectors of entities and relationships are extracted and involved in the diffusion model, effectively improving the quality of generated images. On top of that, we also introduce a Swin-Transformer-based UNet architecture, called Swinv2-Unet, which can address the problems stemming from the CNN convolution operations. Extensive experiments are conducted to evaluate the performance of the proposed model by using three real-world datasets, i.e., MSCOCO, CUB and MM-CelebA-HQ. The experimental results show that the proposed Swinv2-Imagen model outperforms several popular state-of-the-art methods.
Exploring the complex structure of the human brain is crucial for understanding its functionality and diagnosing brain disorders. Thanks to advancements in neuroimaging technology, a novel approach has emerged that involves modeling the human brain as a graph-structured pattern, with different brain regions represented as nodes and the functional relationships among these regions as edges. Moreover, graph neural networks (GNNs) have demonstrated a significant advantage in mining graph-structured data. Developing GNNs to learn brain graph representations for brain disorder analysis has recently gained increasing attention. However, there is a lack of systematic survey work summarizing current research methods in this domain. In this paper, we aim to bridge this gap by reviewing brain graph learning works that utilize GNNs. We first introduce the process of brain graph modeling based on common neuroimaging data. Subsequently, we systematically categorize current works based on the type of brain graph generated and the targeted research problems. To make this research accessible to a broader range of interested researchers, we provide an overview of representative methods and commonly used datasets, along with their implementation sources. Finally, we present our insights on future research directions. The repository of this survey is available at \url{this https URL}.
Context: It is not uncommon for a new team member to join an existing Agile software development team, even after development has started. This new team member faces a number of challenges before they are integrated into the team and can contribute productively to team progress. Ideally, each newcomer should be supported in this transition through an effective team onboarding program, although prior evidence suggests that this is challenging for many organisations. Objective: We seek to understand how Agile teams address the challenge of team onboarding in order to inform future onboarding design. Method: We conducted an interview survey of eleven participants from eight organisations to investigate what onboarding activities are common across Agile software development teams. We also identify common goals of onboarding from a synthesis of literature. A repertory grid instrument is used to map the contributions of onboarding techniques to onboarding goals. Results: Our study reveals that a broad range of team onboarding techniques, both formal and informal, are used in practice. It also shows that particular techniques that have high contributions to a given goal or set of goals. Conclusions: In presenting a set of onboarding goals to consider and an evidence-based mechanism for selecting techniques to achieve the desired goals it is expected that this study will contribute to better-informed onboarding design and planning. An increase in practitioner awareness of the options for supporting new team members is also an expected outcome.
We present the third data release from the Parkes Pulsar Timing Array (PPTA) project. The release contains observations of 32 pulsars obtained using the 64-m Parkes "Murriyang" radio telescope. The data span is up to 18 years with a typical cadence of 3 weeks. This data release is formed by combining an updated version of our second data release with 3\sim 3 years of more recent data primarily obtained using an ultra-wide-bandwidth receiver system that operates between 704 and 4032 MHz. We provide calibrated pulse profiles, flux-density dynamic spectra, pulse times of arrival, and initial pulsar timing models. We describe methods for processing such wide-bandwidth observations, and compare this data release with our previous release.
Online deception and financial scams represent a pervasive threat in the digital age, yet a quantitative analysis and understanding of their propagation is lacking. This study introduces a novel model based on the framework of epidemiological models to describe the interaction between scammers and their victims. We propose a five-compartment deterministic model (SVRAsRsS-V-R-A_s-R_s) calibrated using longitudinal data in fraud reports from the Canadian Anti-Fraud Centre. The model's theoretical properties are established, including the non-negativity of the state variables and the stability threshold defined by the basic reproduction number (R0\mathcal{R}_0). A non-standard finite difference scheme is developed for the numerical simulations to ensure dynamical consistency between the continuous deterministic model and its discrete equivalent. A key finding of the model sensitivity analysis indicates that the proliferation of scams is overwhelmingly driven by the lifecycle of scammers, their recruitment, attrition, and arrest, rather than the susceptibility of the victim population. The results of this study provide strong quantitative evidence that the most effective control strategies are those that directly disrupt the scammers' population. Overall, this study provides a crucial model for designing and evaluating evidence-based policies to combat the scourge of cybercrime.
Among the multitude of software development processes available, hardly any is used by the book. Regardless of company size or industry sector, a majority of project teams and companies use customized processes that combine different development methods -- so-called hybrid development methods. Even though such hybrid development methods are highly individualized, a common understanding of how to systematically construct synergetic practices is missing. In this paper, we make a first step towards devising such guidelines. Grounded in 1,467 data points from a large-scale online survey among practitioners, we study the current state of practice in process use to answer the question: What are hybrid development methods made of? Our findings reveal that only eight methods and few practices build the core of modern software development. This small set allows for statistically constructing hybrid development methods. Using an 85% agreement level in the participants' selections, we provide two examples illustrating how hybrid development methods are characterized by the practices they are made of. Our evidence-based analysis approach lays the foundation for devising hybrid development methods.
Research around Spiking Neural Networks has ignited during the last years due to their advantages when compared to traditional neural networks, including their efficient processing and inherent ability to model complex temporal dynamics. Despite these differences, Spiking Neural Networks face similar issues than other neural computation counterparts when deployed in real-world settings. This work addresses one of the practical circumstances that can hinder the trustworthiness of this family of models: the possibility of querying a trained model with samples far from the distribution of its training data (also referred to as Out-of-Distribution or OoD data). Specifically, this work presents a novel OoD detector that can identify whether test examples input to a Spiking Neural Network belong to the distribution of the data over which it was trained. For this purpose, we characterize the internal activations of the hidden layers of the network in the form of spike count patterns, which lay a basis for determining when the activations induced by a test instance is atypical. Furthermore, a local explanation method is devised to produce attribution maps revealing which parts of the input instance push most towards the detection of an example as an OoD sample. Experimental results are performed over several image classification datasets to compare the proposed detector to other OoD detection schemes from the literature. As the obtained results clearly show, the proposed detector performs competitively against such alternative schemes, and produces relevance attribution maps that conform to expectations for synthetically created OoD instances.
4
Increasingly, the combination of clinical judgment and predictive risk modelling have been assisting social workers to segregate children at risk of maltreatment and recommend potential interventions of authorities. A critical concern among governments and research communities worldwide is that misinterpretations due to poor modelling techniques will often result in biased outcomes for people with certain characteristics (e.g., race, socioeconomic status). In the New Zealand care and protection system, the over-representation of Māori might be incidentally intensified by predictive risk models leading to possible cycles of bias towards Māori, ending disadvantaged or discriminated against, in decision-making policies. Ensuring these models can identify the risk as accurately as possible and do not unintentionally add to an over-representation of Māori becomes a crucial matter. In this article we address this concern with the application of predictive risk modelling in the New Zealand care and protection system. We study potential factors that might impact the accuracy and fairness of such statistical models along with possible approaches for improvement.
Flexible and efficient noise characterization is crucial for the precise estimation of gravitational wave parameters. We introduce a fast and accurate Bayesian method for estimating the power spectral density (PSD) of long, stationary time series tailored specifically for LISA data analysis. Our approach models the PSD as a geometric mean of a parametric and a nonparametric component, combining the computational efficiency of parametric models with the flexibility to capture deviations from theoretical expectations. The nonparametric component is expressed by a mixture of penalized B-splines. Adaptive, data-driven knot placement performed once during initialization eliminates computationally expensive reversible-jump Markov Chain Monte Carlo, while hierarchical roughness penalty priors prevent overfitting. This design yields stable, flexible PSD estimates with runtimes of minutes instead of hours. Validation on simulated autoregressive AR(4) data demonstrates estimator consistency. It shows that well-matched parametric components reduce the integrated absolute error compared to an uninformative baseline, requiring fewer spline knots to achieve comparable accuracy. Applied to a year of simulated LISA XX-channel noise, our method achieves relative integrated absolute errors of O(102)\mathcal{O}(10^{-2}) with computation times less than three minutes, which makes it suitable for iterative analysis pipelines and multi-year mission datasets.
Highly pathogenic avian influenza (HPAI), especially the H5N1 strain, remains a major threat to animal health, food security, and public health. Recent spillover events in dairy cattle in the United States, linked to wild birds, highlight the critical importance of understanding transmission pathways at the cattle--wild bird--environment interface. In this work, we formulate and analyze a deterministic compartmental model that captures the transmission of HPAI between dairy cattle and wild birds, incorporating both direct and indirect (environmental) routes. The model combines an SEIRSEIR framework for cattle with an SIRSIR structure for wild birds, coupled through an environmental compartment. We derive the basic reproduction number, R0\mathcal{R}_{0}, using the next-generation matrix approach, decomposing it into cattle-to-cattle, bird-to-bird, and environmental contributions. Qualitative analysis establishes positivity, boundedness, and global stability of equilibria through Lyapunov functions. Numerical simulations confirm the results of the theoretical analyses, illustrating outbreak trajectories, extinction thresholds, and persistence dynamics. A global sensitivity analysis, based on Latin hypercube sampling and partial rank correlation coefficients, identifies key parameters, particularly transmission among cattle, environmental contamination, and recovery rate as critical drivers of epidemic outcomes. Our results show that disease elimination is achievable when \mathcal{R}_{0} < 1, while persistence is inevitable for \mathcal{R}_{0} > 1. These findings provide a comprehensive mathematical framework for assessing HPAI risks and offer guidance for biosecurity strategies aimed at mitigating spillover and controlling outbreaks in livestock populations.
Recent advances in Gaussian Splatting have significantly boosted the reconstruction of head avatars, enabling high-quality facial modeling by representing an 3D avatar as a collection of 3D Gaussians. However, existing methods predominantly rely on frontal-view images, leaving the back-head poorly constructed. This leads to geometric inconsistencies, structural blurring, and reduced realism in the rear regions, ultimately limiting the fidelity of reconstructed avatars. To address this challenge, we propose AvatarBack, a novel plug-and-play framework specifically designed to reconstruct complete and consistent 3D Gaussian avatars by explicitly modeling the missing back-head regions. AvatarBack integrates two core technical innovations,i.e., the Subject-specific Generator (SSG) and the Adaptive Spatial Alignment Strategy (ASA). The former leverages a generative prior to synthesize identity-consistent, plausible back-view pseudo-images from sparse frontal inputs, providing robust multi-view supervision. To achieve precise geometric alignment between these synthetic views and the 3D Gaussian representation, the later employs learnable transformation matrices optimized during training, effectively resolving inherent pose and coordinate discrepancies. Extensive experiments on NeRSemble and K-hairstyle datasets, evaluated using geometric, photometric, and GPT-4o-based perceptual metrics, demonstrate that AvatarBack significantly enhances back-head reconstruction quality while preserving frontal fidelity. Moreover, the reconstructed avatars maintain consistent visual realism under diverse motions and remain fully animatable.
In this paper, we investigate the extent to which features derived from bank statements provided by loan applicants, and which are not declared on an application form, can enhance a credit scoring model for a New Zealand lending company. Exploring the potential of such information to improve credit scoring models in this manner has not been studied previously. We construct a baseline model based solely on the existing scoring features obtained from the loan application form, and a second baseline model based solely on the new bank statement-derived features. A combined feature model is then created by augmenting the application form features with the new bank statement derived features. Our experimental results using ROC analysis show that a combined feature model performs better than both of the two baseline models, and show that a number of the bank statement-derived features have value in improving the credit scoring model. The target data set used for modelling was highly imbalanced, and Naive Bayes was found to be the best performing model, and outperformed a number of other classifiers commonly used in credit scoring, suggesting its potential for future use on highly imbalanced data sets.
This research paper investigates the influence of industry on electronic customer relationship management (e-CRM) performance. A case study approach with two cases was applied to evaluate the influence of e-CRM on customer behavioral and attitudinal loyalty along with customer pyramid. The cases covered two industries consisting of computer and automotive industries. For investigating customer behavioral loyalty and customer pyramid companies database were computed while for examining customer attitudinal loyalty a survey was conducted. The results show that e-CRM has significantly different impacts on customer behavioral and attitudinal loyalty and customer pyramid in two industries. This research provides new approach for organizations to evaluate their e-CRM performance and compare it with other companies in order to formulate stronger policies for customer relationship management.
There are no more papers matching your filters at the moment.