Jeonbuk National University
The past decade has witnessed the rapid development and adoption of ML & DL methodologies in agricultural systems, showcased by great successes in agricultural applications. However, these conventional ML/DL models have certain limitations: they heavily rely on large, costly-to-acquire labeled datasets for training, require specialized expertise for development and maintenance, and are mostly tailored for specific tasks, thus lacking generalizability. Recently, large pre-trained models, also known as FMs, have demonstrated remarkable successes in language, vision, and decision-making tasks across various domains. These models are trained on a large amount of data from multiple domains and modalities. Once trained, they can accomplish versatile tasks with just minor fine-tuning and minimal task-specific labeled data. Despite their proven effectiveness and huge potential, there has been little exploration of applying FMs to agriculture AI. Thus, this study aims to explore the potential of FMs in the field of smart agriculture. In particular, conceptual tools and technical background are presented to help the understanding of the problem space and uncover new research directions. To this end, recent FMs in the general CS domain are reviewed, and the models are categorized into four categories: language FMs, vision FMs, multimodal FMs, and reinforcement learning FMs. Then, the steps of developing agriculture FMs (AFMs) are outlined and potential applications in smart agriculture are discussed. Moreover, challenges and risks associated with developing AFMs are discussed, including model training, validation, and deployment. In summary, the advancement of AI in agriculture is explored by introducing AFMs as a promising paradigm that can significantly mitigate the reliance on extensive labeled datasets and enhance the efficiency, effectiveness, and generalization of agricultural AI systems.
43
Just-in-Time software defect prediction (JIT-SDP) plays a critical role in prioritizing risky code changes during code review and continuous integration. However, existing datasets often suffer from noisy labels and low precision in identifying bug-inducing commits. To address this, we present ReDef (Revert-based Defect dataset), a high-confidence benchmark of function-level modifications curated from 22 large-scale C/C++ projects. Defective cases are anchored by revert commits, while clean cases are validated through post-hoc history checks. Ambiguous instances are conservatively filtered out via a GPT-assisted triage process involving multiple votes and audits. This pipeline yields 3,164 defective and 10,268 clean modifications, offering substantially more reliable labels than prior existing resources. Beyond dataset construction, we provide the first systematic evaluation of how pre-trained language models (PLMs) reason about code modifications -- specifically, which input encodings most effectively expose change information, and whether models genuinely capture edit semantics. We fine-tune CodeBERT, CodeT5+, and UniXcoder under five encoding strategies, and further probe their sensitivity through counterfactual perturbations that swap added/deleted blocks, invert diff polarity, or inject spurious markers. Our results show that compact diff-style encodings consistently outperform whole-function formats across all PLMs, with statistical tests confirming large, model-independent effects. However, under counterfactual tests, performance degrades little or not at all -- revealing that what appears to be robustness in fact reflects reliance on superficial cues rather than true semantic understanding. These findings indicate that, unlike in snapshot-based tasks, current PLMs remain limited in their ability to genuinely comprehend code modifications.
We show that a simple supersymmetric U(1)BLU(1)_{B-L} extension of the standard model can explain simultaneously the large electron neutrino asymmetry hinted by the recent EMPRESS data as well as the observed tiny baryon number asymmetry via the resonant leptogenesis mechanism. The condensation of BLB-L Higgs dominating the universe at its decay is the sole source for these generation processes. Here, the infrequent decays of the BLB-L Higgs to heavy right handed neutrinos and successive prompt decays of these right handed neutrinos around the electroweak phase transition produce the observed baryon number asymmetry, while the complete decay of the same BLB-L Higgs at a later epoch leads to a large lepton number asymmetry. The right amounts of both asymmetries are found to be obtained for the symmetry-breaking scale vϕ1010 GeVv_\phi \sim 10^{10}~{\rm GeV}. Moreover, in a close connection to the positivity of both asymmetries, seemingly only the normal mass hierarchy of light neutrino species works. Finally, the gravitational wave background from the topologically stable strong type-I cosmic strings, generated from the breaking of U(1)BLU(1)_{B-L} symmetry, can be within the reach of future experiments such as ultimate DECIGO.
We demonstrate electrically switchable, non-volatile dipoles in graphene/thin hBN/α\alpha-RuCl3_3 heterostructures, stabilized purely by interfacial charge transfer across an atomically thin dielectric barrier. This mechanism requires no sliding or twisting to explicitly break inversion symmetry and produces robust ferroelectric-like hysteresis loops that emerge prominently near 30~K. Systematic measurements under strong in-plane and out-of-plane magnetic fields reveal negligible effects on the hysteresis characteristics, confirming that the primary mechanism driving the dipole switching is electrostatic. Our findings establish a distinct and robust route to electrically tunable ferroelectric phenomena in van der Waals heterostructures, opening opportunities to explore the interplay between interfacial charge transfer and temperature-tuned barrier crossing of dipole states at the atomic scale.
Software Reliability Growth Models (SRGMs) are widely used to predict software reliability based on defect discovery data collected during testing or operational phases. However, their predictive accuracy often degrades in data-scarce environments, such as early-stage testing or safety-critical systems. Although cross-project transfer learning has been explored to mitigate this issue by leveraging data from past projects, its applicability remains limited due to the scarcity and confidentiality of real-world datasets. To overcome these limitations, we propose Deep Synthetic Cross-project SRGM (DSC-SRGM), a novel approach that integrates synthetic data generation with cross-project transfer learning. Synthetic datasets are generated using traditional SRGMs to preserve the statistical characteristics of real-world defect discovery trends. A cross-correlation-based clustering method is applied to identify synthetic datasets with patterns similar to the target project. These datasets are then used to train a deep learning model for reliability prediction. The proposed method is evaluated on 60 real-world datasets, and its performance is compared with both traditional SRGMs and cross-project deep learning models trained on real-world datasets. DSC-SRGM achieves up to 23.3% improvement in predictive accuracy over traditional SRGMs and 32.2% over cross-project deep learning models trained on real-world datasets. However, excessive use of synthetic data or a naive combination of synthetic and real-world data may degrade prediction performance, highlighting the importance of maintaining an appropriate data balance. These findings indicate that DSC-SRGM is a promising approach for software reliability prediction in data-scarce environments.
To facilitate the widespread adoption of renewable energy, dispatchable, zero-emission power sources are essential for grid stability. This work performs a comprehensive techno-economic analysis of a self-sustainable thermophotovoltaic (TPV) system, an architecture that integrates solar charging to function as a standalone power generation asset. Using theory-based models for air-bridge InGaAs and Si diode cells, our analysis reveals that while the system is not currently competitive from a pure levelized of storage cost (LCOS) perspective due to the high capital expenditure for thermal battery materials, its primary value lies in its competitive levelized cost of electricity (LCOE). The results demonstrate that the LCOE of this self-sustaining system can be competitive with conventional dispatchable generators, such as gas turbines. Furthermore, at scales exceeding the gigawatt-hour level, a Si-based system can also achieve an LCOE comparable to that of traditional gas-turbine power plants, despite having a lower conversion efficiency than its InGaAs counterpart. This highlights a practical engineering pathway for leveraging silicon's immense manufacturing scalability, offering a lower-risk route to deployment compared to III-V materials. Ultimately, this work establishes the self-sustainable TPV architecture as a compelling pathway toward providing grid-scale, on-demand, zero-emission power.
The growing demand for precision agriculture necessitates efficient and accurate crop-weed recognition and classification systems. Current datasets often lack the sample size, diversity, and hierarchical structure needed to develop robust deep learning models for discriminating crops and weeds in agricultural fields. Moreover, the similar external structure and phenomics of crops and weeds complicate recognition tasks. To address these issues, we present the CWD30 dataset, a large-scale, diverse, holistic, and hierarchical dataset tailored for crop-weed recognition tasks in precision agriculture. CWD30 comprises over 219,770 high-resolution images of 20 weed species and 10 crop species, encompassing various growth stages, multiple viewing angles, and environmental conditions. The images were collected from diverse agricultural fields across different geographic locations and seasons, ensuring a representative dataset. The dataset's hierarchical taxonomy enables fine-grained classification and facilitates the development of more accurate, robust, and generalizable deep learning models. We conduct extensive baseline experiments to validate the efficacy of the CWD30 dataset. Our experiments reveal that the dataset poses significant challenges due to intra-class variations, inter-class similarities, and data imbalance. Additionally, we demonstrate that minor training modifications like using CWD30 pretrained backbones can significantly enhance model performance and reduce convergence time, saving training resources on several downstream tasks. These challenges provide valuable insights and opportunities for future research in crop-weed recognition. We believe that the CWD30 dataset will serve as a benchmark for evaluating crop-weed recognition algorithms, promoting advancements in precision agriculture, and fostering collaboration among researchers in the field.
40
We performed a search for the KLπ0ννˉK_L \to \pi^{0} \nu \bar{\nu} decay using the data taken in 2021 at the J-PARC KOTO experiment. With newly installed counters and new analysis method, the expected background was suppressed to 0.252±0.055stat0.252\pm0.055_{\mathrm{stat}}0.067+0.052^{+0.052}_{-0.067}syst_{\mathrm{syst}}. With a single event sensitivity of $(9.33 \pm 0.06_{\rm stat} \pm 0.84_{\rm syst})\times 10^{-10}$, no events were observed in the signal region. An upper limit on the branching fraction for the decay was set to be 2.2×1092.2\times10^{-9} at the 90% confidence level (C.L.), which improved the previous upper limit from KOTO by a factor of 1.4. With the same data, a search for $K_L \to \pi^{0} X^{0}wasalsoperformed,where was also performed, where X^{0}$ is an invisible boson with a mass ranging from 1 MeV/c2c^{2} to 260 MeV/c2c^{2}. For X0X^{0} with a mass of 135 MeV/c2c^{2}, an upper limit on the branching fraction of $K_L \to \pi^{0} X^{0}wassettobe was set to be 1.6\times10^{-9}$ at the 90% C.L.
We present updated observational constraints on the spatially flat ϕ\phiCDM model, where dark energy is described by a minimally coupled scalar field ϕ\phi with an inverse power-law potential V=V0ϕαV=V_0 \phi^{-\alpha}. Using Planck 2018 CMB temperature, polarization (P18), and lensing power spectra (lensing), along with a compilation of non-CMB data including baryon acoustic oscillation, type Ia supernova, Hubble parameter, and growth rate measurements, we constrain ϕ\phiCDM and ϕ\phiCDM+ALA_L models where ALA_L is the CMB lensing consistency parameter. The scalar field parameter α\alpha, which governs dark energy dynamics, is more tightly constrained by non-CMB data than by CMB data alone. For the full dataset, we obtain α=0.055±0.041\alpha = 0.055 \pm 0.041 in the ϕ\phiCDM model and α=0.095±0.056\alpha = 0.095 \pm 0.056 in the ϕ\phiCDM+ALA_L model, mildly favoring evolving dark energy over a cosmological constant by 1.3σ1.3\sigma and 1.7σ1.7\sigma. The Hubble constant is H0=67.550.46+0.53H_0=67.55_{-0.46}^{+0.53} km s1^{-1} Mpc1^{-1} in the ϕ\phiCDM model, consistent with median statistics and some local determinations, but in tension with other local determinations. The constraints for matter density and clustering amplitude (Ωm=0.3096±0.0055\Omega_m = 0.3096 \pm 0.0055, σ8=0.80130.0067+0.0077\sigma_8 = 0.8013_{-0.0067}^{+0.0077}) of the flat ϕ\phiCDM model statistically agree with Λ\LambdaCDM model values. Allowing ALA_L to vary reduces tensions between CMB and non-CMB data, although we find AL=1.105±0.037A_L = 1.105 \pm 0.037, 2.8σ2.8\sigma higher than unity, consistent with the excess smoothing seen in Planck data. Model comparison using AIC and DIC indicates that the ϕ\phiCDM model provides a fit comparable to Λ\LambdaCDM, with the ϕ\phiCDM+ALA_L slightly preferred. Overall, while the Λ\LambdaCDM model remains an excellent fit, current data leave open the possibility of mildly evolving quintessence-like dynamical dark energy.
Human health can be critically affected by cardiovascular diseases, such as hypertension, arrhythmias, and stroke. Heart rate and blood pressure are important biometric information for the monitoring of cardiovascular system and early diagnosis of cardiovascular diseases. Existing methods for estimating the heart rate are based on electrocardiography and photoplethyomography, which require contacting the sensor to the skin surface. Moreover, catheter and cuff-based methods for measuring blood pressure cause inconvenience and have limited applicability. Therefore, in this thesis, we propose a vision-based method for estimating the heart rate and blood pressure. This thesis proposes a 2-stage deep learning framework consisting of a dual remote photoplethysmography network (DRP-Net) and bounded blood pressure network (BBP-Net). In the first stage, DRP-Net infers remote photoplethysmography (rPPG) signals for the acral and facial regions, and these phase-shifted rPPG signals are utilized to estimate the heart rate. In the second stage, BBP-Net integrates temporal features and analyzes phase discrepancy between the acral and facial rPPG signals to estimate SBP and DBP values. To improve the accuracy of estimating the heart rate, we employed a data augmentation method based on a frame interpolation model. Moreover, we designed BBP-Net to infer blood pressure within a predefined range by incorporating a scaled sigmoid function. Our method resulted in estimating the heart rate with the mean absolute error (MAE) of 1.78 BPM, reducing the MAE by 34.31 % compared to the recent method, on the MMSE-HR dataset. The MAE for estimating the systolic blood pressure (SBP) and diastolic blood pressure (DBP) were 10.19 mmHg and 7.09 mmHg. On the V4V dataset, the MAE for the heart rate, SBP, and DBP were 3.83 BPM, 13.64 mmHg, and 9.4 mmHg, respectively.
35
Semiconductor wafer defect classification is critical for ensuring high precision and yield in manufacturing. Traditional CNN-based models often struggle with class imbalances and recognition of the multiple overlapping defect types in wafer maps. To address these challenges, we propose ViT-Tiny, a lightweight Vision Transformer (ViT) framework optimized for wafer defect classification. Trained on the WM-38k dataset. ViT-Tiny outperforms its ViT-Base counterpart and state-of-the-art (SOTA) models, such as MSF-Trans and CNN-based architectures. Through extensive ablation studies, we determine that a patch size of 16 provides optimal performance. ViT-Tiny achieves an F1-score of 98.4%, surpassing MSF-Trans by 2.94% in four-defect classification, improving recall by 2.86% in two-defect classification, and increasing precision by 3.13% in three-defect classification. Additionally, it demonstrates enhanced robustness under limited labeled data conditions, making it a computationally efficient and reliable solution for real-world semiconductor defect detection.
Fault localization, the process of identifying the software components responsible for failures, is essential but often time-consuming. Recent advances in Large Language Models (LLMs) have enabled fault localization without extensive defect datasets or model fine-tuning. However, existing LLM-based methods rely only on general LLM capabilities and lack integration of project-specific knowledge, resulting in limited effectiveness, especially for complex software. We introduce MemFL, a novel approach that enhances LLM-based fault localization by integrating project-specific knowledge via external memory. This memory includes static summaries of the project and dynamic, iterative debugging insights gathered from previous attempts. By leveraging external memory, MemFL simplifies debugging into three streamlined steps, significantly improving efficiency and accuracy. Iterative refinement through dynamic memory further enhances reasoning quality over time. Evaluated on the Defects4J benchmark, MemFL using GPT-4o-mini localized 12.7% more bugs than current LLM-based methods, achieving this improvement with just 21% of the execution time (17.4 seconds per bug) and 33% of the API cost (0.0033 dollars per bug). On complex projects, MemFL's advantage increased to 27.6%. Additionally, MemFL with GPT-4.1-mini outperformed existing methods by 24.4%, requiring only 24.7 seconds and 0.0094 dollars per bug. MemFL thus demonstrates significant improvements by effectively incorporating project-specific knowledge into LLM-based fault localization, delivering high accuracy with reduced time and cost.
Given a signed social graph, how can we learn appropriate node representations to infer the signs of missing edges? Signed social graphs have received considerable attention to model trust relationships. Learning node representations is crucial to effectively analyze graph data, and various techniques such as network embedding and graph convolutional network (GCN) have been proposed for learning signed graphs. However, traditional network embedding methods are not end-to-end for a specific task such as link sign prediction, and GCN-based methods suffer from a performance degradation problem when their depth increases. In this paper, we propose Signed Graph Diffusion Network (SGDNet), a novel graph neural network that achieves end-to-end node representation learning for link sign prediction in signed social graphs. We propose a random walk technique specially designed for signed graphs so that SGDNet effectively diffuses hidden node features. Through extensive experiments, we demonstrate that SGDNet outperforms state-of-the-art models in terms of link sign prediction accuracy.
Transition metal dichalcogenides provide a platform for exploring spin-valley physics, offering a promising approach to electric-field-driven spin control for low-power spintronic and quantum devices. Here, we demonstrate electric-field-induced spin splitting in the Q and Q' valleys of multilayer n-type WSe2 using quantum-point-contact spectroscopy. Systematic modulations in four distinct conductance quantization steps, providing direct evidence of spin-valley-layer coupling-driven spin-resolved density of states, were achieved by tuning the out-of-plane gate voltage. Notably, the electric-field-induced spin splitting significantly dominated the magnetic-field-induced Zeeman effect (e.g., ~ 6 meV for a displacement field change of ~ 0.04 V/nm vs. ~ 1 meV for a magnetic field of 9 T), demonstrating a powerful, non-magnetic manipulation of spin states. This ability to manipulate spin states by gate voltage is crucial for advancing next-generation low-power spintronic and quantum information technologies.
Federated learning (FL) often suffers from performance degradation due to key challenges such as data heterogeneity and communication constraints. To address these limitations, we present a novel FL framework called FedWSQ, which integrates weight standardization (WS) and the proposed distribution-aware non-uniform quantization (DANUQ). WS enhances FL performance by filtering out biased components in local updates during training, thereby improving the robustness of the model against data heterogeneity and unstable client participation. In addition, DANUQ minimizes quantization errors by leveraging the statistical properties of local model updates. As a result, FedWSQ significantly reduces communication overhead while maintaining superior model accuracy. Extensive experiments on FL benchmark datasets demonstrate that FedWSQ consistently outperforms existing FL methods across various challenging FL settings, including extreme data heterogeneity and ultra-low-bit communication scenarios.
This paper presents a novel and efficient image enhancement method based on pigment representation. Unlike conventional methods where the color transformation is restricted to pre-defined color spaces like RGB, our method dynamically adapts to input content by transforming RGB colors into a high-dimensional feature space referred to as \textit{pigments}. The proposed pigment representation offers adaptability and expressiveness, achieving superior image enhancement performance. The proposed method involves transforming input RGB colors into high-dimensional pigments, which are then reprojected individually and blended to refine and aggregate the information of the colors in pigment spaces. Those pigments are then transformed back into RGB colors to generate an enhanced output image. The transformation and reprojection parameters are derived from the visual encoder which adaptively estimates such parameters based on the content in the input image. Extensive experimental results demonstrate the superior performance of the proposed method over state-of-the-art methods in image enhancement tasks, including image retouching and tone mapping, while maintaining relatively low computational complexity and small model size.
Accurate prediction of surface energies and stabilities is essential for materials design, yet first-principles calculations remain computationally expensive and most existing interatomic potentials are trained only on bulk systems. Here, we demonstrate that fine-tuning foundation machine learning potentials (MLPs) significantly improves both computational efficiency and predictive accuracy for surface modeling. While existing universal interatomic potentials (UIPs) have been solely trained and validated on bulk datasets, we extend their applicability to complex and scientifically significant unary, binary, and ternary surface systems. We systematically compare models trained from scratch, zero-shot inference, conventional fine-tuning, and multi-head fine-tuning approach that enhances transferability and mitigates catastrophic forgetting. Fine-tuning consistently reduces prediction errors with orders-of-magnitude fewer training configurations, and multi-head fine-tuning delivers robust and generalizable predictions even for materials beyond the initial training domain. These findings offer practical guidance for leveraging pre-trained MLPs to accelerate surface modeling and highlight a scalable path toward data-efficient, next-generation atomic-scale simulations in computational materials science.
Artificial intelligence has advanced global weather forecasting, outperforming traditional numerical models in both accuracy and computational efficiency. Nevertheless, extending predictions beyond subseasonal timescales requires the development of deep learning (DL)-based ocean-atmosphere coupled models that can realistically simulate complex oceanic responses to atmospheric forcing. This study presents KIST-Ocean, a DL-based global three-dimensional ocean general circulation model using a U-shaped visual attention adversarial network architecture. KIST-Ocean integrates partial convolution, adversarial training, and transfer learning to address coastal complexity and predictive distribution drift in auto-regressive models. Comprehensive evaluations confirmed the model's robust ocean predictive skill and efficiency. Moreover, it accurately captures realistic ocean response, such as Kelvin and Rossby wave propagation in the tropical Pacific, and vertical motions induced by cyclonic and anticyclonic wind stress, demonstrating its ability to represent key ocean-atmosphere coupling mechanisms underlying climate phenomena, including the El Nino-Southern Oscillation. These findings reinforce confidence in DL-based global weather and climate models and their extending DL-based approaches to broader Earth system modeling, offering potential for enhancing climate prediction capabilities.
In this paper, we consider the following attraction repulsion chemotaxis model with nonlinear signal term: \begin{align*} &u_{t}=\nabla \cdot(\nabla u-\xi_{1} u \nabla v +\xi_{2} u \nabla w), \quad &0=\Delta v -\lambda_{1}v +f_{1}(u), \quad &0=\Delta w -\lambda_{2}w +f_{2}(u), \quad x \in \mathbb{R}^{N}, t>0, \end{align*} where ξ1,ξ2,λ1,λ2\xi_{1},\xi_{2},\lambda_{1},\lambda_{2} are for some positive constants, and \begin{equation*} f_{1} \in C^{1}([0,\infty)) \; \text{satisfying} \; 0 \leqslant f_{1}(s) \leqslant c_{1}s^{l}, \; \forall s \geqslant 0 \ \text{and} \ l> 0, \end{equation*} \begin{equation*} f_{2} \in C^{1}([0,\infty))\; \text{satisfying} \; 0 \leqslant f_{2}(s) \leqslant c_{2}s^{m}, \; \forall s \geqslant 0 \ \text{and} \ m> 0. \end{equation*} We will show that this problem has a unique global bounded solution when $ l>\frac{2}{N}, l
The KOTO II experiment is proposed to measure the branching ratio of the decay KLπ0ννˉK_L\to\pi^0\nu\bar{\nu} at J-PARC. With a beamline to extract long-lived neutral kaons at 5 degrees from a production target, the single event sensitivity of the decay is 8.5×10138.5\times 10^{-13}, which is much smaller than the Standard Model prediction 3×10113\times 10^{-11}. This allows searches for new physics beyond the Standard Model and the first discovery of the decay with a significance exceeding 5σ5\sigma. As the only experiment proposed in the world dedicated to rare kaon decays, KOTO II will be indispensable in the quest for a complete understanding of flavor dynamics in the quark sector. Moreover, by combining efforts from the kaon community worldwide, we plan to develop the KOTO II detector further and expand the physics reach of the experiment to include measurements of the branching ratio of the KLπ0+K_L\to\pi^0\ell^+\ell^- decays, studies of other KLK_L decays, and searches for dark photons, axions, and axion-like particles. KOTO II will therefore obtain a comprehensive understanding of KLK_L decays, providing further constraints on new physics scenarios with existing K+K^+ results.
There are no more papers matching your filters at the moment.