University of Missouri
Gliomas are the most common malignant primary brain tumors in adults and one of the deadliest types of cancer. There are many challenges in treatment and monitoring due to the genetic diversity and high intrinsic heterogeneity in appearance, shape, histology, and treatment response. Treatments include surgery, radiation, and systemic therapies, with magnetic resonance imaging (MRI) playing a key role in treatment planning and post-treatment longitudinal assessment. The 2024 Brain Tumor Segmentation (BraTS) challenge on post-treatment glioma MRI will provide a community standard and benchmark for state-of-the-art automated segmentation models based on the largest expert-annotated post-treatment glioma MRI dataset. Challenge competitors will develop automated segmentation models to predict four distinct tumor sub-regions consisting of enhancing tissue (ET), surrounding non-enhancing T2/fluid-attenuated inversion recovery (FLAIR) hyperintensity (SNFH), non-enhancing tumor core (NETC), and resection cavity (RC). Models will be evaluated on separate validation and test datasets using standardized performance metrics utilized across the BraTS 2024 cluster of challenges, including lesion-wise Dice Similarity Coefficient and Hausdorff Distance. Models developed during this challenge will advance the field of automated MRI segmentation and contribute to their integration into clinical practice, ultimately enhancing patient care.
A comprehensive survey details the evolution of Multi-Task Learning (MTL) from its origins in traditional machine learning to its current state, encompassing deep learning and the advent of pretrained foundation models. The paper categorizes methodologies, distinguishes MTL from related paradigms, and summarizes key advancements across each era, highlighting how large-scale models like Transformers are enabling more unified and versatile multi-task capabilities.
Exact plane gravitational radiation fields are presented within the framework of general relativity and their properties are described. The physics of nonlinear elliptically polarized plane gravitational waves is developed in close analogy with electromagnetic waves. The motion of free test particles in the dynamic gravitational fields of elliptically polarized plane waves is investigated. In particular, we demonstrate the cosmic jet property of these spacetimes, namely, most timelike geodesics tend to line up in the direction of wave propagation and produce a cosmic jet whose speed asymptotically approaches the speed of light.
Large language models (LLMs) have significantly transformed natural language understanding and generation, but they raise privacy concerns due to potential exposure of sensitive information. Studies have highlighted the risk of information leakage, where adversaries can extract sensitive information embedded in the prompts. In this work, we introduce a novel private prediction framework for generating high-quality synthetic text with strong privacy guarantees. Our approach leverages the Differential Privacy (DP) framework to ensure worst-case theoretical bounds on information leakage without requiring any fine-tuning of the underlying models. The proposed method performs inference on private records and aggregates the resulting per-token output distributions. This enables the generation of longer and coherent synthetic text while maintaining privacy guarantees. Additionally, we propose a simple blending operation that combines private and public inference to further enhance utility. Empirical evaluations demonstrate that our approach outperforms previous state-of-the-art methods on in-context-learning (ICL) tasks, making it a promising direction for privacy-preserving text generation while maintaining high utility. Our code is available at this https URL.
The prediction that black holes radiate due to quantum effects is often considered one of the most secure in quantum field theory in curved space-time. Yet this prediction rests on two dubious assumptions: that ordinary physics may be applied to vacuum fluctuations at energy scales increasing exponentially without bound; and that quantum-gravitational effects may be neglected. Various suggestions have been put forward to address these issues: that they might be explained away by lessons from sonic black hole models; that the prediction is indeed successfully reproduced by quantum gravity; that the success of the link provided by the prediction between black holes and thermodynamics justifies the prediction. This paper explains the nature of the difficulties, and reviews the proposals that have been put forward to deal with them. None of the proposals put forward can so far be considered to be really successful, and simple dimensional arguments show that quantum-gravitational effects might well alter the evaporation process outlined by Hawking. Thus a definitive theoretical treatment will require an understanding of quantum gravity in at least some regimes. Until then, no compelling theoretical case for or against radiation by black holes is likely to be made. The possibility that non-radiating "mini" black holes exist should be taken seriously; such holes could be part of the dark matter in the Universe. Attempts to place observational limits on the number of "mini" black holes (independent of the assumption that they radiate) would be most welcome.
Traditional Chinese Medicine (TCM), as an effective alternative medicine, has been receiving increasing attention. In recent years, the rapid development of large language models (LLMs) tailored for TCM has highlighted the urgent need for an objective and comprehensive evaluation framework to assess their performance on real-world tasks. However, existing evaluation datasets are limited in scope and primarily text-based, lacking a unified and standardized multimodal question-answering (QA) benchmark. To address this issue, we introduce TCM-Ladder, the first comprehensive multimodal QA dataset specifically designed for evaluating large TCM language models. The dataset covers multiple core disciplines of TCM, including fundamental theory, diagnostics, herbal formulas, internal medicine, surgery, pharmacognosy, and pediatrics. In addition to textual content, TCM-Ladder incorporates various modalities such as images and videos. The dataset was constructed using a combination of automated and manual filtering processes and comprises over 52,000 questions. These questions include single-choice, multiple-choice, fill-in-the-blank, diagnostic dialogue, and visual comprehension tasks. We trained a reasoning model on TCM-Ladder and conducted comparative experiments against nine state-of-the-art general domain and five leading TCM-specific LLMs to evaluate their performance on the dataset. Moreover, we propose Ladder-Score, an evaluation method specifically designed for TCM question answering that effectively assesses answer quality in terms of terminology usage and semantic expression. To the best of our knowledge, this is the first work to systematically evaluate mainstream general domain and TCM-specific LLMs on a unified multimodal benchmark. The datasets and leaderboard are publicly available at this https URL and will be continuously updated.
We report a new population of objects discovered using the data from the James Webb Space Telescope, which are characterized by their point-like morphology and narrow permitted emission lines. Our sample includes nine objects found in three JWST wide survey fields, which have z=3.624 to 5.378 and M_B ~ -18.3 to -21.4 mag. Their light distributions follow gaussian profiles, with the full-width-at-half-maximum (FWHM) values only 3.7%--34.6% larger than those of the point spread functions. Their sharpest FWHM sizes correspond to only 0.49 to 0.96 kpc. They have very strong [O III] and Halpha lines (median rest-frame equivalent widths of 1610.6 and 1374.0 A, respectively), and the line widths of the latter are only 150--360 km/s. Due to the limitation of the current data, the exact nature of this new population is still uncertain. The emission line diagnostics shows that at least one object is consistent with being AGN. On the other hand, the spectral energy distributions of most of the nine objects can be fitted reasonably by normal galaxy templates, which suggest that they could also be very young (median age of 120 Myrs), star-forming (median star formation rate of 1.8 Msun/yr) galaxies in the early formation stage (having acquired a median stellar mass of only 10^{8.4} Msun). If they are indeed star-forming galaxies, their gas-phase metallicities range from 12+log(O/H) = 7.8 to 8.3. It will be critical to understand this population by deeper medium-resolution spectroscopy in the future. If they are AGNs, they constitute a new kind of type 2 AGNs that are low-luminosity and almost "hostless". If they are star-forming galaxies, they also constitute a new kind whose early formation is likely a secular process starting from a very compact core, which is in line with the picture of a monolithic collapse.
Researchers from the University of Missouri and Duke University developed ODR+, an enhanced open-source Deep Research Agent, achieving 10% exact match accuracy on the challenging BrowseComp-Small benchmark. This work establishes a new open-source baseline and introduced BrowseComp-Small to make DRA evaluation more accessible for academic labs.
Flow matching (FM) models extend ODE sampler based diffusion models into a general framework, significantly reducing sampling steps through learned vector fields. However, the theoretical understanding of FM models, particularly how their sample trajectories interact with underlying data geometry, remains underexplored. A rigorous theoretical analysis of FM ODE is essential for sample quality, stability, and broader applicability. In this paper, we advance the theory of FM models through a comprehensive analysis of sample trajectories. Central to our theory is the discovery that the denoiser, a key component of FM models, guides ODE dynamics through attracting and absorbing behaviors that adapt to the data geometry. We identify and analyze the three stages of ODE evolution: in the initial and intermediate stages, trajectories move toward the mean and local clusters of the data. At the terminal stage, we rigorously establish the convergence of FM ODE under weak assumptions, addressing scenarios where the data lie on a low-dimensional submanifold-cases that previous results could not handle. Our terminal stage analysis offers insights into the memorization phenomenon and establishes equivariance properties of FM ODEs. These findings bridge critical gaps in understanding flow matching models, with practical implications for optimizing sampling strategies and architectures guided by the intrinsic geometry of data.
In the realm of aerial imaging, the ability to detect small objects is pivotal for a myriad of applications, encompassing environmental surveillance, urban design, and crisis management. Leveraging RetinaNet, this work unveils DDR-Net: a data-driven, deep-learning model devised to enhance the detection of diminutive objects. DDR-Net introduces novel, data-driven techniques to autonomously ascertain optimal feature maps and anchor estimations, cultivating a tailored and proficient training process while maintaining precision. Additionally, this paper presents an innovative sampling technique to bolster model efficacy under limited data training constraints. The model's enhanced detection capabilities support critical applications including wildlife and habitat monitoring, traffic flow optimization, and public safety improvements through accurate identification of small objects like vehicles and pedestrians. DDR-Net significantly reduces the cost and time required for data collection and training, offering efficient performance even with limited data. Empirical assessments over assorted aerial avian imagery datasets demonstrate that DDR-Net markedly surpasses RetinaNet and alternative contemporary models. These innovations advance current aerial image analysis technologies and promise wide-ranging impacts across multiple sectors including agriculture, security, and archaeology.
We introduce Advertisement Embedding Attacks (AEA), a new class of LLM security threats that stealthily inject promotional or malicious content into model outputs and AI agents. AEA operate through two low-cost vectors: (1) hijacking third-party service-distribution platforms to prepend adversarial prompts, and (2) publishing back-doored open-source checkpoints fine-tuned with attacker data. Unlike conventional attacks that degrade accuracy, AEA subvert information integrity, causing models to return covert ads, propaganda, or hate speech while appearing normal. We detail the attack pipeline, map five stakeholder victim groups, and present an initial prompt-based self-inspection defense that mitigates these injections without additional model retraining. Our findings reveal an urgent, under-addressed gap in LLM security and call for coordinated detection, auditing, and policy responses from the AI-safety community.
Gual'ero Piccinini from the University of Missouri empirically refutes the Classical Language of Thought hypothesis by demonstrating that biological neural hardware is incompatible with its digital architectural requirements. The research instead supports a Nonclassical Language of Thought, positing that neural representations can mirror natural language structures and are processed nondigitally within the brain.
Christian Heinicke and Friedrich W. Hehl present a comprehensive introduction to the Schwarzschild and Kerr metrics, detailing their mathematical properties and physical interpretations. These fundamental solutions to Einstein's field equations describe the gravitational fields of non-rotating and rotating black holes, predicting phenomena like event horizons, ergospheres, and frame-dragging that have since been observationally confirmed.
Researchers primarily at Kansas State University are developing an interoperable knowledge graph and ontology framework for sustainable wheat production, integrating diverse data sources from farm to table. This modular system aims to provide structured data that enables data-driven insights for optimizing nitrogen and disease management while explicitly linking to global sustainability goals.
We present a structural analysis of 521 galaxy candidates at 6.5 < z < 12.5, with SNR &gt; 10\sigma in the F444W filter, taken from the EPOCHS v1 sample, consisting of uniformly reduced deep JWST NIRCam data, covering the CEERS, JADES GOOD-S, NGDEEP, SMACS0723, GLASS and PEARLS surveys. We use standard software to fit single S\'ersic models to each galaxy in the rest-frame optical and extract their parametric structural parameters (S\'ersic index, half-light radius and axis-ratio), and \texttt{Morfometryka} to measure their non-parametric concentration and asymmetry parameters. We find a wide range of sizes for these early galaxies, but with a strong galaxy-size mass correlation up to z12z \sim 12 such that galaxy sizes continue to get progressively smaller in the high-redshift regime, following $R_{e} = 2.74 \pm 0.49 \left( 1 + z \right) ^{-0.79 \pm 0.08}$ kpc. Using non-parametric methods we find that galaxy merger fractions, classified through asymmetry parameters, at these redshifts remain consistent with those in literature, maintaining a value of fm0.12±0.07f_{m} \sim 0.12 \pm 0.07 showing little dependence with redshift when combined with literature at z &gt; 4. We find that galaxies which are smaller in size also appear rounder, with an excess of high axis-ratio objects. Finally, we artificially redshift a subsample of our objects to determine how robust the observational trends we see are, determining that observed trends are due to real evolutionary effects, rather than being a consequence of redshift effects.
In this paper, we propose Conceptual Codebook Learning (CoCoLe), a novel fine-tuning method for vision-language models (VLMs) to address the challenge of improving the generalization capability of VLMs while fine-tuning them on downstream tasks in a few-shot setting. We recognize that visual concepts, such as textures, shapes, and colors are naturally transferable across domains and play a crucial role in generalization tasks. Motivated by this interesting finding, we learn a conceptual codebook consisting of visual concepts as keys and conceptual prompts as values, which serves as a link between the image encoder's outputs and the text encoder's inputs. Specifically, for a given image, we leverage the codebook to identify the most relevant conceptual prompts associated with the class embeddings to perform the classification. Additionally, we incorporate a handcrafted concept cache as a regularization to alleviate the overfitting issues in low-shot scenarios. We observe that this conceptual codebook learning method is able to achieve enhanced alignment between visual and linguistic modalities. Extensive experimental results demonstrate that our CoCoLe method remarkably outperforms the existing state-of-the-art methods across various evaluation settings, including base-to-new generalization, cross-dataset evaluation, and domain generalization tasks. Detailed ablation studies further confirm the efficacy of each component in CoCoLe.
Recent significant advances in integrating multiple Large Language Model (LLM) systems have enabled Agentic Frameworks capable of performing complex tasks autonomously, including novel scientific research. We develop and demonstrate such a framework specifically for the inverse design of photonic metamaterials. When queried with a desired optical spectrum, the Agent autonomously proposes and develops a forward deep learning model, accesses external tools via APIs for tasks like simulation and optimization, utilizes memory, and generates a final design via a deep inverse method. The framework's effectiveness is demonstrated in its ability to automate, reason, plan, and adapt. Notably, the Agentic Framework possesses internal reflection and decision flexibility, permitting highly varied and potentially novel outputs.
We introduce RNA-FrameFlow, the first generative model for 3D RNA backbone design. We build upon SE(3) flow matching for protein backbone generation and establish protocols for data preparation and evaluation to address unique challenges posed by RNA modeling. We formulate RNA structures as a set of rigid-body frames and associated loss functions which account for larger, more conformationally flexible RNA backbones (13 atoms per nucleotide) vs. proteins (4 atoms per residue). Toward tackling the lack of diversity in 3D RNA datasets, we explore training with structural clustering and cropping augmentations. Additionally, we define a suite of evaluation metrics to measure whether the generated RNA structures are globally self-consistent (via inverse folding followed by forward folding) and locally recover RNA-specific structural descriptors. The most performant version of RNA-FrameFlow generates locally realistic RNA backbones of 40-150 nucleotides, over 40% of which pass our validity criteria as measured by a self-consistency TM-score >= 0.45, at which two RNAs have the same global fold. Open-source code: this https URL
78
Denoising diffusion probabilistic models (DDPMs) have pioneered new state-of-the-art results in disciplines such as computer vision and computational biology for diverse tasks ranging from text-guided image generation to structure-guided protein design. Along this latter line of research, methods have recently been proposed for generating 3D molecules using equivariant graph neural networks (GNNs) within a DDPM framework. However, such methods are unable to learn important geometric properties of 3D molecules, as they adopt molecule-agnostic and non-geometric GNNs as their 3D graph denoising networks, which notably hinders their ability to generate valid large 3D molecules. In this work, we address these gaps by introducing the Geometry-Complete Diffusion Model (GCDM) for 3D molecule generation, which outperforms existing 3D molecular diffusion models by significant margins across conditional and unconditional settings for the QM9 dataset and the larger GEOM-Drugs dataset, respectively, and generates more novel and unique unconditional 3D molecules for the QM9 dataset compared to previous methods. Importantly, we demonstrate that the geometry-complete denoising process of GCDM learned for 3D molecule generation enables the model to generate a significant proportion of valid and energetically-stable large molecules at the scale of GEOM-Drugs, whereas previous methods fail to do so with the features they learn. Additionally, we show that extensions of GCDM can not only effectively design 3D molecules for specific protein pockets but also that GCDM's geometric features can be repurposed to consistently optimize the geometry and chemical composition of existing 3D molecules for molecular stability and property specificity, demonstrating new versatility of molecular diffusion models. Our source code and data are freely available at this https URL.
188
The translation of AI-generated brain metastases (BM) segmentation into clinical practice relies heavily on diverse, high-quality annotated medical imaging datasets. The BraTS-METS 2023 challenge has gained momentum for testing and benchmarking algorithms using rigorously annotated internationally compiled real-world datasets. This study presents the results of the segmentation challenge and characterizes the challenging cases that impacted the performance of the winning algorithms. Untreated brain metastases on standard anatomic MRI sequences (T1, T2, FLAIR, T1PG) from eight contributed international datasets were annotated in stepwise method: published UNET algorithms, student, neuroradiologist, final approver neuroradiologist. Segmentations were ranked based on lesion-wise Dice and Hausdorff distance (HD95) scores. False positives (FP) and false negatives (FN) were rigorously penalized, receiving a score of 0 for Dice and a fixed penalty of 374 for HD95. Eight datasets comprising 1303 studies were annotated, with 402 studies (3076 lesions) released on Synapse as publicly available datasets to challenge competitors. Additionally, 31 studies (139 lesions) were held out for validation, and 59 studies (218 lesions) were used for testing. Segmentation accuracy was measured as rank across subjects, with the winning team achieving a LesionWise mean score of 7.9. Common errors among the leading teams included false negatives for small lesions and misregistration of masks in this http URL BraTS-METS 2023 challenge successfully curated well-annotated, diverse datasets and identified common errors, facilitating the translation of BM segmentation across varied clinical environments and providing personalized volumetric reports to patients undergoing BM treatment.
There are no more papers matching your filters at the moment.