Seattle University
Reliable Text-to-SQL (RTS) introduces a framework to enhance the dependability of natural language interfaces for databases by integrating an adaptive abstention mechanism and human feedback. It uses a Branching Point Predictor (BPP) with conformal prediction to provide probabilistic guarantees on schema linking accuracy, allowing smaller LLMs to achieve performance comparable to much larger models.
We report measurements of the short-range forces between two macroscopic gold-coated plates using a torsion pendulum. The force is measured for separations between 0.7 μ\mum and 7 μ\mum, and is well described by a combination of the Casimir force, including the finite-temperature correction, and an electrostatic force due to patch potentials on the plate surfaces. We use our data to place constraints on the Yukawa-type "new" forces predicted by theories with extra dimensions. We establish a new best bound for force ranges 0.4 μ\mum to 4 μ\mum, and, for forces mediated by gauge bosons propagating in (4+n)(4+n) dimensions and coupling to the baryon number, extract a (4+n)(4+n)-dimensional Planck scale lower limit of M>70M_*>70 TeV.
Image inpainting consists of filling holes or missing parts of an image. Inpainting face images with symmetric characteristics is more challenging than inpainting a natural scene. None of the powerful existing models can fill out the missing parts of an image while considering the symmetry and homogeneity of the picture. Moreover, the metrics that assess a repaired face image quality cannot measure the preservation of symmetry between the rebuilt and existing parts of a face. In this paper, we intend to solve the symmetry problem in the face inpainting task by using multiple discriminators that check each face organ's reality separately and a transformer-based network. We also propose "symmetry concentration score" as a new metric for measuring the symmetry of a repaired face image. The quantitative and qualitative results show the superiority of our proposed method compared to some of the recently proposed algorithms in terms of the reality, symmetry, and homogeneity of the inpainted parts.
15
The nonlinear Schrodinger equation is well known as a universal equation in the study of wave motion. In the context of wave motion at the free surface of an incompressible fluid, the equation accurately predicts the evolution of modulated wave trains with low to moderate wave steepness. While there is an abundance of studies investigating the reconstruction of the surface profile η\eta, and the fidelity of such profiles provided by the nonlinear Schrodinger equation as predictions of real surface water waves, very few works have focused on the associated flow field in the fluid. In the current work, it is shown that the velocity potential ϕ\phi can be reconstructed in a similar way as the free-surface profile. This observation opens up a range of potential applications since the nonlinear Schrodinger equation features fairly simple closed-form solutions and can be solved numerically with comparatively little effort. In particular, it is shown that particle trajectories in the fluid can be described with relative ease not only in the context of the nonlinear Schrodinger equation, but also in higher-order models such as the Dysthe equation, and in models incorporating certain types of viscous effects.
The Center for Exascale Monte Carlo Neutron Transport is developing Monte Carlo / Dynamic Code (MC/DC) as a portable Monte Carlo neutron transport package for rapid numerical methods exploration on CPU- and GPU-based high-performance computers. In this paper, we describe MC/DC's current event-based GPU algorithm as well as the just-in-time (JIT) compilation scheme we use to enable GPU operability on Nvidia and AMD GPUs from MC/DC's Python source. To analyze performance, we conduct runtime tests of the C5G7 k-eigenvalue benchmark problem and a continuous-energy infinite pin cell on Nvidia Tesla V100 GPU, AMD MI250X GPU, and the AMD MI300A APU and make comparison to a dual-socket Intel Xeon Sapphire Rapid CPU node. We found that for the multi-group C5G7 benchmark problem, we respectively see a 15×\times, 0.7×\times, 12×\times speedup on a V100, MI250X, and MI300A over 112 Intel Xeon CPU cores. For the continuous-energy infinite pin-cell benchmark, we found speedups of 5×\times, 3×\times, 4×\times on a V100, MI250X, and MI300A, respectively, over the same CPU node.
Let MM be a closed triangulable manifold, and let Δ\Delta be a triangulation of MM. What is the smallest number of vertices that Δ\Delta can have? How big or small can the number of edges of Δ\Delta be as a function of the number of vertices? More generally, what are the possible face numbers (ff-numbers, for short) that Δ\Delta can have? In other words, what restrictions does the topology of MM place on the possible ff-numbers of triangulations of MM? To make things even more interesting, we can add some combinatorial conditions on the triangulations we are considering (e.g., flagness, balancedness, etc.) and ask what additional restrictions these combinatorial conditions impose. While only a few theorems in this area of combinatorics were known a couple of decades ago, in the last ten years or so, the field simply exploded with new results and ideas. Thus we feel that a survey paper is long overdue. As new theorems are being proved while we are typing this chapter, and as we have only a limited number of pages, we apologize in advance to our friends and colleagues, some of whose results will not get mentioned here.
In today's digital landscape, video content dominates internet traffic, underscoring the need for efficient video processing to support seamless live streaming experiences on platforms like YouTube Live, Twitch, and Facebook Live. This paper introduces a comprehensive framework designed to optimize video transcoding parameters, with a specific focus on preset and bitrate selection to minimize distortion while respecting constraints on bitrate and transcoding time. The framework comprises three main steps: feature extraction, prediction, and optimization. It leverages extracted features to predict transcoding time and rate-distortion, employing both supervised and unsupervised methods. By utilizing integer linear programming, it identifies the optimal sequence of presets and bitrates for video segments, ensuring real-time application feasibility under set constraints. The results demonstrate the framework's effectiveness in enhancing video quality for live streaming, maintaining high standards of video delivery while managing computational resources efficiently. This optimization approach meets the evolving demands of video delivery by offering a solution for real-time transcoding optimization. Evaluation using the User Generated Content dataset showed an average PSNR improvement of 1.5 dB over the default Twitch configuration, highlighting significant PSNR gains. Additionally, subsequent experiments demonstrated a BD-rate reduction of -49.60%, reinforcing the framework's superior performance over Twitch's default configuration.
Despite being an integral tool for finding health-related information online, YouTube has faced criticism for disseminating COVID-19 misinformation globally to its users. Yet, prior audit studies have predominantly investigated YouTube within the Global North contexts, often overlooking the Global South. To address this gap, we conducted a comprehensive 10-day geolocation-based audit on YouTube to compare the prevalence of COVID-19 misinformation in search results between the United States (US) and South Africa (SA), the countries heavily affected by the pandemic in the Global North and the Global South, respectively. For each country, we selected 3 geolocations and placed sock-puppets, or bots emulating "real" users, that collected search results for 48 search queries sorted by 4 search filters for 10 days, yielding a dataset of 915K results. We found that 31.55% of the top-10 search results contained COVID-19 misinformation. Among the top-10 search results, bots in SA faced significantly more misinformative search results than their US counterparts. Overall, our study highlights the contrasting algorithmic behaviors of YouTube search between two countries, underscoring the need for the platform to regulate algorithmic behavior consistently across different regions of the Globe.
Recently, many attempts have been made to construct a transformer base U-shaped architecture, and new methods have been proposed that outperformed CNN-based rivals. However, serious problems such as blockiness and cropped edges in predicted masks remain because of transformers' patch partitioning operations. In this work, we propose a new U-shaped architecture for medical image segmentation with the help of the newly introduced focal modulation mechanism. The proposed architecture has asymmetric depths for the encoder and decoder. Due to the ability of the focal module to aggregate local and global features, our model could simultaneously benefit the wide receptive field of transformers and local viewing of CNNs. This helps the proposed method balance the local and global feature usage to outperform one of the most powerful transformer-based U-shaped models called Swin-UNet. We achieved a 1.68% higher DICE score and a 0.89 better HD metric on the Synapse dataset. Also, with extremely limited data, we had a 4.25% higher DICE score on the NeoPolyp dataset. Our implementations are available at: this https URL
40
Accurate and early detection of breast cancer is essential for successful treatment. This paper introduces a novel deep-learning approach for improved breast cancer classification in histopathological images, a crucial step in diagnosis. Our method hinges on the Dense Residual Dual-Shuffle Attention Network (DRDA-Net), inspired by ShuffleNet's efficient architecture. DRDA-Net achieves exceptional accuracy across various magnification levels on the BreaKHis dataset, a breast cancer histopathology analysis benchmark. However, for real-world deployment, computational efficiency is paramount. We integrate a pre-trained MobileNet model renowned for its lightweight design to address computational. MobileNet ensures fast execution even on devices with limited resources without sacrificing performance. This combined approach offers a promising solution for accurate breast cancer diagnosis, paving the way for faster and more accessible screening procedures.
A previously unknown form of compromising emanations has been discovered. LED status indicators on data communication equipment, under certain conditions, are shown to carry a modulated optical signal that is significantly correlated with information being processed by the device. Physical access is not required; the attacker gains access to all data going through the device, including plaintext in the case of data encryption systems. Experiments show that it is possible to intercept data under realistic conditions at a considerable distance. Many different sorts of devices, including modems and Internet Protocol routers, were found to be vulnerable. A taxonomy of compromising optical emanations is developed, and design changes are described that will successfully block this kind of "Optical TEMPEST" attack.
Data science for social justice (DS4SJ) is data-scientific work that supports the liberation of oppressed and marginalized people. By nature, this work lies at the intersection of technical scholarship and activist practice. We discuss this growing efforts in DS4SJ within the broad mathematics community. We begin by defining terms and offering a series of guiding principles for engaging in critical data science work, providing examples of how these principles play out in practice. We then highlight the roles that DS4SJ can play in the scholarship and pedagogy of practicing mathematicians. We focus in particular on the engagement of early-career mathematicians in DS4SJ, which we illustrate through a series of four personal vignettes. While the primary aim of DS4SJ is to achieve impact for marginalized communities, we also argue that engagement with DS4SJ can benefit the entire mathematical ecosystem, including researchers, instructors, students, departments, institutes, and professional societies. We close with reflections on how these various actors can support ongoing efforts in data science for social justice.
Not all nations on earth have previously been surveyed accurately enough to know for certain which peak is the national highpoint, the highest peak in the country. Knowledge of these peaks is important for understanding the physical geography of these countries in terms of natural resource availability, watershed management, and tourism potential. For this study, ground surveys were conducted between 2018-2025 with modern professional surveying equipment, including differential GPS units and Abney levels, to accurately determine the national highpoints in five African and Asian countries where uncertainty existed. New national highpoints were determined for Saudi Arabia (Jabal Ferwa), Uzbekistan (Alpomish), Gambia (Sare Firasu Hill), Guinea-Bissau (Mt Ronde), and Togo (Mt Atilakoutse). Elevations were measured with sub-meter vertical accuracy for candidate peaks in Saudi Arabia, Gambia, Guinea-Bissau, and Togo. Relative elevations were measured between contender peaks in Uzbekistan with sufficient accuracy to determine the highpoint.
Pseudodiagrams are knot or link diagrams where some of the crossing information is missing. Pseudoknots are equivalence classes of pseudodiagrams, where equivalence is generated by a natural set of Reidemeister moves. In this paper, we introduce a Gauss-diagrammatic theory for pseudoknots which gives rise to the notion of a virtual pseudoknot. We provide new, easily computable isotopy and homotopy invariants for classical and virtual pseudodiagrams. We also give tables of unknotting numbers for homotopically trivial pseudoknots and homotopy classes of homotopically nontrivial pseudoknots. Since pseudoknots are closely related to singular knots, this work also has implications for the classification of classical and virtual singular knots.
Research on optical TEMPEST has moved forward since 2002 when the first pair of papers on the subject emerged independently and from widely separated locations in the world within a week of each other. Since that time, vulnerabilities have evolved along with systems, and several new threat vectors have consequently appeared. Although the supply chain ecosystem of Ethernet has reduced the vulnerability of billions of devices through use of standardised PHY solutions, other recent trends including the Internet of Things (IoT) in both industrial settings and the general population, High Frequency Trading (HFT) in the financial sector, the European General Data Protection Regulation (GDPR), and inexpensive drones have made it relevant again for consideration in the design of new products for privacy. One of the general principles of security is that vulnerabilities, once fixed, sometimes do not stay that way.
The cubic-vortical Whitham equation is a model for wave motion on a vertically sheared current of constant vorticity in a shallow inviscid fluid. It generalizes the classical Whitham equation by allowing constant vorticity and by adding a cubic nonlinear term. The inclusion of this extra nonlinear term allows the equation to admit periodic, traveling-wave solutions with larger amplitude than the Whitham equation. Increasing vorticity leads to solutions with larger amplitude as well. The stability of these solutions is examined numerically. All moderate- and large-amplitude solutions, regardless of wavelength, are found to be unstable. A formula for a stability cutoff as a function of vorticity and wavelength for small-amplitude solutions is presented. In the case with zero vorticity, small-amplitude solutions are unstable with respect to the modulational instability if kh > 1.252, where k is the wavenumber and h is the mean fluid depth.
There is a growing demand for transparency in search engines to understand how search results are curated and to enhance users' trust. Prior research has introduced search result explanations with a focus on how to explain, assuming explanations are beneficial. Our study takes a step back to examine if search explanations are needed and when they are likely to provide benefits. Additionally, we summarize key characteristics of helpful explanations and share users' perspectives on explanation features provided by Google and Bing. Interviews with non-technical individuals reveal that users do not always seek or understand search explanations and mostly desire them for complex and critical tasks. They find Google's search explanations too obvious but appreciate the ability to contest search results. Based on our findings, we offer design recommendations for search engines and explanations to help users better evaluate search results and enhance their search experience.
The relation between the pion's quark distribution function, q(x)q(x), its light-front wave function, and the elastic charge form factor, F(Δ2)F(\Delta^2) is explored. The square of the leading-twist pion wave function at a special probe scale, ζH\zeta_H, is determined using models and Poincare covariance from realistic results for q(x)q(x). This wave function is then used to compute form factors with the result that the Drell-Yan-West and quark counting relationships are not satisfied. A new relationship between q(x)q(x) and F(Δ2)F(\Delta^2) is proposed.
Finding a software engineering approach that allows for portability, rapid development, and open collaboration for high-performance computing on GPUs and CPUs is a challenge. We implement a portability scheme using the Numba compiler for Python in Monte Carlo / Dynamic Code (MC/DC), a new neutron transport application for rapidly developing Monte Carlo. Using this scheme, we have built MC/DC as an application that can run as a pure Python, compiled CPU, or compiled GPU solver. In GPU mode, we use Numba paired with an asynchronous GPU scheduler called Harmonize to increase GPU performance. We present performance results (including weak scaling up to 256 nodes) for a time-dependent problem on both CPUs and GPUs and compare favorably to a production C++ code.
The enhancement of image luminosity is especially critical in endoscopic images. Underexposed endoscopic images often suffer from reduced contrast and uneven brightness, significantly impacting diagnostic accuracy and treatment planning. Internal body imaging is challenging due to uneven lighting and shadowy regions. Enhancing such images is essential since precise image interpretation is crucial for patient outcomes. In this paper, we introduce BrightVAE, an architecture based on the hierarchical Vector Quantized Variational Autoencoder (hierarchical VQ-VAE) tailored explicitly for enhancing luminosity in low-light endoscopic images. Our architecture is meticulously designed to tackle the unique challenges inherent in endoscopic imaging, such as significant variations in illumination and obscured details due to poor lighting conditions. The proposed model emphasizes advanced feature extraction from three distinct viewpoints-incorporating various receptive fields, skip connections, and feature attentions to robustly enhance image quality and support more accurate medical diagnoses. Through rigorous experimental analysis, we demonstrate the effectiveness of these techniques in enhancing low-light endoscopic images. To evaluate the performance of our architecture, we employ three widely recognized metrics-SSIM, PSNR, and LPIPS-specifically on Endo4IE dataset, which consists of endoscopic images. We evaluated our method using the Endo4IE dataset, which consists exclusively of endoscopic images, and showed significant advancements over the state-of-the-art methods for enhancing luminosity in endoscopic imaging.
There are no more papers matching your filters at the moment.