SCITIX (SGP) TECH PTE. LTD.
The Chain-of-Alpha framework automates the discovery of interpretable, formulaic alpha factors for quantitative trading by employing a dual-chain architecture driven by Large Language Models. This system consistently outperformed both traditional and prior LLM-based methods on China A-share market data, yielding higher annualized returns (e.g., 0.1324 on CSI 500) and information ratios (e.g., 1.4178 on CSI 500) while demonstrating superior efficiency.
Mask-Enhanced Autoregressive Prediction (MEAP) improves the in-context retrieval and long-context reasoning of decoder-only Large Language Models by integrating masked tokens into the standard next-token prediction objective. This method achieves significant data efficiency, outperforming standard next-token prediction by up to 3x on retrieval tasks, while maintaining computational efficiency and reducing contextual hallucinations.
18
The ongoing evolution of language models has led to the development of large-scale architectures that demonstrate exceptional performance across a wide range of tasks. However, these models come with significant computational and energy demands, as well as potential privacy implications. In this context, Small Reasoning Language Models (SRLMs) with approximately 0.5 billion parameters present a compelling alternative due to their remarkable computational efficiency and cost-effectiveness, particularly in resource-constrained environments. Despite these advantages, the limited capacity of 0.5 billion parameter models poses challenges in handling complex tasks such as mathematical reasoning. This research investigates various training strategies, including supervised fine-tuning (SFT), knowledge distillation (KD), and reinforcement learning (RL), as well as their hybrid implementations, to enhance the performance of 0.5B SRLMs. We analyze effective methodologies to bridge the performance gap between SRLMS and larger models and present insights into optimal training pipelines tailored for these smaller architectures. Through extensive experimental validation and analysis, our work aims to provide actionable recommendations for maximizing the reasoning capabilities of 0.5B models.
The HeteroSpec framework, developed by researchers from Peking University and SCITIX, improves large language model inference by dynamically managing speculative decoding resources based on contextual predictability. It achieves an average 4.24x decoding speedup over vanilla autoregressive decoding and reduces target model verification tokens by up to 22.79% compared to state-of-the-art methods like EAGLE-3.
Large Language Models (LLMs) have demonstrated remarkable performance across a wide range of tasks by understanding input information and predicting corresponding outputs. However, the internal mechanisms by which LLMs comprehend input and make effective predictions remain poorly understood. In this paper, we explore the working mechanism of LLMs in information processing from the perspective of Information Bottleneck Theory. We propose a non-training construction strategy to define a task space and identify the following key findings: (1) LLMs compress input information into specific task spaces (e.g., sentiment space, topic space) to facilitate task understanding; (2) they then extract and utilize relevant information from the task space at critical moments to generate accurate predictions. Based on these insights, we introduce two novel approaches: an Information Compression-based Context Learning (IC-ICL) and a Task-Space-guided Fine-Tuning (TS-FT). IC-ICL enhances reasoning performance and inference efficiency by compressing retrieved example information into the task space. TS-FT employs a space-guided loss to fine-tune LLMs, encouraging the learning of more effective compression and selection mechanisms. Experiments across multiple datasets validate the effectiveness of task space construction. Additionally, IC-ICL not only improves performance but also accelerates inference speed by over 40\%, while TS-FT achieves superior results with a minimal strategy adjustment.
There are no more papers matching your filters at the moment.