Researchers from NARLabs, NTU, and NCU introduce the "chat vector" method, which transfers instruction-following and human alignment capabilities to large language models in new languages by applying a pre-computed weight difference to a continually pre-trained model. This approach effectively bypasses the complex Reinforcement Learning from Human Feedback (RLHF) process while largely preserving existing language and knowledge proficiencies.
View blogThe Event Horizon Telescope Collaboration conducted the first multi-epoch polarimetric imaging of M87* at event-horizon scales, observing a stable black hole shadow diameter while detecting substantial year-to-year variability in the ring's azimuthal brightness and linear polarization patterns, along with initial constraints on extended jet emission.
View blogA novel cross-layer parameter sharing strategy, BasisSharing, is introduced for compressing large language models by enabling weight matrices across different layers to share common basis vectors while retaining unique functionality via layer-specific coefficients. This method outperformed existing compression techniques, achieving up to 25% lower perplexity and 4% higher accuracy on downstream tasks, along with a 1.57x inference throughput improvement at 50% compression.
View blog