Superposition-Transformer
Publica novel architecture that leverages Autoencoders to superimpose the hidden representations of a base model and a fine-tuned model within a shared parameter space. Using B-spline-based blending coefficients and autoencoders that adaptively reconstruct the original hidden states based on the input data distribution.
Creat:2024-10-22T23:22:11
Update:2025-01-08T16:46:59
132
Stars
88
Stars Increase