Jul 24, 2023
Řečník · 0 sledujících
Řečník · 0 sledujících
Řečník · 0 sledujících
Efficiently capturing the long-range patterns in sequential data sources salient to a given task—such as classification—poses a fundamental challenge. Popular approaches in the space tradeoff between the memory burden of remembering the entire past, as in transformers, the computational burden of complicated sequential dependencies in the memory computation, as in recurrent neural networks, or the parameter burden of convolutional networks with many or large filters. We instead take inspiration from wavelet-based multiresolution analysis to define a new building block for sequence modeling, which we call a MultiresLayer. The key component of our model is a *multiresolution convolution*, capturing multiscale trends in the input sequence. Our MultiresConv is defined by *shared* filters across a dilated causal convolution tree. Thus our MultiresNet garners the computational advantages of convolutional networks and the principled theoretical motivation of wavelet decompositions. Our MultiresLayer is simple to implement, requires significantly fewer parameters, and maintains at most a 𝒪(Nlog N) memory footprint for a length N sequence. Yet, MultiresNet yields state-of-the-art sequence classification performance on a number of benchmark datasets (Sequential CIFAR-10, ListOps, PTB-XL).Efficiently capturing the long-range patterns in sequential data sources salient to a given task—such as classification—poses a fundamental challenge. Popular approaches in the space tradeoff between the memory burden of remembering the entire past, as in transformers, the computational burden of complicated sequential dependencies in the memory computation, as in recurrent neural networks, or the parameter burden of convolutional networks with many or large filters. We instead take inspiration…
Professional recording and live streaming, delivered globally.
Presentations on similar topic, category or speaker
Yu Yang, …
Eunji Kim, …
Jacob Imola, …