Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Table Representation Learning Workshop (TRL)

Exploration of autoregressive models for in-context learning on tabular data

Stefan Baur · Sohyeong Kim

Keywords: [ In-context learning ] [ Tabular data ] [ TabPFN ] [ Mamba ]


Abstract:

We explore different auto-regressive model architectures for in-context learning on tabular datasets trained in a similar manner to TabPFN.Namely, we compare transformer based models with a structured state-space model architecture (Mamba) and a hybrid architecture (Jamba), mixing transformer and Mamba layers.We find that auto-regressive transformer models perform similarly to the original TabPFN transformer architectures, albeit at the cost of a doubled context length.Mamba performs worse than similar sized transformer models, while hybrid models show promise in harnessing some advantages of state-space models such as supporting long input context length and fast inference.

Chat is not available.