transformers in tabular tiny survey 2024.4.8

推荐阅读

TabLLM

pmlr2023,
Few-shot Classification of Tabular Data with Large Language Models

方法

使用把tabular数据序列化成文字的方法进行classification。
在这里插入图片描述使用的序列化方法有几个,有人工也有AI生成。
在这里插入图片描述

效果

做few shot learning的效果
在这里插入图片描述看上去一般。

TransTab

Learning Transferable Tabular Transformers Across Tables

方法

属于transfer learning的方法。对category、binary和numeric值进行embedding后再进行transformers最后进行classification。
在这里插入图片描述

使用场景

原文:
- S(1) Transfer learning. We collect data tables from multiple cancer trials for testing the efficacy
of the same drug on different patients. These tables were designed independently with overlapping
columns. How do we learn ML models for one trial by leveraging tables from all trials?
- S(2) Incremental learning. Additional columns might be added over time. For example, additional
features are collected across different trial phases. How do we update the ML models using tables
from all trial phases?
- S(3) Pretraining+Finetuning. The trial outcome label (e.g., mortality) might not be always available
from all table sources. Can we benefit pretraining on those tables without labels? How do we finetune
the model on the target table with labels?
- S(4) Zero-shot inference. We model the drug efficacy based on our trial records. The next step is to
conduct inference with the model to find patients that can benefit from the drug. However, patient
tables do not share the same columns as trial tables so direct inference is not possible.

效果

具体看原文吧,与当时的baseline比有提升。

MET

Masked Encoding for Tabular Data

tabtransformer

2020年,arxiv,TabTransformer: Tabular Data Modeling Using Contextual Embeddings

方法

transformer无监督训练,mlp监督训练。

原文

we introduce a pre-training procedure to train the Transformer layers using unlabeled data. This is followed by fine-tuning of the pre-trained Transformer layers along with the top MLP layer using the labeled data

效果

跟mlp

在这里插入图片描述

跟其他模型

在这里插入图片描述

tabnet

2020, arxiv,Google Cloud AI,Attentive Interpretable Tabular Learning, 封装的非常好,都可以当工具包使用了。

方法

跟transformer没关系的。
在这里插入图片描述feature selection用的是17年的某个选择模型,最后agg一下做predict。

相关推荐

  1. 【Leetcode】【2024048】1544. Make The String Great

    2024-04-10 10:42:03       35 阅读

最近更新

  1. docker php8.1+nginx base 镜像 dockerfile 配置

    2024-04-10 10:42:03       98 阅读
  2. Could not load dynamic library ‘cudart64_100.dll‘

    2024-04-10 10:42:03       106 阅读
  3. 在Django里面运行非项目文件

    2024-04-10 10:42:03       87 阅读
  4. Python语言-面向对象

    2024-04-10 10:42:03       96 阅读

热门阅读

  1. 未来AI技术创业机会

    2024-04-10 10:42:03       41 阅读
  2. CentOS 7详细介绍。

    2024-04-10 10:42:03       39 阅读
  3. 文心一言和GPT-4全面比较

    2024-04-10 10:42:03       40 阅读
  4. 三种芯片:ASIC、FPGA、SoC

    2024-04-10 10:42:03       38 阅读
  5. Vue将iconfont封装成全局组件,自定义类名

    2024-04-10 10:42:03       38 阅读
  6. List接口(1) | ArrayList

    2024-04-10 10:42:03       42 阅读
  7. 关于yolov8x-p2.yaml

    2024-04-10 10:42:03       38 阅读
  8. 链表,栈,队列的区别及其应用

    2024-04-10 10:42:03       34 阅读
  9. 如何生成 ssh 秘钥?

    2024-04-10 10:42:03       35 阅读
  10. 7-11完全二叉树的层序遍历

    2024-04-10 10:42:03       37 阅读