t4-full / README.md
jpgard's picture
Update README.md
bc354d7 verified
---
license: other
license_name: other
license_link: >-
https://huggingface.co/datasets/approximatelabs/tablib-v1-sample/blob/main/README.md
task_categories:
- tabular-classification
- tabular-regression
language:
- en
pretty_name: T4 (The Tremendous TabLib Trawl)
size_categories:
- 1B<n<10B
---
The Tremendous TabLib Trawl (T4) is a dataset for training tabular foundation models.
The dataset is described in detail in our paper, ["Large Scale Transfer Learning for Tabular Data via Language Modeling."](https://arxiv.org/abs/2406.12031)
The paper also includes a datasheet for this dataset.
T4 consists of a set of Parquet files (described below). For examples and infrastructure showing how to train a lannguage model
on T4, see our open-source Python library, [rtfm](https://github.com/mlfoundations/rtfm), which was used to train TabuLa-8B on T4.
# Files and Directory Structure
The T4 dataset contains approximately 3.1M tables. Each table is a separate Parquet file, named according to the `content_hash` of the dataset in TabLib.
The dataset is stored in "chunk" subdirectores, which represent batches of tables from the preprocessing phase.
Each chunk directory (e.g. `chunk-0000`) is stored as a single .zip file; unzip these files to access the underlying Parquet files.
The dataset occupies a total of 219GB compressed (1.34TB uncompressed) on disk.
# License and Acceptable Use
We release this dataset under the same license as the original corpuse from which it was derived, TabLib.
**By using this dataset, you are acknowledging that you have permission to access the TabLib dataset,
and you agree to abide by the terms of use and license of TabLib.**
TabLib can be accessed on [HF Datasets](https://huggingface.co/datasets/approximatelabs/tablib-v1-full), and you can read more about TabLib in the associated [paper](https://arxiv.org/abs/2310.07875) and [blog post](https://www.approximatelabs.com/blog/tablib).
We claim no affiliation with the original creators of TabLib, and this dataset release is not associated with Approximate Labs
(but we are grateful to the original TabLib authors for their contributions to the research community and for releasing TabLib).