Rubin-Wei/MLPMemory-Mistral-wikipedia
2B β’ Updated
β’ 1
input_ids listlengths 2.05k 2.05k | attention_mask listlengths 2.05k 2.05k | labels listlengths 2.05k 2.05k | dstore_range listlengths 2 2 |
|---|---|---|---|
[
1,
1183,
1744,
2554,
29478,
5792,
26078,
1070,
1161,
2016,
29515,
29473,
29508,
29491,
1188,
3421,
27678,
11429,
1155,
2218,
3494,
29493,
12568,
29473,
29518,
29491,
1188,
3421,
1133,
2400,
11429,
1155,
16572,
29493,
15568,
29473,
29538,
2949... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
1183,
1744,
2554,
29478,
5792,
26078,
1070,
1161,
2016,
29515,
29473,
29508,
29491,
1188,
3421,
27678,
11429,
1155,
2218,
3494,
29493,
12568,
29473,
29518,
29491,
1188,
3421,
1133,
2400,
11429,
1155,
16572,
29493,
15568,
29473,
29538,
2949... | [
0,
1023
] |
[
1,
1183,
1744,
2554,
29478,
5792,
26078,
1070,
1161,
2016,
29515,
29473,
29508,
29491,
1188,
3421,
27678,
11429,
1155,
2218,
3494,
29493,
12568,
29473,
29518,
29491,
1188,
3421,
1133,
2400,
11429,
1155,
16572,
29493,
15568,
29473,
29538,
2949... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... | [
1023,
2047
] |
[
930,
906,
954,
32586,
19011,
1430,
3271,
25664,
3521,
29473,
1011,
930,
906,
954,
32586,
28645,
29476,
7931,
5063,
29668,
3521,
29473,
32586,
1011,
930,
906,
947,
11302,
28962,
29497,
1149,
3521,
29473,
32586,
1011,
930,
906,
947,
1119,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... | [
2047,
3071
] |
[
1051,
3933,
29478,
13433,
2515,
1032,
22397,
2684,
1122,
6238,
18513,
22817,
1892,
4408,
29491,
3248,
1051,
3933,
29478,
2252,
3041,
1054,
4112,
10596,
1066,
1924,
1146,
1507,
1168,
6946,
3711,
29491,
1113,
2081,
29510,
29481,
1505,
19293,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... | [
3071,
4095
] |
[
29508,
29542,
29552,
29555,
1325,
11710,
1521,
29501,
12956,
8093,
1072,
25140,
3521,
1186,
1199,
1037,
18747,
1093,
7131,
29473,
29508,
29542,
29552,
29502,
1325,
5177,
2723,
24900,
3521,
1186,
1199,
1037,
21656,
1093,
7131,
29473,
29508,
29... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... | [
4095,
5119
] |
[
1050,
7130,
1213,
1761,
10189,
14155,
1066,
1312,
2873,
5099,
29491,
3677,
1447,
1184,
1137,
2328,
1040,
2517,
25565,
3546,
1050,
29473,
1,
8008,
26950,
2515,
1284,
1633,
6167,
1926,
29476,
6709,
1158,
1032,
4677,
29501,
29480,
1329,
23061,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... | [
5119,
6143
] |
[
2404,
25718,
1896,
4888,
2404,
29473,
29538,
29549,
29493,
29508,
29549,
29518,
2404,
29473,
29538,
29584,
29550,
2404,
1161,
29508,
3521,
1155,
4006,
1503,
1540,
29501,
4505,
29515,
6084,
29513,
6182,
29501,
4224,
29515,
29539,
2522,
1264,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... | [
6143,
7167
] |
[1233,3466,1228,2216,28081,29493,1930,29501,3690,2506,1072,1603,17327,1314,1065,1040,3627,6371,29491(...TRUNCATED) | [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1(...TRUNCATED) | [-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100(...TRUNCATED) | [
7167,
8191
] |
[1040,3576,5358,3712,29491,1,1328,4767,29493,6620,1054,29483,1059,28174,1206,1040,7160,20531,1065,91(...TRUNCATED) | [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1(...TRUNCATED) | [-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100(...TRUNCATED) | [
8191,
9215
] |
[1603,4994,1122,1040,2218,3494,7390,2317,1072,17700,15820,1489,1894,29491,1,1328,21512,1070,1040,171(...TRUNCATED) | [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1(...TRUNCATED) | [-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100,-100(...TRUNCATED) | [
9215,
10239
] |
This dataset is a preprocessed version of the English Wikipedia snapshot from December 2021, processed using the preprocess_dataset.py script provided in the repository below.
Paper: MLP Memory: A Retriever-Pretrained Memory for Large Language Models
Dataset Source: English Wikipedia (December 2021)
Tokenizer: Mistral-7B-v0.3
Two key preprocessing parameters used are:
block_size: 2048 stride: 1024The dataset is organized into train and test splits as follows:
βββ test
βββ train
βββ dataset_dict.json
βββ dstore_summary.json
The train/test split was created using the following code:
from datasets import load_dataset
dataset = load_dataset('json', data_files=dataset_path)
split_dataset = dataset['train'].train_test_split(
test_size=0.002,
seed=42,
shuffle=True
)
train_dataset = split_dataset['train']
test_dataset = split_dataset['test']
raw_datasets = DatasetDict({"train": train_dataset, "test": test_dataset})
@inproceedings{Wei2025MLPMA,
title={MLP Memory: A Retriever-Pretrained Memory for Large Language Models},
author={Rubin Wei and Jiaqi Cao and Jiarui Wang and Jushi Kai and Qipeng Guo and Bowen Zhou and Zhouhan Lin},
year={2025},
url={https://api.semanticscholar.org/CorpusID:281658735}
}
For questions and discussions, feel free to email: weirubinn@gmail.com