DATASET = {
const tokenizer = GPT2Tokenizer;
const [nl] = tokenizer.encode('\n');
const tokens = tokenizer.encode(TEXT);
const ret = [];
for (let i=0, n=tokens.length; i<n+CONTEXT_SIZE+1; ++i) {
const context = tokens.slice(i, i+CONTEXT_SIZE);
const input = tokenizer.decode(context);
const output = tokens[i+CONTEXT_SIZE] === nl;
ret.push({ input, output });
}
return ret;
}