diff --git a/tokenise+train.py b/tokenise+train.py index 5cfa4c9..4c24966 100644 --- a/tokenise+train.py +++ b/tokenise+train.py @@ -48,7 +48,7 @@ def train(filepath: str, ouputdir: Path, blocksize: int, vocabsize: int, num_ste # data = TokenDataset(dat, tokenizer_file=tok, block_size=blocksize, from_cache=True) - ai.train(filepath, output_dir=str(ouputdir), line_by_line=False, from_cache=False, learning_rate=1e-3, batch_size=128, num_steps=num_steps, generate_every=1000, save_every=1000, num_workers=4) + ai.train(filepath, output_dir=str(ouputdir), line_by_line=False, from_cache=False, learning_rate=1e-3, batch_size=256, num_steps=num_steps, generate_every=1000, save_every=1000) return "Done!"