mirror of
https://github.com/karpathy/nanochat.git
synced 2025-12-06 20:32:14 +00:00
fix merge conflict
This commit is contained in:
parent
5ca0950c9c
commit
e86f8fc030
|
|
@ -141,13 +141,9 @@ adamw_optimizer, muon_optimizer = optimizers
|
||||||
# Initialize the DataLoaders for train/val
|
# Initialize the DataLoaders for train/val
|
||||||
base_dir = get_base_dir()
|
base_dir = get_base_dir()
|
||||||
tokens_dir = os.path.join(base_dir, "tokenized_data")
|
tokens_dir = os.path.join(base_dir, "tokenized_data")
|
||||||
<<<<<<< HEAD
|
|
||||||
train_loader = tokenizing_distributed_data_loader(device_batch_size, max_seq_len, split="train")
|
|
||||||
build_val_loader = lambda: tokenizing_distributed_data_loader(device_batch_size_val, max_seq_len, split="val")
|
|
||||||
=======
|
|
||||||
train_loader = tokenizing_distributed_data_loader(device_batch_size, max_seq_len, split="train", device=device)
|
train_loader = tokenizing_distributed_data_loader(device_batch_size, max_seq_len, split="train", device=device)
|
||||||
build_val_loader = lambda: tokenizing_distributed_data_loader(device_batch_size, max_seq_len, split="val", device=device)
|
build_val_loader = lambda: tokenizing_distributed_data_loader(device_batch_size, max_seq_len, split="val", device=device)
|
||||||
>>>>>>> 722da4f... trying to add basic cpu support, will try mps too
|
|
||||||
x, y = next(train_loader) # kick off load of the very first batch of data
|
x, y = next(train_loader) # kick off load of the very first batch of data
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
# -----------------------------------------------------------------------------
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue
Block a user