Skip to content

Instantly share code, notes, and snippets.

@hamelsmu
Created May 14, 2019 21:09
Show Gist options
  • Save hamelsmu/0aad67c7719a52b9868c5c2e59af0dda to your computer and use it in GitHub Desktop.
Save hamelsmu/0aad67c7719a52b9868c5c2e59af0dda to your computer and use it in GitHub Desktop.
For troubleshooting wandb
from fastai.text import TextLMDataBunch as lmdb
from fastai.text.transform import Tokenizer
import pandas as pd
from pathlib import Path
small_df = pd.read_csv('https://storage.googleapis.com/issue_label_bot/pre_processed_data/processed_part0000.csv').head(1000)
stokenizer = Tokenizer(pre_rules=[pass_through], n_cpus=30)
spath = Path('lang_model_test/')
sdata_lm = lmdb.from_df(path=spath,
train_df=small_df,
valid_df=small_df,
text_cols='text',
tokenizer=stokenizer)
slearn = language_model_learner(data=sdata_lm,
arch=AWD_LSTM,
drop_mult=.5,
pretrained=False)
wandb.init()
escb = EarlyStoppingCallback(learn=slearn, patience=5)
smcb = SaveModelCallback(learn=slearn)
rpcb = ReduceLROnPlateauCallback(learn=slearn, patience=3)
sgcb = ShowGraph(learn=slearn)
wandcb = WandbCallback(learn=slearn, log='all', save_model=True, monitor='valid_loss')
scallbacks = [escb, smcb, rpcb, sgcb, wandcb]
slearn.fit_one_cycle(cyc_len=1,
max_lr=1e-2,
tot_epochs=10,
callbacks=scallbacks)
@borisdayma
Copy link

Thanks I just created wandb/wandb#355

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment