代码拉取完成,页面将自动刷新
from src.utils import init_experiment
from src.slu.baseline_loader import get_dataloader
from src.slu.baseline_model import ConceptTagger
from src.slu.baseline_trainer import BaselineTrainer
from config import get_params
import numpy as np
from tqdm import tqdm
def run_baseline(params):
# initialize experiment
logger = init_experiment(params, logger_filename=params.logger_filename)
# get dataloader
dataloader_tr, dataloader_val, dataloader_test, vocab = get_dataloader(params.tgt_dm, params.batch_size, params.n_samples)
# build model
concept_tagger = ConceptTagger(params, vocab)
concept_tagger.cuda()
# build trainer
baseline_trainer = BaselineTrainer(params, concept_tagger)
for e in range(params.epoch):
logger.info("============== epoch {} ==============".format(e+1))
loss_list = []
pbar = tqdm(enumerate(dataloader_tr), total=len(dataloader_tr))
for i, (X, lengths, y, y_dm) in pbar:
X, lengths = X.cuda(), lengths.cuda()
loss = baseline_trainer.train_step(X, lengths, y, y_dm)
loss_list.append(loss)
pbar.set_description("(Epoch {}) LOSS:{:.4f}".format((e+1), np.mean(loss_list)))
logger.info("Finish training epoch {}. LOSS:{:.4f}".format((e+1), np.mean(loss_list)))
logger.info("============== Evaluate Epoch {} ==============".format(e+1))
bin_f1_score, f1_score, stop_training_flag = baseline_trainer.evaluate(e, dataloader_val, istestset=False)
logger.info("Eval on dev set. Bin-F1: {:.4f}. Slot-F1: {:.4f}.".format(bin_f1_score, f1_score))
bin_f1_score, f1_score, _ = baseline_trainer.evaluate(e, dataloader_test, istestset=True)
logger.info("Eval on test set. Bin-F1: {:.4f}. Slot-F1: {:.4f}.".format(bin_f1_score, f1_score))
if stop_training_flag == True:
break
if __name__ == "__main__":
params = get_params()
run_baseline(params)
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。