Commit 58bebe8a authored by Vít Starý Novotný's avatar Vít Starý Novotný
Browse files

Add learning rate (5e-5, the default value) to config

parent 176559bb
Loading
Loading
Loading
Loading
Loading
+1 −0
Original line number Diff line number Diff line
@@ -94,6 +94,7 @@ root_path = /nlp/projekty/ahisto/public_html/named-entity-search/results/
base_model = xlm-roberta-base
batch_size = 4
gradient_accumulation_steps = 4
learning_rate = 0.00005
log_every_n_steps = 100
evaluate_every_n_steps = 1000
save_every_n_steps = 1000
+2 −0
Original line number Diff line number Diff line
@@ -40,6 +40,7 @@ class NerModel:
    BASE_MODEL = CONFIG['base_model']
    BATCH_SIZE = CONFIG.getint('batch_size')
    GRADIENT_ACCUMULATION_STEPS = CONFIG.getint('gradient_accumulation_steps')
    LEARNING_RATE = CONFIG.getfloat('learning_rate')
    EVAL_STEPS = CONFIG.getint('evaluate_every_n_steps')
    SAVE_STEPS = CONFIG.getint('save_every_n_steps')
    LOGGING_STEPS = CONFIG.getint('log_every_n_steps')
@@ -248,6 +249,7 @@ class NerModel:
            save_total_limit=cls.STOPPING_PATIENCE + 1,
            load_best_model_at_end=True,
            metric_for_best_model='eval_{ner_objective}_{ner_evaluator}',
            learning_rate=cls.LEARNING_RATE,
            logging_strategy='steps',
            logging_steps=cls.LOGGING_STEPS,
            do_train=True,