Loading ahisto_named_entity_search/default.ini +1 −0 Original line number Diff line number Diff line Loading @@ -94,6 +94,7 @@ root_path = /nlp/projekty/ahisto/public_html/named-entity-search/results/ base_model = xlm-roberta-base batch_size = 4 gradient_accumulation_steps = 4 learning_rate = 0.00005 log_every_n_steps = 100 evaluate_every_n_steps = 1000 save_every_n_steps = 1000 Loading ahisto_named_entity_search/recognition/model.py +2 −0 Original line number Diff line number Diff line Loading @@ -40,6 +40,7 @@ class NerModel: BASE_MODEL = CONFIG['base_model'] BATCH_SIZE = CONFIG.getint('batch_size') GRADIENT_ACCUMULATION_STEPS = CONFIG.getint('gradient_accumulation_steps') LEARNING_RATE = CONFIG.getfloat('learning_rate') EVAL_STEPS = CONFIG.getint('evaluate_every_n_steps') SAVE_STEPS = CONFIG.getint('save_every_n_steps') LOGGING_STEPS = CONFIG.getint('log_every_n_steps') Loading Loading @@ -248,6 +249,7 @@ class NerModel: save_total_limit=cls.STOPPING_PATIENCE + 1, load_best_model_at_end=True, metric_for_best_model='eval_{ner_objective}_{ner_evaluator}', learning_rate=cls.LEARNING_RATE, logging_strategy='steps', logging_steps=cls.LOGGING_STEPS, do_train=True, Loading Loading
ahisto_named_entity_search/default.ini +1 −0 Original line number Diff line number Diff line Loading @@ -94,6 +94,7 @@ root_path = /nlp/projekty/ahisto/public_html/named-entity-search/results/ base_model = xlm-roberta-base batch_size = 4 gradient_accumulation_steps = 4 learning_rate = 0.00005 log_every_n_steps = 100 evaluate_every_n_steps = 1000 save_every_n_steps = 1000 Loading
ahisto_named_entity_search/recognition/model.py +2 −0 Original line number Diff line number Diff line Loading @@ -40,6 +40,7 @@ class NerModel: BASE_MODEL = CONFIG['base_model'] BATCH_SIZE = CONFIG.getint('batch_size') GRADIENT_ACCUMULATION_STEPS = CONFIG.getint('gradient_accumulation_steps') LEARNING_RATE = CONFIG.getfloat('learning_rate') EVAL_STEPS = CONFIG.getint('evaluate_every_n_steps') SAVE_STEPS = CONFIG.getint('save_every_n_steps') LOGGING_STEPS = CONFIG.getint('log_every_n_steps') Loading Loading @@ -248,6 +249,7 @@ class NerModel: save_total_limit=cls.STOPPING_PATIENCE + 1, load_best_model_at_end=True, metric_for_best_model='eval_{ner_objective}_{ner_evaluator}', learning_rate=cls.LEARNING_RATE, logging_strategy='steps', logging_steps=cls.LOGGING_STEPS, do_train=True, Loading