| | import configparser |
| | import logging |
| |
|
| | def getconfig(configfile_path:str): |
| | """ |
| | configfile_path: file path of .cfg file |
| | """ |
| |
|
| | config = configparser.ConfigParser() |
| |
|
| | try: |
| | config.read_file(open(configfile_path)) |
| | return config |
| | except: |
| | logging.warning("config file not found") |
| |
|
| |
|
| | |
| | def get_classifier_params(model_name): |
| | config = getconfig('paramconfig.cfg') |
| | params = {} |
| | params['model_name'] = config.get(model_name,'MODEL') |
| | params['split_by'] = config.get(model_name,'SPLIT_BY') |
| | params['split_length'] = int(config.get(model_name,'SPLIT_LENGTH')) |
| | params['split_overlap'] = int(config.get(model_name,'SPLIT_OVERLAP')) |
| | params['remove_punc'] = bool(int(config.get(model_name,'REMOVE_PUNC'))) |
| | params['split_respect_sentence_boundary'] = bool(int(config.get(model_name,'RESPECT_SENTENCE_BOUNDARY'))) |
| | params['threshold'] = float(config.get(model_name,'THRESHOLD')) |
| | params['top_n'] = int(config.get(model_name,'TOP_KEY')) |
| |
|
| | return params |
| |
|
| | |
| | model_dict = { |
| | 'Llama3.1-8B': 'meta-llama/Meta-Llama-3.1-8B-Instruct', |
| | 'Llama3.1-70B': 'meta-llama/Meta-Llama-3.1-70B-Instruct', |
| | 'Llama3.1-405B': 'meta-llama/Meta-Llama-3.1-405B-Instruct', |
| | 'Zephyr 7B β': 'HuggingFaceH4/zephyr-7b-beta', |
| | 'Mistral-7B': 'mistralai/Mistral-7B-Instruct-v0.2', |
| | 'Mixtral-8x7B': 'mistralai/Mixtral-8x7B-Instruct-v0.1', |
| | 'Llama3.2-3B':'meta-llama/Llama-3.2-3B-Instruct', |
| | 'Llama3.2-1B': 'meta-llama/Llama-3.2-1B-Instruct', |
| | 'Gemma2-2B': 'google/gemma-2-2b-it', |
| | 'Llama-3.1-8B': 'meta-llama/Llama-3.1-8B-Instruct' |
| | } |