File size: 4,209 Bytes
cb2428f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
from typing import Any, List

from swift.llm import MODEL_MAPPING, TEMPLATE_MAPPING, ModelType, TemplateType
from swift.utils import is_megatron_available


def get_url_suffix(model_id):
    if ':' in model_id:
        return model_id.split(':')[0]
    return model_id


def get_cache_mapping(fpath):
    with open(fpath, 'r', encoding='utf-8') as f:
        text = f.read()
    idx = text.find('| Model ID |')
    text = text[idx:]
    text_list = text.split('\n')[2:]
    cache_mapping = {}
    for text in text_list:
        if not text:
            continue
        items = text.split('|')
        if len(items) < 6:
            break
        cache_mapping[items[1]] = items[5]
    return cache_mapping


def get_model_info_table():
    fpaths = ['docs/source/Instruction/支持的模型和数据集.md', 'docs/source_en/Instruction/Supported-models-and-datasets.md']
    cache_mapping = get_cache_mapping(fpaths[0])
    end_words = [['### 多模态大模型', '## 数据集'], ['### Multimodal large models', '## Datasets']]
    result = [
        '| Model ID | Model Type | Default Template | '
        'Requires | Support Megatron | Tags | HF Model ID |\n'
        '| -------- | -----------| ---------------- | '
        '-------- | ---------------- | ---- | ----------- |\n'
    ] * 2
    res_llm: List[Any] = []
    res_mllm: List[Any] = []
    mg_count = 0
    for template in TemplateType.get_template_name_list():
        assert template in TEMPLATE_MAPPING

    for model_type in ModelType.get_model_name_list():
        model_meta = MODEL_MAPPING[model_type]
        template = model_meta.template
        for group in model_meta.model_groups:
            for model in group.models:
                ms_model_id = model.ms_model_id
                hf_model_id = model.hf_model_id
                if ms_model_id:
                    ms_model_id = f'[{ms_model_id}](https://modelscope.cn/models/{get_url_suffix(ms_model_id)})'
                else:
                    ms_model_id = '-'
                if hf_model_id:
                    hf_model_id = f'[{hf_model_id}](https://huggingface.co/{get_url_suffix(hf_model_id)})'
                else:
                    hf_model_id = '-'
                tags = ', '.join(group.tags or model_meta.tags) or '-'
                requires = ', '.join(group.requires or model_meta.requires) or '-'
                if is_megatron_available():
                    from swift.megatron import model
                    support_megatron = getattr(model_meta, 'support_megatron', False)
                    for word in ['gptq', 'awq', 'bnb', 'aqlm', 'int', 'nf4', 'fp8']:
                        if word in ms_model_id.lower():
                            support_megatron = False
                            break
                    support_megatron = '&#x2714;' if support_megatron else '&#x2718;'
                else:
                    support_megatron = cache_mapping.get(ms_model_id, '&#x2718;')
                if support_megatron == '&#x2714;':
                    mg_count += 1
                r = f'|{ms_model_id}|{model_type}|{template}|{requires}|{support_megatron}|{tags}|{hf_model_id}|\n'
                if model_meta.is_multimodal:
                    res_mllm.append(r)
                else:
                    res_llm.append(r)
    print(f'LLM总数: {len(res_llm)}, MLLM总数: {len(res_mllm)}, Megatron支持模型: {mg_count}')
    text = ['', '']  # llm, mllm
    for i, res in enumerate([res_llm, res_mllm]):
        for r in res:
            text[i] += r
        result[i] += text[i]

    for i, fpath in enumerate(fpaths):
        with open(fpath, 'r', encoding='utf-8') as f:
            text = f.read()
        llm_start_idx = text.find('| Model ID |')
        mllm_start_idx = text[llm_start_idx + 1:].find('| Model ID |') + llm_start_idx + 1
        llm_end_idx = text.find(end_words[i][0])
        mllm_end_idx = text.find(end_words[i][1])
        output = text[:llm_start_idx] + result[0] + '\n\n' + text[llm_end_idx:mllm_start_idx] + result[
            1] + '\n\n' + text[mllm_end_idx:]
        with open(fpath, 'w', encoding='utf-8') as f:
            f.write(output)


if __name__ == '__main__':
    get_model_info_table()