import os os.environ['CUDA_VISIBLE_DEVICES'] = '0,1' def test_llama3(): from swift.llm import infer_main, InferArguments infer_main( InferArguments( model='LLM-Research/Meta-Llama-3.1-8B-Instruct', max_batch_size=2, val_dataset='AI-ModelScope/alpaca-gpt4-data-en#2')) if __name__ == '__main__': test_llama3()