Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Commit
·
6e6bac8
1
Parent(s):
6a53027
refactor: add logging for CUDA and PyTorch versions
Browse files- generate_summaries_uv.py +13 -0
generate_summaries_uv.py
CHANGED
|
@@ -36,6 +36,19 @@ logging.basicConfig(
|
|
| 36 |
)
|
| 37 |
logger = logging.getLogger(__name__)
|
| 38 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 39 |
|
| 40 |
def format_prompt(content: str, card_type: str, tokenizer) -> str:
|
| 41 |
"""Format content as a prompt for the model."""
|
|
|
|
| 36 |
)
|
| 37 |
logger = logging.getLogger(__name__)
|
| 38 |
|
| 39 |
+
# log cuda version
|
| 40 |
+
import torch
|
| 41 |
+
if torch.cuda.is_available():
|
| 42 |
+
logger.info(f"CUDA version: {torch.version.cuda}")
|
| 43 |
+
logger.info(f"CUDA device: {torch.cuda.get_device_name(0)}")
|
| 44 |
+
logger.info(f"CUDA capability: {torch.cuda.get_device_capability(0)}")
|
| 45 |
+
else:
|
| 46 |
+
logger.info("CUDA is not available")
|
| 47 |
+
|
| 48 |
+
# log torch version
|
| 49 |
+
logger.info(f"PyTorch version: {torch.__version__}")
|
| 50 |
+
logger.info(f"vLLM version: {vllm.__version__}")
|
| 51 |
+
|
| 52 |
|
| 53 |
def format_prompt(content: str, card_type: str, tokenizer) -> str:
|
| 54 |
"""Format content as a prompt for the model."""
|