Update inference-cache-config/trn1/llama3.json
Browse files
inference-cache-config/trn1/llama3.json
CHANGED
|
@@ -125,7 +125,7 @@
|
|
| 125 |
{
|
| 126 |
"batch_size": 1,
|
| 127 |
"sequence_length": 16384,
|
| 128 |
-
"tensor_parallel_size":
|
| 129 |
"instance_type" : "trn1"
|
| 130 |
},
|
| 131 |
{
|
|
|
|
| 125 |
{
|
| 126 |
"batch_size": 1,
|
| 127 |
"sequence_length": 16384,
|
| 128 |
+
"tensor_parallel_size": 24,
|
| 129 |
"instance_type" : "trn1"
|
| 130 |
},
|
| 131 |
{
|