Upload test file from inferences on LLM

#33
by Nivose - opened
data/report_SSG_inference_question-answering_privateCloud_35348ab0-1abf-4839-9f2c-33c5e9b10fe4.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"b502f765-2c20-4f4a-9265-5126c5b4eba0",
4
+ "reportDatetime":"2024-09-30T09:10:41",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":104,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"llama2-13b",
21
+ "parametersNumber":13.0,
22
+ "framework":"vllm",
23
+ "quantization":"q16"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":3225,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":88326,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"Simple question answer with 104 questions from simple ones to contextualised one. Dataset : 92 simple questions (Average : 11 Tokens) and 12 contextualized questions (On average question of 15 Tokens and contexts of 160 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0048764027775579,
52
+ "measurementDuration":34.46843481063843,
53
+ "measurementDateTime":"2024-09-30T09:10:41"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-116-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.30359268188477
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.8582,
88
+ "longitude":2.3387
89
+ },
90
+ "quality":"high"
91
+ }
data/report_SSG_inference_question-answering_privateCloud_372309f6-196b-4286-98a2-67ddb959ebe5.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"a3342146-0d82-4501-acf4-47cb7d9fda17",
4
+ "reportDatetime":"2024-10-22T05:44:53",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":104,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"phi3med-14b",
21
+ "parametersNumber":14.0,
22
+ "framework":"vllm",
23
+ "quantization":"q16"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":3533,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":202614,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"Simple question answer with 104 questions from simple ones to contextualised one. Dataset : 92 simple questions (Average : 11 Tokens) and 12 contextualized questions (On average question of 15 Tokens and contexts of 160 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0111332190261356,
52
+ "measurementDuration":84.0584135055542,
53
+ "measurementDateTime":"2024-10-22T05:44:53"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-116-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.30360412597656
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.8582,
88
+ "longitude":2.3387
89
+ },
90
+ "quality":"high"
91
+ }
data/report_SSG_inference_question-answering_privateCloud_3adf12f2-31a3-454a-a9d7-6e1710a412f2.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"35348ab0-1abf-4839-9f2c-33c5e9b10fe4",
4
+ "reportDatetime":"2024-09-30T08:28:17",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":104,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"llama2-13b",
21
+ "parametersNumber":13.0,
22
+ "framework":"vllm",
23
+ "quantization":"q16"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":3225,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":39539,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"Simple question answer with 104 questions from simple ones to contextualised one. Dataset : 92 simple questions (Average : 11 Tokens) and 12 contextualized questions (On average question of 15 Tokens and contexts of 160 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0016147225036887,
52
+ "measurementDuration":11.497305870056152,
53
+ "measurementDateTime":"2024-09-30T08:28:17"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-116-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.30359268188477
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.8582,
88
+ "longitude":2.3387
89
+ },
90
+ "quality":"high"
91
+ }
data/report_SSG_inference_question-answering_privateCloud_3d11ef38-e5f9-43f0-ac25-fcdd61115133.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"1493516d-828f-496f-84e1-2eb5b24845ff",
4
+ "reportDatetime":"2024-11-07T09:45:29",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":104,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"meta-llama-3-13b-instruct-gguf",
21
+ "parametersNumber":13.0,
22
+ "framework":"llamacpp",
23
+ "quantization":"q6k"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":5930,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":28163,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"Simple question answer with 104 questions from simple ones to contextualised one. Dataset : 92 simple questions (Average : 11 Tokens) and 12 contextualized questions (On average question of 15 Tokens and contexts of 160 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0010697891100869,
52
+ "measurementDuration":15.013213634490969,
53
+ "measurementDateTime":"2024-11-07T09:45:29"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-124-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.30359649658203
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.8582,
88
+ "longitude":2.3387
89
+ },
90
+ "quality":"high"
91
+ }
data/report_SSG_inference_question-answering_privateCloud_66b9b016-e026-4d02-ae96-b83e978897c7.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"3d11ef38-e5f9-43f0-ac25-fcdd61115133",
4
+ "reportDatetime":"2024-11-06T10:15:23",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":104,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"meta-llama-3-13b-instruct-gguf",
21
+ "parametersNumber":13.0,
22
+ "framework":"llamacpp",
23
+ "quantization":"8bit0"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":5930,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":28559,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"Simple question answer with 104 questions from simple ones to contextualised one. Dataset : 92 simple questions (Average : 11 Tokens) and 12 contextualized questions (On average question of 15 Tokens and contexts of 160 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0007014954565571,
52
+ "measurementDuration":10.449934482574465,
53
+ "measurementDateTime":"2024-11-06T10:15:23"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-124-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.30358123779297
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.8582,
88
+ "longitude":2.3387
89
+ },
90
+ "quality":"high"
91
+ }
data/report_SSG_inference_question-answering_privateCloud_70d6ebc5-7244-49c4-9603-6522d78db536.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"372309f6-196b-4286-98a2-67ddb959ebe5",
4
+ "reportDatetime":"2024-10-16T07:26:12",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":104,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"llama3-8b",
21
+ "parametersNumber":8.0,
22
+ "framework":"vllm",
23
+ "quantization":"q16"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":2987,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":63066,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"Simple question answer with 104 questions from simple ones to contextualised one. Dataset : 92 simple questions (Average : 11 Tokens) and 12 contextualized questions (On average question of 15 Tokens and contexts of 160 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0006877453643107,
52
+ "measurementDuration":5.442811965942383,
53
+ "measurementDateTime":"2024-10-16T07:26:12"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-116-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.3035888671875
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.8582,
88
+ "longitude":2.3387
89
+ },
90
+ "quality":"high"
91
+ }
data/report_SSG_inference_question-answering_privateCloud_9237143f-fd40-45e6-bf72-51d45f5401fa.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"70d6ebc5-7244-49c4-9603-6522d78db536",
4
+ "reportDatetime":"2024-10-10T09:20:16",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":104,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"mistral-7b",
21
+ "parametersNumber":7.0,
22
+ "framework":"vllm",
23
+ "quantization":"q16"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":3159,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":78793,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"Simple question answer with 104 questions from simple ones to contextualised one. Dataset : 92 simple questions (Average : 11 Tokens) and 12 contextualized questions (On average question of 15 Tokens and contexts of 160 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0005677864944745,
52
+ "measurementDuration":4.518536329269409,
53
+ "measurementDateTime":"2024-10-10T09:20:16"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-116-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.3035888671875
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.8582,
88
+ "longitude":2.3387
89
+ },
90
+ "quality":"high"
91
+ }
data/report_SSG_inference_question-answering_privateCloud_a3342146-0d82-4501-acf4-47cb7d9fda17.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"bd01b8bb-5c3f-44ef-b1f8-81e7bb3305ed",
4
+ "reportDatetime":"2024-10-23T08:02:59",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":104,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"phi3med-14b",
21
+ "parametersNumber":14.0,
22
+ "framework":"vllm",
23
+ "quantization":"q16"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":3533,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":329401,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"Simple question answer with 104 questions from simple ones to contextualised one. Dataset : 92 simple questions (Average : 11 Tokens) and 12 contextualized questions (On average question of 15 Tokens and contexts of 160 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0157068207062144,
52
+ "measurementDuration":119.17674136161804,
53
+ "measurementDateTime":"2024-10-23T08:02:59"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-116-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.30359268188477
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.7898,
88
+ "longitude":2.5447
89
+ },
90
+ "quality":"high"
91
+ }
data/report_SSG_inference_question-answering_privateCloud_b502f765-2c20-4f4a-9265-5126c5b4eba0.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"9237143f-fd40-45e6-bf72-51d45f5401fa",
4
+ "reportDatetime":"2024-10-10T06:55:22",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":10,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"llama2-13b",
21
+ "parametersNumber":13.0,
22
+ "framework":"vllm",
23
+ "quantization":"q16"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":1817,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":1008,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"10 Simple question answer with (Average : 11 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0006318521200859,
52
+ "measurementDuration":4.509745121002197,
53
+ "measurementDateTime":"2024-10-10T06:55:22"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-116-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.3035888671875
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.8582,
88
+ "longitude":2.3387
89
+ },
90
+ "quality":"high"
91
+ }
data/report_SSG_inference_question-answering_privateCloud_bd01b8bb-5c3f-44ef-b1f8-81e7bb3305ed.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "header": {
3
+ "reportId":"66b9b016-e026-4d02-ae96-b83e978897c7",
4
+ "reportDatetime":"2024-10-30T10:12:49",
5
+ "reportStatus":"final",
6
+ "publisher": {
7
+ "name":"SopraSteriaGroup",
8
+ "division":"sustAIn",
9
+ "projectName":"GA2I",
10
+ "confidentialityLevel":"public"
11
+ }
12
+ },
13
+ "task": {
14
+ "taskFamily":"question-answering",
15
+ "taskStage":"inference",
16
+ "nbRequest":104,
17
+ "algorithms": [
18
+ {
19
+ "algorithmType":"llm",
20
+ "foundationModelName":"mistral-7b-instruct",
21
+ "parametersNumber":7.0,
22
+ "framework":"vllm",
23
+ "quantization":"q16"
24
+ }
25
+ ],
26
+ "dataset": [
27
+ {
28
+ "dataUsage":"input",
29
+ "dataType":"token",
30
+ "dataQuantity":4101,
31
+ "source":"private",
32
+ "owner":"SSG"
33
+ },
34
+ {
35
+ "dataUsage":"output",
36
+ "dataType":"token",
37
+ "dataQuantity":39458,
38
+ "source":"private",
39
+ "owner":"SSG"
40
+ }
41
+ ],
42
+ "estimatedAccuracy":"good",
43
+ "taskDescription":"Simple question answer with 104 questions from simple ones to contextualised one. Dataset : 92 simple questions (Average : 11 Tokens) and 12 contextualized questions (On average question of 15 Tokens and contexts of 160 Tokens). Test on a single GPU Even if the node have 2 available GPU."
44
+ },
45
+ "measures": [
46
+ {
47
+ "measurementMethod":"codecarbon",
48
+ "version":"2.5.0",
49
+ "cpuTrackingMode":"machine",
50
+ "gpuTrackingMode":"machine",
51
+ "powerConsumption":0.0008431335305842,
52
+ "measurementDuration":6.305508852005005,
53
+ "measurementDateTime":"2024-10-30T10:12:49"
54
+ }
55
+ ],
56
+ "system": {
57
+ "os":"Linux-5.15.0-124-generic-x86_64-with-glibc2.35"
58
+ },
59
+ "software": {
60
+ "language":"python",
61
+ "version":"3.10.12"
62
+ },
63
+ "infrastructure": {
64
+ "infraType":"privateCloud",
65
+ "cloudProvider":"OVH-v100s",
66
+ "components": [
67
+ {
68
+ "componentName":"Intel(R) Xeon(R) Gold 6226R CPU @ 2.90GHz",
69
+ "componentType":"cpu",
70
+ "nbComponent":30
71
+ },
72
+ {
73
+ "componentName":"2 x Tesla V100S-PCIE-32GB",
74
+ "componentType":"gpu",
75
+ "nbComponent":2
76
+ },
77
+ {
78
+ "componentName":"RAM",
79
+ "componentType":"ram",
80
+ "nbComponent":1,
81
+ "memorySize":86.3035774230957
82
+ }
83
+ ]
84
+ },
85
+ "environment": {
86
+ "country":"France",
87
+ "latitude":48.7898,
88
+ "longitude":2.5447
89
+ },
90
+ "quality":"high"
91
+ }