Skip to content

Commit 9f02efc

Browse files
committed
Adds new metrics calculated
Signed-off-by: Flavia Beo <flavia.beo@ibm.com>
1 parent d77e570 commit 9f02efc

File tree

1 file changed

+15
-4
lines changed

1 file changed

+15
-4
lines changed

tests/models/test_decoders.py

Lines changed: 15 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -42,6 +42,7 @@
4242
GRANITE_3p3_8B_INSTRUCT = "ibm-granite/granite-3.3-8b-instruct"
4343
GRANITE_20B_CODE_INSTRUCT_8K = "ibm-granite/granite-20b-code-instruct-8k"
4444
LLAMA_3p1_70B_INSTRUCT = "meta-llama/Llama-3.1-70B-Instruct"
45+
MISTRAL_0p3_7B_INSTRUCT = "mistralai/Mistral-7B-Instruct-v0.3"
4546

4647
micro_model_mapping = {
4748
LLAMA_3p1_8B_INSTRUCT: os.path.join(MICRO_MODELS_HOME, "llama-3.1-8b-layers-3-step-24000"),
@@ -72,6 +73,7 @@
7273
GRANITE_3p3_8B_INSTRUCT,
7374
GRANITE_20B_CODE_INSTRUCT_8K,
7475
LLAMA_3p1_70B_INSTRUCT,
76+
MISTRAL_0p3_7B_INSTRUCT
7577
],
7678
)
7779
# for validation level 1, the default is a failure rate of 1%
@@ -145,25 +147,34 @@
145147
# threshold key is (model_id, is_tiny_model)
146148
fail_thresholds = {
147149
(LLAMA_3p1_8B_INSTRUCT, False): (
148-
2.6994638133048965,
149-
0.00047589250549208347,
150+
2.7080255031585696,
151+
0.0004068055667448795,
150152
),
151153
(GRANITE_3p2_8B_INSTRUCT, False): (
152154
2.3919514417648315,
153155
0.0005767398688476533,
154156
),
157+
(GRANITE_3p2_8B_INSTRUCT, True): (
158+
2.7449850964546205,
159+
0.00018840670207282534,
160+
),
155161
(GRANITE_3p3_8B_INSTRUCT, False): (
156162
2.4444521379470827,
157163
0.0004970188625156878,
158164
),
159165
(GRANITE_20B_CODE_INSTRUCT_8K, False): (
160-
2.640706129074097,
161-
0.00034344267623964697,
166+
2.646075320243838,
167+
0.0003458251833217223,
162168
),
169+
# TODO: run llama 70B with 1,2,4,8 batches
163170
(LLAMA_3p1_70B_INSTRUCT, False): (
164171
2.841279556751251,
165172
0.0044301633024588115,
166173
),
174+
(MISTRAL_0p3_7B_INSTRUCT, False): (
175+
2.846206340789795,
176+
0.0008768103783950205,
177+
),
167178
}
168179
# custom weight adaptation to be used in future. For instance if we would like to add some other adaptation, we can register it with this custom adapter
169180
# and provide it when converting from an aiu fms model's weights to a cpu fms model's weights. Currently this is only done for gptq, but may be done for other

0 commit comments

Comments
 (0)