@@ -111,7 +111,7 @@ def test_create_load_multimodal_embedding_generator_model(
111
111
"gemini-1.5-pro-preview-0514" ,
112
112
"gemini-1.5-flash-preview-0514" ,
113
113
"gemini-1.5-pro-001" ,
114
- # "gemini-1.5-pro-002",
114
+ "gemini-1.5-pro-002" ,
115
115
"gemini-1.5-flash-001" ,
116
116
"gemini-1.5-flash-002" ,
117
117
"gemini-2.0-flash-exp" ,
@@ -146,7 +146,7 @@ def test_create_load_gemini_text_generator_model(
146
146
"gemini-1.5-pro-preview-0514" ,
147
147
"gemini-1.5-flash-preview-0514" ,
148
148
"gemini-1.5-pro-001" ,
149
- # "gemini-1.5-pro-002",
149
+ "gemini-1.5-pro-002" ,
150
150
"gemini-1.5-flash-001" ,
151
151
"gemini-1.5-flash-002" ,
152
152
"gemini-2.0-flash-exp" ,
@@ -173,7 +173,7 @@ def test_gemini_text_generator_predict_default_params_success(
173
173
"gemini-1.5-pro-preview-0514" ,
174
174
"gemini-1.5-flash-preview-0514" ,
175
175
"gemini-1.5-pro-001" ,
176
- # "gemini-1.5-pro-002",
176
+ "gemini-1.5-pro-002" ,
177
177
"gemini-1.5-flash-001" ,
178
178
"gemini-1.5-flash-002" ,
179
179
"gemini-2.0-flash-exp" ,
@@ -202,7 +202,7 @@ def test_gemini_text_generator_predict_with_params_success(
202
202
"gemini-1.5-pro-preview-0514" ,
203
203
"gemini-1.5-flash-preview-0514" ,
204
204
"gemini-1.5-pro-001" ,
205
- # "gemini-1.5-pro-002",
205
+ "gemini-1.5-pro-002" ,
206
206
"gemini-1.5-flash-001" ,
207
207
"gemini-1.5-flash-002" ,
208
208
"gemini-2.0-flash-exp" ,
@@ -233,7 +233,7 @@ def test_gemini_text_generator_multi_cols_predict_success(
233
233
"gemini-1.5-pro-preview-0514" ,
234
234
"gemini-1.5-flash-preview-0514" ,
235
235
"gemini-1.5-pro-001" ,
236
- # "gemini-1.5-pro-002",
236
+ "gemini-1.5-pro-002" ,
237
237
"gemini-1.5-flash-001" ,
238
238
"gemini-1.5-flash-002" ,
239
239
"gemini-2.0-flash-exp" ,
@@ -756,7 +756,7 @@ def test_text_embedding_generator_retry_no_progress(session, bq_connection):
756
756
@pytest .mark .parametrize (
757
757
"model_name" ,
758
758
(
759
- # "gemini-1.5-pro-002",
759
+ "gemini-1.5-pro-002" ,
760
760
"gemini-1.5-flash-002" ,
761
761
"gemini-2.0-flash-001" ,
762
762
"gemini-2.0-flash-lite-001" ,
@@ -786,7 +786,7 @@ def test_llm_gemini_score(llm_fine_tune_df_default_index, model_name):
786
786
@pytest .mark .parametrize (
787
787
"model_name" ,
788
788
(
789
- # "gemini-1.5-pro-002",
789
+ "gemini-1.5-pro-002" ,
790
790
"gemini-1.5-flash-002" ,
791
791
"gemini-2.0-flash-001" ,
792
792
"gemini-2.0-flash-lite-001" ,
0 commit comments