@@ -134,7 +134,7 @@ const vertexAI = new VertexAI({
134134 location : LOCATION ,
135135} ) ;
136136
137- const TEXT_MODEL_NAME = 'gemini-2.0 -flash' ;
137+ const TEXT_MODEL_NAME = 'gemini-2.5 -flash' ;
138138const generativeTextModel = vertexAI . getGenerativeModel ( {
139139 model : TEXT_MODEL_NAME ,
140140 generationConfig : {
@@ -148,31 +148,31 @@ const generativeTextModelPreview = vertexAI.preview.getGenerativeModel({
148148 } ,
149149} ) ;
150150const generativeTextModelWithPrefix = vertexAI . getGenerativeModel ( {
151- model : 'models/gemini-2.0 -flash' ,
151+ model : 'models/gemini-2.5 -flash' ,
152152 generationConfig : {
153153 maxOutputTokens : 256 ,
154154 } ,
155155} ) ;
156156const generativeTextModelWithPrefixPreview =
157- vertexAI . preview . getGenerativeModel ( {
158- model : 'models/gemini-2.0 -flash' ,
159- generationConfig : {
160- maxOutputTokens : 256 ,
161- } ,
162- } ) ;
157+ vertexAI . preview . getGenerativeModel ( {
158+ model : 'models/gemini-2.5 -flash' ,
159+ generationConfig : {
160+ maxOutputTokens : 256 ,
161+ } ,
162+ } ) ;
163163const generativeVisionModel = vertexAI . getGenerativeModel ( {
164- model : 'gemini-2.0 -flash' ,
164+ model : 'gemini-2.5 -flash' ,
165165} ) ;
166166const generativeVisionModelPreview = vertexAI . preview . getGenerativeModel ( {
167- model : 'gemini-2.0 -flash' ,
167+ model : 'gemini-2.5 -flash' ,
168168} ) ;
169169const generativeVisionModelWithPrefix = vertexAI . getGenerativeModel ( {
170- model : 'gemini-2.0 -flash' ,
170+ model : 'gemini-2.5 -flash' ,
171171} ) ;
172172const generativeVisionModelWithPrefixPreview =
173- vertexAI . preview . getGenerativeModel ( {
174- model : 'models/gemini-2.0 -flash' ,
175- } ) ;
173+ vertexAI . preview . getGenerativeModel ( {
174+ model : 'models/gemini-2.5 -flash' ,
175+ } ) ;
176176describe ( 'generateContentStream' , ( ) => {
177177 beforeEach ( ( ) => {
178178 jasmine . DEFAULT_TIMEOUT_INTERVAL = 30000 ;
@@ -972,7 +972,7 @@ describe('sendMessage', () => {
972972
973973describe ( 'sendMessageStream' , ( ) => {
974974 beforeEach ( ( ) => {
975- jasmine . DEFAULT_TIMEOUT_INTERVAL = 30000 ;
975+ jasmine . DEFAULT_TIMEOUT_INTERVAL = 60000 ;
976976 } ) ;
977977 it ( 'should should return a stream and populate history when generationConfig is passed to startChat' , async ( ) => {
978978 const chat = generativeTextModel . startChat ( {
0 commit comments