diff --git a/__tests__/e2e/scaling/s_149.yaml b/__tests__/e2e/scaling/s_149.yaml index e9df98c4..ae7598bc 100644 --- a/__tests__/e2e/scaling/s_149.yaml +++ b/__tests__/e2e/scaling/s_149.yaml @@ -19,7 +19,7 @@ resources: cpu: 8 customContainerConfig: image: >- - registry.cn-shanghai.aliyuncs.com/serverless_devs/custom-container-http-examples:springboot + registry.cn-hangzhou.aliyuncs.com/serverless_devs/custom-container-http-examples:springboot port: 9000 instanceConcurrency: 20 memorySize: 65536 diff --git a/__tests__/e2e/scaling/scalingConfig_provisionConfig.py b/__tests__/e2e/scaling/scalingConfig_provisionConfig.py index fc12337c..12b08eed 100644 --- a/__tests__/e2e/scaling/scalingConfig_provisionConfig.py +++ b/__tests__/e2e/scaling/scalingConfig_provisionConfig.py @@ -20,11 +20,15 @@ def deploy_function(scalingConfig, provisionConfig, gpuConfig=None, memorySize=N s_yaml = yaml.safe_load(f) s_yaml["resources"]["fcDemo"]["props"]["scalingConfig"] = scalingConfig - s_yaml["resources"]["fcDemo"]["props"]["provisionConfig"] = provisionConfig - if gpuConfig: + if provisionConfig is not None: + s_yaml["resources"]["fcDemo"]["props"]["provisionConfig"] = provisionConfig + elif "provisionConfig" in s_yaml["resources"]["fcDemo"]["props"]: + del s_yaml["resources"]["fcDemo"]["props"]["provisionConfig"] + + if gpuConfig is not None: s_yaml["resources"]["fcDemo"]["props"]["gpuConfig"] = gpuConfig - if memorySize: + if memorySize is not None: s_yaml["resources"]["fcDemo"]["props"]["memorySize"] = memorySize with open(s_yaml_file, "w", encoding="utf-8") as f: @@ -58,7 +62,7 @@ def main(): scalingConfig = { "minInstances": 1, } - provisionConfig = {} + provisionConfig = None deploy_function(scalingConfig, provisionConfig) # 弹性 ----> 常驻 @@ -66,7 +70,7 @@ def main(): "residentPoolId": "fc-pool-5f044a31f87171jkwaraws", "maxInstances": 1, } - provisionConfig = {} + provisionConfig = None deploy_function(scalingConfig, provisionConfig) # 常驻 ----> 常驻 @@ -74,14 +78,14 @@ def main(): "residentPoolId": "fc-pool-16bedd56db9626uva1it08", "maxInstances": 1, } - provisionConfig = {} + provisionConfig = None deploy_function(scalingConfig, provisionConfig) # 常驻 ----> 弹性 scalingConfig = { "minInstances": 1, } - provisionConfig = {} + provisionConfig = None deploy_function(scalingConfig, provisionConfig) cleanup_deployment() @@ -95,7 +99,7 @@ def main(): "gpuType": "fc.gpu.tesla.1", "gpuMemorySize": 1, } - provisionConfig = {} + provisionConfig = None memorySize = 32768 deploy_function(scalingConfig, provisionConfig, gpuConfig, memorySize) @@ -108,7 +112,7 @@ def main(): "gpuType": "fc.gpu.ada.1", "gpuMemorySize": 1, } - provisionConfig = {} + provisionConfig = None memorySize = 65536 deploy_function(scalingConfig, provisionConfig, gpuConfig, memorySize) @@ -121,7 +125,7 @@ def main(): "gpuType": "fc.gpu.ada.2", "gpuMemorySize": 1, } - provisionConfig = {} + provisionConfig = None memorySize = 32768 deploy_function(scalingConfig, provisionConfig, gpuConfig, memorySize) @@ -133,8 +137,11 @@ def main(): "gpuType": "fc.gpu.tesla.1", "gpuMemorySize": 1, } - provisionConfig = {} + provisionConfig = None memorySize = 32768 deploy_function(scalingConfig, provisionConfig, gpuConfig, memorySize) cleanup_deployment() + +if __name__ == "__main__": + main() diff --git a/publish.yaml b/publish.yaml index 70813a24..1785cd29 100644 --- a/publish.yaml +++ b/publish.yaml @@ -3,7 +3,7 @@ Type: Component Name: fc3 Provider: - 阿里云 -Version: 0.1.7 +Version: 0.1.8 Description: 阿里云函数计算全生命周期管理 HomePage: https://github.com/devsapp/fc3 Organization: 阿里云函数计算(FC) diff --git a/src/resources/fc/index.ts b/src/resources/fc/index.ts index d95638c5..571f3fd7 100644 --- a/src/resources/fc/index.ts +++ b/src/resources/fc/index.ts @@ -304,8 +304,8 @@ export default class FC extends FC_Client { */ const { project, logstore } = (config.logConfig || {}) as ILogConfig; const retrySls = slsAuto && isSlsNotExistException(project, logstore, ex); - const localGPUType = config.gpuConfig?.gpuType; - const remoteGPUType = remoteConfig.gpuConfig?.gpuType; + const localGPUType = config?.gpuConfig?.gpuType; + const remoteGPUType = remoteConfig?.gpuConfig?.gpuType; if (retrySls) { if (calculateRetryTime(3)) { throw ex;