From f15d6dfb2c7fc6a8a2b1dac8a66a972c0be7ab7b Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Pawe=C5=82=20P=C4=99czek?= Date: Fri, 23 Feb 2024 14:08:30 +0100 Subject: [PATCH 1/3] Prepare for 0.9.14 release --- inference/core/version.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/inference/core/version.py b/inference/core/version.py index 805233dd02..21cbaccbb2 100644 --- a/inference/core/version.py +++ b/inference/core/version.py @@ -1,4 +1,4 @@ -__version__ = "0.9.14rc2" +__version__ = "0.9.14" if __name__ == "__main__": From 72e0786e8a820ca95114de6f30a465304bdce0da Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Pawe=C5=82=20P=C4=99czek?= Date: Fri, 23 Feb 2024 14:31:31 +0100 Subject: [PATCH 2/3] Fix issue with cogvlm --- .../enterprise/workflows/complier/steps_executors/models.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/inference/enterprise/workflows/complier/steps_executors/models.py b/inference/enterprise/workflows/complier/steps_executors/models.py index a2531048a2..a3d9d65ff0 100644 --- a/inference/enterprise/workflows/complier/steps_executors/models.py +++ b/inference/enterprise/workflows/complier/steps_executors/models.py @@ -1096,7 +1096,7 @@ async def get_cogvlm_generations_locally( core_model="cogvlm", api_key=api_key, ) - result = await model_manager.model_manager( + result = await model_manager.infer_from_request( yolo_world_model_id, inference_request ) serialised_result.append( From e6c7d1639a7c83e24dc2989c45d428b8f9e6effe Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Pawe=C5=82=20P=C4=99czek?= Date: Fri, 23 Feb 2024 14:34:55 +0100 Subject: [PATCH 3/3] Fix bug with cogvlm prompting and adjust tests --- .../workflows/compiler/steps_executors/test_models.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/tests/inference/unit_tests/enterprise/workflows/compiler/steps_executors/test_models.py b/tests/inference/unit_tests/enterprise/workflows/compiler/steps_executors/test_models.py index aeff002176..083072f225 100644 --- a/tests/inference/unit_tests/enterprise/workflows/compiler/steps_executors/test_models.py +++ b/tests/inference/unit_tests/enterprise/workflows/compiler/steps_executors/test_models.py @@ -536,7 +536,7 @@ async def test_get_cogvlm_generations_from_remote_api( async def test_get_cogvlm_generations_locally() -> None: # given model_manager = AsyncMock() - model_manager.model_manager.side_effect = [ + model_manager.infer_from_request.side_effect = [ CogVLMResponse.parse_obj({"response": "Response 1: 42"}), CogVLMResponse.parse_obj({"response": "Response 2: 42"}), CogVLMResponse.parse_obj({"response": "Response 3: 42"}), @@ -569,7 +569,7 @@ async def test_run_cog_vlm_prompting_when_local_execution_chosen_and_no_expected ): # given model_manager = AsyncMock() - model_manager.model_manager.side_effect = [ + model_manager.infer_from_request.side_effect = [ CogVLMResponse.parse_obj({"response": "Response 1: 42"}), CogVLMResponse.parse_obj({"response": "Response 2: 42"}), CogVLMResponse.parse_obj({"response": "Response 3: 42"}), @@ -606,7 +606,7 @@ async def test_run_cog_vlm_prompting_when_local_execution_chosen_and_json_output ): # given model_manager = AsyncMock() - model_manager.model_manager.side_effect = [ + model_manager.infer_from_request.side_effect = [ CogVLMResponse.parse_obj({"response": json.dumps({"value": 42})}), CogVLMResponse.parse_obj({"response": json.dumps({"value": 43})}), CogVLMResponse.parse_obj({"response": json.dumps({"value": 44})}),