Skip to content

Commit

Permalink
Fix bug with cogvlm prompting and adjust tests
Browse files Browse the repository at this point in the history
  • Loading branch information
PawelPeczek-Roboflow committed Feb 23, 2024
1 parent 72e0786 commit e6c7d16
Showing 1 changed file with 3 additions and 3 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -536,7 +536,7 @@ async def test_get_cogvlm_generations_from_remote_api(
async def test_get_cogvlm_generations_locally() -> None:
# given
model_manager = AsyncMock()
model_manager.model_manager.side_effect = [
model_manager.infer_from_request.side_effect = [
CogVLMResponse.parse_obj({"response": "Response 1: 42"}),
CogVLMResponse.parse_obj({"response": "Response 2: 42"}),
CogVLMResponse.parse_obj({"response": "Response 3: 42"}),
Expand Down Expand Up @@ -569,7 +569,7 @@ async def test_run_cog_vlm_prompting_when_local_execution_chosen_and_no_expected
):
# given
model_manager = AsyncMock()
model_manager.model_manager.side_effect = [
model_manager.infer_from_request.side_effect = [
CogVLMResponse.parse_obj({"response": "Response 1: 42"}),
CogVLMResponse.parse_obj({"response": "Response 2: 42"}),
CogVLMResponse.parse_obj({"response": "Response 3: 42"}),
Expand Down Expand Up @@ -606,7 +606,7 @@ async def test_run_cog_vlm_prompting_when_local_execution_chosen_and_json_output
):
# given
model_manager = AsyncMock()
model_manager.model_manager.side_effect = [
model_manager.infer_from_request.side_effect = [
CogVLMResponse.parse_obj({"response": json.dumps({"value": 42})}),
CogVLMResponse.parse_obj({"response": json.dumps({"value": 43})}),
CogVLMResponse.parse_obj({"response": json.dumps({"value": 44})}),
Expand Down

0 comments on commit e6c7d16

Please sign in to comment.