Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
29 changes: 23 additions & 6 deletions src/agentunit/generators/llm_generator.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,12 +4,15 @@

import asyncio
import json
import logging
from dataclasses import dataclass
from typing import Any

from agentunit.datasets.base import DatasetCase, DatasetSource


logger = logging.getLogger(__name__)

try:
import openai

Expand Down Expand Up @@ -114,6 +117,8 @@ async def generate(
if constraints:
prompt += "\n\nAdditional constraints:\n" + "\n".join(f"- {c}" for c in constraints)

logger.debug("Llama generated prompt:\n%s", prompt)

# Generate with Llama
response = await asyncio.to_thread(
self.client.text_generation,
Expand All @@ -123,6 +128,7 @@ async def generate(
temperature=self.config.temperature,
return_full_text=False,
)
logger.debug("Llama raw response:\n%s", response)

# Parse JSON response
try:
Expand Down Expand Up @@ -154,9 +160,14 @@ async def generate(
cases, name=f"llama_generated_{domain.replace(' ', '_')}"
)

except json.JSONDecodeError as e:
msg = f"Failed to parse generated dataset: {e}\nResponse: {response}"
raise ValueError(msg)
except json.JSONDecodeError:
# msg = f"Failed to parse generated dataset: {e}\nResponse: {response}"
logger.error(
"Failed to parse Llama response JSON. Raw response:\n%s",
response,
exc_info=True,
)
raise

def generate_sync(
self, domain: str, task_description: str, constraints: list[str] | None = None
Expand Down Expand Up @@ -251,6 +262,7 @@ async def generate(
if seed_examples:
messages[1]["content"] += f"\n\nSeed examples:\n{json.dumps(seed_examples, indent=2)}"

logger.debug("OpenAI generated prompt (messages):\n%s", json.dumps(messages, indent=2))
# Generate with GPT
response = await self.client.chat.completions.create(
model=self.model,
Expand All @@ -261,6 +273,7 @@ async def generate(
)

response_text = response.choices[0].message.content
logger.debug("OpenAI raw response text:\n%s", response_text)

# Parse JSON response
try:
Expand Down Expand Up @@ -294,9 +307,13 @@ async def generate(
cases, name=f"openai_generated_{domain.replace(' ', '_')}"
)

except json.JSONDecodeError as e:
msg = f"Failed to parse generated dataset: {e}\nResponse: {response_text}"
raise ValueError(msg)
except json.JSONDecodeError:
logger.error(
"Failed to parse OpenAI response JSON. Raw response:\n%s",
response_text,
exc_info=True,
)
raise

def generate_sync(
self,
Expand Down