Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions .pre-commit-config.yaml
Original file line number Diff line number Diff line change
@@ -1,15 +1,15 @@
repos:
- repo: https://github.com/pre-commit/pre-commit-hooks
rev: v2.5.0
rev: v4.4.0
hooks:
- id: check-yaml
- id: end-of-file-fixer
- id: trailing-whitespace
- id: detect-aws-credentials
args: [--allow-missing-credentials]

- repo: https://github.com/humitos/mirrors-autoflake.git
rev: v1.3
- repo: https://github.com/PyCQA/autoflake
rev: v2.2.1
hooks:
- id: autoflake
args: ['--in-place', '--expand-star-imports', '--ignore-init-module-imports', '--remove-all-unused-imports']
Expand Down
3,004 changes: 1,598 additions & 1,406 deletions poetry.lock

Large diffs are not rendered by default.

2 changes: 1 addition & 1 deletion pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ classifiers=[
[tool.poetry.dependencies]
python = "^3.10"
urllib3 = ">=2.3.0"
ray = "2.44.0"
ray = ">=2.44.0"
semantic-version = "2.10.0"
pyarrow = "20.0.0"
pyfunctional = "1.5.0"
Expand Down
1 change: 0 additions & 1 deletion src/fmeval/data_loaders/json_data_loader.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,6 @@

from ray.data.datasource.file_based_datasource import (
FileBasedDatasource,
_resolve_kwargs,
)

from fmeval.exceptions import EvalAlgorithmInternalError
Expand Down
1 change: 0 additions & 1 deletion src/fmeval/eval_algorithms/qa_accuracy.py
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,6 @@
require,
create_shared_resource,
cleanup_shared_resource,
assert_condition,
)

F1_SCORE = "f1_score"
Expand Down
3 changes: 0 additions & 3 deletions src/fmeval/eval_algorithms/util.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,6 @@
from typing import Any, Dict, List, Optional, Tuple, Union
from fmeval.constants import (
DatasetColumns,
EVAL_OUTPUT_RECORDS_BATCH_SIZE,
MEAN,
NUM_ROWS_DETERMINISTIC,
DATASET_COLUMNS,
Expand All @@ -23,8 +22,6 @@
CategoryScore,
DATASET_CONFIGS,
EVAL_DATASETS,
EvalOutput,
get_default_prompt_template,
)
from fmeval.exceptions import EvalAlgorithmInternalError
from fmeval.model_runners.composers.composers import PromptComposer
Expand Down
1 change: 0 additions & 1 deletion test/unit/data_loaders/test_json_data_loader.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,6 @@
from fmeval.data_loaders.json_data_loader import (
JsonDataLoader,
JsonDataLoaderConfig,
CustomJSONDatasource,
)
from fmeval.data_loaders.util import DataConfig
from typing import Any, Dict, List, NamedTuple, Optional, Union
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,19 +8,14 @@

from fmeval.constants import (
DatasetColumns,
MIME_TYPE_JSON,
BUTTER_FINGER,
RANDOM_UPPER_CASE,
WHITESPACE_ADD_REMOVE,
MEAN,
)
from fmeval.eval_algorithms import (
EvalScore,
EvalOutput,
CategoryScore,
BUILT_IN_DATASET_DEFAULT_PROMPT_TEMPLATES,
DEFAULT_PROMPT_TEMPLATE,
WOMENS_CLOTHING_ECOMMERCE_REVIEWS,
)
from fmeval.eval_algorithms.classification_accuracy_semantic_robustness import (
ClassificationAccuracySemanticRobustnessConfig,
Expand Down
5 changes: 0 additions & 5 deletions test/unit/eval_algorithms/test_prompt_stereotyping.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,16 +8,11 @@

from fmeval.constants import (
DatasetColumns,
MIME_TYPE_JSON,
DEFAULT_EVAL_RESULTS_PATH,
)
from fmeval.eval_algorithms import (
EvalOutput,
CategoryScore,
EvalScore,
EvalAlgorithm,
DEFAULT_PROMPT_TEMPLATE,
CROWS_PAIRS,
)
from fmeval.eval_algorithms.prompt_stereotyping import (
PromptStereotyping,
Expand Down
7 changes: 0 additions & 7 deletions test/unit/eval_algorithms/test_qa_accuracy.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,11 +18,6 @@
EvalOutput,
CategoryScore,
EvalScore,
BUILT_IN_DATASET_DEFAULT_PROMPT_TEMPLATES,
TRIVIA_QA,
BOOLQ,
NATURAL_QUESTIONS,
DEFAULT_PROMPT_TEMPLATE,
)
from fmeval.eval_algorithms.helper_models.helper_model import BertscoreHelperModel
from fmeval.eval_algorithms.qa_accuracy import (
Expand All @@ -41,8 +36,6 @@
_split,
_quasi_exact_match_score,
SCORE_NAMES,
SplitWithDelimiter,
BertScore,
)
from fmeval.exceptions import EvalAlgorithmClientError

Expand Down
8 changes: 0 additions & 8 deletions test/unit/eval_algorithms/test_qa_toxicity.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,17 +7,9 @@

from fmeval.constants import (
DatasetColumns,
MIME_TYPE_JSON,
)
from fmeval.eval_algorithms import (
EvalScore,
EvalOutput,
CategoryScore,
NATURAL_QUESTIONS,
BUILT_IN_DATASET_DEFAULT_PROMPT_TEMPLATES,
TRIVIA_QA,
BOOLQ,
DEFAULT_PROMPT_TEMPLATE,
)
from fmeval.eval_algorithms.helper_models.helper_model import (
TOXIGEN_SCORE_NAME,
Expand Down
1 change: 0 additions & 1 deletion test/unit/eval_algorithms/test_toxicity.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,6 @@

from fmeval.constants import (
DatasetColumns,
MIME_TYPE_JSON,
MEAN,
)
from fmeval.data_loaders.data_config import DataConfig
Expand Down
2 changes: 0 additions & 2 deletions test/unit/eval_algorithms/test_util.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,6 @@
from fmeval.eval_algorithms import (
EvalAlgorithm,
DATASET_CONFIGS,
EVAL_DATASETS,
BOOLQ,
TRIVIA_QA,
NATURAL_QUESTIONS,
Expand All @@ -35,7 +34,6 @@
WOMENS_CLOTHING_ECOMMERCE_REVIEWS,
REAL_TOXICITY_PROMPTS,
REAL_TOXICITY_PROMPTS_CHALLENGING,
EvalOutput,
CategoryScore,
)
from fmeval.eval_algorithms.eval_algorithm import EvalScore
Expand Down
Loading