Skip to content

Conversation

@rwtarpit
Copy link
Contributor

Reference Issues/PRs

Fixes #2985

What does this implement/fix? Explain your changes.

Implements a new file for loading Monster datasets from hugginngface_hub.
Creates methods for loading list of available monster datasets in huggingface_hub and downloading the datasets.

Does your contribution introduce a new dependency? If yes, which one?

hugging_face

Any other comments?

I am yet to implement the test_monster_loaders file.

PR checklist

For all contributions
  • I've added myself to the list of contributors. Alternatively, you can use the @all-contributors bot to do this for you after the PR has been merged.
  • The PR title starts with either [ENH], [MNT], [DOC], [BUG], [REF], [DEP] or [GOV] indicating whether the PR topic is related to enhancement, maintenance, documentation, bugs, refactoring, deprecation or governance.
For new estimators and functions
  • I've added the estimator/function to the online API documentation.
  • (OPTIONAL) I've added myself as a __maintainer__ at the top of relevant files and want to be contacted regarding its maintenance. Unmaintained files may be removed. This is for the full file, and you should not add yourself if you are just making minor changes or do not want to help maintain its contents.
For developers with write access
  • (OPTIONAL) I've updated aeon's CODEOWNERS to receive notifications about future changes to these files.

@aeon-actions-bot aeon-actions-bot bot added datasets Datasets and data loaders enhancement New feature, improvement request or other non-bug code enhancement labels Nov 27, 2025
@aeon-actions-bot
Copy link
Contributor

Thank you for contributing to aeon

I have added the following labels to this PR based on the title: [ enhancement ].
I have added the following labels to this PR based on the changes made: [ datasets ]. Feel free to change these if they do not properly represent the PR.

The Checks tab will show the status of our automated tests. You can click on individual test runs in the tab or "Details" in the panel below to see more information if there is a failure.

If our pre-commit code quality check fails, any trivial fixes will automatically be pushed to your PR unless it is a draft.

Don't hesitate to ask questions on the aeon Slack channel if you have any.

PR CI actions

These checkboxes will add labels to enable/disable CI functionality for this PR. This may not take effect immediately, and a new commit may be required to run the new configuration.

  • Run pre-commit checks for all files
  • Run mypy typecheck tests
  • Run all pytest tests and configurations
  • Run all notebook example tests
  • Run numba-disabled codecov tests
  • Stop automatic pre-commit fixes (always disabled for drafts)
  • Disable numba cache loading
  • Regenerate expected results for testing
  • Push an empty commit to re-run CI checks

@rwtarpit
Copy link
Contributor Author

@baraline pls have a look.
and also can you guide me on testing file for this.
thank you.

@baraline
Copy link
Member

At first glance the file looks good. I think you can remove the univariate dataset name list, as you pull all the dataset names already.

Concerning the tests, you can look at existing tests file from other loaders, for exemple https://github.com/aeon-toolkit/aeon/blob/main/aeon/datasets/tests/test_data_loaders.py
More specifically, the xxxx_from_repo tests. You can perform a similar test with similar function tags but trying to fecth from Monster, and validate that what you fetched is in the expected format and not empty.

Also, you have some pre-commit failure, you can refer to the developer documentation it explains how to install it, it will correct most of the issues by itself when you try to commit and signal you the others.

@rwtarpit
Copy link
Contributor Author

@baraline how do i resolve FAILED aeon/datasets/tests/test_monster_loader.py::test_load_monster_dataset - ModuleNotFoundError: No module named 'huggingface_hub'.
i tried removing _check_soft_dependencies but still same error.

@baraline
Copy link
Member

baraline commented Nov 27, 2025

The check soft dep should stay, it is used to check if the dependency is installed (as it is optional) and raide appropriate messages to the users. Don't have access to my computer rn, so add the check soft dep back and i'll check the errors throwed when I can later

@rwtarpit rwtarpit marked this pull request as ready for review November 29, 2025 13:41
Copy link
Member

@baraline baraline left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Test are failing because the test skip is wrongly parameterized. The wrong check is applied to validate if the test should be run or not. You should do

    @pytest.mark.skipif(
        not _check_soft_dependencies("hugging-face", severity="none"),
        reason="required soft dependency hugging-face not available",
    )

Instead of

@pytest.mark.skipif(
    PR_TESTING,
    reason="Only run on overnights because of read from internet.",
)

@baraline
Copy link
Member

baraline commented Dec 2, 2025

You should test with huggingface-hub (the dependency you added in the project.toml), not with hugging-face. You correctly did it when you created the loader like this :

_check_soft_dependencies("huggingface-hub", severity="none")
from huggingface_hub import hf_hub_download

My example didn't take the correct name.

@rwtarpit
Copy link
Contributor Author

rwtarpit commented Dec 2, 2025

You should test with huggingface-hub (the dependency you added in the project.toml), not with hugging-face. You correctly did it when you created the loader like this :

_check_soft_dependencies("huggingface-hub", severity="none") from huggingface_hub import hf_hub_download

My example didn't take the correct name.

my bad!!
did it in a hurry

@baraline
Copy link
Member

baraline commented Dec 2, 2025

The remaining failure is not linked to what you did, so we should be good to go. We'll just wait for it to be resolved. Thanks !

@rwtarpit
Copy link
Contributor Author

rwtarpit commented Dec 2, 2025

The remaining failure is not linked to what you did, so we should be good to go. We'll just wait for it to be resolved. Thanks !

thank you for guiding me!!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

datasets Datasets and data loaders enhancement New feature, improvement request or other non-bug code enhancement

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants