Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

High changepoint number prediction and penalty value #338

Closed
ClaretJeanLoup opened this issue Feb 27, 2025 · 0 comments
Closed

High changepoint number prediction and penalty value #338

ClaretJeanLoup opened this issue Feb 27, 2025 · 0 comments

Comments

@ClaretJeanLoup
Copy link

Hi,

I'm running ruptures on biological data: I'm trying to detect genomic duplication breakpoints from depth of coverage variation in genome mapping data. I tried to use:
algo = rpt.Pelt(model="rbf").fit(d_subset['norm'].values)
refined_result = algo.predict(pen=X)
Making the penalty value vary from 1 to values in the thousands, but no matter the penalty value I still en up with more than 2k change points on a dataset of 12K values.
Is my data too noisy to reduce the number of changepoints or do I not get how penalty should be set? I expected higher penalty values to result in longer computing time and a smaller number of predicted change points.
I tried other algorithms like BottomUp and Window and got some nice predictions on smoothed (mean value of slidding windows of a thousand points) and normalised data but I would like to see if I can get more precise ones with Pelt algorithm.

Thanks for the amazing package you designed!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant