-
Notifications
You must be signed in to change notification settings - Fork 152
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[DOC] Rework getting started guide and single problem forecasting loaders #2248
Conversation
Thank you for contributing to
|
Check out this pull request on See visual diffs & provide feedback on Jupyter Notebooks. Powered by ReviewNB |
Where in anomaly detection? I am referring just to inputs, dont really care what format they want the data internally
cant you just transpose the dataframe? seems very odd if that removes the indicies |
I guess I can, seems odd to do so and I thought the examples made it clear, but sure, Need to rewrite the tests that use col_names |
I remember some previous changes where we tried to remove Seems better to completely follow one format and leave changing that to the |
yes but you cannot then extract series X["channel_name"], anyway, I have done it and removed the test for column names |
You would just use https://pandas.pydata.org/docs/reference/api/pandas.DataFrame.loc.html |
It was more I didn't want to rewrite the notebooks that used plot_series. But I've ditched that now |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
just small questions, if you want to keep them for later i dont mind
View / edit / reply to this conversation on ReviewNB hadifawaz1999 commented on 2024-11-03T12:51:53Z i would think its better to add a section per task to use the load any dataset function, load_classification load_regression etc. what do u think ? TonyBagnall commented on 2024-11-03T16:46:53Z yes I agree, but maybe not in this PR? Really only wanted to do the getting_started.md, then will work through notebooks module by module, starting with datasets |
yes I agree, but maybe not in this PR? Really only wanted to do the getting_started.md, then will work through notebooks module by module, starting with datasets View entire conversation on ReviewNB |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm
fixes #2246 part of #1518
This has expanded to tidying up the single problem data loaders for forecasting. So its in two related parts
datasets._single_problem_loaders
There are seven baked in forecasting data sets and were eight loaders.
firstly returning y, X is opposite to collections, and secondly there seems no need to split the data in the loader, the user can surely do that themselves. Changed to
returns numpy array with axis == 0, i.e.
n_channels, n_timepoints
and
returns a data frame with axis == 1 and all the column names set as before.
Read me
Split along series/collection estimators and adding an example for each module, including experimental. It makes it longer, maybe we dont want it, but its a good top level intro imo, will link for further details
First version done, highlighted there is no anomaly detection notebook, see #1960 and the transformers notebooks need an overhaul, but thats future work. The main goal is to get things ready for the new forecasting base class