Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

keras model - CV #3

Open
germayneng opened this issue Jun 27, 2018 · 3 comments
Open

keras model - CV #3

germayneng opened this issue Jun 27, 2018 · 3 comments

Comments

@germayneng
Copy link

Hi CPMP,

I have read your solutions explanation at https://www.kaggle.com/c/web-traffic-time-series-forecasting/discussion/39395 and would like to say thank you for giving the solution.

I do have one question and hope you can clarify for me. Your final model is an xgb model using keras as well as your GLM's predictor as features along with some features you did.

But in order use the keras and GLM model, you would require OOF predictions for September 13th, 2017 and November 13th, 2017. (which you did output)

For the keras model, to generate the OOF, you need to perform some form of CV, and you did explain that you did via group Kfold. My question would be when doing so, there will be some folds where you will train the future to predict on the past. Is this aspect alright?

@jfpuget
Copy link
Owner

jfpuget commented Jun 27, 2018 via email

@germayneng
Copy link
Author

germayneng commented Jun 27, 2018

Hi ,

I re-read and understood. It seems you are using feed forward NN, which to my understand, allows you to output multiple target. Each target are the dates (w0d1,w0d2....) and each row of your train data are unique pages. So when you define your folds via pages, the inputs which you built using the past will predict multiple output (w0d1...)

is my understand correct?

Regards
Germayne

@germayneng
Copy link
Author

germayneng commented Jun 28, 2018

Hi CPMP,

  1. for the final xgb ensemble, you did 5 fold cv split. In this case, wont there be case where some of the folds are trained on the 'future' data and predicted on the past data?

Essentially each rows are different page views of a particular day in 2016. Is it in this case, you decided to not treat each row as a time ordered data but just as individual samples in this ensembling step ( since you are training on residual as targets)

  1. notice that you fitted back your training data into the model that you trained with for your GLM. what is the story behind this? I understand this is the OOF, but why fit it back to the original training data ( after u done it, then u proceed to add .ALLvisits)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants