Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add loc argument to distribution output classes #540

Merged
merged 4 commits into from
Jan 6, 2020

Conversation

lostella
Copy link
Contributor

Description of changes: This PR adds the loc argument to the distribution method of DistributionOutput classes. Assertions are added in case the specific distributions don't support either loc or scale. A bug in the TransformedPiecewiseLinear distribution was also fixed: this went unnoticed since it only appears when the base distribution is composed with more than one affine transformation, which never happens when using PiecewiseLinearOutput.

By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice.

@lostella lostella requested a review from vafl December 20, 2019 15:44
@codecov-io
Copy link

codecov-io commented Dec 20, 2019

Codecov Report

Merging #540 into master will decrease coverage by <.01%.
The diff coverage is 80%.

Impacted file tree graph

@@            Coverage Diff             @@
##           master     #540      +/-   ##
==========================================
- Coverage   83.18%   83.18%   -0.01%     
==========================================
  Files         177      177              
  Lines        9877     9882       +5     
==========================================
+ Hits         8216     8220       +4     
- Misses       1661     1662       +1
Impacted Files Coverage Δ
...c/gluonts/distribution/transformed_distribution.py 95.29% <100%> (+0.17%) ⬆️
src/gluonts/model/forecast.py 71.92% <33.33%> (-0.2%) ⬇️

@lostella lostella requested a review from jgasthaus December 20, 2019 19:24
Copy link
Contributor

@mbohlkeschneider mbohlkeschneider left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This PR reminds me that we actually use the wrong abstraction for the AffineTransformation transformed output. I think the right abstraction would be to have some kind of ConditionalBijection to model this (not for this PR, but we should think about this in the future).

return self.distr_cls(*distr_args)
else:
distr = self.distr_cls(*distr_args)
return TransformedDistribution(
distr, [AffineTransformation(scale=scale)]
distr, [AffineTransformation(loc=loc, scale=scale)]
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

loc is actually never passed in any of our models, right? See DeepAR here for example.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It's not indeed, and will default to None so long as it's not used

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

To expand on this a bit: the reason for having loc alongside scale is to allow using a more general scaling, like min-max scaling or normalizing data (in the true sense of the word) instead of just dividing by some scaling factor. And if this is doable in a minimally invasive, 100% backwards compatible manner why not

@lostella
Copy link
Contributor Author

lostella commented Jan 6, 2020

This PR reminds me that we actually use the wrong abstraction for the AffineTransformation transformed output. I think the right abstraction would be to have some kind of ConditionalBijection to model this (not for this PR, but we should think about this in the future).

@mbohlkeschneider yes, I think we could adjust a bit how distributions and transformations interact. Especially in the ubiquitous case of scaling, where we "manually" transform data before the model input layer, and use transformed distributions in the likelihood instead.

@lostella lostella merged commit a4e5584 into awslabs:master Jan 6, 2020
@lostella lostella deleted the distribution-output-loc branch January 6, 2020 16:53
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants