What are some situations when normalizing input data to zero mean, unit variance is not appropriate or not beneficial?
Clash Royale CLAN TAG#URR8PPP
up vote
7
down vote
favorite
I have seen normalization of input data to zero mean, unit variance many times in machine learning. Is this a good practice to be done all the time or are there times when it is not appropriate or not beneficial?
machine-learning feature-scaling normalization
add a comment |Â
up vote
7
down vote
favorite
I have seen normalization of input data to zero mean, unit variance many times in machine learning. Is this a good practice to be done all the time or are there times when it is not appropriate or not beneficial?
machine-learning feature-scaling normalization
add a comment |Â
up vote
7
down vote
favorite
up vote
7
down vote
favorite
I have seen normalization of input data to zero mean, unit variance many times in machine learning. Is this a good practice to be done all the time or are there times when it is not appropriate or not beneficial?
machine-learning feature-scaling normalization
I have seen normalization of input data to zero mean, unit variance many times in machine learning. Is this a good practice to be done all the time or are there times when it is not appropriate or not beneficial?
machine-learning feature-scaling normalization
machine-learning feature-scaling normalization
asked Sep 3 at 6:02
user781486
2627
2627
add a comment |Â
add a comment |Â
1 Answer
1
active
oldest
votes
up vote
6
down vote
accepted
A detailed answer to the question can be found here.
[...]are there times when it is not appropriate or not beneficial?
Short answer: Yes and No. Yes in the terms, that it can significantly change your output of e.g. clustering algorithms. No, on the other hand, if these changes are what you want to achieve. Or to put it in the words of the author of the mentioned source:
Scaling features for clustering algorithms can substantially change the outcome. Imagine four clusters around the origin, each one in a different quadrant, all nicely scaled. Now, imagine the y-axis being stretched to ten times the length of the the x-axis. instead of four little quadrant-clusters, you're going to get the long squashed baguette of data chopped into four pieces along its length! (And, the important part is, you might prefer either of these!)
The take-home-message of this is: always think carefully about what you want to achieve and what kind of data your algorithms prefer - it does matter!
PCA would, by the way, be one of the algorithms that do not want to be operated without normalization - just to highlight the other side of the story.
â André
Sep 3 at 8:42
add a comment |Â
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
6
down vote
accepted
A detailed answer to the question can be found here.
[...]are there times when it is not appropriate or not beneficial?
Short answer: Yes and No. Yes in the terms, that it can significantly change your output of e.g. clustering algorithms. No, on the other hand, if these changes are what you want to achieve. Or to put it in the words of the author of the mentioned source:
Scaling features for clustering algorithms can substantially change the outcome. Imagine four clusters around the origin, each one in a different quadrant, all nicely scaled. Now, imagine the y-axis being stretched to ten times the length of the the x-axis. instead of four little quadrant-clusters, you're going to get the long squashed baguette of data chopped into four pieces along its length! (And, the important part is, you might prefer either of these!)
The take-home-message of this is: always think carefully about what you want to achieve and what kind of data your algorithms prefer - it does matter!
PCA would, by the way, be one of the algorithms that do not want to be operated without normalization - just to highlight the other side of the story.
â André
Sep 3 at 8:42
add a comment |Â
up vote
6
down vote
accepted
A detailed answer to the question can be found here.
[...]are there times when it is not appropriate or not beneficial?
Short answer: Yes and No. Yes in the terms, that it can significantly change your output of e.g. clustering algorithms. No, on the other hand, if these changes are what you want to achieve. Or to put it in the words of the author of the mentioned source:
Scaling features for clustering algorithms can substantially change the outcome. Imagine four clusters around the origin, each one in a different quadrant, all nicely scaled. Now, imagine the y-axis being stretched to ten times the length of the the x-axis. instead of four little quadrant-clusters, you're going to get the long squashed baguette of data chopped into four pieces along its length! (And, the important part is, you might prefer either of these!)
The take-home-message of this is: always think carefully about what you want to achieve and what kind of data your algorithms prefer - it does matter!
PCA would, by the way, be one of the algorithms that do not want to be operated without normalization - just to highlight the other side of the story.
â André
Sep 3 at 8:42
add a comment |Â
up vote
6
down vote
accepted
up vote
6
down vote
accepted
A detailed answer to the question can be found here.
[...]are there times when it is not appropriate or not beneficial?
Short answer: Yes and No. Yes in the terms, that it can significantly change your output of e.g. clustering algorithms. No, on the other hand, if these changes are what you want to achieve. Or to put it in the words of the author of the mentioned source:
Scaling features for clustering algorithms can substantially change the outcome. Imagine four clusters around the origin, each one in a different quadrant, all nicely scaled. Now, imagine the y-axis being stretched to ten times the length of the the x-axis. instead of four little quadrant-clusters, you're going to get the long squashed baguette of data chopped into four pieces along its length! (And, the important part is, you might prefer either of these!)
The take-home-message of this is: always think carefully about what you want to achieve and what kind of data your algorithms prefer - it does matter!
A detailed answer to the question can be found here.
[...]are there times when it is not appropriate or not beneficial?
Short answer: Yes and No. Yes in the terms, that it can significantly change your output of e.g. clustering algorithms. No, on the other hand, if these changes are what you want to achieve. Or to put it in the words of the author of the mentioned source:
Scaling features for clustering algorithms can substantially change the outcome. Imagine four clusters around the origin, each one in a different quadrant, all nicely scaled. Now, imagine the y-axis being stretched to ten times the length of the the x-axis. instead of four little quadrant-clusters, you're going to get the long squashed baguette of data chopped into four pieces along its length! (And, the important part is, you might prefer either of these!)
The take-home-message of this is: always think carefully about what you want to achieve and what kind of data your algorithms prefer - it does matter!
edited Sep 5 at 17:37
answered Sep 3 at 8:40
André
3508
3508
PCA would, by the way, be one of the algorithms that do not want to be operated without normalization - just to highlight the other side of the story.
â André
Sep 3 at 8:42
add a comment |Â
PCA would, by the way, be one of the algorithms that do not want to be operated without normalization - just to highlight the other side of the story.
â André
Sep 3 at 8:42
PCA would, by the way, be one of the algorithms that do not want to be operated without normalization - just to highlight the other side of the story.
â André
Sep 3 at 8:42
PCA would, by the way, be one of the algorithms that do not want to be operated without normalization - just to highlight the other side of the story.
â André
Sep 3 at 8:42
add a comment |Â
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f37734%2fwhat-are-some-situations-when-normalizing-input-data-to-zero-mean-unit-variance%23new-answer', 'question_page');
);
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password