Why would one crossvalidate the random state number?Linear kernel in SVM performing much worse than RBF or PolyWhy is the number of samples smaller than the number of values in my decision tree?How does one fine-tune parameters and weights at the same time?Predicting contract churn/cancellation: Great model results does not work in the real worldHow to choose the random seed?Why would a fake feature with random numbers get selected in feature importance?Random state in machine learning modelsRandom Forest, Duplicating Data increases Accuracy. Why?Why is the reported loss different from the mean squared error calculated on the train data?Why is my MLP with 2 features is doing worse than MLP with 1 feature where the one feature is a combination of feature1*feature2?

How to explain intravenous drug abuse to a 6-year-old?

Should one save up to purchase a house/condo or maximize their 401(k) first?

Why is there a cap on 401k contributions?

Using mean length and mean weight to calculate mean BMI?

Expl3 and recent xparse on overleaf: No expl3 loader detected

How long can fsck take on a 30 TB volume?

Trying to understand a summation

Company stopped paying my salary. What are my options?

Is it a good idea to copy a trader when investing?

What are these pads?

What is the Ancient One's mistake?

Employee is self-centered and affects the team negatively

How to avoid making self and former employee look bad when reporting on fixing former employee's work?

Is the tensor product (of vector spaces) commutative?

Align a table column at a specific symbol

Do these creatures from the Tomb of Annihilation campaign speak Common?

Why doesn't a particle exert force on itself?

How do I give a darkroom course without negatives from the attendees?

Why doesn't Dany protect her dragons better?

Are there vaccine ingredients which may not be disclosed ("hidden", "trade secret", or similar)?

Names of the Six Tastes

Every group the homology of some space?

Examples where existence is harder than evaluation

Why is it wrong to *implement* myself a known, published, widely believed to be secure crypto algorithm?



Why would one crossvalidate the random state number?


Linear kernel in SVM performing much worse than RBF or PolyWhy is the number of samples smaller than the number of values in my decision tree?How does one fine-tune parameters and weights at the same time?Predicting contract churn/cancellation: Great model results does not work in the real worldHow to choose the random seed?Why would a fake feature with random numbers get selected in feature importance?Random state in machine learning modelsRandom Forest, Duplicating Data increases Accuracy. Why?Why is the reported loss different from the mean squared error calculated on the train data?Why is my MLP with 2 features is doing worse than MLP with 1 feature where the one feature is a combination of feature1*feature2?













3












$begingroup$


Still learning about machine learning, I've stumbled across a kaggle (link), which I cannot understand.



Here are lines 72 and 73:



parameters = 'solver': ['lbfgs'], 
'max_iter': [1000,1100,1200,1300,1400,1500,1600,1700,1800,1900,2000 ],
'alpha': 10.0 ** -np.arange(1, 10),
'hidden_layer_sizes':np.arange(10, 15),
'random_state':[0,1,2,3,4,5,6,7,8,9]
clf = GridSearchCV(MLPClassifier(), parameters, n_jobs=-1)


As you can see, the random_state parameter is been tested across 10 values.



What is the point of doing this?



If one model perform better with some random_state, does it make any sense to use this particular parameter on other models?










share|improve this question











$endgroup$
















    3












    $begingroup$


    Still learning about machine learning, I've stumbled across a kaggle (link), which I cannot understand.



    Here are lines 72 and 73:



    parameters = 'solver': ['lbfgs'], 
    'max_iter': [1000,1100,1200,1300,1400,1500,1600,1700,1800,1900,2000 ],
    'alpha': 10.0 ** -np.arange(1, 10),
    'hidden_layer_sizes':np.arange(10, 15),
    'random_state':[0,1,2,3,4,5,6,7,8,9]
    clf = GridSearchCV(MLPClassifier(), parameters, n_jobs=-1)


    As you can see, the random_state parameter is been tested across 10 values.



    What is the point of doing this?



    If one model perform better with some random_state, does it make any sense to use this particular parameter on other models?










    share|improve this question











    $endgroup$














      3












      3








      3





      $begingroup$


      Still learning about machine learning, I've stumbled across a kaggle (link), which I cannot understand.



      Here are lines 72 and 73:



      parameters = 'solver': ['lbfgs'], 
      'max_iter': [1000,1100,1200,1300,1400,1500,1600,1700,1800,1900,2000 ],
      'alpha': 10.0 ** -np.arange(1, 10),
      'hidden_layer_sizes':np.arange(10, 15),
      'random_state':[0,1,2,3,4,5,6,7,8,9]
      clf = GridSearchCV(MLPClassifier(), parameters, n_jobs=-1)


      As you can see, the random_state parameter is been tested across 10 values.



      What is the point of doing this?



      If one model perform better with some random_state, does it make any sense to use this particular parameter on other models?










      share|improve this question











      $endgroup$




      Still learning about machine learning, I've stumbled across a kaggle (link), which I cannot understand.



      Here are lines 72 and 73:



      parameters = 'solver': ['lbfgs'], 
      'max_iter': [1000,1100,1200,1300,1400,1500,1600,1700,1800,1900,2000 ],
      'alpha': 10.0 ** -np.arange(1, 10),
      'hidden_layer_sizes':np.arange(10, 15),
      'random_state':[0,1,2,3,4,5,6,7,8,9]
      clf = GridSearchCV(MLPClassifier(), parameters, n_jobs=-1)


      As you can see, the random_state parameter is been tested across 10 values.



      What is the point of doing this?



      If one model perform better with some random_state, does it make any sense to use this particular parameter on other models?







      python scikit-learn mlp randomized-algorithms






      share|improve this question















      share|improve this question













      share|improve this question




      share|improve this question








      edited 2 days ago









      n1k31t4

      7,0862423




      7,0862423










      asked May 4 at 19:09









      Dan ChaltielDan Chaltiel

      1808




      1808




















          1 Answer
          1






          active

          oldest

          votes


















          8












          $begingroup$

          I personally think that the general idea of optimising your model with different random seeds is not a good idea. There are many other, more important, aspects of the modelling process that you can worry about, tweak and compare before spending time on the effects of random initialisation.



          That being said, if you just want to test the effect of random initialisation of model weights on a final validation metric, this could be an approach to do so. Kind of the reverse argument to my point above. If you can show for different random seeds (ceteris paribus: with all other parameters equal) that the final model performs differently, it shows maybe that their is either inconsistency in the model, or a bug in the code even. I would not expect a well-validated model to give hugely differing results if being run with a different random seed, so if it does, it tells me something weird is going on!






          share|improve this answer









          $endgroup$













            Your Answer








            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "557"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: false,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: null,
            bindNavPrevention: true,
            postfix: "",
            imageUploader:
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            ,
            onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );













            draft saved

            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f51397%2fwhy-would-one-crossvalidate-the-random-state-number%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown

























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            8












            $begingroup$

            I personally think that the general idea of optimising your model with different random seeds is not a good idea. There are many other, more important, aspects of the modelling process that you can worry about, tweak and compare before spending time on the effects of random initialisation.



            That being said, if you just want to test the effect of random initialisation of model weights on a final validation metric, this could be an approach to do so. Kind of the reverse argument to my point above. If you can show for different random seeds (ceteris paribus: with all other parameters equal) that the final model performs differently, it shows maybe that their is either inconsistency in the model, or a bug in the code even. I would not expect a well-validated model to give hugely differing results if being run with a different random seed, so if it does, it tells me something weird is going on!






            share|improve this answer









            $endgroup$

















              8












              $begingroup$

              I personally think that the general idea of optimising your model with different random seeds is not a good idea. There are many other, more important, aspects of the modelling process that you can worry about, tweak and compare before spending time on the effects of random initialisation.



              That being said, if you just want to test the effect of random initialisation of model weights on a final validation metric, this could be an approach to do so. Kind of the reverse argument to my point above. If you can show for different random seeds (ceteris paribus: with all other parameters equal) that the final model performs differently, it shows maybe that their is either inconsistency in the model, or a bug in the code even. I would not expect a well-validated model to give hugely differing results if being run with a different random seed, so if it does, it tells me something weird is going on!






              share|improve this answer









              $endgroup$















                8












                8








                8





                $begingroup$

                I personally think that the general idea of optimising your model with different random seeds is not a good idea. There are many other, more important, aspects of the modelling process that you can worry about, tweak and compare before spending time on the effects of random initialisation.



                That being said, if you just want to test the effect of random initialisation of model weights on a final validation metric, this could be an approach to do so. Kind of the reverse argument to my point above. If you can show for different random seeds (ceteris paribus: with all other parameters equal) that the final model performs differently, it shows maybe that their is either inconsistency in the model, or a bug in the code even. I would not expect a well-validated model to give hugely differing results if being run with a different random seed, so if it does, it tells me something weird is going on!






                share|improve this answer









                $endgroup$



                I personally think that the general idea of optimising your model with different random seeds is not a good idea. There are many other, more important, aspects of the modelling process that you can worry about, tweak and compare before spending time on the effects of random initialisation.



                That being said, if you just want to test the effect of random initialisation of model weights on a final validation metric, this could be an approach to do so. Kind of the reverse argument to my point above. If you can show for different random seeds (ceteris paribus: with all other parameters equal) that the final model performs differently, it shows maybe that their is either inconsistency in the model, or a bug in the code even. I would not expect a well-validated model to give hugely differing results if being run with a different random seed, so if it does, it tells me something weird is going on!







                share|improve this answer












                share|improve this answer



                share|improve this answer










                answered May 4 at 19:18









                n1k31t4n1k31t4

                7,0862423




                7,0862423



























                    draft saved

                    draft discarded
















































                    Thanks for contributing an answer to Data Science Stack Exchange!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid


                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.

                    Use MathJax to format equations. MathJax reference.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f51397%2fwhy-would-one-crossvalidate-the-random-state-number%23new-answer', 'question_page');

                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Category:9 (number) SubcategoriesMedia in category "9 (number)"Navigation menuUpload mediaGND ID: 4485639-8Library of Congress authority ID: sh85091979ReasonatorScholiaStatistics

                    Circuit construction for execution of conditional statements using least significant bitHow are two different registers being used as “control”?How exactly is the stated composite state of the two registers being produced using the $R_zz$ controlled rotations?Efficiently performing controlled rotations in HHLWould this quantum algorithm implementation work?How to prepare a superposed states of odd integers from $1$ to $sqrtN$?Why is this implementation of the order finding algorithm not working?Circuit construction for Hamiltonian simulationHow can I invert the least significant bit of a certain term of a superposed state?Implementing an oracleImplementing a controlled sum operation

                    Magento 2 “No Payment Methods” in Admin New OrderHow to integrate Paypal Express Checkout with the Magento APIMagento 1.5 - Sales > Order > edit order and shipping methods disappearAuto Invoice Check/Money Order Payment methodAdd more simple payment methods?Shipping methods not showingWhat should I do to change payment methods if changing the configuration has no effects?1.9 - No Payment Methods showing upMy Payment Methods not Showing for downloadable/virtual product when checkout?Magento2 API to access internal payment methodHow to call an existing payment methods in the registration form?