“Best practices” for formulating MIPsBranching rules in commercial MIP solversRecommended books/materials for practical applications of Operations Research in industryWhat are the tradeoffs between “exact” and Reinforcement Learning methods for solving optimization problemsHow to decide to write an objective function?When to use indicator constraints versus big-M approaches in solving (mixed-)integer programsWhy is it important to choose big-M carefully and what are the consequences of doing it badly?Java source code for branch and priceHow to get the best bound of large LP problems in CPLEX?Correct way to get a dual extreme ray for an infeasible LP in CPLEX / C++Heuristics for mixed integer linear and nonlinear programs

Don't the events of "Forest of the Dead" contradict the fixed point in "The Wedding of River Song"?

When I press the space bar it deletes the letters in front of it

What was this character's plan?

When an electron changes its spin, or any other intrinsic property, is it still the same electron?

What happens to unproductive professors?

Is it okay to use open source code to do an interview task?

Generalized Behrend version for Grothendieck-Lefschetz trace formula

A sequence that changes sign finally at infinity?

Why is the ladder of the LM always in the dark side of the LM?

Distance between horizontal tree levels

How do I explain that I don't want to maintain old projects?

Non-Chromatic Orchestral Instruments?

A Logic Puzzle—One of These Things Doesn’t Quite Fit

How does one acquire an undead eyeball encased in a gem?

How should I ask for a "pint" in countries that use metric?

My previous employer committed a severe violation of the law and is also being sued by me. How do I explain the situation to future employers?

This LM317 diagram doesn't make any sense to me

No Torah = Revert to Nothingness?

Why is the Cauchy Distribution is so useful?

Why different specifications for telescopes and binoculars?

What's it called when the bad guy gets eaten?

Can a landlord force all residents to use the landlord's in-house debit card accounts?

Is it ok for parents to kiss and romance with each other while their 2- to 8-year-old child watches?

Why is a mixture of two normally distributed variables only bimodal if their means differ by at least two times the common standard deviation?



“Best practices” for formulating MIPs


Branching rules in commercial MIP solversRecommended books/materials for practical applications of Operations Research in industryWhat are the tradeoffs between “exact” and Reinforcement Learning methods for solving optimization problemsHow to decide to write an objective function?When to use indicator constraints versus big-M approaches in solving (mixed-)integer programsWhy is it important to choose big-M carefully and what are the consequences of doing it badly?Java source code for branch and priceHow to get the best bound of large LP problems in CPLEX?Correct way to get a dual extreme ray for an infeasible LP in CPLEX / C++Heuristics for mixed integer linear and nonlinear programs













20












$begingroup$


Often there are many alternatives ways for formulating a MIP. For example:



  1. The model contains inequality constraints that must hold with equality in an optimal solution.

  2. The model contains continuous variables that will necessarily be integer in an optimal (or just integer feasible) solution.

  3. The model contains variables that have a value of at most 1 in an optimal solution.

In these cases, the modeler can decide whether to (1) include the constraints as equality constraints or not, (2) declare the variables as integer or continuous variables and (3) to declare the variables with upper bounds of 1 or not. When using commercial solvers, I have noticed that the performance may vary significantly dependent on such choices. Hence, I am wondering whether there are any rules of thumb for formulating MIPs in solvers. More generally, what information is valuable to solvers and what information may even hurt performance?










share|improve this question









$endgroup$
















    20












    $begingroup$


    Often there are many alternatives ways for formulating a MIP. For example:



    1. The model contains inequality constraints that must hold with equality in an optimal solution.

    2. The model contains continuous variables that will necessarily be integer in an optimal (or just integer feasible) solution.

    3. The model contains variables that have a value of at most 1 in an optimal solution.

    In these cases, the modeler can decide whether to (1) include the constraints as equality constraints or not, (2) declare the variables as integer or continuous variables and (3) to declare the variables with upper bounds of 1 or not. When using commercial solvers, I have noticed that the performance may vary significantly dependent on such choices. Hence, I am wondering whether there are any rules of thumb for formulating MIPs in solvers. More generally, what information is valuable to solvers and what information may even hurt performance?










    share|improve this question









    $endgroup$














      20












      20








      20


      2



      $begingroup$


      Often there are many alternatives ways for formulating a MIP. For example:



      1. The model contains inequality constraints that must hold with equality in an optimal solution.

      2. The model contains continuous variables that will necessarily be integer in an optimal (or just integer feasible) solution.

      3. The model contains variables that have a value of at most 1 in an optimal solution.

      In these cases, the modeler can decide whether to (1) include the constraints as equality constraints or not, (2) declare the variables as integer or continuous variables and (3) to declare the variables with upper bounds of 1 or not. When using commercial solvers, I have noticed that the performance may vary significantly dependent on such choices. Hence, I am wondering whether there are any rules of thumb for formulating MIPs in solvers. More generally, what information is valuable to solvers and what information may even hurt performance?










      share|improve this question









      $endgroup$




      Often there are many alternatives ways for formulating a MIP. For example:



      1. The model contains inequality constraints that must hold with equality in an optimal solution.

      2. The model contains continuous variables that will necessarily be integer in an optimal (or just integer feasible) solution.

      3. The model contains variables that have a value of at most 1 in an optimal solution.

      In these cases, the modeler can decide whether to (1) include the constraints as equality constraints or not, (2) declare the variables as integer or continuous variables and (3) to declare the variables with upper bounds of 1 or not. When using commercial solvers, I have noticed that the performance may vary significantly dependent on such choices. Hence, I am wondering whether there are any rules of thumb for formulating MIPs in solvers. More generally, what information is valuable to solvers and what information may even hurt performance?







      mixed-integer-programming linear-programming solver cplex






      share|improve this question













      share|improve this question











      share|improve this question




      share|improve this question










      asked Jun 30 at 12:10









      Rolf van LieshoutRolf van Lieshout

      2059 bronze badges




      2059 bronze badges




















          4 Answers
          4






          active

          oldest

          votes


















          12












          $begingroup$

          This is an extremely interesting question. I agree with @Richard that you have to try it out. I have seen that tiny changes to a model can make huge differences, but in my experience, more general changes to a model may have more impact in the end. There are, I think, some guidelines that may help, from algorithmics and theory.



          Why do we choose "big $M$ as small as possible" (so, in fact, we look for small $M$)? Because we know that the LP based B&B algorithm which solves our MILPs can benefit from a better bound. Then again, since we know that big $M$ usually gives horrible relaxations, we try to avoid it altogether. A reformulation may help, maybe at the expense of investing more variables. Maybe too many, in that case, you would need a different algorithm (like dynamic variable or constraint generation). In my view, the model and the algorithm (may) go hand in hand. General observations:



          • use binary variables instead of general integers

          • avoid symmetric formulations

          • think of completely different representations of a solution (eg in the cutting stock problem, variables can represent (a) how often an item is cut from a particular roll, or (b) an entire cutting pattern given as non-negative integer numbers, or (c) an entire cutting pattern given as a path in an acyclic network, or or or.

          • look at the LP solution in the root node: anything you "see" that the LP is allowed to do that would be forbidden for the IP? May give an extra constraint. Or a cutting plane.

          It may happen that the vanialla solver on an ugly model is still faster than a specialized solver on the elaborate model, and this is frustrating at first, but then we realize that this is what drives solver development.






          share|improve this answer











          $endgroup$








          • 1




            $begingroup$
            Are there any guidelines for trading off between binary vs general integer variables? Let's say I have two formulations, one with $n$ general integer variables and the other with $N$ binary variables ($N > n$). Are there guidelines for which formulation is preferred, maybe as a function of $N/n$ or the range of the general integers (say, $0,ldots,k$)? The answer likely depends on the number of constraints, so for simplicity, let's say the two formulations have around the same number of constraints.
            $endgroup$
            – Aleksandr Kazachkov
            Jul 2 at 18:38











          • $begingroup$
            since for binary variables so much logic kicks in (propagation etc.) I would assume you always choose the binary program; do you have an example where an integer formulation is better, regardless of size?
            $endgroup$
            – Marco Lübbecke
            Jul 2 at 19:54






          • 1




            $begingroup$
            This decision arose recently for me, but I did not test the two formulations against each other, and I am not sure it will make sense to do so in my setting, so the question was more from curiosity / theoretical interest. I formulated my problem with binary variables, but later found an earlier paper that worked on the same problem, had very experienced authors, and used the general integer formulation. I probably will try to implement the general integer version and see how it goes, though in the end I designed a bunch of heuristics that make the IPs typically trivial after presolve anyway.
            $endgroup$
            – Aleksandr Kazachkov
            Jul 2 at 21:10


















          15












          $begingroup$

          I’m assuming that we want our models to be solved as quickly as possible. If that is the case, then the honest answer is: you need to try the models out and see.



          To give you a concrete example (see here): through what I thought was a super-clever reformulation, I was able to remove 85% of the variables in the problem, and I thought that this would make it super fast. However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not.



          I always try to put the tightest upper bound I have, make my big-M as small as possible, use continuous over binary/integer when I can, but honestly, you cannot know anything for sure until you test it.






          share|improve this answer









          $endgroup$












          • $begingroup$
            On However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not. Vs or.stackexchange.com/users/354/marco-l%c3%bcbbecke's answer avoid symmetric formulations I am extremely confused.
            $endgroup$
            – abhiieor
            Jul 1 at 8:57










          • $begingroup$
            That’s exactly why you should test: sometimes symmetry really kills your performance, because the B&B doesn’t “know” where to go, another time it helps unearth underlying structures and speed up the performance. I was totally confused by that result (which is why I wrote the blog post about it) .Basically, you have to try.
            $endgroup$
            – Richard
            Jul 1 at 10:58










          • $begingroup$
            Maybe another rule of thumb: symmetry in the objective function is bad. Symmetry in variables might be good.
            $endgroup$
            – Richard
            Jul 1 at 11:01


















          8












          $begingroup$

          Perhaps you are only talking about MILPs, but you don't say, so I will say something of interest for MINLPs.



          Regarding point 1:



          • Suppose your model has a convex nonlinear inequality constraint, $f(x) le c$, i.e., $f(x)$ is convex.

          and



          • It is known that the constraint will be active at the optimum, of any continuous relaxation i.e., $f(x) = c$, for instance because the continuous relaxations are concave programs with compact constraints, thereby having global optimum at an extreme of the constraints

          Then you are usually better off entering the constraint as an inequality rather than as an equality. That is because the inequality constraint is convex, and generally much easier for an optimizer to deal with than a nonlinear equality constraint, which is a non-convex constraint. You might think you'd be doing the optimizer a favor by specifying the equality constraint, thereby cleaving off $f(x) < c$ from consideration, but such a "favor" will usually backfire.



          However, in the case of a linear equality constraint known to hold with equality at the optimum of the original problem, or a continuous relaxation, I think it would generally be better, and in any event, not worse (except for (un)lucky "stars aligning" cases), to specify such a linear constraint as an equality constraint.



          Regarding point 3, I have a pretty hard time imagining it would be better to not provide an upper bound of 1 than to provide it when you know the optimal variable value is at most 1. But I suppose for some particular model and solver, the stars could always align such that specifying such an upper bound slowed things down.



          I will leave others to comment on point 2.






          share|improve this answer











          $endgroup$








          • 2




            $begingroup$
            Mark is dead on. Also, regarding the inequality v. equality question, sometimes rounding error becomes a hazard. With an equality constraint, deviating from the RHS by more than the solver's tolerance in either direction can make a feasible solution look infeasible. With an inequality constraint, the hazard is only one one side, so you have possibly cut the risk of it.
            $endgroup$
            – prubin
            Jun 30 at 17:51










          • $begingroup$
            Another argument I have come along in LP contexts for using inequality constraints rather than equality constraints is that by using inequality constraints, the dual variables are restricted to the positive domain, which may accelerate the solution process. I think this is why people often use a set covering formulation rather than a set partitioning formulation, even though all covering constraints hold with equality in an optimal solution.
            $endgroup$
            – Rolf van Lieshout
            Jul 1 at 12:34


















          6












          $begingroup$

          In addition to previous answers, I like to cite the following book, which discuss in details several different options (and kind of good practices) to model different common applications:



          Model Building in Mathematical Programming, by H. Paul Williams. March, 2013.
          ISBN: 978-1-118-44333-0.



          Every chapter is very interesting and is worth of (re)reading several times.






          share|improve this answer








          New contributor



          Stefano Gualandi is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
          Check out our Code of Conduct.





          $endgroup$















            Your Answer








            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "700"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: false,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: null,
            bindNavPrevention: true,
            postfix: "",
            imageUploader:
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            ,
            noCode: true, onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );













            draft saved

            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2for.stackexchange.com%2fquestions%2f834%2fbest-practices-for-formulating-mips%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown

























            4 Answers
            4






            active

            oldest

            votes








            4 Answers
            4






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            12












            $begingroup$

            This is an extremely interesting question. I agree with @Richard that you have to try it out. I have seen that tiny changes to a model can make huge differences, but in my experience, more general changes to a model may have more impact in the end. There are, I think, some guidelines that may help, from algorithmics and theory.



            Why do we choose "big $M$ as small as possible" (so, in fact, we look for small $M$)? Because we know that the LP based B&B algorithm which solves our MILPs can benefit from a better bound. Then again, since we know that big $M$ usually gives horrible relaxations, we try to avoid it altogether. A reformulation may help, maybe at the expense of investing more variables. Maybe too many, in that case, you would need a different algorithm (like dynamic variable or constraint generation). In my view, the model and the algorithm (may) go hand in hand. General observations:



            • use binary variables instead of general integers

            • avoid symmetric formulations

            • think of completely different representations of a solution (eg in the cutting stock problem, variables can represent (a) how often an item is cut from a particular roll, or (b) an entire cutting pattern given as non-negative integer numbers, or (c) an entire cutting pattern given as a path in an acyclic network, or or or.

            • look at the LP solution in the root node: anything you "see" that the LP is allowed to do that would be forbidden for the IP? May give an extra constraint. Or a cutting plane.

            It may happen that the vanialla solver on an ugly model is still faster than a specialized solver on the elaborate model, and this is frustrating at first, but then we realize that this is what drives solver development.






            share|improve this answer











            $endgroup$








            • 1




              $begingroup$
              Are there any guidelines for trading off between binary vs general integer variables? Let's say I have two formulations, one with $n$ general integer variables and the other with $N$ binary variables ($N > n$). Are there guidelines for which formulation is preferred, maybe as a function of $N/n$ or the range of the general integers (say, $0,ldots,k$)? The answer likely depends on the number of constraints, so for simplicity, let's say the two formulations have around the same number of constraints.
              $endgroup$
              – Aleksandr Kazachkov
              Jul 2 at 18:38











            • $begingroup$
              since for binary variables so much logic kicks in (propagation etc.) I would assume you always choose the binary program; do you have an example where an integer formulation is better, regardless of size?
              $endgroup$
              – Marco Lübbecke
              Jul 2 at 19:54






            • 1




              $begingroup$
              This decision arose recently for me, but I did not test the two formulations against each other, and I am not sure it will make sense to do so in my setting, so the question was more from curiosity / theoretical interest. I formulated my problem with binary variables, but later found an earlier paper that worked on the same problem, had very experienced authors, and used the general integer formulation. I probably will try to implement the general integer version and see how it goes, though in the end I designed a bunch of heuristics that make the IPs typically trivial after presolve anyway.
              $endgroup$
              – Aleksandr Kazachkov
              Jul 2 at 21:10















            12












            $begingroup$

            This is an extremely interesting question. I agree with @Richard that you have to try it out. I have seen that tiny changes to a model can make huge differences, but in my experience, more general changes to a model may have more impact in the end. There are, I think, some guidelines that may help, from algorithmics and theory.



            Why do we choose "big $M$ as small as possible" (so, in fact, we look for small $M$)? Because we know that the LP based B&B algorithm which solves our MILPs can benefit from a better bound. Then again, since we know that big $M$ usually gives horrible relaxations, we try to avoid it altogether. A reformulation may help, maybe at the expense of investing more variables. Maybe too many, in that case, you would need a different algorithm (like dynamic variable or constraint generation). In my view, the model and the algorithm (may) go hand in hand. General observations:



            • use binary variables instead of general integers

            • avoid symmetric formulations

            • think of completely different representations of a solution (eg in the cutting stock problem, variables can represent (a) how often an item is cut from a particular roll, or (b) an entire cutting pattern given as non-negative integer numbers, or (c) an entire cutting pattern given as a path in an acyclic network, or or or.

            • look at the LP solution in the root node: anything you "see" that the LP is allowed to do that would be forbidden for the IP? May give an extra constraint. Or a cutting plane.

            It may happen that the vanialla solver on an ugly model is still faster than a specialized solver on the elaborate model, and this is frustrating at first, but then we realize that this is what drives solver development.






            share|improve this answer











            $endgroup$








            • 1




              $begingroup$
              Are there any guidelines for trading off between binary vs general integer variables? Let's say I have two formulations, one with $n$ general integer variables and the other with $N$ binary variables ($N > n$). Are there guidelines for which formulation is preferred, maybe as a function of $N/n$ or the range of the general integers (say, $0,ldots,k$)? The answer likely depends on the number of constraints, so for simplicity, let's say the two formulations have around the same number of constraints.
              $endgroup$
              – Aleksandr Kazachkov
              Jul 2 at 18:38











            • $begingroup$
              since for binary variables so much logic kicks in (propagation etc.) I would assume you always choose the binary program; do you have an example where an integer formulation is better, regardless of size?
              $endgroup$
              – Marco Lübbecke
              Jul 2 at 19:54






            • 1




              $begingroup$
              This decision arose recently for me, but I did not test the two formulations against each other, and I am not sure it will make sense to do so in my setting, so the question was more from curiosity / theoretical interest. I formulated my problem with binary variables, but later found an earlier paper that worked on the same problem, had very experienced authors, and used the general integer formulation. I probably will try to implement the general integer version and see how it goes, though in the end I designed a bunch of heuristics that make the IPs typically trivial after presolve anyway.
              $endgroup$
              – Aleksandr Kazachkov
              Jul 2 at 21:10













            12












            12








            12





            $begingroup$

            This is an extremely interesting question. I agree with @Richard that you have to try it out. I have seen that tiny changes to a model can make huge differences, but in my experience, more general changes to a model may have more impact in the end. There are, I think, some guidelines that may help, from algorithmics and theory.



            Why do we choose "big $M$ as small as possible" (so, in fact, we look for small $M$)? Because we know that the LP based B&B algorithm which solves our MILPs can benefit from a better bound. Then again, since we know that big $M$ usually gives horrible relaxations, we try to avoid it altogether. A reformulation may help, maybe at the expense of investing more variables. Maybe too many, in that case, you would need a different algorithm (like dynamic variable or constraint generation). In my view, the model and the algorithm (may) go hand in hand. General observations:



            • use binary variables instead of general integers

            • avoid symmetric formulations

            • think of completely different representations of a solution (eg in the cutting stock problem, variables can represent (a) how often an item is cut from a particular roll, or (b) an entire cutting pattern given as non-negative integer numbers, or (c) an entire cutting pattern given as a path in an acyclic network, or or or.

            • look at the LP solution in the root node: anything you "see" that the LP is allowed to do that would be forbidden for the IP? May give an extra constraint. Or a cutting plane.

            It may happen that the vanialla solver on an ugly model is still faster than a specialized solver on the elaborate model, and this is frustrating at first, but then we realize that this is what drives solver development.






            share|improve this answer











            $endgroup$



            This is an extremely interesting question. I agree with @Richard that you have to try it out. I have seen that tiny changes to a model can make huge differences, but in my experience, more general changes to a model may have more impact in the end. There are, I think, some guidelines that may help, from algorithmics and theory.



            Why do we choose "big $M$ as small as possible" (so, in fact, we look for small $M$)? Because we know that the LP based B&B algorithm which solves our MILPs can benefit from a better bound. Then again, since we know that big $M$ usually gives horrible relaxations, we try to avoid it altogether. A reformulation may help, maybe at the expense of investing more variables. Maybe too many, in that case, you would need a different algorithm (like dynamic variable or constraint generation). In my view, the model and the algorithm (may) go hand in hand. General observations:



            • use binary variables instead of general integers

            • avoid symmetric formulations

            • think of completely different representations of a solution (eg in the cutting stock problem, variables can represent (a) how often an item is cut from a particular roll, or (b) an entire cutting pattern given as non-negative integer numbers, or (c) an entire cutting pattern given as a path in an acyclic network, or or or.

            • look at the LP solution in the root node: anything you "see" that the LP is allowed to do that would be forbidden for the IP? May give an extra constraint. Or a cutting plane.

            It may happen that the vanialla solver on an ugly model is still faster than a specialized solver on the elaborate model, and this is frustrating at first, but then we realize that this is what drives solver development.







            share|improve this answer














            share|improve this answer



            share|improve this answer








            edited Jul 1 at 8:33

























            answered Jul 1 at 7:30









            Marco LübbeckeMarco Lübbecke

            1,3151 silver badge18 bronze badges




            1,3151 silver badge18 bronze badges







            • 1




              $begingroup$
              Are there any guidelines for trading off between binary vs general integer variables? Let's say I have two formulations, one with $n$ general integer variables and the other with $N$ binary variables ($N > n$). Are there guidelines for which formulation is preferred, maybe as a function of $N/n$ or the range of the general integers (say, $0,ldots,k$)? The answer likely depends on the number of constraints, so for simplicity, let's say the two formulations have around the same number of constraints.
              $endgroup$
              – Aleksandr Kazachkov
              Jul 2 at 18:38











            • $begingroup$
              since for binary variables so much logic kicks in (propagation etc.) I would assume you always choose the binary program; do you have an example where an integer formulation is better, regardless of size?
              $endgroup$
              – Marco Lübbecke
              Jul 2 at 19:54






            • 1




              $begingroup$
              This decision arose recently for me, but I did not test the two formulations against each other, and I am not sure it will make sense to do so in my setting, so the question was more from curiosity / theoretical interest. I formulated my problem with binary variables, but later found an earlier paper that worked on the same problem, had very experienced authors, and used the general integer formulation. I probably will try to implement the general integer version and see how it goes, though in the end I designed a bunch of heuristics that make the IPs typically trivial after presolve anyway.
              $endgroup$
              – Aleksandr Kazachkov
              Jul 2 at 21:10












            • 1




              $begingroup$
              Are there any guidelines for trading off between binary vs general integer variables? Let's say I have two formulations, one with $n$ general integer variables and the other with $N$ binary variables ($N > n$). Are there guidelines for which formulation is preferred, maybe as a function of $N/n$ or the range of the general integers (say, $0,ldots,k$)? The answer likely depends on the number of constraints, so for simplicity, let's say the two formulations have around the same number of constraints.
              $endgroup$
              – Aleksandr Kazachkov
              Jul 2 at 18:38











            • $begingroup$
              since for binary variables so much logic kicks in (propagation etc.) I would assume you always choose the binary program; do you have an example where an integer formulation is better, regardless of size?
              $endgroup$
              – Marco Lübbecke
              Jul 2 at 19:54






            • 1




              $begingroup$
              This decision arose recently for me, but I did not test the two formulations against each other, and I am not sure it will make sense to do so in my setting, so the question was more from curiosity / theoretical interest. I formulated my problem with binary variables, but later found an earlier paper that worked on the same problem, had very experienced authors, and used the general integer formulation. I probably will try to implement the general integer version and see how it goes, though in the end I designed a bunch of heuristics that make the IPs typically trivial after presolve anyway.
              $endgroup$
              – Aleksandr Kazachkov
              Jul 2 at 21:10







            1




            1




            $begingroup$
            Are there any guidelines for trading off between binary vs general integer variables? Let's say I have two formulations, one with $n$ general integer variables and the other with $N$ binary variables ($N > n$). Are there guidelines for which formulation is preferred, maybe as a function of $N/n$ or the range of the general integers (say, $0,ldots,k$)? The answer likely depends on the number of constraints, so for simplicity, let's say the two formulations have around the same number of constraints.
            $endgroup$
            – Aleksandr Kazachkov
            Jul 2 at 18:38





            $begingroup$
            Are there any guidelines for trading off between binary vs general integer variables? Let's say I have two formulations, one with $n$ general integer variables and the other with $N$ binary variables ($N > n$). Are there guidelines for which formulation is preferred, maybe as a function of $N/n$ or the range of the general integers (say, $0,ldots,k$)? The answer likely depends on the number of constraints, so for simplicity, let's say the two formulations have around the same number of constraints.
            $endgroup$
            – Aleksandr Kazachkov
            Jul 2 at 18:38













            $begingroup$
            since for binary variables so much logic kicks in (propagation etc.) I would assume you always choose the binary program; do you have an example where an integer formulation is better, regardless of size?
            $endgroup$
            – Marco Lübbecke
            Jul 2 at 19:54




            $begingroup$
            since for binary variables so much logic kicks in (propagation etc.) I would assume you always choose the binary program; do you have an example where an integer formulation is better, regardless of size?
            $endgroup$
            – Marco Lübbecke
            Jul 2 at 19:54




            1




            1




            $begingroup$
            This decision arose recently for me, but I did not test the two formulations against each other, and I am not sure it will make sense to do so in my setting, so the question was more from curiosity / theoretical interest. I formulated my problem with binary variables, but later found an earlier paper that worked on the same problem, had very experienced authors, and used the general integer formulation. I probably will try to implement the general integer version and see how it goes, though in the end I designed a bunch of heuristics that make the IPs typically trivial after presolve anyway.
            $endgroup$
            – Aleksandr Kazachkov
            Jul 2 at 21:10




            $begingroup$
            This decision arose recently for me, but I did not test the two formulations against each other, and I am not sure it will make sense to do so in my setting, so the question was more from curiosity / theoretical interest. I formulated my problem with binary variables, but later found an earlier paper that worked on the same problem, had very experienced authors, and used the general integer formulation. I probably will try to implement the general integer version and see how it goes, though in the end I designed a bunch of heuristics that make the IPs typically trivial after presolve anyway.
            $endgroup$
            – Aleksandr Kazachkov
            Jul 2 at 21:10











            15












            $begingroup$

            I’m assuming that we want our models to be solved as quickly as possible. If that is the case, then the honest answer is: you need to try the models out and see.



            To give you a concrete example (see here): through what I thought was a super-clever reformulation, I was able to remove 85% of the variables in the problem, and I thought that this would make it super fast. However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not.



            I always try to put the tightest upper bound I have, make my big-M as small as possible, use continuous over binary/integer when I can, but honestly, you cannot know anything for sure until you test it.






            share|improve this answer









            $endgroup$












            • $begingroup$
              On However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not. Vs or.stackexchange.com/users/354/marco-l%c3%bcbbecke's answer avoid symmetric formulations I am extremely confused.
              $endgroup$
              – abhiieor
              Jul 1 at 8:57










            • $begingroup$
              That’s exactly why you should test: sometimes symmetry really kills your performance, because the B&B doesn’t “know” where to go, another time it helps unearth underlying structures and speed up the performance. I was totally confused by that result (which is why I wrote the blog post about it) .Basically, you have to try.
              $endgroup$
              – Richard
              Jul 1 at 10:58










            • $begingroup$
              Maybe another rule of thumb: symmetry in the objective function is bad. Symmetry in variables might be good.
              $endgroup$
              – Richard
              Jul 1 at 11:01















            15












            $begingroup$

            I’m assuming that we want our models to be solved as quickly as possible. If that is the case, then the honest answer is: you need to try the models out and see.



            To give you a concrete example (see here): through what I thought was a super-clever reformulation, I was able to remove 85% of the variables in the problem, and I thought that this would make it super fast. However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not.



            I always try to put the tightest upper bound I have, make my big-M as small as possible, use continuous over binary/integer when I can, but honestly, you cannot know anything for sure until you test it.






            share|improve this answer









            $endgroup$












            • $begingroup$
              On However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not. Vs or.stackexchange.com/users/354/marco-l%c3%bcbbecke's answer avoid symmetric formulations I am extremely confused.
              $endgroup$
              – abhiieor
              Jul 1 at 8:57










            • $begingroup$
              That’s exactly why you should test: sometimes symmetry really kills your performance, because the B&B doesn’t “know” where to go, another time it helps unearth underlying structures and speed up the performance. I was totally confused by that result (which is why I wrote the blog post about it) .Basically, you have to try.
              $endgroup$
              – Richard
              Jul 1 at 10:58










            • $begingroup$
              Maybe another rule of thumb: symmetry in the objective function is bad. Symmetry in variables might be good.
              $endgroup$
              – Richard
              Jul 1 at 11:01













            15












            15








            15





            $begingroup$

            I’m assuming that we want our models to be solved as quickly as possible. If that is the case, then the honest answer is: you need to try the models out and see.



            To give you a concrete example (see here): through what I thought was a super-clever reformulation, I was able to remove 85% of the variables in the problem, and I thought that this would make it super fast. However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not.



            I always try to put the tightest upper bound I have, make my big-M as small as possible, use continuous over binary/integer when I can, but honestly, you cannot know anything for sure until you test it.






            share|improve this answer









            $endgroup$



            I’m assuming that we want our models to be solved as quickly as possible. If that is the case, then the honest answer is: you need to try the models out and see.



            To give you a concrete example (see here): through what I thought was a super-clever reformulation, I was able to remove 85% of the variables in the problem, and I thought that this would make it super fast. However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not.



            I always try to put the tightest upper bound I have, make my big-M as small as possible, use continuous over binary/integer when I can, but honestly, you cannot know anything for sure until you test it.







            share|improve this answer












            share|improve this answer



            share|improve this answer










            answered Jun 30 at 13:45









            RichardRichard

            7851 silver badge10 bronze badges




            7851 silver badge10 bronze badges











            • $begingroup$
              On However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not. Vs or.stackexchange.com/users/354/marco-l%c3%bcbbecke's answer avoid symmetric formulations I am extremely confused.
              $endgroup$
              – abhiieor
              Jul 1 at 8:57










            • $begingroup$
              That’s exactly why you should test: sometimes symmetry really kills your performance, because the B&B doesn’t “know” where to go, another time it helps unearth underlying structures and speed up the performance. I was totally confused by that result (which is why I wrote the blog post about it) .Basically, you have to try.
              $endgroup$
              – Richard
              Jul 1 at 10:58










            • $begingroup$
              Maybe another rule of thumb: symmetry in the objective function is bad. Symmetry in variables might be good.
              $endgroup$
              – Richard
              Jul 1 at 11:01
















            • $begingroup$
              On However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not. Vs or.stackexchange.com/users/354/marco-l%c3%bcbbecke's answer avoid symmetric formulations I am extremely confused.
              $endgroup$
              – abhiieor
              Jul 1 at 8:57










            • $begingroup$
              That’s exactly why you should test: sometimes symmetry really kills your performance, because the B&B doesn’t “know” where to go, another time it helps unearth underlying structures and speed up the performance. I was totally confused by that result (which is why I wrote the blog post about it) .Basically, you have to try.
              $endgroup$
              – Richard
              Jul 1 at 10:58










            • $begingroup$
              Maybe another rule of thumb: symmetry in the objective function is bad. Symmetry in variables might be good.
              $endgroup$
              – Richard
              Jul 1 at 11:01















            $begingroup$
            On However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not. Vs or.stackexchange.com/users/354/marco-l%c3%bcbbecke's answer avoid symmetric formulations I am extremely confused.
            $endgroup$
            – abhiieor
            Jul 1 at 8:57




            $begingroup$
            On However the performance was orders(!!) of magnitude worse than before, because the original problem was symmetric, and the new one was not. Vs or.stackexchange.com/users/354/marco-l%c3%bcbbecke's answer avoid symmetric formulations I am extremely confused.
            $endgroup$
            – abhiieor
            Jul 1 at 8:57












            $begingroup$
            That’s exactly why you should test: sometimes symmetry really kills your performance, because the B&B doesn’t “know” where to go, another time it helps unearth underlying structures and speed up the performance. I was totally confused by that result (which is why I wrote the blog post about it) .Basically, you have to try.
            $endgroup$
            – Richard
            Jul 1 at 10:58




            $begingroup$
            That’s exactly why you should test: sometimes symmetry really kills your performance, because the B&B doesn’t “know” where to go, another time it helps unearth underlying structures and speed up the performance. I was totally confused by that result (which is why I wrote the blog post about it) .Basically, you have to try.
            $endgroup$
            – Richard
            Jul 1 at 10:58












            $begingroup$
            Maybe another rule of thumb: symmetry in the objective function is bad. Symmetry in variables might be good.
            $endgroup$
            – Richard
            Jul 1 at 11:01




            $begingroup$
            Maybe another rule of thumb: symmetry in the objective function is bad. Symmetry in variables might be good.
            $endgroup$
            – Richard
            Jul 1 at 11:01











            8












            $begingroup$

            Perhaps you are only talking about MILPs, but you don't say, so I will say something of interest for MINLPs.



            Regarding point 1:



            • Suppose your model has a convex nonlinear inequality constraint, $f(x) le c$, i.e., $f(x)$ is convex.

            and



            • It is known that the constraint will be active at the optimum, of any continuous relaxation i.e., $f(x) = c$, for instance because the continuous relaxations are concave programs with compact constraints, thereby having global optimum at an extreme of the constraints

            Then you are usually better off entering the constraint as an inequality rather than as an equality. That is because the inequality constraint is convex, and generally much easier for an optimizer to deal with than a nonlinear equality constraint, which is a non-convex constraint. You might think you'd be doing the optimizer a favor by specifying the equality constraint, thereby cleaving off $f(x) < c$ from consideration, but such a "favor" will usually backfire.



            However, in the case of a linear equality constraint known to hold with equality at the optimum of the original problem, or a continuous relaxation, I think it would generally be better, and in any event, not worse (except for (un)lucky "stars aligning" cases), to specify such a linear constraint as an equality constraint.



            Regarding point 3, I have a pretty hard time imagining it would be better to not provide an upper bound of 1 than to provide it when you know the optimal variable value is at most 1. But I suppose for some particular model and solver, the stars could always align such that specifying such an upper bound slowed things down.



            I will leave others to comment on point 2.






            share|improve this answer











            $endgroup$








            • 2




              $begingroup$
              Mark is dead on. Also, regarding the inequality v. equality question, sometimes rounding error becomes a hazard. With an equality constraint, deviating from the RHS by more than the solver's tolerance in either direction can make a feasible solution look infeasible. With an inequality constraint, the hazard is only one one side, so you have possibly cut the risk of it.
              $endgroup$
              – prubin
              Jun 30 at 17:51










            • $begingroup$
              Another argument I have come along in LP contexts for using inequality constraints rather than equality constraints is that by using inequality constraints, the dual variables are restricted to the positive domain, which may accelerate the solution process. I think this is why people often use a set covering formulation rather than a set partitioning formulation, even though all covering constraints hold with equality in an optimal solution.
              $endgroup$
              – Rolf van Lieshout
              Jul 1 at 12:34















            8












            $begingroup$

            Perhaps you are only talking about MILPs, but you don't say, so I will say something of interest for MINLPs.



            Regarding point 1:



            • Suppose your model has a convex nonlinear inequality constraint, $f(x) le c$, i.e., $f(x)$ is convex.

            and



            • It is known that the constraint will be active at the optimum, of any continuous relaxation i.e., $f(x) = c$, for instance because the continuous relaxations are concave programs with compact constraints, thereby having global optimum at an extreme of the constraints

            Then you are usually better off entering the constraint as an inequality rather than as an equality. That is because the inequality constraint is convex, and generally much easier for an optimizer to deal with than a nonlinear equality constraint, which is a non-convex constraint. You might think you'd be doing the optimizer a favor by specifying the equality constraint, thereby cleaving off $f(x) < c$ from consideration, but such a "favor" will usually backfire.



            However, in the case of a linear equality constraint known to hold with equality at the optimum of the original problem, or a continuous relaxation, I think it would generally be better, and in any event, not worse (except for (un)lucky "stars aligning" cases), to specify such a linear constraint as an equality constraint.



            Regarding point 3, I have a pretty hard time imagining it would be better to not provide an upper bound of 1 than to provide it when you know the optimal variable value is at most 1. But I suppose for some particular model and solver, the stars could always align such that specifying such an upper bound slowed things down.



            I will leave others to comment on point 2.






            share|improve this answer











            $endgroup$








            • 2




              $begingroup$
              Mark is dead on. Also, regarding the inequality v. equality question, sometimes rounding error becomes a hazard. With an equality constraint, deviating from the RHS by more than the solver's tolerance in either direction can make a feasible solution look infeasible. With an inequality constraint, the hazard is only one one side, so you have possibly cut the risk of it.
              $endgroup$
              – prubin
              Jun 30 at 17:51










            • $begingroup$
              Another argument I have come along in LP contexts for using inequality constraints rather than equality constraints is that by using inequality constraints, the dual variables are restricted to the positive domain, which may accelerate the solution process. I think this is why people often use a set covering formulation rather than a set partitioning formulation, even though all covering constraints hold with equality in an optimal solution.
              $endgroup$
              – Rolf van Lieshout
              Jul 1 at 12:34













            8












            8








            8





            $begingroup$

            Perhaps you are only talking about MILPs, but you don't say, so I will say something of interest for MINLPs.



            Regarding point 1:



            • Suppose your model has a convex nonlinear inequality constraint, $f(x) le c$, i.e., $f(x)$ is convex.

            and



            • It is known that the constraint will be active at the optimum, of any continuous relaxation i.e., $f(x) = c$, for instance because the continuous relaxations are concave programs with compact constraints, thereby having global optimum at an extreme of the constraints

            Then you are usually better off entering the constraint as an inequality rather than as an equality. That is because the inequality constraint is convex, and generally much easier for an optimizer to deal with than a nonlinear equality constraint, which is a non-convex constraint. You might think you'd be doing the optimizer a favor by specifying the equality constraint, thereby cleaving off $f(x) < c$ from consideration, but such a "favor" will usually backfire.



            However, in the case of a linear equality constraint known to hold with equality at the optimum of the original problem, or a continuous relaxation, I think it would generally be better, and in any event, not worse (except for (un)lucky "stars aligning" cases), to specify such a linear constraint as an equality constraint.



            Regarding point 3, I have a pretty hard time imagining it would be better to not provide an upper bound of 1 than to provide it when you know the optimal variable value is at most 1. But I suppose for some particular model and solver, the stars could always align such that specifying such an upper bound slowed things down.



            I will leave others to comment on point 2.






            share|improve this answer











            $endgroup$



            Perhaps you are only talking about MILPs, but you don't say, so I will say something of interest for MINLPs.



            Regarding point 1:



            • Suppose your model has a convex nonlinear inequality constraint, $f(x) le c$, i.e., $f(x)$ is convex.

            and



            • It is known that the constraint will be active at the optimum, of any continuous relaxation i.e., $f(x) = c$, for instance because the continuous relaxations are concave programs with compact constraints, thereby having global optimum at an extreme of the constraints

            Then you are usually better off entering the constraint as an inequality rather than as an equality. That is because the inequality constraint is convex, and generally much easier for an optimizer to deal with than a nonlinear equality constraint, which is a non-convex constraint. You might think you'd be doing the optimizer a favor by specifying the equality constraint, thereby cleaving off $f(x) < c$ from consideration, but such a "favor" will usually backfire.



            However, in the case of a linear equality constraint known to hold with equality at the optimum of the original problem, or a continuous relaxation, I think it would generally be better, and in any event, not worse (except for (un)lucky "stars aligning" cases), to specify such a linear constraint as an equality constraint.



            Regarding point 3, I have a pretty hard time imagining it would be better to not provide an upper bound of 1 than to provide it when you know the optimal variable value is at most 1. But I suppose for some particular model and solver, the stars could always align such that specifying such an upper bound slowed things down.



            I will leave others to comment on point 2.







            share|improve this answer














            share|improve this answer



            share|improve this answer








            edited Jun 30 at 12:52

























            answered Jun 30 at 12:29









            Mark L. StoneMark L. Stone

            2,1745 silver badges23 bronze badges




            2,1745 silver badges23 bronze badges







            • 2




              $begingroup$
              Mark is dead on. Also, regarding the inequality v. equality question, sometimes rounding error becomes a hazard. With an equality constraint, deviating from the RHS by more than the solver's tolerance in either direction can make a feasible solution look infeasible. With an inequality constraint, the hazard is only one one side, so you have possibly cut the risk of it.
              $endgroup$
              – prubin
              Jun 30 at 17:51










            • $begingroup$
              Another argument I have come along in LP contexts for using inequality constraints rather than equality constraints is that by using inequality constraints, the dual variables are restricted to the positive domain, which may accelerate the solution process. I think this is why people often use a set covering formulation rather than a set partitioning formulation, even though all covering constraints hold with equality in an optimal solution.
              $endgroup$
              – Rolf van Lieshout
              Jul 1 at 12:34












            • 2




              $begingroup$
              Mark is dead on. Also, regarding the inequality v. equality question, sometimes rounding error becomes a hazard. With an equality constraint, deviating from the RHS by more than the solver's tolerance in either direction can make a feasible solution look infeasible. With an inequality constraint, the hazard is only one one side, so you have possibly cut the risk of it.
              $endgroup$
              – prubin
              Jun 30 at 17:51










            • $begingroup$
              Another argument I have come along in LP contexts for using inequality constraints rather than equality constraints is that by using inequality constraints, the dual variables are restricted to the positive domain, which may accelerate the solution process. I think this is why people often use a set covering formulation rather than a set partitioning formulation, even though all covering constraints hold with equality in an optimal solution.
              $endgroup$
              – Rolf van Lieshout
              Jul 1 at 12:34







            2




            2




            $begingroup$
            Mark is dead on. Also, regarding the inequality v. equality question, sometimes rounding error becomes a hazard. With an equality constraint, deviating from the RHS by more than the solver's tolerance in either direction can make a feasible solution look infeasible. With an inequality constraint, the hazard is only one one side, so you have possibly cut the risk of it.
            $endgroup$
            – prubin
            Jun 30 at 17:51




            $begingroup$
            Mark is dead on. Also, regarding the inequality v. equality question, sometimes rounding error becomes a hazard. With an equality constraint, deviating from the RHS by more than the solver's tolerance in either direction can make a feasible solution look infeasible. With an inequality constraint, the hazard is only one one side, so you have possibly cut the risk of it.
            $endgroup$
            – prubin
            Jun 30 at 17:51












            $begingroup$
            Another argument I have come along in LP contexts for using inequality constraints rather than equality constraints is that by using inequality constraints, the dual variables are restricted to the positive domain, which may accelerate the solution process. I think this is why people often use a set covering formulation rather than a set partitioning formulation, even though all covering constraints hold with equality in an optimal solution.
            $endgroup$
            – Rolf van Lieshout
            Jul 1 at 12:34




            $begingroup$
            Another argument I have come along in LP contexts for using inequality constraints rather than equality constraints is that by using inequality constraints, the dual variables are restricted to the positive domain, which may accelerate the solution process. I think this is why people often use a set covering formulation rather than a set partitioning formulation, even though all covering constraints hold with equality in an optimal solution.
            $endgroup$
            – Rolf van Lieshout
            Jul 1 at 12:34











            6












            $begingroup$

            In addition to previous answers, I like to cite the following book, which discuss in details several different options (and kind of good practices) to model different common applications:



            Model Building in Mathematical Programming, by H. Paul Williams. March, 2013.
            ISBN: 978-1-118-44333-0.



            Every chapter is very interesting and is worth of (re)reading several times.






            share|improve this answer








            New contributor



            Stefano Gualandi is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
            Check out our Code of Conduct.





            $endgroup$

















              6












              $begingroup$

              In addition to previous answers, I like to cite the following book, which discuss in details several different options (and kind of good practices) to model different common applications:



              Model Building in Mathematical Programming, by H. Paul Williams. March, 2013.
              ISBN: 978-1-118-44333-0.



              Every chapter is very interesting and is worth of (re)reading several times.






              share|improve this answer








              New contributor



              Stefano Gualandi is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
              Check out our Code of Conduct.





              $endgroup$















                6












                6








                6





                $begingroup$

                In addition to previous answers, I like to cite the following book, which discuss in details several different options (and kind of good practices) to model different common applications:



                Model Building in Mathematical Programming, by H. Paul Williams. March, 2013.
                ISBN: 978-1-118-44333-0.



                Every chapter is very interesting and is worth of (re)reading several times.






                share|improve this answer








                New contributor



                Stefano Gualandi is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
                Check out our Code of Conduct.





                $endgroup$



                In addition to previous answers, I like to cite the following book, which discuss in details several different options (and kind of good practices) to model different common applications:



                Model Building in Mathematical Programming, by H. Paul Williams. March, 2013.
                ISBN: 978-1-118-44333-0.



                Every chapter is very interesting and is worth of (re)reading several times.







                share|improve this answer








                New contributor



                Stefano Gualandi is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
                Check out our Code of Conduct.








                share|improve this answer



                share|improve this answer






                New contributor



                Stefano Gualandi is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
                Check out our Code of Conduct.








                answered Jul 1 at 17:49









                Stefano GualandiStefano Gualandi

                1815 bronze badges




                1815 bronze badges




                New contributor



                Stefano Gualandi is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
                Check out our Code of Conduct.




                New contributor




                Stefano Gualandi is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
                Check out our Code of Conduct.





























                    draft saved

                    draft discarded
















































                    Thanks for contributing an answer to Operations Research Stack Exchange!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid


                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.

                    Use MathJax to format equations. MathJax reference.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2for.stackexchange.com%2fquestions%2f834%2fbest-practices-for-formulating-mips%23new-answer', 'question_page');

                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Category:9 (number) SubcategoriesMedia in category "9 (number)"Navigation menuUpload mediaGND ID: 4485639-8Library of Congress authority ID: sh85091979ReasonatorScholiaStatistics

                    Circuit construction for execution of conditional statements using least significant bitHow are two different registers being used as “control”?How exactly is the stated composite state of the two registers being produced using the $R_zz$ controlled rotations?Efficiently performing controlled rotations in HHLWould this quantum algorithm implementation work?How to prepare a superposed states of odd integers from $1$ to $sqrtN$?Why is this implementation of the order finding algorithm not working?Circuit construction for Hamiltonian simulationHow can I invert the least significant bit of a certain term of a superposed state?Implementing an oracleImplementing a controlled sum operation

                    Magento 2 “No Payment Methods” in Admin New OrderHow to integrate Paypal Express Checkout with the Magento APIMagento 1.5 - Sales > Order > edit order and shipping methods disappearAuto Invoice Check/Money Order Payment methodAdd more simple payment methods?Shipping methods not showingWhat should I do to change payment methods if changing the configuration has no effects?1.9 - No Payment Methods showing upMy Payment Methods not Showing for downloadable/virtual product when checkout?Magento2 API to access internal payment methodHow to call an existing payment methods in the registration form?