Dynamic Programming

Preetish Simhadri
5 min readMar 24, 2021

Before we start with sorting algorithms let us first know what the word algorithm is meant.

What is an Algorithm?

The set of instructions which are designed to perform certain task is called an algorithm.

Numerous individuals, including myself, need to run their algorithms appropriately and compose codes that are effectively understandable. In any case, as I get familiar with planning and investigating algorithms when all is said in done, it’s getting progressively significant for me to compose codes that are computationally successful, as confirmed by manager inquiries just as inquiries in codathons and coding challenge sites. The captivating part of this subject is that it isn’t important to comprehend it during the beginning phases of calculation plan.

Much of the time, there are various approaches to take care of an issue or complete an assignment in programming. In any case, one ought to figure out how to analyze various algorithms and pick the best code dependent on its proficiency. Profoundly productive algorithms burn-through less framework assets and work at a quicker rate. The intricacy of a algorithms decides its effectiveness.

Why we study Algorithm?

As the speed of processor expands, execution is oftentimes supposed to be less focal than other programming quality attributes (for example security, extensibility, reusability etc.) Notwithstanding, huge issue sizes are ordinary nearby computational science, which makes execution a vital factor. This is on the grounds that more drawn-out Algorithm time, to give some examples mean more slow outcomes, less through research and greater expense of calculation (if purchasing CPU Hours from an outside party). The investigation of Algorithm, in this manner, gives us a language to communicate execution as a component of issue size.

Introduction of Dynamic Programming:

Dynamic Programming is the most powerful design technique for solving optimization problems.

Divide & Conquer algorithm partition the problem into disjoint subproblems solve the subproblems recursively and then combine their solution to solve the original problems.

Dynamic Programming is used when the subproblems are not independent, e.g., when they share the same subproblems. In this case, divide and conquer may do more work than necessary, because it solves the same sub problem multiple times.

Dynamic Programming solves each subproblems just once and stores the result in a table so that it can be repeatedly retrieved if needed again.

Dynamic Programming is a Bottom-up approach- we solve all possible small problems and then combine to obtain solutions for bigger problems.

Dynamic Programming is a paradigm of algorithm design in which an optimization problem is solved by a combination of achieving sub-problem solutions and appearing to the “principle of optimality”.

Characteristics of Dynamic Programming:

Dynamic programming works on two different characteristics:

Optimal Substructure: If an optimal solution contains optimal sub solutions then a problem exhibits optimal substructure.

If a problem has optimal substructure, then we can recursively define an optimal solution.

If a problem doesn’t have optimal substructure, there is no basis for defining a recursive algorithm to find the optimal solutions.

Overlapping subproblems: When a recursive algorithm would visit the same subproblems repeatedly, then a problem has overlapping subproblems.

If a problem has overlapping subproblems, then we can improve on a recursive implementation by computing each subproblem only once.

If a problem doesn’t have overlapping sub problems, we don’t have anything to gain by using dynamic programming.

Elements of Dynamic Programming:

There are basically three elements of dynamic programming:

1. Substructure: Decompose the given problem into smaller subproblems. Express the solution of the original problem in terms of the solution for smaller problems.

2. Table Structure: After solving the sub-problems, store the results to the sub problems in a table. This is done because subproblem solutions are reused many times, and we do not want to repeatedly solve the same problem repeatedly.

3. Bottom-up Computation: Using table, combine the solution of smaller subproblems to solve larger subproblems and eventually arrives at a solution to complete problem.

Some more important points regarding Bottom-up Computation:

i. Start with smallest subproblems.

ii. Combining their solutions obtain the solution to sub-problems of increasing size.

iii. Until solving at the solution of the original problem.

Components of Dynamic Programming:

There are four components of Dynamic Programming:

1. Stages: The problem can be divided into several subproblems, which are called stages. A stage is a small portion of a given problem. For example, in the shortest path problem, they were defined by the structure of the graph.

2. States: Each stage has several states associated with it. The states for the shortest path problem were the node reached.

3. Decision: At each stage, there can be multiple choices out of which one of the best decisions should be taken. The decision taken at every stage should be optimal; this is called a stage decision.

4. Optimal policy: It is a rule which determines the decision at each stage; a policy is called an optimal policy if it is globally optimal. This is known as Bellman principle of optimality.

5. Given the current state, the optimal choices for each of the remaining states does not depend on the previous states or decisions. In the shortest path problem, it was not necessary to know how we got a node only that we did.

6. There exists a recursive relationship that identify the optimal decisions for stage x, given that stage x+1, has already been solved.

7. The final stage must be solved by ourselves.

Development of Dynamic Programming Algorithm:

Development of Dynamic Programming Algorithm can be divided into four steps:

1. Characterize the structure of an optimal solution.

2. Recursively defined the value of the optimal solution. Like Divide and Conquer, divide the problem into two or more optimal parts recursively. This helps to determine what the solution will look like.

3. Compute the value of the optimal solution from the bottom up (starting with the smallest subproblems)

4. Construct the optimal solution for the entire problem form the computed values of smaller subproblems.

Applications of Dynamic Programming:

There are several applications of Dynamic Programming I have mentioned some of them:

1. 0/1 knapsack problem

2. Mathematical optimization problem

3. All pair Shortest path problem

4. Reliability design problem

5. Longest common subsequence (LCS)

6. Flight control and robotics control

7. Time-sharing: It schedules the job to maximize CPU usage.

--

--