Lesson 49 of 107 6 min

Problem: Combination Sum

Learn how to find all unique combinations that sum to a target with pruning.

Reading Mode

Hide the curriculum rail and keep the lesson centered for focused reading.

Problem Statement

Mental Model

Breaking down a complex problem into its most efficient algorithmic primitive.

Given an array of distinct integers candidates and a target, return a list of all unique combinations where the numbers sum to target. You may use the same number an unlimited number of times.

Approach: Backtracking with Pruning

Because we can reuse numbers, our decision tree doesn't shrink in width, but it must shrink in remaining target.

  1. State: Current path, current sum, and starting index (to avoid duplicate combinations like [2,3] and [3,2]).
  2. Base Case:
    • sum == target: Success! Add to results.
    • sum > target: Fail! Stop exploring this branch (Pruning).
  3. Recursion: Start the loop from index to allow reusing the current number but prevent using previous numbers.

Java Implementation

public List<List<Integer>> combinationSum(int[] candidates, int target) {
    List<List<Integer>> results = new ArrayList<>();
    backtrack(results, new ArrayList<>(), candidates, target, 0);
    return results;
}

private void backtrack(List<List<Integer>> res, List<Integer> path, int[] nums, int remain, int start) {
    if (remain < 0) return; // Pruning
    if (remain == 0) {
        res.add(new ArrayList<>(path));
        return;
    }
    
    for (int i = start; i < nums.length; i++) {
        path.add(nums[i]);
        // We pass 'i' as start index because we can reuse the same element
        backtrack(res, path, nums, remain - nums[i], i);
        path.remove(path.size() - 1);
    }
}

Complexity Analysis

  • Time Complexity: $O(N^{\frac{T}{M} + 1})$ where $N$ is candidates, $T$ is target, and $M$ is the smallest candidate.
  • Space Complexity: $O(\frac{T}{M})$ for the recursion stack.

Interview Tips

  • Why pass i as the next start? "To ensure we only move forward in the array, preventing duplicate combinations in different orders."
  • Mention sorting the array first to prune even earlier (if remain - nums[i] < 0, the rest of the loop can be skipped).

5. Verbal Interview Script (Staff Tier)

Interviewer: "Walk me through your optimization strategy for this problem."

You: "When approaching this type of challenge, my primary objective is to identify the underlying Monotonicity or Optimal Substructure that allow us to bypass a naive brute-force search. In my implementation of 'Problem: Combination Sum', I focused on reducing the time complexity by leveraging a Dynamic Programming state transition. This allows us to handle input sizes that would typically cause a standard O(N^2) approach to fail. Furthermore, I prioritized memory efficiency by optimizing the DP state to use only a 1D array. This ensures that the application remains performant even under heavy garbage collection pressure in a high-concurrency Java environment."

6. Staff-Level Interview Follow-Ups

Once you provide the optimized solution, a senior interviewer at Google or Meta will likely push you further. Here is how to handle the most common follow-ups:

Follow-up 1: "How does this scale to a Distributed System?"

If the input data is too large to fit on a single machine (e.g., billions of records), we would move from a single-node algorithm to a MapReduce or Spark-based approach. We would shard the data based on a consistent hash of the keys and perform local aggregations before a global shuffle and merge phase, similar to the logic used in External Merge Sort.

Follow-up 2: "What are the Concurrency implications?"

In a multi-threaded Java environment, we must ensure that our state (e.g., the DP table or the frequency map) is thread-safe. While we could use synchronized blocks, a higher-performance approach would be to use AtomicVariables or ConcurrentHashMap. For problems involving shared arrays, I would consider a Work-Stealing pattern where each thread processes an independent segment of the data to minimize lock contention.

7. Performance Nuances (The Java Perspective)

  1. Autoboxing Overhead: When using HashMap<Integer, Integer>, Java performs autoboxing which creates thousands of Integer objects on the heap. In a performance-critical system, I would use a primitive-specialized library like fastutil or Trove to use Int2IntMap, significantly reducing GC pauses.
  2. Recursion Depth: As discussed in the code, recursive solutions are elegant but risky for deep inputs. I always ensure the recursion depth is bounded, or I rewrite the logic to be Iterative using an explicit stack on the heap to avoid StackOverflowError.

6. Staff-Level Verbal Masterclass (Communication)

Interviewer: "How would you defend this specific implementation in a production review?"

You: "In a mission-critical environment, I prioritize the Big-O efficiency of the primary data path, but I also focus on the Predictability of the system. In this implementation, I chose a recursive approach with memoization. While a recursive solution is more readable, I would strictly monitor the stack depth. If this were to handle skewed inputs, I would immediately transition to an explicit stack on the heap to avoid a StackOverflowError. From a memory perspective, I leverage primitive arrays to ensure that we minimize the garbage collection pauses (Stop-the-world) that typically plague high-throughput Java applications."

7. Global Scale & Distributed Pivot

When a problem like this is moved from a single machine to a global distributed architecture, the constraints change fundamentally.

  1. Data Partitioning: We would shard the input space using Consistent Hashing. This ensures that even if our dataset grows to petabytes, any single query only hits a small subset of our cluster, maintaining logarithmic lookup times.
  2. State Consistency: For problems involving state updates (like DP or Caching), we would use a Distributed Consensus protocol like Raft or Paxos to ensure that all replicas agree on the final state, even in the event of a network partition (The P in CAP theorem).

8. Performance Nuances (The Staff Perspective)

  1. Cache Locality: Accessing a 2D matrix in row-major order (reading [i][j] then [i][j+1]) is significantly faster than column-major order in modern CPUs due to L1/L2 cache pre-fetching. I always structure my loops to align with how the memory is physically laid out.
  2. Autoboxing and Generics: In Java, using List<Integer> instead of int[] can be 3x slower due to the overhead of object headers and constant wrapping. For the most performance-sensitive sections of this algorithm, I advocate for primitive specialized structures.

Key Takeaways

  • sum == target: Success! Add to results.
  • sum > target: Fail! Stop exploring this branch (Pruning).
  • Time Complexity: $O(N^{\frac{T}{M} + 1})$ where $N$ is candidates, $T$ is target, and $M$ is the smallest candidate.

Want to track your progress?

Sign in to save your progress, track completed lessons, and pick up where you left off.