You are given two arrays of integers a, b both with length n <= 10^5. For each 0 <= x <= n print the sum of a[i]*b[x-i] for all 0 <= i <= x.
It's obvious this can be done in quadratic time, but can we do any better?
Help with problem
You are given two arrays of integers a, b both with length n <= 10^5. For each 0 <= x <= n print the sum of a[i]*b[x-i] for all 0 <= i <= x.
It's obvious this can be done in quadratic time, but can we do any better?