**1. The problem statement, all variables and given/known data**

**The L=100 mH inductor in the following figure has an initial current of Io=10 mA. If the voltage is, v(t)=1e−10t+2e−5t V, what is the current, i(t), through the inductor?**

**Express your answer as a function of time with units of mA.**

The figure shows an independent voltage source connected to an inductor.

**2. Relevant equations**

v = L(di/dt)

i dt = (v dt)/L

[itex]i = \int_{t_0}^{t} v dt + i_0[/itex]

**3. The attempt at a solution**

I solved the following equation:

[itex]i(t) = \frac{1}{0.1 H}(\int_{t_0}^{t} (e^{-10T}+2e^{-5T}) dT)*\frac{10^3 mA}{1 A} + 10 mA[/itex]

and obtained the following:

[itex]i(t) = (e^{-0.1} + 4e^{-0.05} – e^{-10t} – e^{-5t})*10^3 + 10 mA[/itex]

My homework software rejects this answer, saying that my "answer either contains an incorrect additive numerical constant or is missing one."

I can think of nothing that I might have omitted, and I know that this software is very picky. For example, when rounding, it rejects the "round 5 to even" rule that was instilled in me early on as completely incorrect, always expecting 5 to be rounded up even when followed only by zeroes. Given how picky the software is, I am not altogether convinced that my answer is actually wrong.

Does anyone see any obvious oversights in my work?

http://ift.tt/1kF7odd