We derive an upper bound on the local Rademacher complexity of -norm multiple kernel learning, which yields a tighter excess risk bound than global approaches. Previous local approaches analyzed the case only while our analysis covers all cases assuming the different feature mappings corresponding to the different kernels to be uncorrelated. We also show a lower bound that shows that the bound is tight, and derive consequences regarding ex-cess loss, namely fast convergence rates of the order where is the minimum eigenvalue decay rate of the individual kernels.