Jump-Start RL (JSRL) performs well in a few scenarios (5 out of 18) but significantly worse in the rest. It doesn't improve the base policy but learns a new one, failing to preserve desired traits like smooth and natural motion. In contrast, our refined policies maintain smooth, natural behavior by staying close to the base policy through the bounded residual action strategy.