Selective KV-Cache Sharing to Mitigate Timing Side-Channels in LLM Inference