Intel Unveils New Low-Latency LLM Inference Solution Optimized for Intel GPUs