Speculative Decoding: Accelerating LLM Inference | Skills Pool