Running On-Device LLMs: A Technical Deep Dive Exploring the state of Apple Silicon and the tools that make local AI inference possible today.