Pure Go hardware accelerated local inference on Vision Language Models/LLMs/TLMs using llama.cpp