If your graphics card has the capability and sufficient vRAM to fit the model, I believe GPT4ALL and ollama will detect that and use it automatically. My gaming laptop has an NVidia RTX 2090 with 6GB vRAM. GPT4ALL uses it automatically when the model is small enough to fit, and otherwise runs on CPU, in my experience.