@mos_8502 There's a lot of it depends, but most of the cost is AI is in training the models rather than the queries themselves. If you have a machine with a newish GPU and you download a model like granite or phi, the cost is your time and a nominal amount of electricity.
@mos_8502 Mainly get a capable model (like granite) locally on your box and call it with pytorch. I have a python script on my laptop for calling llm queries completely locally. Having a newish GPU will help.