As usual Linux-fans can´t think of other things than operating systems while totally ignoring the local applications or the impact from the Internet. This is really fascinating and when it comes to pictures a lot of pictures contains GPS-info and XMP and/or IPTC-metadata that can make a lot of difference to whether you get exposed or not using your computer.
Using Linux is not at all a vaccin from file realated problems or dependenses coming with using all sorts of resources on the Internet - or are you Linux-folks livingb in a local sandbox not using the Internet or any modern cloud-services at all?? What about your use of social media?
The big divide when it comes to the computer world will not at all be betwen using Linux or Windows but to run things locally or using cloud-services on the Internet and it is a fact that the higher demands locally run AI-models will put on local computer resources the more people will be forced to use the cloud-services of pure economical reasons.
BUT, buying the hardware that will make it possible to isolate yourselves from the AI-cloud-services dependencies and run more powerful local AI-models locally is quickly becoming a very exclusive option that less users can afford.
I have just made some tests comparing if my new tailormade computer with an Nvidia RTX 5070 Ti card with 16 GB better can meet my demands running local AI-models than my older 3060 Ti-system just 8GB.
My old system could not handle Google Gemma 3 12b that demanded at least 16 GB VRAM at the GPU in order to meet my demands that was to be able to use iMatch Autotagger to identify animal and plant species in my pictures. The smaller Gemma 3 4b that was the biggest Gemma-model I could use earlier on 8 GB just could not do the job. So earlier I was forced to use either Google Gemini Flash 2.5 or Open AI GPT 4.1 to meet those requirements.
Well, yesterday I installed the bigger AI-model Gemma 4 12B on my new computer, more than twice as powerful as the old one. That model really put those 16 GB VRAM to work BUT it really managed to solve these problems while running my 160 test pictures locally instaed via the cloud. I must say I didn´t expect that since the big French cloud AI-service Mistral 3.1 to a large extent failed to solve the problems.
I had chosen 100 example pictures with animals and plants to see if Gemma 3 12b had what it takes to identify and classify those pictures and write both relevant Descriptions and Keywords. I also processed 60 pictures with global architecture pictures where the task was to try to identify landmarks and write Descripotions and Keywords using the concepts of an architect.
I have published these pictures in two portfolios on the biggest photo site in Sweden called Fotosidan (translates to “The Photo Page”) so people can see for themselves how it looks and what to expect when it comes to these results.
It is also important to stress that this is not an absolute measurement of what to expect because what is produced and achieved is a result of my prompting in iMatch that has four promts that has to be developed and fine tuned to your liking. There are different promts for Descriptions, Keywords, Landmarks and also an Ad-hoc-prompts where specific data for special selections of pictures can be added.
Animals and Plants in East-Africa:
Sten-Åke Sändh - Portfolio
Global Architecture
Sten-Åke Sändh - Portfolio
What is not reflected in these picture texts in the Portfolios is that the data also is formatted and structured to increase readability. That formatting has been stripped by Fotosidan but is visible för exemple in Photolab 9 or Capture One and of cource even in Imatch DAM where I have processed these pictures metadata. This is how this structure looks in other applications than Photosidan.
You also need alocal AI-platform like Ollama or LM Studio to run these local models ands even if I use Googles Gemma in this case Google can never get access to these data because the process is entirely local and this way you are free to use lots of local free AI-models without paying a cent to any american AI cloud platform owner.
Since Gemma 3 12b meets all my demands I can now abort my Open AI API if I want without any problems when it comes to my picture metadata.
So this is ONE way to go if you don´t want to depend on the big American cloud services and has access to a software like iMatch DAM that is really open to a great variety of both American, Chinese or European cloud services or local ones of your liking and it has nothing at all to do with Linux. Even DXO, Adobe and Capture One could offer the same kind of AI-interfaces but so far they are strictly proprietary.