This is a free / open source software that like QGIS we can download an installer, double-click and install it on our system.
Once done, we have a program on our computer that looks like ChatGPT.
Now, one thing to learn : this software is a container. The actual AI is a "model" that we have to download, like we do with plugins in QGIS, to work in this container. And there's many models available - i'll come to this later.
The first thing we have to do after downloading, is Go on "Models" tab, click + Add Model, browse and download one.
You can ignore the ones titled "ChatGPT" etc which have "install" button and a text field for adding an API key. These are 3rd party services, with limits and costs.
Go for the entries with "Download" button instead. Download one model. The GPT4All site quotes some that we can start with.
Then, go to "Chats" section and here you can select the one you downloaded and start a conversation just like we do in chatgpt. Be prepared for slowness in response. This thing is now running on your machine instead of a supercomputer somewhere.
After this, go to "LocalDocs" section. Here's where our core problem statement gets acted on.
Start a collection, add some documents and proceed. It will take some time to do some "embedding", and then will be done/ready.
Now go to Chats section again, and this time, open the "LocalDocs" sidebar on right and check-on your newly setup document collection.
Now you can ask it questions and it'll dig through your documents and then answer based on the docs.
Screenshot:
In my limited attempt it was even able to tell me a page number it got the answer from. I had to include the instruction in my prompt for it. "Write the doc filename and page numbers for reference."
I think this is a good place to start with. The whole thing happened on my computer, without internet, no usage limits. Took time, but hey, some folks have more time than money.
----
Models market
As I mentioned earlier, there's many models available. That's where you come in.
- I've only just started reading into this stuff, so don't know what differentiates the various models being published by different folks.
- It's possible that we might get some models doing our specific job very nicely, and others not so much.
- What performs well for one use case, may not do so for another.
- There's also smaller models published which run better on normal laptops and what they can do is uncharted.
- And then, there's the Settings. There's a ton of technical options (go to "Model" tab in Settings), and it can wildly vary what outputs we get.
- So there's a lot of room for model-shopping and settings fine-tuning here.
- What can help is: different folks trying out different combinations and reporting back what worked and what didn't, with their specific use case.
- So, inviting you to jump in and start tinkering with this, apply it to your work area, and share your experiences.
Size vs usefulness of AI models
(note: don't confuse with the instagram kind of AI models :D ):
- The "everything" AI models which grab the headlines, have ingested all sorts of myriad info like science and history textbooks, past news, TV serial plots, food recipes, coding libraries, religious texts, celebrity gossip, social media noise - you name it. That's what makes them huge and resource-hungry.
- But you might have a very specific use case for which all the other info is utterly useless.
- So, there is an opportunity here : if we can find specific small models that do our specific work, we can do a lot with much lesser resources required.