Super Quick: RAG Comparison between GPT4 and Open-Source
Who reads PDF files better?

As we noted in the last OpenAI dev day (more like RAGgers doomsday, [kidding]) GPT4 now has the ability to glean information from pdf files (that you can just drag and drop) and answer questions about it. I wanted to check how well it performs. I also wanted to compare how open source options fare against it.
The open source option
We are going to use the same configuration I had worked on in my last article. Details are here. Basically I create a folder in which I put my research papers. Then I create embeddings using GPT4AllEmbeddings and save a local copy using Chroma. All the code for that is here. Then I use ollama and load a zephyr 7b parameters model. Consequently I run a chainlit interface to interact with the model.

Let me put the text in for better readability.
User: What is KNNOR?
Chatbot: KNNOR, short for K Nearest Neighbor OverSampling, is a method used to create new data points in the given dataset. Its practicality lies in three main activities: identifying critical data that will be utilized to generate new points. This approach is helpful in situations where there is a lack of data or imbalanced distribution of data, which leads to poor generalization capabilities in machine learning models. KNNOR’s pre-augmentation step involves calculating the optimized distance between existing data points, followed by creating new data points based on these identified critical ones. While this method results in a longer processing time compared to other oversampling methods like SMOTE, it has been found to be more effective in certain scenarios, particularly when dealing with abstract art data, where each image of a painting has distinct features that make the model prone to overfitting.
Not bad. The information is correct. It gives some pros and cons as well. Lets take a look under the hood to see what is the source of this infomation.

It is important to know the source document because I am going to upload the same pdf into GPT4 and ask the same question.
GPT4 has entered the building
I need to buy a subscription first.
Once you buy a subscription with 20USD per month, you get the option to switch between GPT3.5 and 4.

Next we test GPT4 by asking the same question without giving the pdf file.

So when I ask what is KNNOR, GPT4 gives a made up full-form and then proceeds to explain K Nearest Neighbors algorithm (which is just KNN) then it gives more made up stuff about KNN for Ordinal Regression (KNNOR). I searched high and low but could not find anything. This is expected I believe.
Moment of truth
Next i dragged and dropped the same pdf file (Fast_and_Efficient_Image_Generation_Using_Variational_Autoencoders_and_K-Nearest_Neighbor_OveRsampling_Approach.pdf) into GPT4’s chat window.

Once I write the question and press enter, it shows a nice little animation that it is reading the file.

Finally I get the answer.

So GPT4 gets it absolutely right. It gives the overview, the steps in the algorithm, the advantages and the applications. The answer is longer and better organized than ollama (Zephyr). The only edge that Zephyr had is that it had mentioned a limitation of the algorithm, whereas GPT4 only sings praises.
Conclusion
Well, this is not an exhaustive comparison. We need to check with bigger and better models to see what answers they give.
If I was a rich business, I would perhaps prefer GPT4 for ease of use and also structure of response.
But where’s the fun in that!
Stay tuned.
That is all for today. Below are links to my other llm-based articles:
Open Source LLM:
**Textfiles on MemGPT** (Keep your eyes on this)
Paradigm Shift in Retrieval Augmented Generation: MemGPT [Not really open source as the article mentions use of OpenAI]
Multimodal Image Chat
Super Quick Visual Conversations: Unleashing LLaVA 1.5
PDF Related
Super Quick: Retrieval Augmented Generation Using Ollama
Chatbot Document Retrieval: Asking Non-Trivial Questions
Super Quick: Retrieval Augmented Generation (RAG) with Llama 2.0 on Company Information using CPU
Super Quick: Fine-tuning LLAMA 2.0 on CPU with personal data
Database Related
Super Quick: LLAMA2 on CPU Machine to Generate SQL Queries from Schema
Close Source LLM (OpenAI):
PDF Related
Super Quick PDF-based ChatGPT Tutorial in Python
Chatbot Document Retrieval: Asking Non-Trivial Questions
Database Related
Super Quick: Connecting ChatGPT to a PostgreSQL database
General Info:
OpenAI Dev Day: 6 Announcements and why us RAGgers need to be worried

PlainEnglish.io 🚀
Thank you for being a part of the In Plain English community! Before you go:
- Be sure to clap and follow the writer️
- Learn how you can also write for In Plain English️
- Follow us: X | LinkedIn | YouTube | Discord | Newsletter
- Visit our other platforms: Stackademic | CoFeed | Venture