Assessing models Method 3 Using qroq Fast AI Inference

This page is to learn the 3rd way of accessing LLM on OpenWEB UI using groq (note that is different from grok). Method 1 is here. Method 2 is here.

Method 1 is likely for LLMs that are smaller in size and your computer has adequate computing resources (e.g., GPU? Ram?) to handle the models.
Method 2 will be for calling on direct APIs of large models like Open AI models.

Method 3 is like Method 2 except we call on Groq API which is hosting the large models.

So WHY introduce Method 3 here if it is the same as Method 2?

Setup

The setup is VERY EASY. See video Video 1 below.

Remember while it is free, do NOT share your API key with anyone and do note that there will be rate limits since it is free.


If you prefer to watch a video to understand how to download and use a model, check this Youtube below:

Video 1