This post is part of the GenAI Series The only way we can interact with an LLM model is by passing an instruction to retrieve a response. That instruction is called a prompt. In this post, we are going to discuss what a prompt is and what prompt engineering entails What is a Prompt A prompt is simply the instruction or question you give to an AI or language model to get a response. Think of it as a way to guide the AI toward the answer or output you want. Imagine you’re talking to a super-smart friend who’s always eager to help but needs a little guidance on what…
-
-
Getting started with OpenAI Assistant APIs and Python
Earlier I had discussed using OpenAI Embeddings API to come up with a product recommendation system. In this post, I am going to discuss Assistant APIs, The APIs OpenAI provided to integrate custom GPTs. I am going to create a blog title generator in Flask that will be generating blog titles in the tech/gadget space. In case you are in a hurry you may want to watch the demo video of it: What is a custom GPT? According to OpenAI: …ChatGPT that combine instructions, extra knowledge, and any combination of skills. Custom GPTs, as the name says, are the customized versions of chatGPT that rely on your own data and…
-
Automate Your Finance Blog with WordPress and Google Bard in Python
Introduction Earlier I gave you an overview of how you can you GPT APIs to build a stock sentiment analysis system. In this post, I am going to introduce you to Google Bard, Google’s answer to OpenAI’s chatGPT. Bard has been around for a few months but they recently offered APIs to access their LLM. Initially, they introduced PaLM APIs but very recently they have offered Gemini Pro and Gemini Pro Vision APIs which means that like GPT4 Vision, you can now use Bard APIs to process images. We are not going to discuss image processing capabilities at the moment. I am going to offer a use case where I…