prompt-ai 0.2.0

Creator: railscoder56

Last updated:

Add to Cart

Description:

promptai 0.2.0

Prompt-AI Library
Prompt-AI is a powerful library designed to optimize AI-driven prompt handling and response generation using the Gemini API. By introducing structured database management and efficient embedding retrieval, Prompt-AI significantly enhances performance, reduces response times, and provides a seamless solution for integrating AI models into various applications.
Key Features

Efficient Embedding Management: Prompt-AI stores pre-generated embeddings in a structured database, significantly reducing computational overhead and improving response times.
Real-Time Updates: Manage datasets and dataframes efficiently, ensuring that embeddings are generated once and reused across multiple sessions.
Performance and Scalability: The streamlined approach enhances performance and scalability, making Prompt-AI ideal for chatbots, recommendation systems, and other AI-powered tools.
Versatile Integration: Seamlessly integrates with Node.js endpoint servers, bridging different technologies and workflows.

Installation
To install Prompt-AI, use pip:
pip install prompt-ai

To upgrade to latest version:
pip install prompt-ai --upgrade


After Installation follow these steps to use promp-ai
1. Generate an API Key
To begin, you’ll need to generate an API key. Follow the link below to generate your API key:
Generate API Key
Brief Summary of Gemini Model
The Gemini model is a powerful AI-driven model designed for generating contextually relevant responses to user prompts. Unlike traditional approaches where embeddings are generated on each run, Prompt-AI integrates a more efficient workflow by storing pre-generated embeddings in a NoSQL database. This allows for faster response times and reduces computational overhead, making it ideal for applications like chatbots, recommendation systems, and other AI-powered tools.
3. Setting up MongoDB (In later versions: SQL and Cloud database will be added)

Create a Database in Mongo Atlas or MongoDB Compass (Which you feel good).
Create collection and Documents.
Set the document in this structure:

{
"id": 1,
"title": "Gork vs Chat-gpt",
"text": "In the rapidly evolving landscape of artificial inte...",
}

4. Using Prompt-AI to manage prompts and generate response
Prompt-AI provides two core functions to help you manage prompts and generate responses:
1) configure(mongo_uri: string, db_name: string, collection_name: string, columns: array, API_KEY: string, embeddings: bool)
configure(mongo_uri, db_name, collection_name, columns, API_KEY, embeddings)

This function configures the connection to your MongoDB Atlas and sets up the necessary parameters for generating embeddings.

mongo_uri: string
This should contain your MongoDB Atlas connection string.

mongo_uri = 'mongoDB connection string'


db_name: string
The name of your MongoDB database.

db_name = 'Database Name'


collection_name: string
The name of the collection in your database where the data is stored.

collection_name = 'collection name'


columns: array
An array of strings, each representing a field name present in each document of the collection. The field which contains ANSWER data must be named with 'text'.

columns = ['id', 'title', 'text']


API_KEY: string
The API key generated in the first step.

API_KEY = 'key generated in first step'


embeddings: bool
A boolean flag indicating whether embeddings need to be created (true) or if they already exist (false).

embeddings = True or False

This function call will return datasets in form of tabular dataframe.
id title text embeddings
1 "chat-gpt features" "chat-gpt has..." [4.0322, 2.3344, 1.09...]
2 "Gork vs Chat-gpt" "Gork have plenty..." [1.1702, 0.4184, 5.19...]

overall configure function will look like this...
uri = 'your connection string'
db = 'database_name'
col = 'collection_name'
API_KEY = 'api_key generated in first step'
column = ['id', 'title', 'text']
embeddings = True
dataframe = configure(uri, db, col, API_KEY, column, embeddings)

2) generate(user_prompt, dataframe)
generate(user_prompt, dataframe)

This function processes the user’s prompt, interacts with the database, and returns the AI-generated response.
dataframe will be used here inside generate() function.
@app.post("/api")
async def generate_response(request: PromptRequest):
# Extract the prompt from the request body
user_prompt = request.prompt

# calling generate() function with prompt and dataframe as parameter

response = generate(user_prompt, df)

return {"response": response}

Now on making post request to this endpoint keep prompt: string inside body of the post request in JSON format.
And for handling response you can have it in this way:
const response = await axios.post('endpoint_url/api', {prompt});
res = response.data.response;

Here I have used axios, but you can also use fetch api to make post request and fetch the response in the same way.

With Prompt-AI, you can efficiently manage AI-driven prompt handling, leveraging the Gemini model's capabilities with enhanced performance and scalability. Whether you’re building a chatbot, a recommendation system, or any other AI-powered application, Prompt-AI provides a streamlined and powerful solution.
Happy Coding :)

License

For personal and professional use. You cannot resell or redistribute these repositories in their original state.

Customer Reviews

There are no reviews.