Question:
AI-powered Search with pgvector and Vultr Managed Database for PostgreSQL
Introduction
Among the most challenging aspects of relational databases is the ability to work with vector data at scale. Luckily, the PostgreSQL database server supports a pgvector extension that allows you to efficiently store and query data over Machine Learning (ML) generated embeddings.
ML Embeddings contain arrays of floating point numbers that represent objects such as images, text, video, and audio. These numerical representations express objects in a high-dimensional vector space making search similarities searches possible. Below are sample real-life applications of embeddings:
- Online retail industry: A similarity search recommends related products to customers when they add items to a shopping cart
- Audio and video streaming services: An ML embedding search helps customers find top picks based on other visiting-customer preferences
- Digital image analysis: Similarity searches help in studying images at pixel level for classification
- Web-based chatbot applications: Embedding models simulate human-like conversations to answer instant questions
- Finance industry: An embedding model detects and blocks fraud based on transaction patterns
This guide implements the PostgreSQL pgvector extension to run an AI-powered search application that answers customer Frequently Asked Questions (FAQs) using Python on a Ubuntu 22.04 server. You are to use sample data from the Vultr FAQ section to simulate common queries.
Prerequisites
Before your begin:
- Deploy a Ubuntu 22.04 management server on Vultr
- Deploy a Vultr Managed Database for PostgreSQL cluster
- Using SSH, access the server
- Create a standard user with sudo privileges.
- Create a free OpenAI account and create a secret key
This guide uses the OpenAI API to generate real embeddings to test the pgvector extensions. A free OpenAI account offers three API requests per minute and works well for this guide. In a production environment, add a payment method to your account and increase the limit
Activate the pgvector PostgreSQL Extension
1.Update the server packages
$ sudo apt update
2.Install the Python pip package manager
$ sudo apt install -y python3-pip
3.Using pip, install the Python PostgreSQL driver for Python and the OpenAI modules
$ pip install psycopg2-binary
4.Install the OpenAI modules
$ pip install openai numpy
5.Install the postgresql-client package
$ sudo apt install -y postgresql-client
The above command installs the PostgreSQL psql CLI tool used to access your managed database.
6.Using psql, log in to your Vultr Managed Database for PostgreSQL
$ psql postgres://vultradmin:example-password@prod-db.vultrdb.com:16751/defaultdb
The above command connects to your Vultr Managed Database for PostgreSQL using a connection sting. Replace the following details with your actual Vultr database credentials:
- username: vultradmin
- password: example-password
- host: prod-db.vultrdb.com
- port: 16751
7.Create a new sample company_db database
=> CREATE DATABASE company_db;
8.Switch to the database
=> \c company_db;
Output:
You are now connected to database "company_db" as user "vultradmin".
9.Enable the pgvector extension on each database that requires the extension
company_db=> CREATE EXTENSION vector;
When successful, your output should look like the one below:
CREATE EXTENSION
10.Query the pg_type table to verify the availability of a new VECTOR data type
company_db=> SELECT typname FROM pg_type WHERE typname = 'vector';
Output:
typname
---------
vector
(1 row)
As displayed in the above output, the new vector data type is ready for use.
11.Create a resource_base table
company_db=> CREATE TABLE resource_base (
resource_id BIGSERIAL PRIMARY KEY,
resource_description TEXT,
embedding VECTOR(1536)
);
The above command creates a table with the following columns:
- resource_id is a PRIMARY KEY that uniquely identifies records and uses the BIGSERIAL data type.
- resource_description is a text-based column that stores answers to questions that customers are likely to ask in the FAQ question.
- embedding uses the VECTOR data type with 1536 dimensions to store embeddings for the resource_description values. Later in this guide, you are to generate embeddings using the OpenAI API.
12.Exit the PostgreSQL console
company_db=> \q
Simulate the Project's Logic FlowCreate a Database Gateway
Based on the project's logic flow, create a separate database gateway file for the PostgreSQL database as described in the steps below.
1.Create a new project directory
$ mkdir project
2.Switch to the directory
$ cd project
3.Using a text editor such as Nano, create a new postgresql_gateway.py file
$ nano postgresql_gateway.py
4.Add the following contents to the file. Replace all db_... values with your actual Vultr Managed Database for PostgreSQL details
import psycopg2
class PostgresqlGateway:
def __init__(self):
db_host = 'prod-db.vultrdb.com'
db_port = 16751
db_name = 'company_db'
db_user = 'vultradmin'
db_pass = 'example-password'
self.db_conn = psycopg2.connect(host = db_host, database = db_name, user = db_user, password = db_pass, port = db_port)
def insert_resource(self, resource_description, embedding):
db_cursor = self.db_conn.cursor()
query_string = 'insert into resource_base (resource_description, embedding) values (%s, %s)'
db_cursor.execute(query_string, (resource_description, str(embedding)))
self.db_conn.commit()
return {'message': "Success"}
def get_resources(self, embedding):
db_cursor = self.db_conn.cursor()
query_string = 'select resource_id, resource_description from resource_base ORDER BY embedding <=> (%s::vector(1536)) LIMIT 1;'
db_cursor.execute(query_string, (embedding,))
rows = db_cursor.fetchall()
return list(rows)
Save and close the file
Create an Embeddings Generator
In this section, set up an application that creates embeddings during the following operations:
When populating the resource_base table
When passing clients' queries to the resource_base table for querying purposes
Create a central module to generate the embeddings instead of rewriting the logic on each file as described in the steps below.
1.Create a new embeddings_generator.py file
$ nano embeddings_generator.py
2.Add the following contents to the file. Replace the openai.api_key value with your actual OpenAPI key
import openai
class EmbeddingsGenerator:
def create_embedding(self, user_query):
try:
openai.organization = ""
openai.api_key = "YOUR-OPEN-API-KEY"
open_ai_model_id = "text-embedding-ada-002"
embedding = openai.Embedding.create(input = user_query, model = open_ai_model_id)['data'][0]['embedding']
self.embedding = embedding
self.resp_error = ""
except openai.error.RateLimitError as error:
self.resp_error = {'error': str(error)}
Save and close the file
The above application uses the OpenAI text-embedding-ada-002 model to generate the embeddings. The model is suitable for text similarity searches. It accepts text inputs and converts them to numerical representations (embeddings).
Create an Index File
To run the Python application, create an entry point to the application as described in the steps below
1.Create a new index.py file
$ nano index.py
2.Add the following contents to the file
import http.server
from http import HTTPStatus
import socketserver
from urllib.parse import urlparse, parse_qs
import json
import postgresql_gateway
import embeddings_generator
class HttpHandler(http.server.SimpleHTTPRequestHandler):
def do_POST(self):
self.send_response(HTTPStatus.OK)
self.send_header('Content-type', 'application/json')
self.end_headers()
content_length = int(self.headers['Content-Length'])
http_post_data = json.loads(self.rfile.read(content_length))
resource_description = http_post_data['resource_description']
eg = embeddings_generator.EmbeddingsGenerator()
req = eg.create_embedding(resource_description)
if eg.resp_error == "":
embedding = eg.embedding
pg = postgresql_gateway.PostgresqlGateway()
resp = pg.insert_resource(resource_description, embedding)
else:
resp = eg.resp_error
self.wfile.write(bytes(json.dumps(resp , indent = 2) + "\r\n", "utf8"))
def do_GET(self):
self.send_response(HTTPStatus.OK)
self.send_header('Content-type', 'application/json')
self.end_headers()
parsed_url = urlparse(self.path)
params = parse_qs(parsed_url.query)
query = params['query'][0]
eg = embeddings_generator.EmbeddingsGenerator()
req = eg.create_embedding(query)
if eg.resp_error == "":
embedding = eg.embedding
pg = postgresql_gateway.PostgresqlGateway()
resp = pg.get_resources(embedding)
else:
resp = eg.resp_error
self.wfile.write(bytes(json.dumps(resp , indent = 2) + "\r\n", "utf8"))
socketserver.TCPServer.allow_reuse_address = True
httpd = socketserver.TCPServer(('', 8080), HttpHandler)
print("The HTTP server is running at port 8080...")
try:
httpd.serve_forever()
except KeyboardInterrupt:
httpd.server_close()
print("The HTTP server has stopped running.")
Save and close the file
Test the Application
1.Start the application
$ python3 index.py
2.In a new terminal window, establish another SSH connection to your server
$ ssh example_user@SERVER-IP
3.Populate the resource_base table using the following curl POST commands
$ curl -X POST http://localhost:8080/ -H "Content-Type: application/json" -d '{"resource_description": "Yes, instances in a stopped state continue to reserve dedicated system resources (RAM, SSD storage, IP aliases, vCPU) and therefore incur charges until you destroy the instance. If you wish to no longer accumulate charges for a virtual machine, please use the DESTROY button in the customer portal."}'
$ curl -X POST http://localhost:8080/ -H "Content-Type: application/json" -d '{"resource_description": "We have not charged your card. What you have observed is a temporary authorization in order to validate the card provided. The hold will automatically expire based on your banks policy, generally within a few days."}'
$ curl -X POST http://localhost:8080/ -H "Content-Type: application/json" -d '{"resource_description": "At Vultr, we attempt to keep cloud pricing as consistent as possible globally. However, due to variation in regional costs (such as networking, hardware, labor, taxes, and fees), certain global data center locations may have different per hour and per month pricing for Vultr cloud services. Current data center location pricing is always available in the Control Panel."}'
$ curl -X POST http://localhost:8080/ -H "Content-Type: application/json" -d '{"resource_description": "All servers on your account are billed hourly up to the monthly rate cap. The hourly rate is determined by dividing the monthly rate by 672 hours (28 days). If your server is online for more than 672 hours in a calendar month, you will only be billed the monthly rate. Accumulated charges are invoiced to your account on the 1st of every month."}'
$ curl -X POST http://localhost:8080/ -H "Content-Type: application/json" -d '{"resource_description": "We accept Visa, Mastercard, American Express, Discover, JCB, BitPay (BTC, BCH, ETH, DOGE, PAX, BUSD, LTC, USDC, GUSD), Alipay, UnionPay, and PayPal. Depending on your geographic region, Paypal supports additional payment networks, including Giropay and bank transfers."}'
$ curl -X POST http://localhost:8080/ -H "Content-Type: application/json" -d '{"resource_description": "Vultr.com is required to collect tax in several countries around the world. This tax is generally referred to as VAT (Value Added Tax), consumption tax, or sales tax. Prices listed on our website do not include tax. Tax will be added to your invoice as a separate line item."}'
This guide uses data samples from the Vultr FAQ section. When using a free OpenAI developer account, you must send one query every 20 seconds to avoid the rate-limiting error
Output:
...
{
"message": "Success"
}
4.Send HTTP GET queries to perform a similarity search on the PostgreSQL server. For example:
- Do you charge for stopped instances?
$ curl -G http://localhost:8080/ --data-urlencode "query=Do you charge for stopped instances?"
Output:
[
[
1,
"Yes, instances in a stopped state continue to reserve dedicated system resources (RAM, SSD storage, IP aliases, vCPU) and therefore incur charges until you destroy the instance. If you wish to no longer accumulate charges for a virtual machine, please use the DESTROY button in the customer portal."
]
]
- I linked my credit card but I see a small charge on my card! What gives?
$ curl -G http://localhost:8080/ --data-urlencode "query=I linked my credit card but I see a small charge on my card! What gives?"
Output:
[
[
2,
"We have not charged your card. What you have observed is a temporary authorization in order to validate the card provided. The hold will automatically expire based on your banks policy, generally within a few days."
]
]
$ curl -G http://localhost:8080/ --data-urlencode "query=Full pricing list?"
Output:
[
[
3,
"At Vultr, we attempt to keep cloud pricing as consistent as possible globally. However, due to variation in regional costs (such as networking, hardware, labor, taxes, and fees), certain global data center locations may have different per hour and per month pricing for Vultr cloud services. Current data center location pricing is always available in the Control Panel."
]
]
- Is pricing the same in all data center locations?
$ curl -G http://localhost:8080/ --data-urlencode "query=Is pricing the same in all data center locations?"
Output.
[
[
3,
"At Vultr, we attempt to keep cloud pricing as consistent as possible globally. However, due to variation in regional costs (such as networking, hardware, labor, taxes, and fees), certain global data center locations may have different per hour and per month pricing for Vultr cloud services. Current data center location pricing is always available in the Control Panel."
]
]
- What payment methods do you accept?
$ curl -G http://localhost:8080/ --data-urlencode "query=What payment methods do you accept?"
Output:
[
[
5,
"We accept Visa, Mastercard, American Express, Discover, JCB, BitPay (BTC, BCH, ETH, DOGE, PAX, BUSD, LTC, USDC, GUSD), Alipay, UnionPay, and PayPal. Depending on your geographic region, Paypal supports additional payment networks, including Giropay and bank transfers."
]
]
Based on the above results, the application returns the most relevant answer for each query to the user
Check the Embeddings
After using the curl commands to send HTTP POST requests, the application populates the resource_base table with AI-generated embeddings. Access the company_db database to verify the embeddings as described in this section.
1.Access the PostgreSQL database server. Append the company_db database to your connection string to access the database directly
$ psql postgres://vultradmin:example-password@prod-db.vultrdb.com:16751/company_db
2.Query the resource_base table using the SQL command below. Apply the PostgreSQL SUBSTRING() and RIGHT() functions to return only the first and last few characters from the resource_description and embedding columns
company_db=> SELECT
resource_id,
CONCAT(SUBSTRING(resource_description, 0, 15), '...', RIGHT(resource_description, 15)) as resource_description,
CONCAT(SUBSTRING(embedding::VARCHAR, 0, 30), '...', RIGHT(embedding::VARCHAR, 15)) as embedding
FROM resource_base;
Your output should look like the one below:
resource_id | resource_description | embedding
-------------+----------------------------------+-------------------------------------------------
1 | Yes, instances...ustomer portal. | [0.00080605154,-0.04708257,0....7,0.0064484123]
2 | We have not ch...hin a few days. | [-0.023812525,-0.011136047,0....,-0.0033343954]
3 | At Vultr, we a... Control Panel. | [-0.0018340687,-0.028246619,0...24,0.015378715]
4 | All servers on...of every month. | [-0.003029692,-0.016905943,0....,0.00010902301]
5 | We accept Visa...bank transfers. | [0.009505584,0.0031462372,0.0...-0.00023975992]
6 | Vultr.com is r...rate line item. | [-0.00049098214,-0.039759535,...5,0.0070797624]
(6 rows)
3.Create an index on the resource_base table. This is necessary when scaling your application and have more records in the table. The lists parameter in the ivfflat index sets the number of clusters that PostgreSQL creates when building the index. PostgreSQL uses the index clusters in its algorithms to find the relation between vectors. Apply the following formula when setting the list value:
- For a table with less than one million rows use:
lists = rows / 1000
- For tables with more than one million rows use:
lists = squareroot(rows)
- Verify that you have a minimum of ten clusters. When records in the sample application are still few, use the minimum value of 10
e_commerce=> CREATE INDEX ON resource_base USING ivfflat (embedding vector_cosine_ops) WITH (lists = 10);
Conclusion
In this guide, you implemented the PostgreSQL pgvector extension that generates and queries data over ML-generated embeddings. You created a sample database that stores a company knowledge base using vector data, and used the PostgreSQL cosine distance operator <=> to query data to display the most relevant results. For more information, visit the PgVector extension repository.