Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

AI model servers- Docker container

The main hardware cost of the BusinessGPT deployment are the AI servers responsible for answering questions. These utilize GPUs.

These servers will also be used to create embeddings for processing and preparing the data for AI.

Server spec:

Linux server Ubuntu 2CPU, 8GB RAM HDD/SSD with a R/W speed of at least 100MB/s.

GPU: CUDA 11.8+, Min 12GB RAM.
The disk size should be 30% larger than the original content base file size.

Multiple docker containers with a load balancer may be used for higher performance / High availability.

Graphics Cards

Below are GPU options that are available for purchase for standard servers. The difference between the cards is the number of questions per minute that can be processed.

...

Linux server Ubuntu 4CPU, 8GB RAM. HDD/SSD with a R/W speed of at least 100MB/s.
The disk size should be 30% larger than the original content file size.

AI Server - Docker container

Linux server Ubuntu 2CPU, 8GB RAM HDD/SSD with a R/W speed of at least 100MB/s.

GPU: CUDA 11.8+, Min 12GB RAM.
The disk size should be 30% larger than the original content base file size.

Multiple docker containers may be used for higher performance / High availability with a load balancer.