Generative AI models require significant amounts of computational power, particularly during the training phase because they typically involve large-scale matrix multiplication and other computationally intensive operations. For training, it is common to use many powerful GPUs to accelerate the process. For inferencing, less powerful hardware can be used but a significant amount of compute power to provide fast responses is required.