Emrah: To some extent, yes. Being able to scale these resources dynamically and efficiently without needing to log into each machine individually is an important aspect of cloud. However, it is more than just a data center with a lot of machines. ‘Cloud’ is more of a catch-all. It is often a term that is used to encompass many technologies that provide computing power, storage, and a network infrastructure that is scalable and flexible.
Emrah: The difference between the two is basically who owns the infrastructure. A third party owns the infrastructure of public cloud and offers services and resources to the customer. There is a lot of standardization of technology that is easier for the less experienced user, so that is a great choice if you don’t have any other concerns about public clouds. Many companies are satisfied with the public cloud option.
However, when it comes to customization or a desire to use different technology, public cloud can present challenges. When you need flexibility, the containerized, moveable and portable technology is better. My recommendation would be to have a private cloud on premises so you can use containerization and Kubernetes to orchestrate your payload and configure your work according to your needs.
As long as you have an effective flow of communication between your cloud and on prem infrastructure, private versus public cloud is really a matter of preference. The choice comes down to whether you want to own the infrastructure and the responsibility that comes with it, or you want to outsource that so you can focus on your main function. The choice is ultimately use-case specific, driven by the factory’s needs and requirements
Emrah: Most of what we understand as cloud now is derived from the cloud revolution of the early 2000s, which introduced virtual machines as representations of physical machines.
The container is a simple, lightweight virtual machine and containerization refers to a method of packaging an application and its dependencies in a single unit. Docker, Kubernetes and Helm charts are all tools of container management.
Docker is the most common platform that enables you to do containerization by automating the process. You’ll see containerization and the use of Docker at the development stage.
Kubernetes and Helm charts, on the other hand, are used during the deployment phase to manage the deployment at scale. Kubernetes is basically an orchestration tool for managing containerized applications and the Helm charts are pre-configured Kubernetes resources.
Emrah: Cost plays a big role in this. For instance, if you already have the infrastructure – servers with a good lifetime of 5 to 10 years left—you might not want to incur additional cost for the next five years for new technology. You’d only want to move to the cloud if there are obvious advantages from the standpoint of scaling, flexibility or such. However, it’s really based on a particular use case.
If, for example, you are running a very GPU-heavy application, you’ll probably want to be closer to an on-prem solution rather than a cloud solution because running GPUs on the cloud is pretty expensive. If I’m going to be doing this on a continuous basis for an extended period of time, I would probably rather have a GPU data center on my premises or rented in another physical location because it could be half as much as running this on the cloud. Most often, in a case like this, these are AI companies who are training large language models (LLMs) and others. They have their own data centers or rent them to do that training.
Emrah: First, we need to define what we mean by the AI boom. In certain areas, such as training large language models, there has definitely been significant progress. However, this progress often relies on scaling up existing technologies like graphics processing units (GPUs), which isn’t necessarily tied to the cloud.
When we talk about revolutionary changes in AI training techniques and methodologies, like the advent of transformers, we’re building on an existing understanding of these models. This has created a high demand for compute resources, which the current supply of GPUs is needed. From this perspective, the cloud isn’t a bottleneck; it’s the underlying technology, like GPUs, that plays a crucial role.
However, if there’s an inflection point where AI no longer relies on GPUs or we can utilize a different type of computing tool, then the cloud might need to adapt to embrace this new technology. For example, at that point, we could leverage all the connected resources globally to improve our training processes.
One area where the cloud might need to catch up is Kubernetes. While Kubernetes works with GPUs, it doesn’t do so as efficiently as we’d like for large-scale AI training. This is why many prefer non-Kubernetes deployments for significant AI training tasks. So, in that sense, the cloud might need to evolve, but in other areas it’s a different story.
Cloud and AI deployment
About the Authors

Samantha is the Global Product Manager overseeing SmartFactory AI™ Productivity, Simulation AutoSched® and Simulation AutoMod®. Prior to joining Applied Materials Automation Product Group Samantha was Manager of Industry 4.0 at Bosch, where she also was previously a Data Scientist. She also has experience as a Research Associate for the Geographic Information Science and Technology Group of Oak Ridge National Laboratory. She holds a M.S. in Mathematics from the University of Tennessee, Knoxville, and a B.S. in Mathematics from University of North Georgia, Dahlonega.

Emrah leads the team delivering AI/ML solutions and cloud transformation of APG software for semiconductor manufacturers. He is a pioneer in developing SaaS applications, cloud transformation efforts and building optimization and analytics frameworks for cloud computing. He holds patents in semiconductor manufacturing, cloud analytics and retail science. He also has a well-established research record in planning and scheduling in semiconductor manufacturing. His work is published in IEEE and INFORMS and has been presented in IERC and INFORMS. He earned his Ph.D. in Operations Research and Industrial Engineering from University of Texas at Austin. He holds B.S. and M.S. degrees in Industrial Engineering from Bilkent University, Turkey.