Skip to content

d.run: The Ideal Platform Supporting Generative AI

Download PDF

d.run is an integrated intelligent computing platform launched by DaoCloud, specifically designed for model development, model training, inference services, and intelligent applications based on K8s and AI frameworks. d.run is not just a tool for running workloads like web applications and microservices. For artificial intelligence (AI) and machine learning (ML) workloads, such as large language models (LLM), d.run is the ideal platform for end-to-end lifecycle management.

In 2021, an authoritative report pointed out that 42% of respondents indicated they had used platforms like d.run based on K8s for AI/ML workflows. Last year, this proportion increased to 65%, with expectations for even higher numbers this year.

This widespread application spans various industries: from cutting-edge innovative companies like OpenAI to AI cloud service providers like CoreWeave, and well-known brands such as Shell and Spotify. In the domestic market, organizations ranging from retail e-commerce to financial government enterprises, from large and medium-sized state-owned enterprises to confidential units, are beginning to rely on platforms like d.run to support their AI/ML distributed workloads.

This article will explore why d.run provides unique support in every lifecycle stage of AI/ML research and engineering.

Introduction

It is well known that K8s is an efficient container orchestration and management platform in distributed computing environments. K8s was initially developed as an orchestration project by Google to manage its internal computing clusters and massive applications. After becoming open source, K8s has become the practical standard for deploying, scaling, and managing containerized applications in various environments.

A recent series of cases show that K8s platforms like d.run are very useful for some emerging use cases: Organizations both domestically and internationally that seek efficient development, training, and deployment of LLMs have begun to leverage platforms like d.run. d.run offers numerous advantages for comprehensive support throughout the entire lifecycle of LLMs, eliminating the need to integrate complex frameworks across different technology stacks.

Advantages of d.run at Each Stage

From model pre-training to model deployment, to fine-tuning experiments and application building, d.run can be utilized at every stage of the LLM lifecycle.

Model Pre-training

Model Pre-training

During the model pre-training phase, d.run provides a solid foundation for model training with its unparalleled scalability and resilience. One of d.run's biggest advantages is its ability to automatically scale resources based on demand, which is a critical feature for AI/ML workloads facing enormous computational needs. d.run achieves this by automating the lifecycle management of Pods; if a Pod encounters an error, it will be automatically terminated and restarted. In other words, Pods have self-healing capabilities.

d.run also allows for easy addition or removal of Pods and nodes on demand, enabling dynamic scaling to meet evolving workload requirements. Its declarative infrastructure facilitates users in communicating their needs, thereby simplifying management processes. These are powerful development features that cannot be obtained using other tools like Slurm. This means you can achieve higher output, train models more efficiently, without worrying about the limitations of the underlying infrastructure.

Tools like Jupyter Notebooks and VSCode are essential for LLM experiments and prompt engineering, and d.run's built-in network abstraction allows data scientists to easily create development environments and integrate with these development tools. Additionally, port forwarding and configuration management occur automatically, simplifying the configuration of end-user workspaces (tenants) and the environment and network management for cluster administrators.

Model Fine-tuning

Model Fine-tuning

Although d.run has all the tools needed for developing LLMs, many companies today do not start from scratch to build large language models, often opting for existing models and then customizing and fine-tuning them based on their specific environments. This scenario of fine-tuning existing models is also very suitable for platforms like d.run due to its dynamic adaptability. Unlike Slurm, d.run can handle multiple workloads in parallel, making the training process more efficient. Another advantage lies in the rich tool ecosystem that d.run builds for model training, which includes Kubeflow (Operators designed for Pytorch, Tensorflow, and MPI), Kueue, HwameiStor, and Spiderpool, among other efficient specialized tools.

Model Deployment

Model Deployment

When it comes to LLM model deployment or model inference services, d.run provides a simplified process: you simply present an endpoint to data scientists. The network stack simplifies the process of releasing models to the outside world, easily pushing the models to the consumer side. d.run offers a comprehensive toolkit and a rich ecosystem for model deployment, including load balancing, Ingress controllers, and network policies. This aids in the seamless deployment of LLM endpoints and their integration with services and applications.

Infrastructure abstraction further simplifies the deployment process, ensuring scalability and automatic scaling capabilities. d.run abstracts all underlying infrastructure into a universal API for managing various containerized applications. Therefore, no matter where the workload runs, you can use the same tools and processes. This greatly simplifies the management and monitoring of production environments.

Prompt Engineering

Prompt Engineering

The advantages do not stop there. After deploying LLM models, d.run can enhance user experience when developing applications or engaging users in model experiments. For example, hosting applications on platforms like Gradio or Streamlit using d.run is almost effortless, as the d.run community has a complete toolkit specifically for cross-platform application hosting. This simplifies the deployment process, while service endpoints and automatic scaling capabilities ensure smooth and scalable experiments.

Security

At every stage, d.run provides robust security to ensure the safety of your data and intellectual property. For example, d.run's built-in global role-based access control (RBAC) enables fine-grained access control, granting appropriate permissions to users or service accounts; Pod security contexts allow you to set security attributes at the Pod level, thereby reducing the attack surface within the cluster. These features ensure the environmental security of containers, models, and datasets throughout the entire AI/ML lifecycle.

Real Success Stories

These advantages are not just theoretical; many of today's most innovative cutting-edge companies are running and managing the entire lifecycle of LLMs on K8s platforms like d.run, including leading tech companies operating large-scale clusters (such as OpenAI) and emerging AI cloud service providers (Core Weave, Lambda Cloud Services).

For example, OpenAI's cluster consists of over 7,500 nodes, supporting its large language models and distributed machine learning workloads. Despite alternatives like Slurm, K8s provides OpenAI engineers with a superior development experience and a cloud-native integrated environment. With K8s, they can also easily and flexibly deploy containers, manage heterogeneous nodes, and handle dynamic infrastructure components.

Christopher Berner, Infrastructure Lead at OpenAI says

The research team can now leverage the framework we built on K8s to easily initiate model experiments and scale experiments by 10x or 50x without spending too much effort on management.

OpenAI runs K8s across multiple data centers in Azure, benefiting from a cluster-wide MPI communication domain that supports cross-node parallel jobs and batch operations. As a batch scheduling system, K8s' autoscaler ensures dynamic scaling, reducing idle node costs while maintaining low latency. Moreover, K8s is incredibly fast, allowing those researching distributed training systems to initiate and scale experiments in days rather than months.

By adopting K8s, OpenAI has found that model portability is excellent, allowing model experiments to be easily migrated between clusters. K8s provides a consistent API that simplifies this migration process. Additionally, while leveraging Azure's infrastructure, OpenAI can also fully utilize its own data centers, saving costs while enhancing availability.

Of course, it is not only large companies like OpenAI that can benefit: K8s platforms like d.run have become mainstream platforms for building, training, and deploying language models, completely revolutionizing the AI landscape. Hosting AI/ML workloads on d.run offers multiple advantages: scalability, flexibility, network abstraction, and a better user experience during experimentation. With d.run, you can use the best tools and technologies to meet your needs, easily building, training, and deploying AI/ML workloads.