Cluster Engine을 중심으로 PyTorch, Hugging Face 같은 프레임워크와 Kubernetes, Docker 같은 강력한 환경을 통합하세요.
클러스터 전체에 걸쳐 컨테이너화된 워크로드를 자동으로 확장하고 관리하여, GPU 활용도와 가동 시간을 극대화합니다.
Seamlessly orchestrate complex tasks with Kubernetes, optimized for AI/ML, HPC, and cloud-native applications in a GPU cloud environment.
Get Started NowRun AI workloads with secure, high-performance GPU-optimized containers or bring your own configurations into our scalable container management system.
Containers are automatically deployed with minimal setup, reducing manual engineering and packaging time in GPU cloud operations.
Get Started NowMonitor GPU usage and system performance in real-time with custom alerts, ensuring stability across clustered GPU environments.
Track every container’s performance from start to finish, with full visibility into resource usage and job health.
Get Started NowGrant fine-grained permissions to teams working on AI projects using GPU cloud infrastructure, managing access with IAM policies.
Easily manage GPU and cluster access per team or project through role-based user groups—essential for scaling AI deployments securely.
Get Started NowIsolated VPCs for each customer to ensure secure, separate network and compute resources.
Dedicated private subnets and secure messaging for end-to-end data integrity and safety.
Ensure fast and secure access to your GPU cloud platform via private connections and dedicated virtual gateways.
Get Started Now자주 묻는 질문에 대한 빠른 답변을 저희 사이트에서 확인하세요 자주 묻는 질문.
The Cluster Engine is GMI Cloud’s on-demand compute power offering platform. There are at least three types of compute services available, which are CE-CaaS(Container), CE-BMaaS(Bare-metal) and CE-Cluster(Managed K8S/Slurm). By leveraging kubernetes, openstack orchestration softwares, and by deploying RDMA networks, Cluster Engine is designed to automate different compute workloads with fine-grained control.
The CE-CaaS service, which offers prebuilt, GPU‑optimized containers for rapid deployment of AI application workloads, uses Native Kubernetes to ensure seamless, secure, and automated orchestration of small compute workload, with the option to bring your own custom image templates.
우리는 pip와 conda를 사용하여 고도로 사용자 정의 가능한 환경을 갖춘 텐서플로우, 파이토치, 케라스, 카페, MXNet 및 ONNX를 지원합니다.
The CE leverages the design of organizations to isolate tenants, incorporating organizational user management with fine-grained role-based access control (RBAC). As for the network isolation and access control, the virtual private network (VPC) mechanism is applied to the internal network isolation with elastic ip for public access, while firewall rules are also introduced to assure the public network security.
The CE provides real‑time monitoring with customizable alerts to maintain visibility over resource usage and container health. It also includes proprietary high‑performance storage filesystem shared between containers and bare-metals, which is an ideal solution for both AI training or generative AI Inferencing workloads.