Create a GPU Instance
Procedures
Log in to zenConsole, go to Products > Virtual Machine > GPU & AI > GPU Cloud, and click Create GPU Instance.
Select a location. Choose the closest region for optimal performance.
Choose your use case and image. Choose an available pre-built environment to quickly set up the instance with essential tools.
AI/ML Inference Ollama image preinstalled with:
Llama 3.1:8b
,NvidiaDriver 550.90.07
,CUDA 12.6
,cuDNN 9.3.0
,Python 3.12.3
,Ollama 0.3.9
,JupyterLab 4.2.5
. Used for deploying models and running inference tasks.Common Use Ubuntu Server 24.04 LTS image preinstalled with:
NvidiaDriver 550.90.07
,CUDA 12.6
,cuDNN 9.3.0
. Used for common computing needs.
Select a GPU model. Opt for the NVIDIA GeForce RTX 4090 with 24 GiB of vRAM, ensuring high-performance computation with 82+ TFLOPS of FP32 compute power, perfect for demanding AI tasks.
Configure storage.
Set your storage requirements, basic and standard NVMe SSDs for fast I/O operations.
Manage instance access.
Add an SSH key pair to securely access your instance.
Use the default user "ubuntu" or create a custom password for login.
Choose quantity. Set the number of instances you want to create and give the instance a resource name.
More settings.
Select your OS time zone.
Assign your resources to your desired resource group.
Confirm order. Check the order summary and confirm your order.
Key Considerations
Location and Latency: Choosing the right region is important for reducing latency. Select a region closest to your user base or data source.
Storage Allocation: Ensure you allocate sufficient NVMe SSD storage based on your workload requirements. If working with large datasets, you might need to expand storage.
Security:
Recommended to use SSH key pairs for secure access.
Regularly update your software and patches to maintain security.
Last updated