Use Cases
Common applications and workflows for rented GPU computing power.
AI model training
Section titled “AI model training”Machine learning workflows
Section titled “Machine learning workflows”Deep learning frameworks:
- PyTorch with CUDA acceleration
- TensorFlow/Keras GPU support
- JAX for research workflows
- Hugging Face Transformers
Pre-installed libraries:
# Verify GPU availabilityimport torchprint(torch.cuda.is_available())print(torch.cuda.get_device_name())
# Common ML stackimport tensorflow as tfimport pandas as pdimport numpy as npimport sklearn
Training optimization:
- Mixed precision training for faster performance
- Gradient accumulation for large batch sizes
- Model checkpointing for long training runs
- TensorBoard monitoring for loss tracking
Popular model types
Section titled “Popular model types”Computer vision:
- Image classification (ResNet, EfficientNet)
- Object detection (YOLO, R-CNN)
- Semantic segmentation (U-Net, DeepLab)
- Generative models (StyleGAN, DCGAN)
Natural language processing:
- Transformer fine-tuning (BERT, GPT)
- Language model training
- Text generation and completion
- Sentiment analysis and classification
Time series and forecasting:
- LSTM and GRU networks
- Transformer architectures for sequences
- Financial market prediction
- Demand forecasting models
Training best practices
Section titled “Training best practices”Resource management:
# Monitor GPU memory usageimport torchprint(f"Allocated: {torch.cuda.memory_allocated()}")print(f"Cached: {torch.cuda.memory_reserved()}")
# Clear cache when neededtorch.cuda.empty_cache()
# Use gradient accumulation for large modelsfor i, batch in enumerate(dataloader): loss = model(batch) loss = loss / accumulation_steps loss.backward()
if (i + 1) % accumulation_steps == 0: optimizer.step() optimizer.zero_grad()
Checkpointing:
# Save model checkpoints regularlyif epoch % save_frequency == 0: torch.save({ 'epoch': epoch, 'model_state_dict': model.state_dict(), 'optimizer_state_dict': optimizer.state_dict(), 'loss': loss, }, f'/workspace/checkpoint_epoch_{epoch}.pth')
Stable Diffusion and image generation
Section titled “Stable Diffusion and image generation”Pre-configured environments
Section titled “Pre-configured environments”Automatic1111 WebUI:
- Access via
http://10.77.x.2:7860
- 100+ pre-installed models
- Extensions for enhanced functionality
- ControlNet for guided generation
ComfyUI:
- Node-based workflow interface
- Advanced model chaining
- Custom node support
- Real-time parameter adjustment
Model management
Section titled “Model management”Popular models included:
- Stable Diffusion 1.5 and 2.1
- SDXL (Stable Diffusion XL)
- Realistic Vision
- DreamShaper
- Anything V3/V4/V5
Custom model installation:
# Download models to appropriate directoriescd /workspace/stable-diffusion-webui/models/Stable-diffusion/wget https://huggingface.co/runwayml/stable-diffusion-v1-5/resolve/main/v1-5-pruned.ckpt
# For ComfyUIcd /workspace/ComfyUI/models/checkpoints/wget https://example.com/custom-model.safetensors
Advanced techniques
Section titled “Advanced techniques”ControlNet workflows:
- Pose-guided generation
- Depth map conditioning
- Edge detection guidance
- Style transfer applications
LoRA fine-tuning:
- Character-specific training
- Style adaptation
- Concept reinforcement
- Efficient model customization
Upscaling and enhancement:
- Real-ESRGAN integration
- GFPGAN face restoration
- CodeFormer enhancement
- Custom upscaling models
Cryptocurrency mining
Section titled “Cryptocurrency mining”Pre-configured miners
Section titled “Pre-configured miners”GPU miners included:
- T-Rex (NVIDIA optimized)
- TeamRedMiner (AMD optimized)
- lolMiner (dual mining support)
- NBMiner (LHR bypass)
CPU miners:
- XMRig (Monero optimized)
- CPUMiner (general purpose)
- SRBMiner (multi-algorithm)
Mining setup
Section titled “Mining setup”Auto-configuration:
# Check available minersls /opt/miners/
# T-Rex for NVIDIA cards/opt/miners/t-rex/t-rex -a ethash -o stratum+tcp://pool:4444 -u wallet -w worker
# TeamRedMiner for AMD/opt/miners/teamredminer/teamredminer -a ethash -o stratum+tcp://pool:4444 -u wallet
Pool configuration:
- Pre-configured pool lists
- Automatic failover support
- Regional pool optimization
- SSL/TLS encrypted connections
Profitability optimization
Section titled “Profitability optimization”Algorithm switching:
- Automatic profit switching
- Market condition monitoring
- Real-time profitability calculation
- Multi-pool support
Performance tuning:
# NVIDIA optimizationnvidia-smi -pm 1 # Enable persistence modenvidia-smi -pl 250 # Set power limit
# Memory clock optimizationnvidia-settings -a [gpu:0]/GPUMemoryTransferRate[3]=8000
# Monitor performancewatch -n 1 nvidia-smi
Earnings tracking:
- Real-time hashrate monitoring
- Pool statistics integration
- Profitability calculators
- Historical performance data
3D rendering and visualization
Section titled “3D rendering and visualization”Rendering engines
Section titled “Rendering engines”Blender integration:
- CUDA/OpenCL acceleration
- Cycles and Eevee renderers
- GPU-accelerated viewport
- OptiX denoising support
Other rendering software:
- Cinema 4D with Redshift
- Autodesk Maya with Arnold
- 3ds Max with V-Ray
- Houdini with Mantra/Karma
Rendering workflows
Section titled “Rendering workflows”Animation rendering:
# Command-line Blender renderingblender -b scene.blend -o /workspace/render/frame_#### -s 1 -e 250 -a
# GPU-specific optimizationblender -b scene.blend -P gpu_render_script.py
Batch processing:
# Python script for batch renderingimport bpyimport os
# Set GPU renderingbpy.context.preferences.addons['cycles'].preferences.compute_device_type = 'CUDA'bpy.context.preferences.addons['cycles'].preferences.get_devices()
# Render multiple scenesscene_files = ['/workspace/scene1.blend', '/workspace/scene2.blend']for scene in scene_files: bpy.ops.wm.open_mainfile(filepath=scene) bpy.context.scene.render.filepath = f'/workspace/output/{os.path.basename(scene)}' bpy.ops.render.render(write_still=True)
Optimization techniques
Section titled “Optimization techniques”Memory management:
- Tile-based rendering for large scenes
- Out-of-core geometry handling
- Texture streaming optimization
- GPU memory monitoring
Performance tuning:
- Optimal tile sizes for GPU
- Denoising parameter adjustment
- Sample count optimization
- Light cache configuration
Gaming and streaming
Section titled “Gaming and streaming”Remote gaming setup
Section titled “Remote gaming setup”Steam integration:
- Steam client pre-installed
- GPU-optimized game settings
- Cloud save synchronization
- Workshop content support
Game streaming:
- Parsec client configuration
- OBS Studio for recording
- Low-latency streaming setup
- Hardware encoding optimization
Performance optimization
Section titled “Performance optimization”Graphics settings:
# NVIDIA settings for gamingnvidia-settings --assign [gpu:0]/GPUPowerMizerMode=1nvidia-settings --assign [gpu:0]/GPUFanControlState=1
# Display configurationxrandr --output HDMI-1 --mode 1920x1080 --rate 60
Network optimization:
- Low-latency VPN configuration
- QoS settings for gaming traffic
- Bandwidth allocation optimization
- Jitter reduction techniques
Popular gaming scenarios
Section titled “Popular gaming scenarios”Cloud gaming:
- AAA title streaming
- VR content rendering
- Multi-monitor gaming setups
- Competitive gaming optimization
Content creation:
- Game recording and editing
- Live streaming production
- Highlight compilation
- Social media content creation
Scientific computing
Section titled “Scientific computing”Research applications
Section titled “Research applications”Computational biology:
- Protein folding simulations
- Molecular dynamics
- Bioinformatics pipelines
- Drug discovery workflows
Physics simulations:
- Finite element analysis
- Computational fluid dynamics
- Particle physics calculations
- Climate modeling
Data science:
- Large-scale data processing
- Statistical modeling
- Visualization and plotting
- Interactive analysis
Software environments
Section titled “Software environments”Pre-installed tools:
- MATLAB with GPU acceleration
- Jupyter Lab with scientific kernels
- R with GPU packages
- GNU Octave
Custom installations:
# Install domain-specific toolsconda install -c conda-forge gromacspip install biopython rdkit-pypiapt install quantum-espresso
# GPU-accelerated librariespip install cupy-cuda11x # GPU NumPypip install rapids-cudf # GPU pandas
Workflow optimization
Section titled “Workflow optimization”Parallel processing:
# Multi-GPU accelerationimport torch.nn as nnfrom torch.nn.parallel import DataParallel
model = MyModel()if torch.cuda.device_count() > 1: model = DataParallel(model)model.to('cuda')
Memory management:
- Efficient data loading strategies
- Batch processing optimization
- Out-of-core computing techniques
- Memory mapping for large datasets
Development and testing
Section titled “Development and testing”Software development
Section titled “Software development”IDE access:
- VS Code Server (
http://10.77.x.2:8080
) - JupyterLab development environment
- Git integration and version control
- Package manager access
Testing environments:
- Isolated dependency testing
- Performance benchmarking
- GPU compatibility validation
- Cross-platform testing
DevOps workflows
Section titled “DevOps workflows”Container development:
# Build Docker images with GPU supportdocker build --tag myapp:gpu .docker run --gpus all myapp:gpu
# Kubernetes GPU workload testingkubectl apply -f gpu-workload.yaml
CI/CD integration:
- Automated GPU testing pipelines
- Performance regression detection
- Model validation workflows
- Deployment testing
Cost optimization strategies
Section titled “Cost optimization strategies”Efficient resource usage
Section titled “Efficient resource usage”Right-sizing rentals:
- Match GPU tier to workload requirements
- Use performance monitoring to validate sizing
- Consider memory requirements vs compute needs
- Avoid over-provisioning
Time management:
# Set up automatic completionecho "sudo shutdown -h +240" | at now # 4-hour auto-shutdown
# Monitor remaining timetimeout 3600 python long_training.py # 1-hour timeout
Batch processing
Section titled “Batch processing”Job scheduling:
- Queue multiple tasks for single rental
- Optimize task ordering for efficiency
- Use containerization for reproducible environments
- Implement checkpoint/resume mechanisms
Data preparation:
- Pre-download datasets during off-peak hours
- Use compression for transfer efficiency
- Prepare data locally when possible
- Cache frequently used models and datasets
Cost tracking
Section titled “Cost tracking”Usage monitoring:
# Track training costsimport timestart_time = time.time()# ... training code ...duration = time.time() - start_timecost = duration / 3600 * hourly_rateprint(f"Training cost: ${cost:.2f}")
Performance metrics:
- Samples per second for ML training
- Frames per second for rendering
- Hash rate for mining
- Throughput for data processing