|
| 1 | +--- |
| 2 | +title: Intel GPU / Tiber Cloud |
| 3 | +layout: default |
| 4 | +nav_order: 2.7 |
| 5 | +parent: Deployment |
| 6 | +grand_parent: TrustGraph Documentation |
| 7 | +--- |
| 8 | + |
| 9 | +# Intel Tiber Cloud Deployment |
| 10 | + |
| 11 | +Deploy TrustGraph on Intel Tiber Cloud with Intel GPU and Gaudi accelerated systems for high-performance AI workloads. |
| 12 | + |
| 13 | +## Overview |
| 14 | + |
| 15 | +TrustGraph provides specialized deployment configurations for Intel's advanced hardware platforms, including Intel Tiber Cloud and bare-metal Intel GPU/Gaudi systems. This deployment method is designed for: |
| 16 | + |
| 17 | +- **High-performance AI inference** with Intel accelerators |
| 18 | +- **Self-hosted deployments** on Intel hardware |
| 19 | +- **Research and development** with cutting-edge Intel AI technologies |
| 20 | +- **Enterprise workloads** requiring Intel-optimized performance |
| 21 | + |
| 22 | +⚠️ **Work in Progress**: This deployment method is actively being developed and optimized for Intel's latest hardware platforms. |
| 23 | + |
| 24 | +## What You Get |
| 25 | + |
| 26 | +The Intel Tiber Cloud deployment includes: |
| 27 | + |
| 28 | +- **Intel accelerated systems** (Gaudi, GPU, or GR platforms) |
| 29 | +- **Optimized AI inference** with vLLM or TGI servers |
| 30 | +- **Large language models** like Llama 3.3 70B |
| 31 | +- **Complete TrustGraph stack** deployed via containerization |
| 32 | +- **SSH-based deployment** with automated scripts |
| 33 | +- **Port forwarding setup** for secure access |
| 34 | +- **Monitoring and observability** with Grafana |
| 35 | +- **Web workbench** for document processing and Graph RAG |
| 36 | + |
| 37 | +## Intel Hardware Platforms |
| 38 | + |
| 39 | +### Intel Gaudi Systems |
| 40 | +- **Software**: TrustGraph 1.0.13 + vLLM (HabanaAI fork) |
| 41 | +- **Model**: Llama 3.3 70B |
| 42 | +- **Deployment**: `deploy-gaudi-vllm.zip` |
| 43 | +- **Optimized for**: AI training and inference workloads |
| 44 | + |
| 45 | +### Intel GPU Multi-GPU 1550 Systems |
| 46 | +- **Software**: TrustGraph 1.0.13 + TGI Server 3.3.1-intel-xpu |
| 47 | +- **Deployment**: `deploy-gpu-tgi.zip` |
| 48 | +- **Optimized for**: High-throughput GPU inference |
| 49 | + |
| 50 | +### Intel GR Systems |
| 51 | +- **Software**: TrustGraph 1.0.13 + TGI Server 3.3.1-intel-xpu |
| 52 | +- **Deployment**: `deploy-gr.zip` |
| 53 | +- **Optimized for**: Specialized Intel AI workloads |
| 54 | + |
| 55 | +## Why Choose Intel Tiber Cloud? |
| 56 | + |
| 57 | +### 🚀 **Cutting-Edge AI Hardware** |
| 58 | +- **Intel Gaudi**: Purpose-built for AI training and inference |
| 59 | +- **Intel GPU**: High-performance parallel processing |
| 60 | +- **Specialized Architecture**: Optimized for AI/ML workloads |
| 61 | + |
| 62 | +### ⚡ **Performance Optimization** |
| 63 | +- **Hardware-Accelerated Inference**: Native Intel optimization |
| 64 | +- **Large Model Support**: Handle models like Llama 3.3 70B efficiently |
| 65 | +- **Reduced Latency**: Direct hardware acceleration |
| 66 | + |
| 67 | +### 🔒 **Self-Hosted Control** |
| 68 | +- **Data Sovereignty**: Complete control over data and models |
| 69 | +- **Custom Configuration**: Tailor deployments to specific needs |
| 70 | +- **Enterprise Security**: Self-hosted infrastructure |
| 71 | + |
| 72 | +### 🛠️ **Developer Access** |
| 73 | +- **Research Platform**: Access to latest Intel AI technologies |
| 74 | +- **Experimentation**: Test advanced AI configurations |
| 75 | +- **Direct Hardware Access**: Low-level optimization capabilities |
| 76 | + |
| 77 | +## Deployment Method |
| 78 | + |
| 79 | +The Intel deployment uses automated deployment scripts that: |
| 80 | + |
| 81 | +- Connect via SSH jump host to Intel Tiber Cloud |
| 82 | +- Deploy pre-configured TrustGraph packages |
| 83 | +- Set up Intel-optimized AI inference servers |
| 84 | +- Configure port forwarding for secure access |
| 85 | +- Initialize monitoring and web interfaces |
| 86 | + |
| 87 | +## Quick Process Overview |
| 88 | + |
| 89 | +1. **Obtain access** to Intel Tiber Cloud instance |
| 90 | +2. **Create HuggingFace token** and accept model licenses |
| 91 | +3. **Choose deployment type** (Gaudi, GPU, or GR) |
| 92 | +4. **Deploy via script** with SSH parameters |
| 93 | +5. **Connect via SSH** with port forwarding |
| 94 | +6. **Access services** through forwarded ports |
| 95 | + |
| 96 | +## Access Configuration |
| 97 | + |
| 98 | +Intel Tiber Cloud uses SSH jump host access: |
| 99 | + |
| 100 | +```bash |
| 101 | +# SSH connection format |
| 102 | +ssh -J guest@[jump-host] sdp@[target-host] |
| 103 | + |
| 104 | +# Deployment command |
| 105 | +./deploy-tiber guest@[jump-host] sdp@[target-host] [deployment-package] |
| 106 | + |
| 107 | +# Port forwarding |
| 108 | +./port-forward guest@[jump-host] sdp@[target-host] |
| 109 | +``` |
| 110 | + |
| 111 | +## Access Points |
| 112 | + |
| 113 | +Once deployed, you'll have access to: |
| 114 | + |
| 115 | +- **TrustGraph API**: Port 8089 (forwarded from 8088) |
| 116 | +- **Web Workbench**: Port 8889 (forwarded from 8888) |
| 117 | +- **Grafana Monitoring**: Port 3001 (forwarded from 3000) |
| 118 | + |
| 119 | +## Model Support |
| 120 | + |
| 121 | +**Large Language Models**: Llama 3.3 70B and other HuggingFace models |
| 122 | +**License Requirements**: HuggingFace account with model access |
| 123 | +**Hardware Optimization**: Intel-specific optimizations for inference |
| 124 | +**Inference Engines**: vLLM (Gaudi) and TGI (GPU/GR) |
| 125 | + |
| 126 | +## Complete Documentation |
| 127 | + |
| 128 | +For detailed step-by-step instructions, deployment packages, and troubleshooting, visit: |
| 129 | + |
| 130 | +**[TrustGraph Intel Tiber Cloud Guide](https://github.com/trustgraph-ai/trustgraph-tiber-cloud)** |
| 131 | + |
| 132 | +The repository contains: |
| 133 | +- Deployment automation scripts |
| 134 | +- Intel platform-specific configurations |
| 135 | +- Pre-built deployment packages |
| 136 | +- SSH connection utilities |
| 137 | +- Port forwarding setup |
| 138 | +- Detailed setup instructions |
| 139 | +- Intel hardware optimization guides |
| 140 | + |
| 141 | +## Prerequisites |
| 142 | + |
| 143 | +**Intel Tiber Cloud Access**: Account and instance allocation |
| 144 | +**HuggingFace Token**: For model downloads and licensing |
| 145 | +**SSH Access**: Familiarity with SSH jump host connections |
| 146 | +**Model Licenses**: Acceptance of required model licenses (e.g., Llama) |
| 147 | + |
| 148 | +## Performance Benefits |
| 149 | + |
| 150 | +**Hardware Acceleration**: Native Intel AI acceleration |
| 151 | +**Large Model Efficiency**: Optimized for 70B+ parameter models |
| 152 | +**Reduced Infrastructure Costs**: Efficient hardware utilization |
| 153 | +**Custom Optimization**: Direct access to hardware-level tuning |
| 154 | + |
| 155 | +## Use Cases |
| 156 | + |
| 157 | +**Research & Development**: Access to cutting-edge Intel AI hardware |
| 158 | +**High-Performance Inference**: Large model deployment with optimal performance |
| 159 | +**Self-Hosted AI**: Complete control over AI infrastructure |
| 160 | +**Intel Ecosystem Integration**: Leverage Intel's AI software stack |
| 161 | + |
| 162 | +## Next Steps |
| 163 | + |
| 164 | +After deployment, you can: |
| 165 | +- Load documents through the web workbench |
| 166 | +- Test Graph RAG queries with Llama 3.3 70B |
| 167 | +- Monitor Intel hardware performance through Grafana |
| 168 | +- Experiment with Intel-optimized AI configurations |
| 169 | +- Benchmark performance against other platforms |
| 170 | +- Integrate with Intel's AI development tools |
0 commit comments