💬گفتگو با لاتک

AI Factory

Welcome to the future of intelligent automation with AI Factory – your comprehensive end-to-end platform for building, deploying, and scaling AI solutions across your enterprise. Our integrated ecosystem transforms how organizations harness artificial intelligence, from raw data to production-ready applications that drive measurable business results. AI Factory eliminates the complexity of AI implementation by providing a unified platform where data preparation, infrastructure management, software orchestration, expert services, and proven use cases work seamlessly together. Whether you’re a startup exploring AI possibilities or an enterprise scaling intelligent operations, our factory approach accelerates your journey from concept to deployment in weeks, not months. Experience the power of democratized AI with our low-code tools that enable both technical teams and business users to create sophisticated AI workflows. Join thousands of organizations worldwide who trust AI Factory to power their digital transformation and competitive advantage in the AI-driven economy.

AI Data

Transform raw information into AI-ready fuel with our comprehensive data preparation solutions. Our AI Data module ensures your datasets are clean, structured, and optimized for machine learning applications. From data collection and preprocessing to feature engineering and quality validation, we handle the entire data pipeline. We implement robust data governance frameworks that maintain privacy, security, and compliance standards.

Our automated data pipelines reduce preparation time by up to 80% while ensuring consistency and accuracy. Whether you’re working with structured databases, unstructured text, or real-time streams, our platform adapts to your data landscape. Experience seamless integration with popular data sources and formats, enabling faster time-to-insight for your AI initiatives.

AI Services

Bridge the gap between technology and business success with our expert AI services team. Our AI Services module combines deep technical expertise with industry knowledge to deliver customized AI solutions that drive real business outcomes. Our certified AI engineers and data scientists work closely with your team to understand unique requirements and design tailored implementations. From initial consultation and proof-of-concept development to full-scale deployment and ongoing support, we guide you through every step of your AI journey.

We leverage our comprehensive data infrastructure and software stack to build robust, scalable solutions that integrate seamlessly with your existing systems. Our agile delivery methodology ensures rapid prototyping and iterative refinement based on your feedback. Post-deployment, our team provides continuous monitoring, optimization, and enhancement services to maximize your AI investment ROI.

AI SOftware

Orchestrate intelligent workflows with our comprehensive software stack and agentic AI orchestrator platform. Our AI Software module provides the essential tools and frameworks needed to build, deploy, and manage AI applications at scale. Features include our proprietary Agentic AI Orchestrator that coordinates multiple AI agents for complex task automation. The platform supports popular machine learning frameworks, containerized deployments, and microservices architecture for maximum flexibility.

Built-in MLOps capabilities streamline model versioning, testing, and deployment pipelines for continuous integration. Our low-code interface empowers both technical and business users to create sophisticated AI workflows without extensive programming knowledge. Advanced orchestration capabilities enable seamless coordination between different AI models and business processes.

AI Infrastructure

Power your AI ambitions with enterprise-grade hardware infrastructure designed for scale and performance. Our AI Infrastructure module provides the computational backbone needed for demanding machine learning workloads. From high-performance GPUs and specialized AI chips to distributed computing clusters, we deliver the hardware foundation your AI applications require.

Our infrastructure solutions include automated resource provisioning, load balancing, and fault-tolerant architectures that ensure 99.9% uptime. Scale dynamically from prototype to production with our flexible hardware configurations that adapt to your growing needs. Built-in monitoring and optimization tools maximize resource utilization while minimizing operational costs. Deploy on-premises, in the cloud, or in hybrid environments with our versatile infrastructure solutions.

AI Use Cases

Unlock AI’s potential across industries with our proven use case implementations and success stories. Our AI Use Cases module showcases real-world applications including AI Sales for intelligent lead scoring and automated customer engagement. Transform your marketing with AI Digital Marketer solutions that personalize campaigns and optimize customer journeys in real-time. Empower data-driven decisions with AI BI platforms that provide predictive analytics and automated insights generation.

Strengthen cybersecurity with AI SOC capabilities that detect threats, automate incident response, and enhance security operations. From healthcare diagnostics and financial risk assessment to manufacturing optimization and retail personalization, we demonstrate how AI transforms business operations. Each use case comes with detailed implementation guides, best practices, and measurable success metrics to accelerate your adoption journey.

Complete AI Infrastructure Portfolio: From GPU Foundation to Enterprise Solutions

Build your AI infrastructure on proven GPU foundations with our comprehensive portfolio featuring NVIDIA PCIe and SXM modules alongside Qualcomm processing units as the core computational building blocks for all AI workloads. Our integrated ecosystem scales from these foundational GPU components through enterprise-grade AI computing solutions from Dell, HPE, and Supermicro, paired with high-performance storage systems including Huawei Doradoo and DDN for massive data management capabilities. Complete your AI infrastructure with our advanced networking layer featuring NVIDIA Quantum-2 switches and H3C ROCE-enabled ethernet solutions, topped by ready-to-deploy AI-as-a-Box systems including AI All-in-One platforms and specialized video analytics appliances that leverage the full power of our GPU-centric architecture.

Scalable AI Infrastructure: From Edge to Enterprise

Deploy AI anywhere with our comprehensive hardware ecosystem designed for every stage of your AI journey. Our infrastructure portfolio seamlessly scales from lightweight edge computing to enterprise-grade datacenter operations, ensuring optimal performance at every deployment point. Starting at the edge with our compact SOM GPU modules and PCIe cards for real-time inference, we provide the computational power needed for instant decision-making in resource-constrained environments. Scale to on-premises deployments with our powerful workstations and MGX servers that combine inference capabilities with fine-tuning flexibility for custom model optimization. At the datacenter level, experience unmatched performance with our HGX hyperscale accelerators and DGX deep learning systems that handle the most demanding training, fine-tuning, and inference workloads simultaneously. Our modular AI POD solutions enable rapid datacenter deployment and expansion, providing enterprise-grade infrastructure that grows with your AI ambitions and delivers consistent performance across your entire AI pipeline.

Professional GPU Cards: Unleash Maximum Performance for AI Workloads

Power your most demanding AI applications with our comprehensive GPU card portfolio featuring cutting-edge NVIDIA and Qualcomm GPU Cards designed for technical and creative professionals who require uncompromising performance. Our lineup spans from the flagship NVIDIA B200 and H200 141G for next-generation AI training, to the proven H100 80G/94G and A100 80G series for enterprise-scale machine learning workloads, complemented by specialized cards like the L40S/L40 for mixed AI and graphics tasks. Complete your setup with professional-grade options including the NVIDIA A6000 Pro for AI development, RTX 5090 for AI-accelerated creative workflows, and the innovative Qualcomm A100 Ultra for energy-efficient AI processing. Each GPU card delivers exceptional computational power optimized for the world’s heaviest AI, machine learning, and high-performance computing workloads, ensuring your projects run at maximum efficiency whether you’re training complex neural networks, processing massive datasets, or developing next-generation AI applications.

Graphic Cards , for technical and creative professionals who demand high performance for the world’s heaviest workloads.

NVIDIA B200

NVIDIA H200 141G

NVIDIA H100 80G

NVIDIA H100 94G

NVIDIA A100 80G

QualComm A100 Ultra

NVIDIA L40S

NVIDIA L40

NVIDIA A6000 Pro

NVIDIA RTX 5090

Al Accelerator Comparison Table

Specification Qualcomm AI100 Ultra NVIDIA H100 80GB NVIDIA H100 NVL NVIDIA H200 NVL NVIDIA A100 80GB NVIDIA L40S NVIDIA L40
Architecture Qualcomm Cloud AI 100 Hopper GH100 Hopper GH100 (Dual GPU) Hopper GH100 (Dual GPU) Ampere GA100 Ada Lovelace AD102 Ada Lovelace AD102
Form Factor PCIe Gen4 x16 SXM5 / PCIe Gen5 NVL (Two GPUs with NVLink bridge) NVL (Two GPUs with NVLink bridge) SXM4 / PCIe Gen4 PCIe Gen4 x16 (Dual-slot) PCIe Gen4 x16 (Dual-slot)
Memory Type & Capacity 32GB LPDDR5 80GB HBM3 (SXM) / 80GB HBM2e (PCIe) 2×94GB HBM3 (188GB Total) 2×141GB HBM3e (282GB Total) 80GB HBM2e 48GB GDDR6 ECC 48GB GDDR6 ECC
Memory Bandwidth ~1TB/s 3.35TB/s (SXM) / 2.0TB/s (PCIe) 3.35TB/s per GPU (6.7TB/s aggregate) 4.8TB/s per GPU (9.6TB/s aggregate) 2.04TB/s (SXM) / 1.56TB/s (PCIe) 864GB/s 864GB/s
Interconnect PCIe Gen4 x16 NVLink 4.0 / PCIe Gen5 NVLink 4.0 (GPU-to-GPU + host) NVLink 4.0 (GPU-to-GPU + host) NVLink 3.0 / PCIe Gen4 PCIe Gen4 x16 PCIe Gen4 x16
Power Consumption (TDP) ~150W 700W (SXM) / 350W (PCIe) ~350W per GPU (~700W total) 700W per GPU (~1400W total) 400W (SXM) / 300W (PCIe) 350W 300W
CUDA Cores N/A 16,896 2×16,896 (33,792 total) 2×16,896 (33,792 total) 6912 18,176 18,176
Tensor Cores N/A 456 (4th Gen) 2×456 (912 total) 2×456 (912 total) 432 (3rd Gen) 142 (4th Gen) 142 (4th Gen)
FP64 Performance N/A ~34 TFLOPS ~68 TFLOPS ~68 TFLOPS ~9.7 TFLOPS (SXM) ~0.6 TFLOPS ~0.5 TFLOPS
FP32 Performance N/A ~67 TFLOPS ~134 TFLOPS ~134 TFLOPS ~19.5 TFLOPS (SXM) ~91 TFLOPS ~91 TFLOPS

Deploy AI anywhere with Aetina’s comprehensive MegaEdge platform series, featuring the compact MXM Series, ultra-small MetaEdge M.2 Series, high-performance PCI-e Series, and powerful MGX Training Server for intensive edge training. These expandable AI inference and training platforms seamlessly integrate with sensors, cameras, monitors, and robotic systems through rich I/O interfaces, enabling real-time decision-making at the network’s edge. Transform your operations with enterprise-grade edge computing that brings datacenter-class AI performance directly to where data is generated, reducing latency and enhancing intelligent automation across industries.

MegaEdge MXM Series

MetaEdge M.2 Series

MegaEdge Pci-e Series

MGX Training Server

Data Privacy

AI Accelerator

No Noise Low Power

Open-Source AI Models

AETINA

Edge AI Solution Provider

Complete Product Portfolio Comparison

Aetina Edge AI Computing Solutions

Aetinaoffers acomprehensive portfolioofEdgeAIcomputing solutions spanning from compact DeviceEdge platforms to high-performance MegaEdge workstations and enterprise-grade SuperEdge servers. This unified comparison highlights key specifications and target applications across all product categories.

Product Series Models CPU Architecture AI Performance Memory GPU/Accelerator Form Factor Key Features Target Applications
DeviceEdge Series – Compact ARM-based Edge AI Platforms
Orin Nano Series AIE-PO23-6USB
AIE-PO33-6USB
6-core ARM Cortex-A78AE 1.5GHz 20-40 TOPS 4-8GB LPDDR5 512-1024 CUDA 16-32 Tensor Cores 270×195×80mm
Fanless design
6 USB camera support
Super Mode Industrial I/O
Smart retail
Computer vision
IoT gateways
Orin Nano Series AIE-PN33-6USB
AIE-PN43-6USB
8-core ARM Cortex-A78AE 2.0GHz 70-100 TOPS 8-16GB LPDDR5 1024 CUDA 32 Tensor Cores 270×195×80mm
Fanless design
6 USB camera support
Super Mode Industrial I/O
Smart surveillance
Robotics
Smart
MegaEdge Series – High-Performance x86-based AI Workstations
PCIe Series AIP-KQ67-A1
AIP-FR68-A1
Intel 12th/13th Gen Core i9/i7/i5 GPU dependent (up to 300W) Up to 192GB DDR4/DDR5 NVIDIA RTX series
Qualcomm AI 100 Ultra (up to 870 TOPS)
Desktop/rackmount 340-413mm depth NVIDIA NCS certified
OOB management
TPM 2.0
Industrial AI
Computer vision
AI training
MXM Series AIP-SQ67-A1 Intel 12th Gen Core i7/i5 (up to 65W) GPU dependent Up to 64GB DDR5 SO-DIMM MXM modules PCIe Gen4 x16 270×148×280mm
Compact design
MXM expansion
OOB management
Wall/desk mount
Embedded AI
Space-constrained deployments
SuperEdge Series – Enterprise-Grade AI Training/Inference Servers
Server Series AEX-2UA1 (MGX Server) Intel Xeon 6 Processors (up to 250W) Data center-class performance DDRS RDIMM/LRDIMM 8 Channels, ECC Dual NVIDIA GPUs with NVLink PCIe Gen5 2U server 438×88×420mm Rack-mountable 24/7 reliability
Redundant PSU/fans
Hot-swappable
AI model training
Enterprise deployment
AI On-Prem FR68-based AI Appliance Intel Core i7 13700E 870 TOPS per card 128GB VRAM per AI card Qualcomm Cloud AI100 Ultra (Single/Dual) Desktop form factor
Enterprise-class
70B parameter LLMs
Edge inference
OOB management
Generative AI
LLM inference
Computer vision

Key Product Line Differentiators

DeviceEdge

  • Compact, ARM-based edge inference
  • 20-100 TOPS AI performance
  • Fanless, industrial-grade design
  • Power-efficient (6-47W)
  • NVIDIA Jetson ecosystem

MegaEdge

  • x86 architecture for versatility
  • Support for PCIe GPU cards
  • Flexible MXM expansion options
  • NVIDIA NCS certification
  • Comprehensive I/O connectivity

SuperEdge

  • Enterprise-grade performance
  • Data center reliability features
  • Support for large AI models (70B+)
  • Redundant components
  • Hardware management (BMC)

AETINA

Edge AI Solution Provider

Aetina Product Portfolio Comparison

Entry-level

Aetina DeviceEdge

AIE-PO23/33-6USB | AIE-PN33/43

Mid-range

Aetina MegaEdge

AIP-KQ67 | AIP-FR68 | AIP-SQ67

High-end

Aetina SuperEdge

AEX-2UA1 | MGXServer

Type DeviceEdge MegaEdge PCIe MegaEdge MXM SuperEdge
Form Factor Compact Edge Device Desktop/Rackmount Compact Desktop Server (2U)
CPU Architecture ARM (6–8 core Cortex-A78AE) Intel 12th/13th Gen i9/i7/i5 Intel 12th Gen i7/i5 Intel Xeon (up to 250W TDP)
AI Performance 20–100 TOPS Up to 300W GPU support Variable (MXM module dependent) Data center-class / 870 TOPS
Memory 4–16GB LPDDR5 Up to 192GB DDR5 Up to 64GB DDR5 SO-DIMM ECC RDIMM, ECC LRDIMM
GPU Support NVIDIA Tensor Cores (built-in) NVIDIA RTX Series, Qualcomm AI MXM Type A/B/+ PCIe Gen4 Dual NVIDIA MGX w/NVLink
Storage 128GB NVMe (built-in) 4×2.5" SATA + 2×M.2 2×2.5" SATA + 2×M.2 Hot-swap NVMe, multiple M.2
Networking 2×GbE 3×2.5GbE + 1×10GbE 5×2.5G RJ45 LAN High-speed networking, BMC
Power Consumption 6.4–47W Up to 600W 361W (w/o MXM) Redundant 1600/2000W PSU
Key Applications Smart cameras, IoT, Robotics Industrial AI, Computer Vision Embedded AI, Space-constrained Enterprise AI, LLM Inference

AETINA

Edge AI Solution Provider

MegaEdge Series Comparison

MegaEdge Series Overview

AetinaMegaEdge offersexpandableAIinference/training platforms for x86-based edge computing, featuring Intel Core processors with high-performance GPU options. These systems deliver enhanced AI performance through PCIe expansion (NVIDIA RTX, Qualcomm AI) or MXM modules, designed for industrial AI workstations, computer vision applications, and AI model training at the edge.

300W

GPU Support

Product Model Series CPU Chipset GPU Support Memory Storage Network Power Dimensions
AIP-KQ67-A1 PCIe Series Intel 12th/13th Gen Core i9/i7/i5 Intel Q670E Up to 300W NVIDIA RTX series (A6000/A5000/A4000/A2000) Up to 128GB DDR5 (4×DIMM) 2×2.5" SATA SSD
1×M.2 NVMe
1×1GbE
3×2.5GbE
500W/850W PSU 413×315×159 mm
AIP-FR68-A1 PCIe Series Intel 12th/13th Gen Core i9/i7/i5 (up to 65W) Intel R680E NVIDIA RTX 6000 Ada/A6000
Qualcomm AI 100 Ultra (dual)
4×DDR5 U-DIMM (up to 192GB) 4×2.5" SATA
2×M.2 NVMe
3×2.5GbE
1×10GbE
Up to 600W 340×215×279 mm
AIP-SQ67-A1 MXM Series Intel 12th Gen Core i7/i5 (up to 65W) Intel Q670E MXM Type A/B/B+ PCIe Gen4 x16 2×DDR5 SO-DIMM (up to 64GB) 2×2.5" SATA
2×M.2 NVMe Gen4
5×2.5G RJ45 LAN 361W (w/o MXM) 270×148×280 mm

Key Features & Certifications

  • NVIDIA NCS Certified: PCIe series validated for demanding AI workloads
  • Qualcomm AI Integration: Support for Cloud AI100 Ultra with 870 TOPS OOB
  • Management: Remote device monitoring and management MXM
  • Modularity: Flexible GPU upgrades with MXM 3.1 standard Screwless
  • Design: Tool-free maintenance and servicing
  • TPM 2.0: Hardware-based security features

Target Applications

AETINA

Edge AI Solution Provider

SuperEdge Series Comparison

SuperEdge Enterprise AI Server Overview

Aetina’s SuperEdgeseries delivers data center-classAI processing at the edge with enterprise-grade servers powered by Intel Xeon processorsand high-performance NVIDIA GPU acceleration. These systems provide robust reliability features, NVIDIA MGX architecture compatibility, and are designed for demanding AI training and large-scale inference workloads in enterprise environments.

870

AI Performance (TOPS)

Product Line CPU GPU Support Memory Storage Network Power Form Factor AI Performance Key Features
AEX-2UA1 MGX Server Intel Xeon 6 (up to 250W TDP) 2× Double-width NVIDIA GPUs with NVLink 8-Channel DDR5 6400 MT/s RDIMM 2× M.2 PCIe Gen5 4x E1.S NVMe Hot-swappable BMC Management PCIe Gen5 slot for CX7/BF3 1600W/2000W Redundant (1+1) Hot-swappable 2U Rackmount 438 × 420 × 88mm (Compact 420mm depth) Variable based on installed GPUs NVIDIA MGX architecture Hot-swap components 24/7 reliability
AI On-Prem Solution Intel Core i7-13700E Qualcomm Cloud AI100 Ultra 192GB DDR5 UDIMM (128GB GPU VRAM) 2TB NVMe SSD Multiple LAN ports Desktop form factor (150W per card) Enterprise-class AI in desktop form factor 870 TOPS (INT8)
290 TFLOPS (FP8)
LLM support (70B)
EdgeEye/OOB management
ONNX framework support
NGC-Ready SuperEdge Intel Xeon server processors NVIDIA T4 to A100 GPUs ECC RDIMM ECC LRDIMM support High-speed NVMe storage Mellanox networking support Server-grade redundant power Multi-Access Edge Computing platform Data center-class performance NGC-Ready certification
Enterprise AI acceleration
Multi-tenant support

Enterprise Features

  • Server-grade reliability for 24/7 operation with minimal downtime
  • Redundant, hot-swappable power supplies and cooling systems
  • Hardware BMC management for remote administration NVLink
  • for high-speed GPU-to-GPU communication Support for large
  • memory models (up to 70B parameters) Optimized power
  • efficiency (peak performance/watt)

Target Applications

AI On-Premise Devices: Bring Intelligence to Your Local Infrastructure

Deploy AI directly within your infrastructure with our comprehensive on-premise solutions that eliminate cloud dependency while providing complete data control, reduced latency, and enhanced security for sensitive AI workloads. Our flagship AI Workstations support up to 4 high-performance GPUs for intensive machine learning training and development, while Aetina MegaEdge and SuperEdge series deliver optimized 1-2 GPU performance in compact form factors perfect for real-time inference and edge deployments. Experience the strategic advantage of local AI processing with millisecond response times, regulatory compliance, and scalable capabilities without ongoing cloud costs. Transform manufacturing floors, retail environments, healthcare facilities, and smart city deployments with intelligent automation systems that process data at the point of generation.

AI Workstations, for technical and creative professionals. who demand high performance for the world’s heaviest workloads

Soika Al Workstaon RTX 6000* 4

Soika Al Workstaon RTX 5880* 4

Soika Al Workstaon NVIDIA H100 80G*4

Soika Al Workstaon NVIDIA H200 * 4

Includes :

Nvidia GPU Soika

Mockingjay Soika

Stack

Soika Stack

Soika Mockingjay

RTX 5880 Ada * 4 OPTIMIZED X13 4U 4GPU,X13DEG-QT-P,749TS-R2KO5B,RoHS SYS-741GE-TNRT(x1)
X13 4U 4GPU,X13DEG-QT-P,749TS-R2KO5B,RoHS EWCSC(x1)
0% 3 YRS LABOR, 3 YRS PARTS, 1 YR CRS UNDER LIMITED WRNTY
P4X-EMR6538N-SRN6V-MCC(x2) 6538N 2P 32C2.1G205W(2/205,1.9/185) 60M
S128 1DS.2QA.2DL 7yr
MEM-DR564MC-ER56(x8) 64GB DDR5-5600 2Rx4 ECC RDIMM
HD5-25N4-001T9-E1-TXD-NON-007(x4) SSD 2.5" NVMe PCle4 1.9TB 1DWPD TLC D, 7mm
GPU-NVQRTX5880-ADA(x4) NVIDIA RTX5880 ADA Gen 32GB GDDR6 PCle 4.0-Active Cooling
AOM-TPM-967OV-P(x1) SPI TPM 2.0 using SLB9670,RoHS
AOC-VROCPREMOD(x1) Intel VROC Premium, RAIDO,1,5,10,HF,RoHS
SFT-DCMS-SINGLE(x1) 5MC System Management Software Suite Node License
Soika Enterprise 1 year subscription VLLM support
Interconnection: local super cluster support by Soika Enterprise

Soika Stack

Soika Mockingjay

Hopper 200 4 GPU H200
OPTIMIZED X13 4U
4GPU,X13DEG-QT-P.749TS-R2K05B,RoHS SYS-741GE-TNRT(x1)
X13 4U 4GPU,X13DEG-QT-P.749TS-R2KO5B,RoHS EWCSC(x1)
0% 3 YRS LABOR. 3 YRS PARTS, 1 YR CRS UNDER LIMITED WRNTY
P4X-EMR6538N-SRN6V-MCC(x2) 6538N 2P 32C2.1G205W(2/205.1.9/185) 60M
S128 1DS.2QA.2DL 7yr
MEM-DR564MC-ER56(x8) 64GB DDR5-5600 2Rx4 ECC RDIMM
HDS-25N4-001T9-E1-TXD-NON-007(x4) SSD 2.5* NVMe PCle4 1.9TB 1DWPD TLC D. 7mm
extension GPU-H200 PCle Gen5: 128GB/s
AOM-TPM-967OV-P(x1) SPI TPM 2.0 using SLB9670,RoHS
AOC-VROCPREMOD(x1) Intel VROC Premium, RAIDO,1,5,10,HF,RoHS
SFT-DCMS-SINGLE(x1) SMC System
Soika Enterprise 1 year subscription vLLM support
Interconnection: local super cluster support by Soika Enterprise

Enterprise AI Server Solutions: From Flexible PCIe to Ultimate HGX Performance

Scale your AI infrastructure with our comprehensive server portfolio designed for every stage of your artificial intelligence journey. Our PCIe GPU Servers leverage NVIDIA’s revolutionary MGX modular reference design, delivering the fastest and most flexible path to accelerated computing through partnerships with industry leaders including Supermicro, H3C, HPE, and Xfusion. These modular solutions adapt to any computing need from remote visualization to edge supercomputing, enabling OEM and ODM partners to build tailored solutions while dramatically reducing development time and accelerating market entry with flexible, open-source models.

For the world’s most demanding AI workloads, our HGX Server Series represents the pinnacle of AI reasoning and training inference performance, featuring cutting-edge HGX H100, H200, B200, and B300 systems engineered for technical and creative professionals tackling complex simulations and massive datasets. These powerhouse platforms integrate the full spectrum of NVIDIA technologies including high-performance GPUs, ultra-fast NVLink interconnects, and fully optimized HPC software stacks to deliver unmatched application performance and drive the fastest time-to-insights across enterprise datacenters.

Both solution categories seamlessly integrate with our Soika software ecosystem including Mockingjay orchestration and Stack management tools, ensuring comprehensive AI deployment and management capabilities from flexible edge computing to datacenter-scale artificial intelligence operations

The Fastest, Most Flexible Path to Accelerated Computing

Adapt to any computing need with NVIDIA MGX™, a modular reference design that can be used for a wide variety of use cases, from remote visualization to supercomputing at the edge. MGX provides a new standard for modular server design by improving ROI and reducing time to market. With MGX, OEM and ODM partners can build tailored solutions for different use cases while saving development resources and reducing time to market.

Brands: 

Xfusion MGX Server G5500 V7

H3C Server R5300 G6 PCle

Supermicro AS 4125GS

HPE DL380A Gen12

Includes :

Nvidia GPU Soika

Mockingjay Soika

Stack

Flexible

Open-Source Models

Time to Market

Full Nvidia Stack

AI Reasoning and AI Training Inference

 For technical and creative professionals
Who demand high performance for the world’s heaviest workloads
AI, complex simulations, and massive datasets require multiple GPUs with extremely fast interconnections and a fully accelerated software stack. The NVIDIA HGX™ platform brings together the full power of NVIDIA GPUs, NVIDIA NVLink™, NVIDIA networking, and fully optimized AI and high-performance computing (HPC) software stacks to provide the highest application performance and drive the fastest time to insights for every data center.

Brands: 

HGX H100

HGX H200

HGX B200

HGX B300

Includes :

Nvidia GPU Soika

Mockingjay Soika

Stack

Perforemance

Most Demanding

Time to Market

Full Software Stack

For technical and creative professionals

Who demand high performance for the world’s heaviest workloads

Nvidia

NVIDIA DGX H100

NVIDIA DGX H100 Enterprise License with no code LLM and agent management experience

Nvidia

NVIDIA DGX H200

NVIDIA DGX H200 Enterprise License with no code LLM and agent management experience

Nvidia

NVIDIA DGX B200

NVIDIA DGX B200 Enterprise License with no code LLM and agent management experience

Nvidia

NVIDIA DGX GB200 NVL72

NVIDIA DGX GB200NVL72 Enterpris License with no code LLM and agent management experience

AI Cloud Infrastructure: Scale Beyond Limits with Enterprise-Grade Solutions

Transform your cloud operations with our cutting-edge AI infrastructure featuring NVIDIA’s revolutionary GB300 NVL72 systems and comprehensive AI converged infrastructure racks designed to meet the most demanding enterprise requirements. Our cloud-native AI infrastructure delivers unprecedented computational density and performance optimization, enabling organizations to deploy large-scale AI training, inference, and multi-tenant workloads with maximum efficiency and cost-effectiveness across distributed cloud environments.

The NVIDIA GB300 NVL72 platform represents the apex of cloud AI computing, providing exceptional multi-GPU performance in optimized rack configurations that scale seamlessly from single deployments to massive cloud installations supporting thousands of concurrent AI workloads. Our AI Converged Infrastructure Racks integrate compute, storage, networking, and management layers into unified solutions that eliminate complexity while providing enterprise-grade reliability, automated provisioning, and dynamic resource allocation capabilities essential for modern cloud service providers.

Whether you’re building public cloud AI services, private enterprise clouds, or hybrid multi-cloud architectures, our infrastructure solutions deliver the performance, scalability, and operational efficiency needed to support next-generation AI applications including large language models, computer vision systems, and real-time analytics platforms that require massive parallel processing power and ultra-low latency connectivity across distributed cloud environments.

Nvidia

NVIDIA DGX GB200 NVL72

NVIDIA DGX GB200NVL72 Enterpris License with no code LLM and agent management experience

Nvidia

NVIDIA DGX GB300 NVL72

NVIDIA DGX GB300NVL72 Enterpris License with no code LLM and agent management experience

The Blackwell Rack-Scale Architecture for Real-Time Trillion-Parameter Inference and Training

The NVIDIA GB200 NVL72 is an exascale computer in a single rack. With 36 GB200s interconnected by the largest NVIDIA® NVLink® domain ever offered, NVLink Switch System provides 130 terabytes per second (TB/s) of low-latency GPU communications for AI and high-performance computing (HPC) workloads.

NVIDIA GB200/300 NVL72

CDU

1300kW @ 4ºC Approach Temperature

AI ModPOD: Revolutionary "AI-Ready" Cooling Technology for Cloud Infrastructure

Deploy next-generation AI cloud infrastructure with our comprehensive ModPOD cooling solutions engineered to handle the extreme thermal demands of high-density AI workloads. Our scalable cooling architecture progresses from traditional Air Cooling systems supporting up to 55kW per rack using chilled water and contained cold aisle design, to advanced Hybrid Direct Liquid Cooling with integrated fan systems delivering up to 120kW per rack through dual water circuit configurations that optimize both liquid and air cooling efficiency.

For ultra-high density deployments, our Fanless Direct Liquid Cooling (DLC) technology eliminates traditional cooling bottlenecks by capturing heat directly from servers and connecting it to specialized cooling distribution systems, supporting up to 400kW per rack without fan noise or air circulation dependencies. At the pinnacle of our cooling innovation, Adaptive Cascade Cooling and Heat Reuse Technology combines air and liquid cooling systems into a single hybrid approach that not only maintains optimal data center temperatures but also transfers excess heat to external consumers, creating sustainable energy loops that reduce operational costs.

This progressive cooling ecosystem ensures your AI cloud infrastructure can scale from moderate compute densities to the most demanding GPU-intensive workloads while maintaining optimal performance, energy efficiency, and environmental sustainability across your entire datacenter operation.

AI ModPOD “AI-ready”cooling technology

Air cooling up to 55kW/rack

Chilled water supply from the facility cools down the air coolingsystem (fan cooling coils) positioned above the IT racks in the contained cold aisle)

Hybrid Direct Liquid Cooling with fans

up to 120kW/rack
Combined DLC and air cooling through two water circuits, typically with a 70/30 mix respectively

Fanless DLC up to 400kW/rack

100% server heat is captured through coldplatesand directly connected to the cooling distribution systems

Adaptive Cascade Cooling and heat reuse technology

Combines air and liquid coolingsystems into a single hybrid approach with adaptive control. Heat reuse transfers heat to the consumer while maintaining isolation of the data center loop

Air cooling up to 55kW/rack

Increasing cooling efficiency and capacity (kW/rack)

Huawei Cloud AI Services:
GPU as a Service with Advanced Agentic AI Platform

Accelerate your AI transformation with our comprehensive Huawei Cloud partnership delivering enterprise-grade GPU as a Service solutions enhanced by our proprietary Agentic AI software platform. Through Huawei Cloud’s robust infrastructure, we provide on-demand access to high-performance GPU resources including NVIDIA H100, A100, and V100 instances that scale dynamically with your workload requirements, eliminating the need for massive capital investments in AI hardware while ensuring global accessibility and 99.9% uptime reliability.

Our value-added Agentic AI Software Layer transforms standard GPU cloud services into intelligent, self-managing AI orchestration platforms that automate complex workflows, optimize resource allocation, and coordinate multiple AI agents for sophisticated task execution. This integrated approach combines Huawei Cloud’s world-class infrastructure with our advanced AI orchestration capabilities, enabling enterprises to deploy everything from simple machine learning models to complex multi-agent AI systems with unprecedented ease and efficiency.

Experience the power of cloud-native AI with flexible consumption models, automated scaling, comprehensive monitoring, and enterprise-grade security, all while leveraging our Agentic AI platform that intelligently manages your AI workloads across Huawei Cloud’s global network, ensuring optimal performance, cost efficiency, and seamless integration with your existing business processes and data pipelines

Solutions for Unleashing Digital and Powering Innovation

  • Ready-to-use AI equipment powered by Huawei Cloud
  • High-performance infrastructure for AI workloads
  • Flexible, scalable, and cost-effective solutions
  • Trusted global cloud ecosystem

Enterprise AI Storage: Ultra-High Performance for Demanding AI Workloads

Power your most intensive AI applications with our enterprise-grade storage portfolio featuring Huawei Doradoo, DDN EXAScaler, and DDN-AI IntelliFlash systems engineered for 400Gbps+ InfiniBand and ROCE connectivity with ultra-low latency performance. These advanced storage platforms integrate GPU-direct NVMe SSDs and high-speed object storage capabilities specifically optimized for AI training, inference, and large-scale data analytics workloads that demand massive throughput and minimal data access delays. Our solutions seamlessly handle petabyte-scale datasets with parallel file systems, distributed object storage, and intelligent data tiering that automatically optimizes performance for machine learning pipelines, deep learning training, and real-time AI inference scenarios. Experience unmatched storage performance with enterprise reliability, automated data management, and scalable architectures that grow with your AI infrastructure while maintaining consistent sub-millisecond response times. Each platform delivers the storage foundation essential for next-generation AI applications including large language models, computer vision processing, and high-frequency trading systems that require instant data access across distributed computing environments.

Enterprise-grade AI Storage optimized for 400Gbps In-NiBand,

ultra-low latency, and GPU-direct NVMe SSDs.

Storage Comparison: Huawei Dorado vs DDN for AI Environments

100GB+ Connectivity Models Optimized for AI Training & Inference Workloads

Criteria Dorado 3000 V6 Dorado 5000 V6 Dorado 6000 V6 Dorado 18000 V6 DDN ES/AI400X2 DDN ES/AI200X2 DDN SFA400X2 DDN GRIDScaler
#100GbE Front-End 100GbE; NVMe-oF RoCE at 25Gb 25/100Gb NVMe-oF RoCE; 100GbE 25/100Gb NVMe-oF RoCE; 100GbE 100Gb RoCE; 25Gb NVMe-oF RoCE; up to 104 ports 4x HDR/HDR100 IB or 4x 200/100GbE 8x HDR/HDR100 IB or 8x 200/100GbE 8x HDR/HDR100 IB or 8x 200/100GbE 100Gb EDR IB; 100GbE; newer models with NDR200 IB
Performance 0.05ms latency; 50% higher perf than prev gen 0.05ms latency; E2E NVMe; massive concurrency 0.05ms latency; E2E NVMe; higher SSD counts 0.05ms latency; up to 21M SPC-1 IOPS (family result) High throughput/IOPS; no specific GB/s in data sheet 90+ GB/s read; 65+ GB/s write; ~3M IOPS 115 GB/s read; 75 GB/s write; ~3M IOPS SPEC SFS2014 0.61ms response time; high throughput
AI/GPU Integration NVMe queues; FlashLink; no explicit GDS reference NVMe-oF RoCE with DCB lossless; FlashLink NVMe-oF RoCE with DCB lossless; FlashLink High port counts for fan-out; massively parallel I/O GDS support; Hot Nodes GPU caching GDS support; validated with NVIDIA DGX H100; 1.4 GB/s per H100 GPU Block storage for parallel filesystems; no direct GDS mention IBM Spectrum Scale with GPU integration paths
Architecture SmartMatrix; E2E NVMe; 100Gb back- end RDMA SmartMatrix full- mesh; E2E NVMe; 100Gb RDMA back-end SmartMatrix full-mesh; E2E NVMe; higher density NVMe SmartMatrix full-mesh; 4U controller enclosures; high modularity Parallel Lustre (EXA6); 2U appliance Parallel Lustre (EXA6); end- to-end parallel data paths Pure NVMe block; building block for filesystems Parallel file/NAS based on IBM Spectrum Scale
Scalability Up to 16 controllers; IOPS scales linearly Up to 32 controllers; linear IOPS growth Up to 32 controllers; 1- 16TB cache per dual- controller Up to 32 controllers; 512GB-32TB cache per dual- controller Scale-out architecture; 1- many appliance Linear performance scaling; validated at BasePOD scale Scale with additional arrays; hybrid expansion Enterprise-scale multi-tenant clusters
Best for AI Workloads Edge AI, departmental inference, Dev/Test Medium AI training/inference clusters Larger mid- range training; mixed workloads Enterprise- scale training; multi-tenant AI platforms Compact AI/HPC; smaller GPU clusters DGX H100 BasePOD; validated with 4-16 DGX systems High- throughput AI pipelines; backing filesystems Enterprise NAS for mixed AI/HPC applications

Notes: All Huawei models feature 0.05ms latency with E2E NVMe. DDN A3I solutions validated specifically for NVIDIA DGX H100 environments. GDS = NVIDIA GPUDirect Storage; HDR/HDR100 = 200/100Gb InfiniBand; IB = InfiniBand; RoCE = RDMA over Converged Ethernet.

Enterprise AI Networking: Ultra-High Speed Connectivity for Next-Generation AI Workloads

Transform your AI infrastructure with our comprehensive networking portfolio featuring enterprise-class NVIDIA Quantum-2 architecture switches including the QM9700 and QM9790 InfiniBand systems delivering 400Gbps and 800Gbps+ connectivity for the most demanding HPC fabrics and AI training clusters. Our advanced networking ecosystem combines high-performance InfiniBand and ROCE (RDMA over Converged Ethernet) technologies through H3C’s S9827 and S9855 switches, enabling ultra-low latency communication essential for distributed AI training, large language model processing, and real-time inference workloads.

These enterprise-grade networking solutions excel in modern leaf-spine network topologies, providing the non-blocking, high-bandwidth connectivity required for GPU clusters, AI supercomputing environments, and massive parallel processing systems where microsecond-level latency directly impacts training performance and model accuracy. Our ConnectX adapters, transceivers, and cables complete the networking fabric, ensuring seamless integration and optimal performance across heterogeneous AI infrastructure deployments.

Experience unparalleled network performance with intelligent traffic management, adaptive routing, and congestion control specifically optimized for AI workloads including distributed deep learning, federated learning architectures, and multi-node inference systems that demand consistent, predictable network performance at scale.

Enterprise-class In niBand  Switching with Quantum-2 architecture for next-gen AI workloads and HPC fabrics

Nvidia

NVIDIA Quantum-2 QM9700 (Infiniband)

Nvidia

NVIDIA Quantum-2 QM9790 (Infiniband)

H3C Eth Switch (ROCE) S9827

H3C Eth Switch (ROCE) S9855

Nvidia

ConnectX Adapters & Transceivers and Cables

Features NVIDIA Quantum-2 QM9700
InfiniBand - InternalSM
NVIDIA Quantum-2 QM9790
InfiniBand - ExternalSM
H3C S9827
EthernetRoCE - 800G
H3C S9855
EthernetRoCE - 400G
Port Configuration 64 × 400Gb/s NDR InfiniBand (OSFP) 64 × 400Gb/s NDR InfiniBand (OSFP) Up to 64 × 800G or 128 × 400G ports High-density 400G/200G/100G ports (model dependent)
Fabric Capacity 51.2 Tb/s switching 51.2 Tb/s switching Up to 102.4 Tb/s Up to 25.6 Tb/s (S9855-32D model)
Packet Processing ~66.5 Bpps ~66.5 Bpps Not specified 5346.7 Mpps (S9855-32D model)
Latency Sub-microsecond Sub-microsecond <1.2µs <1.2µs
Protocol & AI Features • SHARPv3 in-network computing
• Adaptive routing
• Advanced congestion control
• Self-healing networking
• SHARPv3 in-network computing
• Adaptive routing
• Advanced congestion control
• Self-healing networking
• RoCEv2 with PFC + ECN
• DCBX for lossless Ethernet
• Next-gen 800G support
• ECMP for scale-out fabrics
• RoCEv2 with DCB
• PFC, ECN, ETS features
• ROCE stream analysis
• PFC deadlock watchdog
Management Onboard subnet manager (up to 2,000 nodes) External UFM/subnet manager control Telemetry, INT, ZTP support Telemetry, INT, ZTP support
Form Factor 1U, 14.5kg 1U, 14.5kg 1U, ≤12.2kg (model dependent) 1U, ≤12.2kg (model dependent)
Power Consumption Typical: 720W with passive cables Typical: 720W with passive cables Max: 739W (dual AC inputs) Max: 713W (S9855-48CD8D model)
Use Cases Large-scale training clusters with collectives, DGX SuperPOD architectures Large centrally-managed fabrics with UFM control Next-gen 800G AI data centers, AIGC workloads High-density 400G/100G deployments with RoCE requirements

InfiniBand Advantages (NVIDIA Quantum-2)

  • Native InfiniBand protocol with RDMA SHARP offloads
  • for collective operations in AI training Lower latency
  • for HPC applications DGX SuperPOD validated
  • architecture
  • Comprehen sive s elf-healing and congestion management

Ethernet Advantages (H3C RoCE)

  • Higher port density with 800G support Standard
  • Ethernet compatibility with RDMA capabilities DCB
  • features for lossless transport Greater versatility and
  • ecosystem compatibility Future-ready with higher
  • bandwidth optiotins

Computational foundation for advanced AI operations.

The Soika Stack Framework serves as the intelligent middleware layer, featuring our comprehensive Model Catalogue for AI model management, advanced Management tools for resource allocation and monitoring, and GPU Cluster orchestration capabilities that seamlessly distribute workloads across your infrastructure. This framework enables seamless model deployment, version control, and performance optimization across heterogeneous computing environments.

At the pinnacle of our platform, Soika Mockingjay revolutionizes AI collaboration through our groundbreaking Agent-to-Agent (A2A) communication platform, where Custom Agents work together in interconnected networks to solve complex business challenges. This no-code platform enables millions of AI agents to collaborate on single missions, creating the world’s largest multi-agent workforce that transforms text-to-action capabilities and accelerates go-to-market strategies.

Whether deploying on-premises with our Enterprise solution for data protection and regional collaboration, or leveraging our Cloud platform for global agent networks and commercial AI agent marketplaces, Soika delivers the complete infrastructure and software stack needed to build, deploy, and scale collaborative AI agents that drive real-world business outcomes.

Soika Mockingjay Platform

Soika Stack Framework

Model Catalogue

Management

GPU Cluster

Soika: Complete AI Infrastructure Platform

Core Components

Stack Module (Infrastructure Management)

Comprehensive AI Infrastructure Control

Real-time GPU and system resource monitoring with live performance tracking

Streamlined model deployment and execution with customizable parameters through intuitive GUI

Secure API key management for agents and services integration

Complete MLOps workflow automation through user-friendly graphical interface

Soika-Mockingjay Module (Agent Creation & Management)

Intelligent Agent Ecosystem

Create, configure, and manage sophisticated AI agents with custom capabilities

Advanced inter-agent communication with granular access control and role-based permissions

Dynamic model switching allowing independent updates for each agent

Knowledge base integration connecting agents to organizational data sources

Web tool connectivity for enhanced agent capabilities

Flexible API integration supporting GET/POST methods and JSON structures

Multi-level access restrictions based on user roles and security requirements

Recommended Use Cases

Enterprise AI Management: Centralized control of organizational AI projects and deployments

Cross-Team Collaboration: Bridge technical and non-technical teams with intuitive interfaces

Secure System Integration: Connect agents to organizational systems with enterprise-grade security

Enhanced Agent Intelligence: Leverage knowledge centers to create more capable and informed AI agents

Soika MOCKINGJAY

SOIKA MOCKINGJAY ENTERPRISE – A software solution to host LLM on premise and design a custom AI assistant by click, flexible to integrations and LLM models switch. All in one AI Workstation & self host software allows to protect data and collaborate at regional or global scale. Manage AI Agents, host it as a part of global AI network, collaborating with the world of Agents at a scale.

Build custom AI Agents with or without GPU servers

Soika Cloud drops limits of AI by creating a collaborative environment between commercial AI Agents of different companies. Our focus at cloud project is on cost reduction, Agent to Agent collaboration at scale and adoption of commercial payments. Cloud platform will open opportunity to sell a product and services through AI Agents at scale. Platform supports National White Label Solution

AI to AI communication at scale

First of its kind A2A platform to enable Agent to Agent collaboration synergy to accomplish complex missions. We let text to action AI Agents to collaborate and combine expertise, integrated tools to make things done. We pay attention to web3 tech to reach some financial automation.

Soika Stack Features

Dashboard & Monitoring
Real-time system overview with active workers, GPUs, loaded models, and replicas. Live system load monitoring with graphical display of GPU, CPU, RAM, and VRAM usage. Average utilization meters and usage analytics including API requests and token consumption

Interactive Playground
Test and experiment with AI models in a controlled environment. Configure model parameters (temperature, max tokens, top P, seed, stop sequences). System and user instruction settings with real-time response testing. Output comparison and code generation for direct project integration

Model Catalog
Comprehensive repository of AI models including LLMs, multimodal, and specialized models. Detailed model information with size, release date, licensing, and I/O specifications. Advanced filtering, search, and comparison capabilities for optimal model selection

Model Management
Complete lifecycle management of deployed AI models. Deploy, monitor, stop, and delete models with replica management. Real-time status monitoring with detailed deployment information. Support for models from Hugging Face and other repositories

Resource Monitoring
Hardware oversight for GPUs and workers with temperature monitoring. Real-time GPU and VRAM utilization tracking with visual indicators. Load detection and performance optimization tools

API Key Management
Secure API key generation and management for external integrations. Key lifecycle control with expiration dates and descriptions. Simple integration with applications and external services

User Administration
Multi-role user management with admin and user permissions. User creation, role assignment, and access control. Comprehensive user listing with creation dates and profile management. This enterprise-grade AI infrastructure platform provides complete control over AI model deployment,. monitoring, and management with robust security and user management features.

Automation Anywhere: Agentic Process Automation System

Intelligent Automation That Thinks and Adapts
Automation Anywhere’s Agentic Process Automation System combines AI’s analytical power with automation’s execution capabilities, creating intelligent agents that think, adapt, and learn like humans while handling complex business processes.

How It Works
Process Reasoning Engine (PRE): Trained on 400M+ enterprise workflows, enabling AI agents to understand context, plan dynamically, and collaborate with humans securely.

Three-Step Process:

  • Discovery: Identify automation opportunities across workflows
  • Design: Create intelligent workflows using AI, RPA, and APIs
  • Deploy: Smart AI agents that analyze data, make decisions, and execute actions autonomously

Key Benefits
Enhanced Productivity

  • Automates repetitive tasks (data entry, reporting, inquiries)
  • Provides real-time insights for faster decision-making
  • Reduces errors through consistent AI-driven processes
  • Integrates seamlessly with existing tools and legacy systems

Universal Application
Works across all departments: IT, Customer Service, Finance, HR, Sales, Marketing, Operations, and Compliance teams can leverage intelligent automation for their specific needs.

Enterprise Security

  • AI guardrails with role-based access controls
  • Complete decision logs and real-time monitoring
  • Defined escalation paths with human oversight
  • Full audit compliance for mission-critical workflows

Advanced Technology
Hybrid AI Models: Combines proprietary Automation Anywhere technology with leading third-party AI (OpenAI, Google, Anthropic) for optimal performance. Legacy Compatible: Works with existing infrastructure without replacement, using APIs and RPA capabilities for seamless integration.

User-Friendly Design
No Technical Skills Required:

  • AI Agent Studio: Low-code workspace for creating agents
  • Automation Co-Pilot: Access automations within existing applications
  • Drag-and-Drop Interface: Visual tools for complex workflows

Why Choose Automation Anywhere?

Beyond Traditional Automation: Extends RPA and BPM investments with AI-driven reasoning, enabling automation of complex, judgment-based workflows previously impossible to automate.

Enterprise-Ready: Industry-leading platform that works safely across any application, team, and environment with comprehensive governance.

Immediate ROI: Start now to gain competitive advantages in the rapidly evolving automation landscape.

Transform your business with Automation Anywhere’s intelligent automation that adapts and scales with your organization’s needs.