AI Workstation of the Deep Learning Elite

This quiet, fast, reliable and universal multi-GPU deep learning machine beats every other solution on the market
VIEW PRODUCTS

AI Server for the Infinite Inference

A cost-effective solution that delivers exceptional performance and scalability for AI inference needs
VIEW PRODUCTS

liquid cooling integration kit for ai multi-gpu server

Upgrades any air-cooled GPU server with liquid-cooling system, boosting performance up to 30% and lowering total facility power consumption
VIEW PRODUCTS
GRANDO SERVER

inference PRODUCT LINE

Comino GRANDO AI INFERENCE Servers: Purpose-Built for Scalable Generative AI

Comino GRANDO AI INFERENCE servers are engineered for high-performance, low-latency inference and fine-tuning of today’s most advanced pre-trained machine learning and deep learning models. Perfectly suited for Generative AI, these systems effortlessly handle workloads based on popular models and platforms such as Stable Diffusion, Midjourney, Hugging Face Transformers, Character.AI, QuillBot, DALL·E 2, DALL·E 3, OpenAI GPT-3.5/4, Mistral, LLaMA, Claude, DeepSeek, Mixtral, Whisper, Falcon, BLOOM, and more.With flexible, cost-optimized multi-GPU configurations, GRANDO servers are designed to scale seamlessly — whether deployed on-premises or within a high-performance data center. Ideal for inference, fine-tuning, or real-time AI applications, Comino GRANDO delivers the versatility, power, and reliability needed to support your most demanding AI workflows.

GRANDO AI INFERENCE BASE

Multi-GPU Server
NVIDIA OPTION: 8x L40S GPUs
AMD OPTION: 8x Radeon RX 7900XTX
1x AMD EPYC 9004/9005 32 cores CPU
Comino Liquid Cooling

Buy with 8x NVIDIA L40S Buy with 8x AMD 7900XTX
GRANDO AI INFERENCE PRO

Multi-GPU Server
NVIDIA OPTION: 8x RTX PRO 6000 (96 GB) GPUs
AMD OPTION: 8x Radeon PRO W7900
1x AMD EPYC 9004/9005 64 cores CPU
Comino Liquid Cooling

Buy with 8x NVIDIA RTX PRO 6000Buy with 8x AMD w7900
GRANDO AI INFERENCE MAX

Multi-GPU Server
NVIDIA OPTION: 8x H200 GPUs
1x AMD EPYC 9004/9005 128 cores CPU
Comino Liquid Cooling

Buy with 8x NVIDIA h200

The GRANDO AI INFERENCE BASE and PRO servers are built for high-efficiency AI workloads, supporting up to eight liquid-cooled NVIDIA RTX PRO 6000 / L40S or AMD W7900 / 7900 XTX GPUs, each with up to 96GB of VRAM — delivering a total of up to 768GB VRAM per server. Optimized for low- and mid-precision inference, these systems are ideal for running large language models (LLMs) such as LLaMA, DeepSeek, GPT, Mixtral, and other transformer-based architectures with ease and consistency.

At the top of the performance spectrum, the GRANDO AI INFERENCE MAX redefines power and scalability. Featuring up to 8x NVIDIA H100 or H200 GPUs, this flagship system offers an incredible 1.12TB of unified VRAM, making it the most powerful inference and fine-tuning solution in the Comino lineup. Designed for enterprise AI deployments and cutting-edge research, INFERENCE MAX delivers the ultimate performance for large-scale generative models and advanced neural networks.

Unmatched Performance, Zero Compromise

Comino’s cutting-edge liquid-cooling technology eliminates thermal throttling entirely, unlocking up to 50% higher sustained performance compared to air-cooled alternatives — even under full load. Built for long-term reliability, GRANDO systems offer up to 3 years of maintenance-free operation, while remaining as simple to maintain as air-cooled systems.With seamless integration via API, the Comino Monitoring System (CMS) provides robust remote diagnostics, performance monitoring, and fleet control—ready to plug into your infrastructure from day one.

GRANDO AI INFERENCE servers are fully optimized and pre-tested for a comprehensive range of industry-standard AI software stacks, including:

  • Toolkits & Runtimes: NVIDIA CUDA Toolkit, cuDNN, AMD ROCm, ONNX Runtime, OpenVINO, TensorRT, DeepSpeed, Hugging Face Transformers, Intel oneAPI
  • Frameworks: PyTorch, TensorFlow, JAX, Keras, MXNet, PaddlePaddle, FastAI, PyTorch Lightning, Flax, Chainer
  • Dev Environments & Libraries: Python, NumPy, SciPy, Dask, Ray, RAPIDS, scikit-learn, Apache TVM, Triton Inference Server

Each server is equipped with up to 8x liquid-cooled NVIDIA (H200 / H100 / RTX PRO 6000 / L40S) or AMD (W7900 / 7900 XTX) GPUs, paired with ultra-fast, high-core-count Threadripper PRO or cost-effective EPYC CPUs. Whether you’re running massive-scale inference, multi-modal AI, LLMs, vision models, or custom pipelines, GRANDO delivers maximum throughput, compatibility, and future-proof scalability for the most demanding AI workloads.

expert review

"INFINITE Inference Power for AI"

Unlock the power of performance with Sendex!

"A lot of inference power comes from this Powerhouse machine from Comino which has not one, not two, not three - it has six GPUs inside!
Harrison Kinsley, the coding maestro aka Sentdex, dives into the ultimate tech thrill with the Comino Grando Workstation featuring a mind-blowing 6x NVIDIA GPUs!"

Talk To Engineer

Let's talk

Grando AI Inference Product Specifications

Please, contact our sales team in case you want a custom setup
Specs
GRANDO AI INFERENCE BASE
GRANDO AI INFERENCE PRO
GRANDO AI INFERENCE MAX
 GPU (NVIDIA)
8X L40S 48 GB
8X RTX PRO 6000 96 GB
8X H200 141 GB
TOTAL GPU MEMORY
VRAM 384 GB
VRAM 768 GB
VRAM 1 128 GB
GPU (AMD)
8x Radeon 7900XTX 24 GB
8X Radeon PRO W7900 48 GB
TOTAL GPU MEMORY
VRAM 192 GB
VRAM 384 GB
CPU
AMD EPYC 9354 (32 cores)
AMD EPYC 9554 (64 cores)
AMD EPYC 9754 (128 cores)
MEMORY
256 GB DDR5
512 GB DDR5
1024 GB DDR5
NETWORKING
DUAL-PORT 10Gb,
Dedicated IPMI
DUAL-PORT 10Gb,
Dedicated IPMI
DUAL-PORT 10Gb,
Dedicated IPMI
STORAGE OS
DUAL 1.92TB M.2 NVME DRIVE
DUAL 1.92TB M.2 NVME DRIVE
DUAL 1.92TB M.2 NVME DRIVE
STORAGE DATA/CACHE
ON REQUEST
ON REQUEST
ON REQUEST
COOLING SYSTEM
CPU & GPU LIQUID COOLING
CPU & GPU LIQUID COOLING
CPU & GPU LIQUID COOLING
SYSTEM ACOUSTICS
HIGH
HIGH
HIGH
OPERATING TEMPERATURE RANGE
UP TO 38ºC
UP TO 38ºC
UP TO 38ºC
OS COMPATIBILITY
UBUNTU / WINDOWS
UBUNTU / WINDOWS
UBUNTU / WINDOWS
SYSTEM POWER USAGE
UP TO 4.5 KW
UP TO 38A @ 120V / 21A @ 220V
UP TO 6.5 KW
UP TO 54A @ 120V / 30A @ 220V
UP TO 6.5 KW
UP TO 54A @ 120V / 30A @ 220V
SIZE
439 x 177 x 681 MM
439 x 177 x 681 MM
439 x 177 x 681 MM
CLASS
SERVER
SERVER
SERVER
GRANDO WORKSTATION

DEEP learning PRODUCT LINE

Comino GRANDO AI Deep Learning Workstations: Built for High-Performance Training and Fine-Tuning

Comino GRANDO AI Deep Learning Workstations are purpose-built for on-premise training and fine-tuning of complex neural networks on large datasets, with a strong emphasis on Generative AI—yet fully capable across a wide range of AI domains.These systems offer best-in-class, customizable multi-GPU configurations designed to accelerate the development of compute-intensive models including Diffusion models, Multimodal systems, Computer Vision pipelines, Large Language Models (LLMs), and other advanced architectures. Whether you're building foundational models or fine-tuning for specialized tasks, GRANDO workstations deliver the raw power, flexibility, and efficiency AI professionals need.

GRANDO AI DL BASE

Multi-GPU Workstation
NVIDIA OPTION: 2x RTX 5090 GPUs
AMD OPTION: 4x Radeon RX 7900XTX
1x AMD Threadripper Pro 7975WX CPU
Comino Liquid Cooling

Buy with 2x NVIDIA 5090 Buy with 4x AMD  7900XTX
GRANDO AI DL PRO

Multi-GPU Workstation
NVIDIA OPTION: 4x L40S GPUs
AMD OPTION: 4x Radeon PRO W7900
1x AMD Threadripper Pro 7985WX CPU
Comino Liquid Cooling

Buy with 4x NVIDIA L40s Buy with 4x AMD W7900
GRANDO AI DL MAX

Multi-GPU Workstation
NVIDIA OPTION: 4x H200 GPUs
1x AMD Threadripper Pro 7995WX CPU
Comino Liquid Cooling

buy with 4x nvidia h200

GRANDO AI DL MAX Workstation: Ultimate On-Premise AI Training PowerhouseThe GRANDO AI DL MAX workstation is engineered for maximum performance, featuring up to four liquid-cooled NVIDIA H100 or H200 GPUs with up to 564GB of combined HBM memory, paired with a 96-core AMD Threadripper PRO CPU running at speeds up to 5.1GHz. This advanced solution delivers up to 50% higher sustained performance compared to traditional air-cooled systems—making it ideal for demanding AI training and fine-tuning workloads.

Beyond raw power, Comino systems are built for long-term reliability and ease of use, offering up to 3 years of maintenance-free operation, with servicing as straightforward as with air-cooled setups. The integrated Comino Monitoring System (CMS) supports remote management and full API integration, allowing seamless deployment into your existing software infrastructure.

GRANDO DL Workstations: Precision-Engineered for Advanced Machine Learning, Deep Learning & Generative AI

GRANDO DL workstations are pre-tested and fully optimized for a wide range of industry-leading AI toolkits and frameworks, including:

  • Toolkits & Runtimes: NVIDIA CUDA Toolkit, cuDNN, AMD ROCm, ONNX Runtime, OpenVINO, TensorRT, Intel oneAPI, DeepSpeed, Triton Inference Server, Hugging Face Transformers
  • Frameworks & Libraries: PyTorch, TensorFlow, JAX, Keras, MXNet, PaddlePaddle, FastAI, PyTorch Lightning, Flax, Chainer, Scikit-learn, XGBoost, LightGBM, Ray, RAPIDS, Dask, NumPy, SciPy, Apache TVM

Each system is equipped with up to four high-performance GPUs NVIDIA H200, H100, RTX PRO 6000, L40S, RTX 5090 or AMD W7900 / 7900 XTX—paired with the latest high-frequency, multi-core Threadripper PRO CPUs for unparalleled training and inference performance.

With silent operation and exceptional thermal efficiency, GRANDO DL workstations are purpose-built for complex and compute-intensive AI workloads including:

  • Text & Language Models: GPT-3, GPT-4, Claude, LLaMA, Mistral, Mixtral, Falcon, DeepSeek, BLOOM, T5, BERT, RoBERTa
  • Vision & Multimodal Models: Stable Diffusion, Midjourney, DALL·E 2, DALL·E 3, ControlNet, CLIP, Segment Anything (SAM), YOLOv8, Detectron2
  • Conversational & Personal AI: Character.AI, QuillBot, Replika, Jasper, Cohere, Open Assistant
  • Audio & Speech Models: Whisper, Bark, Tortoise TTS, VALL-E

Whether you're working on fine-tuning, inference, model development, or real-time deployment, GRANDO DL delivers exceptional performance, reliability, and flexibility for AI professionals, researchers, and creators alike.

Talk To Engineer

Let's talk

Grando AI DL Product Specifications

Please, contact our sales team in case you want a custom setup
Specs
GRANDO AI DL BASE
GRANDO AI DL PRO
GRANDO AI DL MAX
GPU (NVIDIA)
2x geforce 5090 32 GB
4x L40S 48 GB
4x H200 141 GB
TOTAL GPU MEMORY
VRAM 64 GB
VRAM 192 GB
VRAM 564 GB
GPU (AMD)
4x Radeon 7900XTX 24 gb
4x Radeon PRO W7900 48 GB
TOTAL GPU MEMORY
VRAM 96 GB
VRAM 192 GB
CPU
AMD Threadripper PRO 7975WX (32 cores)
AMD Threadripper PRO 7985WX (64 cores)
AMD Threadripper PRO 7995WX (96 cores)
MEMORY
256 GB DDR5
512 GB DDR5
1024 GB DDR5
NETWORKING
DUAL-PORT 10Gb,
Dedicated IPMI
DUAL-PORT 10Gb,
Dedicated IPMI
DUAL-PORT 10Gb,
Dedicated IPMI
STORAGE OS
DUAL 1.92TB M.2 NVME DRIVE
DUAL 1.92TB M.2 NVME DRIVE
DUAL 1.92TB M.2 NVME DRIVE
STORAGE DATA/CACHE
DUAL 3.84TB U.2 NVME DRIVE
DUAL 7.68TB U.2 NVME DRIVE
DUAL 7.68TB U.2 NVME DRIVE
COOLING SYSTEM
CPU & GPU LIQUID COOLING
CPU & GPU LIQUID COOLING
CPU & GPU LIQUID COOLING
SYSTEM ACOUSTICS
LOW
LOW
LOW / MEDIUM
OPERATING TEMPERATURE RANGE
UP TO 30ºC
UP TO 30ºC
UP TO 30ºC
OS COMPATIBILITY
UBUNTU / WINDOWS
UBUNTU / WINDOWS
UBUNTU / WINDOWS
SYSTEM POWER USAGE
UP TO 2.4 KW
UP TO 20A @ 120V / 11A @ 220V
UP TO 2.4 KW
UP TO 20A @ 120V / 11A @ 220V
UP TO 3.6 KW
UP TO 30A @ 120V / 16A @ 220V
SIZE
439 x 177 x 681 MM
439 x 177 x 681 MM
439 x 177 x 681 MM
CLASS
WORKSTATION
WORKSTATION
WORKSTATION

certified to the partners programs

COMINO have established a strong strategic relationship with industry leaders
OE logo
testimonials

Praised by the Top Tech Leaders worldwide

Greg Corke (AEC MAG)

With support for datacentre GPUs, the Comino Grando can potentially transform workflows by giving simulation and Al specialists ready access to vast amounts of computational power on the desktop

Harrison Kinsley

"This is the coolest deep learning machine that I have ever had the opportunity to use. It’s the most power in the smallest form factor also, that I’ve ever used, and finally, it also runs the coolest temperatures, that I’ve ever used"

Linus Sebastian

"God of computers".
"On this machine, compute take such little time, that I've been having trouble getting all GPUs to get fully loaded".
"It appears to be rock freaking solid stable".

Storage Review

The Comino Grando H100 server is an impressive addition to the company’s lineup. Powered by an AMD Threadripper PRO 7995WX CPU and 512GB of DDR5 memory, expandable up to 1TB, the Grando system is highlighted by two NVIDIA H100 NVL GPUs providing exceptional performance for AI-driven workflows.

Jesse Woolston

"The main factor as to why I love the Grando RM is its ability to be diverse with training and modelling, where I can give it any and all assignments and I am able to just utilise the tools and focus on the art".

trusted by
are you ready?

join the elite
of Grando Professionals

order your grando now

Have a media inquiry? Looking for more info about Comino? Contact one of our team members at pr@comino.com

Technology Partners

At Comino, we are dedicated to our flexibility, showcasing a wide array of components to demonstrate our versatility. This expansive range prevents us from being confined by constraints imposed by single vendors. Through custom-tailored solutions that cater to the specific needs of each client, our meticulously selected offerings ensure precise and individualized results. By embracing this multifaceted strategy, we remain committed to delivering exceptional, bespoke solutions that fulfill the unique requirements of our valued clients.