Large Language Models

Learn more

Leverage AI for Content Creation, Data Analytics, Customer Service, and Process Optimisation with DiGiCOR’s GPU-Optimised Systems

The adoption of AI has started as a way to foster innovation, increase efficiency, optimise processes, make informed decisions, and enhance customer experience. This is adopted in various sectors, such as technology, finance, and media.

Enterprises can use open frameworks, libraries, and pre-trained AI models, and customise them for their specific needs with their data. Whether they need AI-powered applications and business models, smart chatbots for customer service, or AI to help with code generation and content creation, they can rely on DiGiCOR’s range of GPU-optimised systems that offer open modular design, vendor flexibility, and easy deployment and upgrade paths for fast-changing technologies.


Inference & Training Workloads to Improve Business Processes and Outcomes with AI

  • Generative AI Inference: leveraging machine learning models such as neural networks to generate new, contextually relevant information based on existing data, allowing the system to make predictions or create content. This is essential for tasks such as natural language understanding, image generation, and creative content production
  • AI-enabled Services/Applications: systems that integrate artificial intelligence to enhance functionality, efficiency, or user experience. These can range from virtual assistants and language translation tools to advanced data analytics platforms
  • Chatbots: conversational agents that engage in natural language conversations. Frequently used for customer support or information retrieval
  • Recommender System: using algorithms to analyse user preferences and behaviours, then suggesting personalised content or product recommendations to enhance user experience, satisfaction, and engagement.
  • Business Automation: streamline and optimise various business processes such as data entry, customer communication, and workflow management.

Use Cases:

  • Content creation (image, audio, video, writing)
  • AI-enabled office applications and services
  • Enterprise business process automation

 Key Technologies:

  • Flexible, modular, highly configurable rackmount servers with different form factors to balance computing, storage, networking, and cost for various enterprise AI workload needs for today and the future
  • PCIe 5.0 supported platforms for future-proofing – GPUs, storage, networking
  • FP8 and FP16 support to boost performance with less resources and cost
  • Intel, AMD, ARM CPU options
  • NVIDIA Certified with NVIDIA AI Enterprise and NGC catalogue to fully leverage pre-trained models and optimised libraries and toolset

Systems

Intel Based GPU Server: SYS-421GE-TNRT

Powered by the dual 4th Generation Intel Xeon Scalable processors with the most built-in accelerators to improve performance in AI, data analytics, networking, storage, and HPC, this system possesses large processing power and memory capacity for your demanding applications. The 4th Generation Intel Xeon Scalable processors support up to 350W TDP each and 32 DDR5 DIMM slots that can support up to 8TB of 4800MHz ECC memory.

This server has 24 hot-swap 2.5-inch drive bays for storage that accommodate NVMe, SATA, or SAS drives. Different types of drives can be mixed and matched to suit the needs. 2 M.2 NVMe slots are also available for additional storage options.

The 421GE-TNRT also has various features to ensure reliability, security, and manageability. It has 4 redundant 2700W titanium level power supplies, 8 hot-swap heavy-duty fans and optional liquid cooling for optimal thermal performance. Additionally, it is equipped with cryptographically signed firmware, a hardware-trusted platform module, and a silicon root of trust for enhanced security.

AMD Based GPU Server: 4125GS-TNRT

The server is powered by dual AMD EPYC™ 9004 Series (Genoa) processors with up to 128 cores/256 threads and up to 400W TDP each. These processors are based on the Zen 4 architecture and offer exceptional performance, scalability and efficiency for your workloads.

Moreover, this server supports up to 24 hot-swap NVMe/SATA/SAS drive bays, including 4 dedicated NVMe bays, giving you ample storage capacity and speed for your data. You can also use the M.2 slot for additional NVMe storage or boot devices.

Equipped with 4000W redundant titanium-level power supplies, this server ensures high efficiency and reliability. It also features IPMI 2.0 with virtual media over LAN and KVM-over-LAN support for easy remote management and monitoring.

SYS-221GE-NR

2U GPU SuperServer that supports up to four NVIDIA H100 PCIe GPUs and dual-socket 4th Gen Intel Xeon Scalable processors designed for high-performance computing, AI/deep learning training, large language model (LLM) natural language processing, and other demanding accelerated computing workloads.

Powered by NVIDIA NVLink, this server has the highest GPU communication which enables high-speed data transfer between GPUs. This server can have up to 8TB of ECC DDR5 memory across 32 DIMM slots, which provides high-capacity and high-speed memory for data-intensive applications. Additionally, 7 PCIe 5.0 x16 FHFL slots are available to offer high-bandwidth and low-latency connectivity for GPUs and other devices.

ARS-221GL-NR

This server is a modular building block platform with an energy-efficient Grace CPU superchip, a high-performance CPU designed for AI and HPC workloads. This system can support up to four NVIDIA H100 PCIe GPUs, which are connected by NVIDIA NVLink for high-speed data transfer12. The system also has 480GB or 240GB of onboard LPDDR5X memory, which offers low latency and high power efficiency.

Moreover, It supports NVIDIA BlueField-3 Data Processing Unit (DPU), which enhances the network performance and security of the system. The system has 8 E1.S hot-swap NVMe drive slots and 7 PCIe 5.0 x16 slots for flexible storage and expansion options. It also has redundant titanium-level power supplies, heavy-duty fans, and optional liquid cooling for reliable and efficient operation.


Ready to Get Started?

If you are ready to take your LLM projects to the next level with DiGiCOR GPU Servers, please contact us today. We are here to help you find the best solution for your needs and budget.

Don’t settle for less. Choose DiGiCOR GPU Servers for LLM today.


A Blueprint for LLM and Generative AI Infrastructure at Scale

What does AI Infrastructure entail, and how does Supermicro offer the blueprints for a future-proof approach?

Learn about the proven approach to AI infrastructure that will carry forward for the foreseeable future