|

Get started building RAG pipelines in your enterprise with Dell Technologies and NVIDIA (Part 1)

A Guest Post by Bertrand Sirodot & Fabricio Bronzati In our previous blog, we showcased running Llama 2 on XE9680 using NVIDIA’s LLM Playground (part of the NeMo framework). It is an innovative platform for experimenting with and deploying large language models (LLMs) for various enterprise applications. The reality is that running straight inference with foundational models in…

| |

Dell Validated Design Guides for Inferencing and for Model Customization – March ’24 Updates

A Guest Post by Damian Erangey & Scott Powers Continuous Innovation with Dell Validated Designs for Generative AI with NVIDIA Since Dell Technologies and NVIDIA introduced what was then known as Project Helix less than a year ago, so much has changed. The rate of growth and adoption of generative AI has been faster than probably any technology…

|

Unveiling the Power of the Dell PowerEdge XE9680 Server on the GPT-J Model from MLPerf™ Inference

A guest post by Frank Han , Rakshith Vasudev & Manpreet Sokhi Abstract For the first time, the latest release of the MLPerf™ inference v3.1 benchmark includes the GPT-J model to represent large language model (LLM) performance on different systems. As a key player in the MLPerf consortium since version 0.7, Dell Technologies is back…

| |

Dell Technologies Validated Design for Generative AI Inferencing with NVIDIA (Interview & Demo)

Join Nomuka and Fabricio, a Senior Principal Systems Development Engineer at Dell, to learn about the world of Generative AI Inferencing and Dell’s Validated Design for Generative AI Inferencing with NVIDIA! Get answers to the following key questions: What is inferencing and why is it essential in the field of AI? What makes the Dell’s…

| |

Dell Technologies Shines in MLPerf™ Stable Diffusion Results

A guest post by Rakshith Vasudev & Frank Han Abstract The recent release of MLPerf Training v3.1 results includes the newly launched Stable Diffusion benchmark. At the time of publication, Dell Technologies leads the OEM market in this performance benchmark for training a Generative AI foundation model, especially for the Stable Diffusion model. With the…

|

Llama 2 (Meta’s Large Language Model) on Dell Technologies PowerEdge XE9680

A guest post by Nomuka Luehr & Fabricio Bronzati NVIDIA’s LLM Playground, part of the NeMo framework is an innovative platform for experimenting with and deploying large language models (LLMs) for various enterprise applications. It’s currently in a private, early access stage and offers the following features: Experimentation with LLMs: The Playground provides an environment to use and experiment…

|

Using Retrieval Augmented Generation (RAG) on a Custom PDF Dataset with Dell Technologies

A guest post by David O’Dell & Bertrand Sirodot The Generative AI transformation Artificial Intelligence is transforming the entire landscape of IT and our digital lives.  We’ve witnessed several major disruptions that have changed the course of technology over the past few decades. The birth of the internet, virtual reality, 3D printing, containerization, and more…

| |

Unlocking the Power of Conversations A Guide to Crafting Effective Prompts for Large Language Models like ChatGPT

A guest post by Navin Mukraj In the realm of conversational AI, the key to engaging and meaningful interactions lies in the art of crafting effective prompts. These prompts serve as the foundation for shaping the responses generated by models like ChatGPT. In this blog post, we will delve into a comprehensive framework that empowers…

| | | |

Dell Project Helix : Generative AI 101 Part 5: The Joint Dell and NVIDIA Solution Architecture

A guest post by Nomuka Luehr Dell and NVIDIA have been collaborating for years to develop innovative solutions for artificial intelligence (AI) computing. They have now created a full-stack solution that allows enterprises to generate and deploy AI models at scale. This joint project builds on their previous work and further strengthens their leadership in the…

| | | | |

Dell Project Helix & Generative AI 101 Part 4: Inferencing (Running your LLM)

A guest post by Nomuka Luehr In the previous parts of this series, we have discussed various aspects of generative AI and language model training, including transformer-based models, LLM training, reinforcement learning, and pre-trained model fine-tuning. In this final part, we will delve into the broader topics of general AI training and inference. Inferencing Once the…