Thursday, May 23, 2019

VxRail is Holding an ACE

I’m here in Las Vegas for Dell Technologies World and I think it’s an absolutely fitting location to let everyone know that VxRail is getting ready to flip over its ACE. However, unlike the mythical ace of spades, VxRail Analytical Consulting Engine, or ACE, gives you the winning hand in your VxRail environment every time. And VxRail ACE, with its cloud connectivity to the VxRail HCI System Software, is a great tie to the Dell Tech Cloud story we announced yesterday.

As IT becomes more complex, it’s time to shift the odds from the house to your favor. Dell EMC ACE actively monitors and provides management guidance for the entire hyperconverged stack. Leveraging artificial intelligence and infrastructure machine learning, VxRail ACE will drive increased operational efficiencies through automation.

As a virtual infrastructure administrator, you’re still playing your hand but you can count on VxRail ACE to be aware of the way things should be done, such as ensuring healthy capacity utilization rates and forecasting resource consumption to keep your HCI stack operating at peak performance; ready for future workloads.



Improving Your Odds


Unlike in Las Vegas, VxRail ACE gives you best practices so you always have more than luck on your side when managing your IT environment. VxRail ACE uses a data lake, with historical data of how customers are using VxRail, and machine learning to identify ways to optimize configurations and streamline infrastructure management. VxRail ACE can help you tackle day-to-day management while keeping an eye out on the horizon so you can make informed decisions – and the more hands that get played, the smarter VxRail ACE gets.

Winning at cards is all about knowing the odds. In Las Vegas, the house always has the advantage. That’s not the case with VxRail ACE in your hand. VxRail ACE can show you the odds of going over capacity. The odds of needing additional performance. The odds that upgrading VxRail to the latest version will enhance your IT environment and business and even detect anomalies in your environment and learn from them.

Doubling Down on ACE


VxRail ACE can help predict future storage capacity requirements


VxRail ACE is an extension of the VxRail HCI System Software that leverages an Adaptive Data Collector (ADC) and using Dell EMC Secure Remote Support (SRS), collects telemetry including events, alarms, utilization and more from the VxRail clusters for processing and analytics in our secure cloud-based data lake. Here our analytics compute engines analyze the data to produce valuable information that you can view for your entire VxRail estate and act on, along with an associated health score for each and every cluster and associated telemetry.

Moreover, as we move forward, we will be doubling down on the functionality delivered in ACE. Our vision for VxRail ACE is to offer action in addition to insights, as future functionality will allow it to troubleshoot issues saving IT valuable time.

And like a professional player coaching you, VxRail ACE is currently in use by Dell EMC Services to ensure system and version accuracy across the install base, including monitoring vitals such as installed OS and physical configurations to start so that you always have another set of eyes helping you play your best hand. Being able to detect anomalies and trend data to create projections reduces manual oversight while increasing service uptime.

ACE Part of a Full House of VxRail Software


VxRail ACE is a key component of VxRail HCI System Software


Since you can’t win the pot with a single card, VxRail gives you a full house with VxRail HCI System Software. Further, the Dell EMC VxRail turnkey experience offers full stack integration – control from top to bottom, of software and hardware together – for a consistent, deeply integrated VMware environment. VxRail goes even further to deliver even more highly differentiated features and benefits based on the VxRail HCI System Software, the orchestration engine for unique automation, integration, management, analytics, and extensions for VxRail environments, which is unique to VxRail and automates deployment, delivers complete lifecycle management, and facilitates key upstream and downstream integration points that create a truly better together experience with VxRail as the foundation.

Best of all, We’re Sharing our ACE with you


VxRail ACE will be available for early access at the end of May. We’re making it available to you to try out at no charge. That’s right, there’s no ante to get access to our ACE. This means there’s no additional software or hardware to install. Everything is delivered via a user-friendly, fully customized portal available through your Service 360 accounts. Why are we doing it this way? We know we have invaluable expertise in our customers. We want to collaborate with you to ensure you’re vested in the development of VxRail ACE so it gives you the value you need. We’ll be working on a continuous innovation/continuous development (CIDC) cadence so you’ll always have the latest and greatest features at your fingertips

Sunday, April 14, 2019

Making AI, HPC and GPUs Easier for Data Scientists

With new GPU-ready containers and server infrastructure, organizations are simplifying the deployment of artificial intelligence applications and the clusters they run on.


Regardless of the domains they work in, researchers, data scientists and developers want to focus on the work they do, not the systems and tools they use to run their workloads. While that’s a simple proposition, the reality on the ground is often far removed from this ideal world.

Those who have built artificial intelligence and other high-performance computing applications from the ground up — starting with bare-metal servers, computing frameworks, software libraries and more — understand that there is a lot of heavy lifting required up front. This work always comes first, before you can focus on the work you really want to do, like training your model, running your inference workloads and seeing the results.

Together with its partners, NVIDIA is working to rewrite this story. It’s making the deployment of GPU-accelerated systems really easy with its NVIDIA GPU Cloud, or NGC.

The NGC container registry


Via the NGC container registry, NVIDIA provides a catalog of GPU-accelerated containers that deliver easy-to-deploy GPU-accelerated software for AI, machine learning and HPC. These containers, which are available to download at no charge, alleviate many of the headaches that come with setting up software. They help you get up and running quickly with tested, optimized and updated frameworks and applications. The containers are designed to take full advantage of NVIDIA GPUs, on-premises or in the cloud, and to work across a wide variety of NVIDIA GPU platforms.

The NGC container registry is simple to use. When you visit the site, you are prompted to answer a simple question: “What are you interested in working on?” To move forward, you simply select one of six options from a catalog: High Performance Computing, Deep Learning, Machine Learning, Inference, Visualization, or Infrastructure. Once a selection is made, you can begin working with, say, the Caffee2 deep-learning framework or the LAMMPS software application for molecular dynamics simulations. After that, you can move forward with the confidence that comes with knowing your software is correctly configured.

The NGC-Ready hardware


Software, of course, is only part of the problem when it comes to deploying large-scale HPC and AI applications. You also have to find the right hardware infrastructure to run your workloads. NVIDIA makes this process easier with its NGC-Ready program. Through this program, hardware vendors validate that the NGC containers run correctly on their servers and workstations.

Dell EMC is an active participant in this program. Our NGC-Ready infrastructure, including the Dell EMC PowerEdge C4140 server, has been tested and validated to run containers from the NVIDIA GPU Cloud. This back end work allows organizations to deploy GPU-accelerated Dell EMC systems with the confidence that comes with knowing they are ready to run NGC containers.

And that’s just one part of the work that Dell EMC does with NVIDIA. My team works closely with the engineers at NVIDIA to optimize systems, perform benchmark testing and take other steps to help ensure that you and others can get the full value of GPU acceleration when deploying NGC containers on Dell EMC hardware.

For example, some users worry that containerizing an application will hurt performance. Through our lab and benchmark testing, we’ve been able to show that little, if any, performance is lost when you containerize your software, as opposed to starting with software that you deploy on bare-metal servers. We put a special focus on proving the potential of using containers in large-scale simulations and deep learning applications. In our lab, we find containers to be portable and efficient as we scale out deep learning workloads.

Even better, we’re demonstrating how the container approach makes life easier for data scientists and other users who are venturing into the brave new world of AI and machine learning. With the right frameworks incorporated into an NGC container and deployed on our NGC-Ready hardware, you can take advantage of GPUs for deep learning by adding just a few lines of code — literally. You add the lines of code and you gain the benefits of GPU acceleration, with no need to port your application code to a new platform.

Ready Solutions for AI


In another ongoing initiative, our engineers at Dell EMC work closely with their counterparts at NVIDIA to bring systems to market that are optimized for GPU-enabled deep learning applications. That’s the case with our Dell EMC Ready Solutions for AI, which are optimized for deep learning with NVIDIA GPUs.

These Ready Solutions provide a GPU‑optimized stack that can shave valuable time off of deep learning projects. If that’s your goal, Dell EMC engineers can help you configure, test and tune your GPU‑enabled hardware and software, so you can get up and running quickly with a top-tier deep learning platform based on a framework that can use both CPUs and GPUs. These solutions even include services to help your data scientists discover insights from data in less time.

Key takeaways


If you take a step back and look at the big picture, you’ll see that there are a lot of resources available to help your organization deploy AI and HPC-driven applications that take advantage of the power of GPU acceleration, and to do so with the confidence that comes with validated and optimized hardware and software solutions.

I think you’ll also see that it’s becoming much easier to get into this game than it was just a handful of years ago. Today, end users no longer need to become experts in the underlying technology to capitalize on GPU-accelerated systems and the power of AI and other HPC-fueled applications. Instead, they can keep their eyes of the real prize — the work they do.