Threadripper Pro! AMD has released the long awaited Threadripper Pro CPUs. I was able to spend a (long) day (and night) running compute performance testing on the flagship 64-core TR Pro 3995WX. In this post I’ve got some HPC workload benchmark results from putting this excellent CPU through its compute paces.
Intel oneAPI AI Analytics Toolkit — Introduction and Install with conda
I recently wrote a post introducing Intel oneAPI that included a simple installation guide of the Base Toolkit. In that post I promised a follow-up about the the oneAPI AI Analytics Toolkit. This is it! I’ll describe what it is and give recommendations for doing an install setup of the AI toolkits using conda with Anaconda Python.
Intel oneAPI Developer Tools — Introduction and Install
Intel oneAPI is a massive collection of very high quality developer tools, and, it’s free to use! In this post I’ll give you a little background on what oneAPI is and my recommendations for doing an install setup to get started exploring the collection of tool-kits.
How To Install TensorFlow 1.15 for NVIDIA RTX30 GPUs (without docker or CUDA install)
In this post I will show you how to install NVIDIA’s build of TensorFlow 1.15 into an Anaconda Python conda environment. This is the same TensorFlow 1.15 that you would have in the NGC docker container, but no docker install required and no local system CUDA install needed either.
Notes on “Notes” (new blog post format)
Starting 2020 off with an addition to my writing, “micro blogging” via GitHub Gists
How To Use MKL with AMD Ryzen and Threadripper CPU’s (Effectively) for Python Numpy (And Other Applications)
In this post I’m going to show you a simple way to significantly speedup Python numpy compute performance on AMD CPU’s when using Anaconda Python.
AMD Ryzen 3900X vs Intel Xeon 2175W Python numpy – MKL vs OpenBLAS
In this post I’ve done more testing with Ryzen 3900X looking at the effect of BLAS libraries on a simple but computationally demanding problem with Python numpy. The results may surprise you! I start with a little bit of history of Intel vs AMD performance to give you what may be a new perspective on the issue.
PyTorch for Scientific Computing – Quantum Mechanics Example Part 4) Full Code Optimizations — 16000 times faster on a Titan V GPU
This is the 16000 times speedup code optimizations for the scientific computing with PyTorch Quantum Mechanics example. The following quote says a lot,
“The big magic is that on the Titan V GPU, with batched tensor algorithms, those million terms are all computed in the same time it would take to compute 1!!!”
PyTorch for Scientific Computing – Quantum Mechanics Example Part 3) Code Optimizations – Batched Matrix Operations, Cholesky Decomposition and Inverse
An amazing result in this testing is that “batched” code ran in constant time on the GPU. That means that doing the Cholesky decomposition on 1 million matrices took the same amount of time as it did with 10 matrices!
In this post we start looking at performance optimization for the Quantum Mechanics problem/code presented in the first 2 posts. This is the start of the promise to make the code over 15,000 times faster! I still find the speedup hard to believe but it turns out little things can make a big difference.
PyTorch for Scientific Computing – Quantum Mechanics Example Part 2) Program Before Code Optimizations
This is the second post on using Pytorch for Scientific computing. I’m doing an example from Quantum Mechanics. In this post we go through the formulas that need to coded and write them up in PyTorch and give everything a test.