Tachyum Presents Software Stack for HPC/AI and Supercomputer Management
Tachyum released a white paper, “Tachyum HPC/AI Software Stack Overview,” detailing the software packages offered by the company and supported by its software team that enable software developers to build HPC applications deployed as part of Prodigy-empowered data centers.
AI Insights: Google Introduces PaLM API & MakerSuite; Simplifies Generative AI Development Lifecycle
“Multiple different processing technologies are combined in today’s HPC systems”
Tachyum’s software designs help meet the high demands of cloud and HPC/AI workloads with Prodigy. The company’s software team cooperates with early customers and partners to create software packages ideally suited for HPC environments primarily using open-source software. With its software stack, Tachyum supports all key software libraries required for HPC applications and addresses current challenges and opportunities.
Tachyum software stack components are comprised of:
- Tachyum firmware: BIOS (UEFI), BMC
- Operating system level: LINUX Kernel, SELinux
- Storage and file-systems options: Ceph
- OS level virtualization: Docker and containerd, Podman based on Kubernetes pods
- Communication stack and runtimes: OpenMPI, MPICH, gPRC, RabbitMQ
- User space utilities and scripting: Tachyum Linux distribution provides all the necessary utilities and scripting languages like Bash, SED, Gawk, Grep, Wget, Rsync, Strace, Git, Subversion, PHP, Perl, Ruby, Tcl, Top, Htop, Iotop, Sar, Nc.
- Parallelism standards: OpenMP, SHMEM
- Programming languages and compilers: Tachyum has ported GNU gcc compiler and go compiler.
- Debugging tools: GDB, KGDB, OpenOCD
- Emulation tools: Tachyum has developed multiple emulators like GDB based emulator for internal use and QEMU based emulator for customers to translate X-86, ARM and RISC-V software to Tachyum’s Prodigy architecture.
- Clustering, management, and job scheduling: SLURM for jobs management, Zabbix and Prometheus with Grafana for monitoring, Kubernetes for automating software deployment, scaling and management of containerized applications, Ansible for IT automation.
- Networking capabilities: Tachyum has developed its own networking infrastructure to support HPC/AI deployments for Prodigy. SONIC was chosen as the network OS.
- AI/HPC frameworks and libraries: BLAS, GEMM, NumPy, Eigen, PyTorchTensorFlow
- AI applications: Resnet and ConvMixer for computer vision, SSD and Yolo for object detection, Fast-SCNN and UNET for instance segmentation, BERT for NPL
- HPC applications: Quantum espresso, DeepMD, LAMMPS
Read More about AI Experiences : Google Brings Generative AI Experiences to Google Workspace
“Multiple different processing technologies are combined in today’s HPC systems,” said Dr. Radoslav Danilak, founder and CEO of Tachyum. “Achieving the required performance and functionality at an affordable power envelope, places a lot of challenges to HPC software environments – especially on scheduling and resource management. Tachyum’s Prodigy, the world’s first Universal Processor, integrates the functionality of CPU, GPGPU and TPU into a single homogeneous architecture so that there is no additional complexity needing to be solved on the software level between hardware components.”
Prodigy delivers unprecedented data center performance, power, and economics, reducing CAPEX and OPEX significantly. Because of its utility for both high-performance and line-of-business applications, Prodigy-powered data center servers can seamlessly and dynamically switch between workloads, eliminating the need for expensive dedicated AI hardware and dramatically increasing server utilization. Tachyum’s Prodigy delivers performance up to 4x that of the highest performing x86 processors (for cloud workloads) and up to 3x that of the highest performing GPU for HPC and 6x for AI applications.
Latest ChatGPT Insights : OpenAI’s ChatGPT 4 Is Here. Is It Time to Forget ChatGPT?
[To share your insights with us, please write to sghosh@martechseries.com]
Comments are closed.