High Performance Computing most generally refers to the practice of aggregating computing power in a way that delivers much higher performance than one could get out of a typical desktop computer or workstation in order to solve large problems in science, engineering, or business.
High-performance computing (HPC) is the use of super computers and parallel processing techniques for solving advanced problems and performing research activities through computer modeling, simulation and analysis. HPC systems have the ability to deliver sustained performance through the concurrent use of computing resources. The terms high-performance computing and supercomputing are sometimes used interchangeably.
High-performance computing (HPC) evolved due to meet increasing demands for processing speed. HPC brings together several technologies such as computer architecture, algorithms, programs and electronics, and system software under a single canopy to solve advanced problems effectively and quickly. A highly efficient HPC system requires a high-bandwidth, low-latency network to connect multiple nodes and clusters. The most common users of HPC systems are scientific researchers, engineers and academic institutions. It is also implemented in Biosciences, Geographical data, Oil and gas industry modeling, Electronic design automation, Climate modeling, Media and entertainment. As demand for processing power and speed grows, HPC will likely interest businesses of all sizes, particularly for transaction processing and data warehouses.
Two of the most popular choices in HPC are Linux (in all the many varieties) and Windows. Linux currently dominates HPC installations, but this in part due to HPC’s legacy in supercomputing, large scale machines, and Unix. Your choice of operating system should really be driven by the kinds of applications you need to run on your high performance computer.
Sabre Informix offers HPC consulting , implementation, performance tuning and training services. HPC Blog