TeraFLOPS (TFLOPS) is a unit that indicates how many trillions of cal­cu­la­tions with floating point numbers a computer can carry out in one second. The value serves as a measure of the per­for­mance of proces­sors, es­pe­cial­ly GPUs and su­per­com­put­ers. TFLOPS are es­pe­cial­ly relevant for ap­pli­ca­tions that involve a lot of cal­cu­la­tion, like ar­ti­fi­cial in­tel­li­gence, sci­en­tif­ic sim­u­la­tions and machine learning.

What are FLOPS and what are they used for?

FLOPS stands for floating point operations per second and is a unit for computing power. A floating point operation is a math­e­mat­i­cal cal­cu­la­tion that involves decimal points. They are es­pe­cial­ly important for cal­cu­la­tion-heavy ap­pli­ca­tions that require a high degree of precision.

FLOPS are mostly used for sci­en­tif­ic cal­cu­la­tions, sim­u­la­tions, ar­ti­fi­cial in­tel­li­gence, machine learning and graphics ap­pli­ca­tions. They play a central role in various areas such as medical image pro­cess­ing and physical sim­u­la­tions. They are also important in finance, for example when it comes to the analysis of market data. In the gaming industry, FLOPS are used to determine the graphics per­for­mance of modern GPUs. With ever in­creas­ing FLOPS capacity, modern computers can deliver more and more realistic physical effects and high-res­o­lu­tion graphics.

FLOPS are typically measured using specially developed benchmark tests that determine the number of floating point op­er­a­tions per second. Fre­quent­ly used bench­marks include LINPACK, which is mostly used for su­per­com­put­ers, and FP32/FP64, which rate the computing power of GPUs. During tests, complex math­e­mat­i­cal cal­cu­la­tions are performed in order to determine how many op­er­a­tions per second a system can handle. Man­u­fac­tur­ers often give the­o­ret­i­cal FLOPS values based on the ar­chi­tec­ture of a computer. However, real-world ap­pli­ca­tions can vary based on workload and ef­fi­cien­cy.

Dedicated Servers
Per­for­mance through in­no­va­tion
  • Dedicated en­ter­prise hardware
  • Con­fig­urable hardware equipment
  • ISO-certified data centers

How many FLOPS are in a teraFLOPS?

One teraFLOPS is equal to one trillion (1,000,000,000,000 or 1012) floating point op­er­a­tions per second. That means that a processor with 1 TFLOPS can execute a trillion math­e­mat­i­cal op­er­a­tions with floating point numbers per second.

By way of com­par­i­son, a computer that only has 1 FLOPS would need 31,000 years to perform a trillion floating point op­er­a­tions. So computers working in TFLOPS are powerful systems capable of modern ap­pli­ca­tions in real time.

What other FLOPS units exist and how do they convert into TFLOPS?

There are many FLOPS units, which differ in how many op­er­a­tions per second they refer to.

Unit FLOPS value Con­ver­sion into TFLOPS
KiloFLOPS 103 FLOPS (1,000) 10-9 TFLOPS
MegaFLOPS 106 FLOPS (1 million) 10-6 TFLOPS
GigaFLOPS 109 FLOPS (1 billion) 10-3 TFLOPS
TeraFLOPS 1012 FLOPS (1 trillion) 1 TFLOP
PetaFLOPS 1015 FLOPS (1 quadrillion) 103 TFLOPS
ExaFLOPS 1018 FLOPS (1 quin­til­lion) 106 TFLOPS

Su­per­com­put­ers’ per­for­mance is measured in petaFLOPS and even exaFLOPS, while high-end graphics cards are usually rated in teraFLOPS.

How many FLOPS do modern computers and GPUs reach?

GPUs and modern computers in the area of high-per­for­mance computing have reached im­pres­sive FLOPS values. The NVIDIA H100, one of the most powerful GPUs for AI and data centers, achieves up to 989 teraFLOPS for FP32 Tensor Core cal­cu­la­tions. That makes it ideal for large neural networks and sim­u­la­tions.

The NVIDIA A30, a GPU that’s optimized for data centers, reaches 10 TFLOPS and is par­tic­u­lar­ly suitable for AI training and in­fer­ences. By com­par­i­son, the gamer-oriented NVIDIA RTX 4090 can overclock to over 100 TFLOPS and enables very realistic graphics.

Su­per­com­put­ers are even more powerful: The Frontier su­per­com­put­er has surpassed the 1 exaFLOPS mark and is used for highly complex sci­en­tif­ic sim­u­la­tions. Other powerful su­per­com­put­ers used in research, like the Japanese computer Fugaku, also operate in this range.

Go to Main Menu