Skip to main content



Archived content

NOTE: this is an archived page and the content is likely to be out of date.

Fujitsu Develops Software that Analyzes Service Quality while Monitoring at World-Record 200 Gbps

Ten-fold jump in analysis performance lowers the cost of providing stable, reliable service infrastructure

Fujitsu Laboratories Ltd.

Kawasaki, Japan, April 02, 2015

Fujitsu Laboratories Ltd. today announced the development of software that monitors transmission packets and analyzes quality in real time at a world-record speed of 200 Gbps.

With the advance of cloud services, brought on by the spread of smartphones and tablets and the expanded use of datacenters, it is increasingly important to have a firm grasp of service quality to ensure the overall operational stability of systems. Until now, however, when monitoring communications packets with commodity hardware, it was difficult to analyze service quality at high speeds because of limits in the performance of CPUs and memory access.

Now Fujitsu has developed software that runs on a single piece of commodity hardware to monitor communications packets and analyze network and application quality in real time at speeds of 200 Gbps, a ten-fold improvement in performance. This was accomplished by technology that reduces computational load when collecting packets, technology that eliminates the need for memory copying and mutual exclusion, and parallel processing technology that prevents contention when using multiple cores.

Together, these technologies make it possible to provide network infrastructure that supports the stable use of services without requiring expensive, specialized equipment, allowing for lower costs.


With the advance of cloud services, brought on by the spread of smartphones and tablets and the expanded use of datacenters, the quality of services using communications networks is increasingly important. Realistically, however, there can be unforeseeable service quality degradation due to increasing volumes of data on the network and complex system architectures. Quickly detecting and recovering from service quality degradation requires real-time analyses to identify whether the cause of the problem was a network-quality issue or an application-level issue, such as response time. To analyze network and application quality, there is a need to take a detailed look at the behavior of each communications packet.

Technological Issues

Communications speeds on networks get faster every year. One hundred Gbps Ethernet, or 200 Gbps when combining upstream and downstream transmissions, has already been deployed on backbone carrier networks, and even in datacenters, systems are being constructed that use multiple 10 Gbps ports, or 20 Gbps with combined upstream and downstream communications. A faster network transmits more packets, which demands more processing power.

But the maximum number of packets that can be received and transmitted per unit of time depends on the limitations of CPUs and other hardware components. Meanwhile, packet collection and network- and application-quality analyses, which depend on memory-access performance, suffer from inter-process memory copying delays. These problems have meant improved processing performance has been difficult to achieve.

About the Technology

Fujitsu has developed technology that makes it possible to collect packets at 200 Gbps and to perform network- and application-quality analyses all with software, eliminating the need for expensive, specialized hardware (Figure 1). For network-quality analysis, this calculates the communications traffic volume on the network and detects packet loss and network latency. For application-quality analysis, this calculates the communications volume for each service and detects the response delay of each application providing a service. Combining these quality analyses together makes it possible to identify sources of service delays and their locations.

Figure 1: Diagram of 200 Gbps service-quality analysisFigure 1: Diagram of 200 Gbps service-quality analysis

Features of the technology are as follows (see Figures 2, 3)

  1. Faster packet collection

    Fujitsu improved the performance of packet-collection processing by coalescing the interrupts that are generated as each packet arrives, which reduces the number of processing iterations, and by distributing interrupt processing over multiple CPU cores, where the CPU halts other processes as new interrupt processes are accepted.

  2. Faster memory access

    Fujitsu refined the methods and timing for looking up data between packet-collection and quality-analysis processes, making it possible to look up data without copying packets or analysis data, while avoiding the problem of writing to regions of memory that were already in the middle of a write or lookup operation. This obviates the need for copying and for mutual exclusion.

  3. Parallel processing

    There is no need for mutual exclusion, so multiple analysis processes running on multiple CPU cores can access the same ring buffer so that, even with multiple CPU cores, there is no contention, and analysis performance can scale in a linear manner in accordance with the number of CPU cores.

Figure 2: Technologies for faster packet collection, faster memory accessFigure 2: Technologies for faster packet collection, faster memory access
Larger View (136 KB)

Figure 3: Parallel processing technology that enables effective use of multiple coresFigure 3: Parallel processing technology that enables effective use of multiple cores
Larger View (100 KB)


This technology was able to boost performance ten-fold, making it possible to monitor packets flowing over high-speed networks at 200 Gbps and to analyze network quality and application quality in real time using only software and commodity hardware, rather than expensive, specialized hardware. This makes it possible to inexpensively identify the causes and locations of service-quality degradations.

In addition, because sender information is kept in logs, this can be used to track the sources of cyber attacks and servers under attack. Moreover, because it is possible to track bandwidth needs on a per-application basis, infrastructure can be built out as needed.

Taken all together, this technology can be used to provide network infrastructure that supports the stable use of services, delivering faster resolution of troubles, stronger security, and faster provision of services.

Future Plans

Fujitsu is proceeding with R&D and verification testing of quality analyses with the goal of including this technology in a commercial product during fiscal 2015.

About Fujitsu

Fujitsu is the leading Japanese information and communication technology (ICT) company offering a full range of technology products, solutions and services. Approximately 162,000 Fujitsu people support customers in more than 100 countries. We use our experience and the power of ICT to shape the future of society with our customers. Fujitsu Limited (TSE: 6702) reported consolidated revenues of 4.8 trillion yen (US$46 billion) for the fiscal year ended March 31, 2014. For more information, please see

About Fujitsu Laboratories

Founded in 1968 as a wholly owned subsidiary of Fujitsu Limited, Fujitsu Laboratories Ltd. is one of the premier research centers in the world. With a global network of laboratories in Japan, China, the United States and Europe, the organization conducts a wide range of basic and applied research in the areas of Next-generation Services, Computer Servers, Networks, Electronic Devices and Advanced Materials. For more information, please see:

Press Contacts

Public and Investor Relations Division

Company:Fujitsu Limited

Technical Contacts

Software Laboratory

E-mail: E-mail:
Company:Fujitsu Laboratories Ltd.

All company or product names mentioned herein are trademarks or registered trademarks of their respective owners. Information provided in this press release is accurate at time of publication and is subject to change without advance notice.

Date: 02 April, 2015
City: Kawasaki, Japan
Company: Fujitsu Laboratories Ltd.