Just as the stage is set for 400G Ethernet (GbE) to roll out in force later this year, mainly in hyperscale, telco and large data-center networks, there is a call to boost that speed to 800GbE or even higher in the coming years.The need for increased speed in data centers and cloud services speeds is driven by many things including the continued growth of hyperscale networks from players like Google, Amazon and Facebook, but also the more distributed and mobile workloads modern networks support. But the reality on the ground is that much lower speeds are what’s commonly in use.
Just a few years ago, many expected all the Internet of Things (IoT) to move to the cloud—and much of the consumer-connected IoT indeed lives there—but one of the key basics of designing and building enterprise-scale IoT solutions is to make a balanced use of edge and cloud computing. Most IoT solutions now require a mix of cloud and edge computing which can alleviate latency, increase scalability, and enhance access to information so that better, faster decisions can be made, and enterprises can become more agile as a result.
Hengtong Rockley has released a 400G QSFP-DD DR4 optical module based on silicon photonics technology geared for cloud data center networks. The new 400G QSFP-DD module is the company’s first 400G silicon photonics optical module product and is intended for use in next generation cloud-scale data center networks for low-cost, low-power consumption optical connections between switches. The deployment of such 400G transceivers will enable data center networks to deliver a 4x increase in network speed compared to existing deployments using 100G.
There are billions of connected devices in use worldwide, and that number is increasing by the millions every year. Unfortunately, many of these IoT devices, as well as those currently being developed and deployed, lack critical security features, making them easy targets for hackers and botnets. Without the proper security measures in place, these devices can lead to catastrophic events.
R&M’s new fiber optic distribution platform Netscale 72 natively supports two parallel optical cabling types, BASE8 and BASE12. That means distribution modules for both applications fit in the same system drawers. Data centers can adapt the trunk cabling within the existing racks and housing. In this way, Netscale 72 facilitates fast migration to new network generations.
In any type of computing environment, the housing, protection and management of network connections is essential for uptime and performance. The methods for providing that protection and management, as well as the products and technologies for doing so, can vary significantly depending on the computing environment in which they will reside. This article looks at options for cabling and network-equipment housing, protection, and management in different environments.
Just a few years ago, many expected all the Internet of Things (IoT) to move to the cloud—and much of the consumer-connected IoT indeed lives there—but one of the key basics of designing and building enterprise-scale IoT solutions is to make a balanced use of edge and cloud computing. Compared to cloud-only solutions, blended solutions that incorporate edge can alleviate latency, increase scalability, and enhance access to information so that better, faster decisions can be made, and enterprises can become more agile as a result.
Edge provides a huge opportunity to host many use cases on one infrastructure, manageable from a single pane of glass. Getting close to end-users not only allows the operator to tap directly into the new revenue streams for ultra-low latency/ultra-reliable services, but also to provide “edge-as-a-service,” and other infrastructure-as-a-service and hosting services to other enterprises.
The following are five technology and market trends in the areas of compute, storage, and network to watch in 2020: The evolution of server architecture; software defined data centers; cloud consolidation; the emergence of edge computing; and advances in server network connectivity.
Containers solve the problem of getting software to run reliably when moved from one computing environment to another — say, from a developer laptop to a test environment, or from a physical machine in a data centre to a virtual machine in a public or private cloud. More and more businesses are joining the container “revolution” — the market around this technology is on track to being a $2.7 billion market by 2020.