Each edge-computing deployment has very specific requirements, but since the need for edge applications over time may shift, it’s important to find an architecture that works now but is flexible enough to meet future needs. Credit: MF3d / Getty Images Many edge-computing deployments are driven by very specific needs, but since new needs may arise down the road with a different set of edge requirements, IT leaders should adopt edge-computing architectures with flexibility and adaptability in mind. The fact that all edge-computing systems have certain things in common – complex combinations of hardware, applications, infrastructure software and networking – doesn’t mean they should all have the same design. Every new project requires highly specialized software and integrated custom networking to deliver on project goals across such diverse applications as industrial control, autonomous transportation, health services, public safety and energy management. Each use case will have its unique requirements in terms of performance, response times, quantity of data to be gathered and processed, and cost. Since projecting the return on investment for these systems remains challenging, it’s best to choose an architecture that is readily customizable in order to minimize the cost of meeting future edge needs. Deploying an edge computing system requires a complex combination of servers, applications, infrastructure software and networking. The server runs the software-control systems and analytics to translate, for example, large amounts of IoT data into actionable information. Open APIs are available on both software and networking systems to customize traffic flows. The network is designed to deliver the latency and reliability required within the cost constraints. Given the variables that come into play with each element of an edge system – devices, local network, gateway, computing – making them all work to meet required specifications may require system-integration skills. Here’s a brief look at these elements. Edge Devices The variety of devices gathering data at the edge of networks is rich. It includes sensors, industrial machines, medical devices, video cameras, actuators, RFID gear, asset monitors, self-order kiosks, PCs, tablets, phones, and a range of smart devices like thermostats, refrigerators, and food preparation systems. Key variables include the number and type of devices, distance from the branch or campus network, amount of data generated, battery life, and security requirements. Connectivity With the variety of devices comes a diversity of connectivity needs, from real-time high-bandwidth control of factory robots to moisture sensors in farm fields that need to upload small amounts of data at wide intervals. Enterprises need to decide the most appropriate links the types of edge devices they have deployed. Networking options include Wi-Fi, Bluetooth, Zigbee, private or public 4/5G, cellular data, and various kinds of wired connectivity, among others. Key variables include the mobility of the devices, cost of the network service, latency requirements, and whether the data flow is one-way or bi-directional. The gateway Edge gateways are aggregation points for data generated by dispersed devices, but they may also filter, process and store data as well as apply security policies. Key variables include the type of device used, which depends on the needs of the system as a whole. The gateway can be a server, a networking device or a specialized appliance. Compute power This is what processes the data and sends it to the centralized cloud location or data center via the wide area network. Edge-computing resources can be deployed at the branch or other on-premises location or can leverage edge-compute resources as a service from cloud or other service providers. All four of these elements must work in unison to provide an edge-computing architecture that works, and to achieve that IT organizations must answer a complex set of questions about of architecture, performance, security/compliance, application software, supplier options and cost. Benefits and challenges The cost and complexity of deployment have slowed early development of edge-computing systems, but the benefits can be significant. Edge computing is driven by the need for very low latency. The current architecture in which data is sent to a centralized data center or public cloud and then returned to the edge location for action creates latency of a second or more. Edge computing can deliver predictable millisecond latency, which is critical for manufacturing, health, and public-safety applications. Edge computing can provide real-time data analysis and vastly reduce the amount and frequency of data required to be sent to distant, centralized locations. It can also provide for high availability and redundancy, and offer security/compliance by keeping sensitive data in local locations and not exposing it to the Internet. Options for Edge Computing IT organizations have several options as they consider how to build out their edge computing capabilities. As edge computing applications are highly specific, such as IoT deployments, IT teams need to consider the ability to customize edge computing architecture to their needs. Many early deployments have been made by sophisticated IT organizations with the skills and resources to deploy edge computing on their own. By doing it themselves, these IT organizations gain the advantage of creating architectures that meet their specific performance, latency and security metrics. But these DIY deployments can be complex and thus time consuming and costly. Many IT departments are leveraging innovative start-ups for technology and larger IT firms for service and support to help jump start their edge-computing deployments. The challenge is to adapt broad edge-computing solutions, especially those from large vendors, to the unique application requirements of individual deployments. Edge computing can also be delivered as a service from cloud providers, telcos, and other service providers. These organizations are slowly building out low-latency computing services in or near the largest cities in the largest countries. Organizations within this limited footprint need to consider latency guarantees, cost and the ability to integrate with the specific service offerings. Recommendations Edge Computing is in its early stages of technological development. Each potential edge-computing application has significantly different requirements, including hardware, software, networking and costs. Architectures should be designed with flexibility and adaptability to meet changing business requirements. Finally, IT organizations should move cautiously with initial deployments, starting with pilot projects that have a clear business case. Related content news Cisco patches actively exploited zero-day flaw in Nexus switches The moderate-severity vulnerability has been observed being exploited in the wild by Chinese APT Velvet Ant. By Lucian Constantin Jul 02, 2024 1 min Network Switches Network Security news Nokia to buy optical networker Infinera for $2.3 billion Customers struggling with managing systems able to handle the scale and power needs of soaring generative AI and cloud operations is fueling the deal. By Evan Schuman Jul 02, 2024 4 mins Mergers and Acquisitions Networking news French antitrust charges threaten Nvidia amid AI chip market surge Enforcement of charges could significantly impact global AI markets and customers, prompting operational changes. By Prasanth Aby Thomas Jul 02, 2024 3 mins Technology Industry GPUs Cloud Computing news Lenovo adds new AI solutions, expands Neptune cooling range to enable heat reuse Lenovo’s updated liquid cooling addresses the heat generated by data centers running AI workloads, while new services help enterprises get started with AI. By Lynn Greiner Jul 02, 2024 4 mins Cooling Systems Generative AI Data Center PODCASTS VIDEOS RESOURCES EVENTS NEWSLETTERS Newsletter Promo Module Test Description for newsletter promo module. Please enter a valid email address Subscribe