In 2018, the world's data was expected to grow 61 percent to 175
zettabytes by 2025. According to research firm Gartner, around 10 percent of enterprise-generated data is created and processed outside a traditional centralized data center or cloud. By 2025, the firm predicts that this figure will reach 75 percent. The increase in
IoT devices at the edge of the network is producing a massive amount of data — storing and using all that data in cloud
data centers pushes network bandwidth requirements to the limit. Despite the improvements in
network technology, data centers cannot guarantee acceptable transfer rates and response times, which often is a critical requirement for many applications. Furthermore, devices at the edge constantly consume data coming from the cloud, forcing companies to decentralize data storage and service provisioning, leveraging physical proximity to the end user. In a similar way, the aim of edge computing is to move the computation away from data centers towards the edge of the network, exploiting
smart objects,
mobile phones, or
network gateways to perform tasks and provide services on behalf of the cloud. By moving
services to the edge, it is possible to provide content
caching, service delivery, persistent data storage, and IoT management resulting in better response times and transfer rates. At the same time, distributing the logic to different network nodes introduces new issues and challenges.
Privacy and security The distributed nature of this paradigm introduces a shift in security schemes used in
cloud computing. In edge computing, data may travel between different distributed nodes connected via the internet, and thus requires special encryption mechanisms independent of the cloud. This approach minimizes latency, reduces bandwidth consumption, and enhances real-time responsiveness for applications. Edge nodes may also be resource-constrained devices, limiting the choice in terms of security methods. Moreover, a shift from centralized top-down infrastructure to a decentralized trust model is required. On the other hand, by keeping and processing data at the edge, it is possible to increase privacy by minimizing the transmission of sensitive information to the cloud. Furthermore, the ownership of collected data shifts from service providers to end-users.
Scalability Scalability in a distributed network must face different issues. First, it must take into account the heterogeneity of the devices, having different performance and energy constraints, the highly dynamic condition, and the reliability of the connections compared to more robust infrastructure of cloud data centers. Moreover, security requirements may introduce further latency in the communication between nodes, which may slow down the scaling process.
Reliability Management of
failovers is crucial in order to keep a service alive. If a single node goes down and is unreachable, users should still be able to access a service without interruptions. Moreover, edge computing systems must provide actions to recover from a failure and alert the user about the incident. To this aim, each device must maintain the
network topology of the entire distributed system, so that detection of errors and recovery become easily applicable. Other factors that may influence this aspect are the connection technologies in use, which may provide different levels of reliability, and the accuracy of the data produced at the edge that could be unreliable due to particular environment conditions. anything health or human / public safety relevant, or involving human perception such as facial recognition, which typically takes a human between 370-620 ms to perform. Edge computing is more likely to be able to mimic the same perception
speed as humans, which is useful in applications such as
augmented reality, where the headset should preferably recognize who a person is at the same time as the wearer does.
Efficiency Due to the nearness of the analytical resources to the end users, sophisticated analytical tools and
artificial intelligence tools can run on the edge of the system. This placement at the edge helps to increase operational efficiency and is responsible for many advantages to the system. In distributed AI systems on the edge, data compression is increasingly recognized as a foundational design layer to mitigate bandwidth constraints caused by the exchange of large models and high-resolution sensor streams. Additionally, the usage of edge computing as an intermediate stage between client devices and the wider internet results in efficiency savings that can be demonstrated in the following example: A client device requires computationally intensive processing on video files to be performed on external servers. By using servers located on a local edge network to perform those computations, the video files only need to be transmitted in the local network. Avoiding transmission over the internet results in significant bandwidth savings and therefore increases efficiency. Another example is
voice recognition. If the recognition is performed locally, it is possible to send the recognized text to the cloud rather than audio recordings, significantly reducing the amount of required bandwidth. == Applications ==