1. Introduction
The telecommunication network carriers have started to move toward Network Function Virtualiztion (NFV) and Software-Defined Networks (SDN) technologies. The architectural design of NFV is meant to deliver new services more efficiently and scale the network effectively [
1]. The NFV and 5G/6G network work together for future generation networks establishment by virtualizing the 5G/6G network infrastructure for effective management and optimization [
2]. When 5G/6G-based network slicing is enabled for the establishment of a distributed cloud, that helps in creating flexible and programmable networks as per users’ future demands. Applying the distributed cloud to an NFV technology enables 5G/6G virtualization, which separates a physical network into multiple virtual networks, which would support various telecom based packet–core radio access networks, leading to greater benefits in the resilience, fault-tolerance, and scalability of the network [
3]. The Border Gateway Protocol (BGP) network routing mishap creates a social impact in articulating this kind of crypto-sensitive problem that has a symmetric role in publishing this future generation telecommunication network research article in the Symmetry section of the
Journal of Computer and Engineering Science that is best available for relevant areas of research and community forums. The research article is also best suited for symmetry, as it contains volumes on mathematical models, simulation with experimental outcomes, conclusions and results that best demonstrate the novelty of a scientific research article.
Symmetry Basel played a major part in providing research publications in the computer networks section that range from first generation networks to sixth generation networks. There is a lot happening in the field of telecommunication network establishment and deployment. The symmetry extended its future plan of automated network management with the help of Artificial Intelligence (AI) and Machine Learning (ML) techniques. The future work of the proposed system mentioned in this article will inherit the AI/ML techniques such as reinforcement learning, deep queue learning, and neural networks for zero touch network automation.
The future 5G and 6G networks tend to offer a great flexibility to the network operators with regard to allocation of speed, coverage, and the capacity of various use cases [
4]. In real time, the operators have to program their networks for the introduction of new services rather than manual configuration [
5]. This adaptability is a key factor for the end users where there is variation in demands from providing a consistent state transition for storage-based distributed systems [
6]. Globally, the data traffic has grown at a faster rate with the rapid deployment of 5G/6G networks. Although the service providers have already started to virtualize their packet core networks, the due benefits of 5G/6G is reaped with the horizontal breaking of NFV silos in a telco cloud to achieve the efficiency in deploying new services with mobile edge computing and network slicing features for end-to-end orchestration of services for 5G mobile system [
7].
NFV decouples Network Functions (NFs) from generic hardware appliances and runs them as software Virtual Network Functions (VNFs) in Virtual Machines (VMs) [
8]. The NFs are the building blocks positioned inside a network infrastructure that include interfaces. It enables virtualization, and VNFs are executed in Commercial-Off-The-Shelf (COTS) hardware servers within the data centers [
9]. The VNFs workload management has created a great buzz in the field of telecommunication for various complex network functions with the availability of necessary optimization tools and frameworks [
10]. The prior work discusses various VNF Manager (VNFM) placement techniques with an elastic search model for minimizing the Capital Expenditure (CapEx) and Operational Expenditure (OpEx) cost. A mixed integer linear programming model is used for solving the VNF service chain composition problem [
11]. This problem is addressed using a non co-operative game theory model for the privacy preservation of VNF management in polynomial time and security of computing environments [
12]. The game theory model focuses on resource distribution among the interaction between users demand and service availability. The Markov chain optimization method is used for the dynamic placement of VNF [
13]. The focus is on VNFM placement that could minimize CapEx and OpEx without compromising on the performance requirements.
The advancement in scaling up faster central processing units and methods to achieve an efficient packet processing in software has spurred NFV [
14]. During NFV deployment, an operations network sets up an NFV Infrastructure (NFVI), which best describes the hardware and software components on which virtual components are built. The VNFM is responsible for the Life Cycle Management (LCM) of virtual instances [
15]. LCM refers a set of functions required to instantiate, maintain, and terminate a network service or a VNF. The purpose of VNFM is to collect the personal information of a virtualized resource from NFVI and VNF indicators from the Element Managers (EMs) or VNF instances [
16]. The EM manages the VNFs in the NFV framework as represented in
Figure 1. The problem in optimal placement of the VNFM component in a distributed environment is new, and it is described as a VNFM Placement Problem (VNFM-PP).
The telecom network management system-based VNFM placement is considered for several heuristic approaches in a distributed platform, and its evaluation gave better results. The European Telecommunication Standards Institute is a consortium of service providers meant for developing a suitable cloud-based platform (i.e., openstack cloud computing platform) to host VNFs (i.e., firewalls, routers, switches), and operators are trying to integrate NFV into openstack [
17]. Development and operation is meant for automating the orchestration services and built tools for managing resources using Application Programming Interface (API). This results in a tremendous uplift of moving the infrastructure to the cloud data centers’. The NFVI Point-of-Presence (PoP) controls the distribution of software instances that result in an increase of NFVI performance with reduction of VNFM placement CapEx and OpEx. To deliver an optimized VNFM placement, the Virtual Topology System (VTS) increases resource prediction for Virtual Extensible Local Area Networks (VXLAN) overlay, provisioning a switch of the networking layer inside the data centers’ programmable fabric [
18]. The progression of VNFs in the cloud-optimized environment brings in attractive automation capabilities of cloud computing and Internet of Things platforms and sets a new standard within the telecommunications industry [
19]. The on-demand scaling of VNF helps with the accommodation of hardware appliances whenever there is an increase in load. The major contributions of this article are as follows:
The existing algorithms such as ILP, Greedy and TSA have not considered VNF characteristics based on the total volume of resources and have a major problem of poor efficiency in the VNFM placement. This paper proposes a node EM calculation method for assigning VNF with periodic information about resources and characteristics before assigning VNF.
The formulation of the problem is termed as VNFM-PP, and it proposed an OEM algorithm for dynamic VNFM placement decisions. The placement uses Monte Carlo estimation for the proposed OEM algorithm instead of calculating the theoretical expectation as in existing static approaches.
The proposed OEM algorithm supports a Border Gateway Protocol–Ethernet Virtual Private Network (BGP-EVPN) control plane system for provisioning VXLAN overlays for placing the VNFM component.
The rest of the article is organized as follows.
Section 2 discusses the prior work.
Section 3 presents the problem statement on telecommunication-based network management based on infrastructure VNFM placement.
Section 4 presents the proposed OEM algorithm for VNFM placement.
Section 5 presents the performance evaluation of the proposed algorithm.
Section 6 summarizes and concludes the paper.
2. Related Work
In the NFV paradigm, traditional middleboxes are managed as single modules of software programmed to perform the role of a particular VNF [
20]. This transformation helps to adapt and capitalize on specific goals such as speed and efficiency enhancement. The virtualization is meant to transform the traditional computing, making it efficient in managing workloads [
21]. During network traffic, in applying the policy or providing network services to traffic, it needs to undergo a particular sequence of VNFs called a service chain [
22]. This transformation is possible only through the consolidation of different VNFs in standard general-purpose computers (servers, storage devices, compute), which can be located either in data centers or network nodes [
23]. In order to improve resource utilization, two main factors are considered: the initial thing is to consider the Fundamental Resource Overheads (FROs) incurred by instantiating VNFs in servers [
24]. The existing work focuses on the VNF placement problem for mapping service function chaining of resources in cloud networks.
The FROs are considered, and past researchers mathematically formulated the resource optimization and guaranteed delay in the VNF placement problem, which is said to be NP-hard. To solve this problem, they devise a polynomial-time approach called Two-Phase Optimization Solution (TPOS) [
25]. The releasing of redundant VNF instances with varying network load could save more server resources and improves resource utilization. A detailed analysis and evaluation of TPOS was conducted; for large-scale problems, it guarantees the delay requirements and outperforms benchmarks in terms of server resource consumption, activation and utilization. The optimal location determination of VNF and VM was determined, and they also focused on the VPS-CD problem as an MILP and proposed a cost-efficient VPS Scheme (CE-VPS) to minimize the cost of VNF placement and scheduling to deploy VNFs in a public cloud [
26]. The MILP minimizes the cost computationally for large networks with the development of an efficient heuristic algorithm.
The ILP based Power-aware and Delay-constrained joint VNF placement and Routing (PD-VDR) problem is formulated [
27]. The problem is addressed based on the consideration of VNF processing time, which changes dynamically based on network traffic. A Holu efficient heuristic algorithm is proposed for solving a PD-VDR problem using heuristic routing and employed a Delay Constrained Least Cost (DCLC) shortest path algorithm [
28]. The problem of dynamic VNFM placement in large-scale distributed NFV system is discussed. A system is proposed with a generic ILP formulation that determines the number of VNFM placement at minimum [
29]. They also proposed a TSA-based meta-heuristic technique to solve the placement problem of VNFM in both static and dynamic environments [
30]. The focus is on the VNF placement problem in data center networks with guaranteed resource optimization and resource availability [
31]. They proposed a novel Joint Path-VNF (JPV) backup model with affinity iterations on algorithms for the reducing the link consumption, and performance evaluation showed that JPV improves better with network availability [
32].
The problem of service chain placement for the VNFs within the cloud is discussed for mixed ILP for VNF allocation and network management in cloud data centers. The proposed system comprises two new resource allocation algorithms based on genetic programming for the initial placement and scaling of VNFs to support traffic changes. The evaluation is carried out in data centers, with an efficient utilization for both computing resources and network configurations being managed concurrently in an openflow infrastructure. The problem of VNF placement and scheduling in a radio access network domain is discussed, as they proposed a slice scheduling mechanism with performance isolation between different slices, and numerical simulation shows that the proposed heuristic algorithm gave an optimal solution. In most of these related works, the coupling between different VNFM placement interface objectives has not been addressed.
Few of the existing works deal with NFV architecture in consideration of a VNF placement problem with multiple objectives such as network resource cost and license cost. However, they do not provide any effective solution to optimize the cost. The existing scheme has not considered VNF characteristics based on the total volume of resources with poor efficiency during VNF placement. The VNF-specific characteristics indicate that the usage of more CPU, memory, and disk resources leads to inefficient placement of VNF. The existing algorithms such as greedy, TSA, ILP, dynamic programming, and Djikstars’ are used for the optimal placement of VNFM. The proposed scheme is concerned with optimized methods for VNFM placement without wasting resources. The VNF is used to predict demands for various VNF resources, analyzing the correlations among those resources such as CPU, memory, and disk.
The optimized VNF placement performance can be increased using ILP as the number of nodes, VNFs to assign, complexity and computation increases with complexity. The demerit of the ILP technique is its extendability; it is not efficient as the size of the problem expands and cannot deal with algorithms that run exponentially. The consequent demerit of increasing complexity is that the performance deteriorates based on optimization algorithms used in computing environments. The issues in addressing the placement of VNFM rely either in data centers or servers inside the network. The issues are further discussed in two different ways: Firstly, the minimal placement of VNFs increases the forwarding cost and leads to traffic churns. Secondly, a maximum number of VNFs placement with a decrease in forwarding cost brings stability to the network. The proposed algorithm follows methods with optimization techniques that increases efficiency with VNFM placement. This paper introduces an Optimized Element Manager (OEM) for VNFM placement as a core technique in NFVI of the NFV framework.
The aim and key outcomes of this article are as follows:
To transform the way network operators design, manage, and deploy their network infrastructure. Earlier, it required low latency for signal processing, wherein now, users require higher throughput for data traffic processing.
The software instances of network functions and services are implemented in the form of VNFs. The placement of VNFM is managed according to their demand in the network, which dynamically changes with time.
The cloud service providers should route their application flows during VNF placement, i.e., through a sequential ordering of VNFs specified in service function chaining.
The distributed infrastructure helps to materialize degradation in performance and type of placement instantiated in line to a set of instances for large-scale environments.
The chaining approach determines the path traversal of VNFs that are dynamically created, embraces a significant decrease in CapEx and OpEx for telecom operators, and makes necessary alterations in the way network functions are developed and deployed.
3. Problem Statement
In large-scale distributed environments, during VNF instance deployment, the VNFM maintains VNF instances without compromising its management function performance. However, VNFs are initiated on a demand basis and scaled elastically for effective maintenance. The number of VNFM adoption is based on VNF instances deployment for optimal performance [
15]. A telecom network-based VNFM management of the NFV ecosystem challenge is addressed for enhancing the NFVI dynamism and efficiency. This function faces challenges such as delays and variations in VNFs workload. The significance in VNFM placement affects the overall system performance and operational cost. The various cost considerations are listed in a summary of key notations, as shown in
Table 1:
In a telecom-based NFV ecosystem, the VNFM is a key component. The placement of VNFM becomes critical, as cost and performance optimization are needed for NFV systems, and so, the placement issue is labeled as VNFM-PP. Hence, we determine the placement function
such that
where
P(
c) is a criterion function [
25]. The computational placement has been represented as placement
P(
c) =
P1,
P2…
Pk; let
V(
Cn,
Cq) denote a set of virtual computes, and the corresponding interconnection between these computes happens through overlays, which is expressed as
O(
Cn,
Cq). The hardware overlay provisioning that happens in L3 fabric using Spine-Leaf (SL) topology results in global error placement can be expressed as:
3.1. Capital Expenditure Cost (CapEx)
Four different computational cost evaluation parameters have been considered for our CapEx, as shown below.
3.1.1. PoPs Site Components Cost
The PoPs site components cost is calculated by representing the number of computing resources allocated for a data center and by calculating the cost of its size for every individual VM, as shown in Equation (
1).
3.1.2. SDN Components Cost
The overall cost for performing an SDN computation is mentioned in Equation (
2). It is incurred with the number of SDN controllers and the cost required for an individual SDN controller.
3.1.3. Installation Cost
The installation cost includes physical hardware resources required for performing compute, storage and network operations in an infrastructure medium for setting up networking services. The cost of software development and first-time installation bind together to obtain an average cost, which gives the total cost spent on installation as shown in Equation (
3).
3.2. Operational Expenditure Cost (OpEx)
Four different operational cost component evaluating parameters have been considered to define our OpEx, which are described below.
3.2.1. Continuous Cost of Infrastructure
The continuous cost of infrastructure is exemplified in managing the infrastructural components such as compute, storage and network by inhibiting several cost parameters such as the controller, network devices and rackspace as represented in Equation (
4).
3.2.2. Cost of Maintenance and Repair
The cost of maintenance and repair characterizes several parameters for hardware failures as shown in Equation (
5) that includes distance, time, wage, software and hardware replacement and time taken to fix these parameters.
3.2.3. Cost of Service Provisioning
The service provisioning cost is exemplified by calculating the time taken for a service to be reconfigured in a particular year and the service connection establishment time being represented in hours as shown in Equation (
6).
3.2.4. Cost of Service Management
The cost of service management characterizes the configuration and management time in Equation (
7). It is being represented in hours required for establishment of the service and managing of resources.
The sum of the weighted cost of
CapEx and
OpEx is expressed in Equation (
8), where
Nc is the number of PoPs per
. The total number of PoPs site cost per
, where
CapEx and
OpEx are the cost per data center, and
i is the interest rate defined for
y years:
The software-based VNF instances spun across geographically distributed areas using NFVI-PoPs that result in an increase of NFVI performance and thereby reducing the CapEx and OpEx costs during VNFM placement. The EM helps in the functional management of VNFs using proprietary interfaces. The placement function is not handled by a single component in an NFV framework. The communication between functional blocks (NFVI, VNFM, VIM) happens through VNF instances. The full architectural framework depicts both the functional blocks and the main reference points between these blocks. Hence, for the optimal VNFM placement, VTS increases prediction of the resource requirements of each component for VXLAN overlay management and provisioning system for data center networks. We proposed an algorithm termed as OEM, which will be foreseen as a contending model in comparison to Greedy, ILP and TSA. Existing algorithms have the potential to resolve the scaling of VNFs based on traffic changes. It is traditionally used to optimize VNFM placement based on VNF allocation and network management in cloud data centers.
5. Performance Evaluation
This section evaluates the OEM algorithm performance. The result obtained after evaluation is compared to first fit greedy algorithms such as TSA and ILP. NFV performance management involves the process of monitoring the NFV performance. A critical challenge in monitoring is to ensure that network performance remains as good as a purpose-built hardware. Two of the most common performance management tools are network performance management and application performance management. An initial obstacle occurs with a visibility gap during NFV deployment as the network management tool runs operations in a periodic manner, where the tools provide visibility to all of these interdependencies. The next challenge is in adapting NFV to the network performance tools that start interacting with the digital infrastructure and orchestration services. The carrier network design choices have a significant impact on the network performance and cost. Critical design choices are categorized into four areas based on the functional design of NFVI architecture; it can be listed as the packet-forwarding method, data-center design, hardware choices, and software efficiency.
The proposed OEM algorithm is implemented with a Monte Carlo estimation model and solved using CPLEX Optimization Studio V12.6.3. These greedy algorithms assess the VNF dynamic sequence on VNFM placement operations. The CPLEX is constructed as a baseline in order for the evaluation of the proposed OEM algorithm for large-scale deployment. The TSA and ILP algorithms iterate on a set of VNF and its instances. During the initial iteration process, the VNF instance is assigned to first VNFM with its adequate capacity requirements. The NFVI-PoPs changes the constraints in a capacity that satisfies with VNF-PP constraints.
5.1. Simulation Tool
The simulation for evaluating OEM is carried out in connecting to the servers of Cloudlab.us, where eight servers are connected together as a chain in a network service function testbed. The data plane development kit is preliminary for performing any kind of traffic steering operation. OpenNetVM Speed Tester NF is a tool which evaluates realistic traffic loads in the management layer with machines of eight 10 Gbps NIC ports. Our goal in this setup is to send traffic from nodes 2 to 10. The bridging between the initiator node1 to executor node11 happens using NIC ports. So, the intermediate nodes from 2 to 10 will act as transparent middle boxes, which is useful for chaining servers. When the web traffic is initiated from node1, a chain of activities happens, which helps in traffic steering activities until before the traffic reaches the descendant node12. Speed tester NFs can automatically balance the traffic across two NFs, which thereby results in the consistent delivery of network services and improves the performance of up to 68 Gbps in comparison to major network speed testers such as Iperf, PRTG network monitor and Wireshark.
5.2. OpenStack Testbed Experimental Setup
Our testbed uses the private cloud that runs on the OpenStack opensource cloud computing platform. The deployment of VMs is carried out using Red Hat Enterprise Linux 7, and the whole experiment follows OpenStack Ocata release. Initially, we performed VLAN (switch) configuration by configuring VLAN elements (VLAN_ID, VLAN_Name) and later assigned IP to VLAN. For our experiment, we used rackserver in which rack 1 is meant for nova compute services. Rack 2 consists of neutron networking controller services. Finally, rack 3 is used for storage services (object swift, block cinder). It uses OVS as a hypervisor agent which acts as a software switch for managing the VMs. A summary list of simulation parameters and values are displayed in
Table 2. The physical interconnection between OpenStack use case happens via a hardware switch that has management, self-service and provider networks. The OpenStack experiment results in minimal production of overlay traffic for self-service networks that best route the traffic via the management network instead of a dedicated network.
5.3. Performance Analysis
The optimality is achieved by calculating the total incurred cost in our proposed OEM algorithm with comparison of the two existing ILP and TSA algorithms. The performance is measured based on cost metrics for obtaining the total incurred cost within 10 time slots, as shown in
Figure 6. In
Figure 6a, the ILP algorithm tends to consolidate the node utilization of resources and reduce the active nodes, as the operational cost values tend to be lowest. In consideration from time slots 7 to 9, the OEM operational cost is lower in comparison to ILP as the cost of ILP increases. In
Figure 6b, the analysis of traffic cost clearly shows that TSA incurs the lowest traffic cost and it involves lot of active nodes. Although TSA looks not so promising at first comparison, it obtains better results with traffic cost comparison. In
Figure 6c, the capital cost comparison proves that our proposed OEM algorithm obtains the lowest CapEx with a good marginal ratio with the changed workload compared with ILP and TSA. In
Figure 6d, in 10 time slots, OEM reduces 21.3% and 10.4% of the total cost been compared with ILP and TSA algorithms, respectively. Lastly, OEM demonstrated the optimal way of procuring the network traffic cost, CapEx cost and OpEx cost. The prefetched result signifies that our proposed algorithm is efficient in element management with NFV-based VNFM placement with better optimality in comparison to ILP and TSA algorithms.
5.4. Algorithmic Performance Analysis
The test bed for our experimental setup is composed of a server with an Intel Core Xeon processor with 128 GB variant memory. The total execution time of the OEM algorithm in comparison to the baseline CPLEX optimizer is represented in
Table 3. The complexity of the proposed OEM algorithm time complexity is O(nlogn), as it involves Monte Carlo estimation as the placement problem is said to be NP-Hard. A better solution for the VNFM placement problem is carried out as VNFM can either proactively or reactively respond to VNF status updates in the network and perform management operations in comparison to other optimization algorithms such as ILP and TSA based on the delay in the performance of the modeling network. The performance evaluation is measured by evaluating the quality of the solution derived from comparison with the existing Greedy, ILP and TSA algorithms based on CPLEX. The OEM algorithm analysis with
= 1 and
= 20 clearly shows that the OpEx cost corresponds to optimality with an average that remains smaller than 8% for 8, 16 and 24 PoPs. The result shows that our proposed OEM algorithm is faster for many of the dynamic VNF class parameters where the total execution time is greater than one second.
5.5. Bandwidth Optimization in EMs
The optimization in the placement of VNFM is addressed in our proposed OEM algorithm with comparison to existing ILP, and TSA algorithms. The performance is measured based on the criterion function, as it includes two metrics: the compute resource cost and LCM cost. The bandwidth consumption of LCM is bigger than the
value, and the placement is favored in minimizing the LCM cost, as represented in
Table 4.
The uniformity in bandwidth cost over communication links in NFVI-PoPs translates the additional placement of VNFM on NFVI for managing VNF instances. This placement method reduces the overall traffic on the communication links and will be circulated to NFVI-PoP without incurring any cost. The evaluation in
Figure 7 describes the number of VNFM placed on NFVI, in relation to the number of VNF instances for 8, 16 and 24 NFVI-PoPs with consideration to
= 1 and
= 20. There is a steady rise in the number of VNFM, if
= 20 is being compared to
= 1. Digging deeper, we observed that the additional VNFMs allocated are of compute C1 type. In our evaluation, the VNF instances with compute C1 utilizes more bandwidth in comparison to C2 VNF instances. Overall, our OEM algorithm significantly outperforms Greedy, ILP, and TSA with average gaps that are smaller than 8%.
5.6. Distributed Large-Scale Deployment
Figure 8 represents the total operational cost for the OEM, ILP, TSA and Greedy solutions in consideration of both VNFM-based generic and specific architecture options for
= 1 and
= 20, respectively. In the generic approach, the two distinct sets of VNFMs are placed for managing compute classes C1 and C2. The utilization rate is relatively small, as the compute classes for the number of VNF instances leave many VNFMs with unutilized capacity, leading to resource wastage. Whenever a generic VNFM is adopted, the VNF instances are managed by a single set of VNFMs. This indeed results in maximizing the VNFMs capacity utilization and placement of fewer VNFMs. For the dynamic VNFM placement, the over-provisioning of the VNF instances problem is solved with 225 VNF instances for every VNF class. The figure reports the PoPs locations as per the VNFM architectural options. However, we use the same number of VNF instances to be placed on NFVI for 8, 16, and 24 PoPs. The differences between these PoPs are already experimented with results of average execution time as shown in
Table 3. The dynamic placement involves the sum of VNFM compute resource cost and LCM cost. The performance gain in dynamic VNFM placement in a distributed environment is calculated based on the comparison of numerical results between OEM and comparative algorithms with dynamic placement scenarios for 8, 16, and 24 PoPs. The numerical results confirm that the quality of the solutions obtained from the OEM algorithm is significantly higher than other algorithms by 36.34% to 164.8%. The superiority of dynamic VNFM placement emerges by its ability to adapt to the system based on its workloads.