CN114173379B - A multi-user computing offloading method based on 5G private network splitter - Google Patents
A multi-user computing offloading method based on 5G private network splitter Download PDFInfo
- Publication number
- CN114173379B CN114173379B CN202111519999.7A CN202111519999A CN114173379B CN 114173379 B CN114173379 B CN 114173379B CN 202111519999 A CN202111519999 A CN 202111519999A CN 114173379 B CN114173379 B CN 114173379B
- Authority
- CN
- China
- Prior art keywords
- objective function
- computing
- service node
- resource
- unloading
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 24
- 238000013468 resource allocation Methods 0.000 claims abstract description 33
- 239000011159 matrix material Substances 0.000 claims description 34
- 239000013598 vector Substances 0.000 claims description 33
- 230000001186 cumulative effect Effects 0.000 claims description 21
- 238000005457 optimization Methods 0.000 claims description 16
- 230000008569 process Effects 0.000 claims description 4
- 238000012545 processing Methods 0.000 abstract description 6
- 230000005540 biological transmission Effects 0.000 abstract description 5
- 230000008901 benefit Effects 0.000 abstract description 4
- 238000004891 communication Methods 0.000 abstract description 2
- 230000006870 function Effects 0.000 description 51
- XNWFRZJHXBZDAG-UHFFFAOYSA-N 2-METHOXYETHANOL Chemical compound COCCO XNWFRZJHXBZDAG-UHFFFAOYSA-N 0.000 description 7
- 238000004364 calculation method Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 2
- 238000007781 pre-processing Methods 0.000 description 2
- 238000012216 screening Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000012417 linear regression Methods 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W28/00—Network traffic management; Network resource management
- H04W28/02—Traffic management, e.g. flow control or congestion control
- H04W28/08—Load balancing or load distribution
- H04W28/09—Management thereof
- H04W28/0925—Management thereof using policies
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W28/00—Network traffic management; Network resource management
- H04W28/02—Traffic management, e.g. flow control or congestion control
- H04W28/08—Load balancing or load distribution
- H04W28/09—Management thereof
- H04W28/0958—Management thereof based on metrics or performance parameters
- H04W28/0967—Quality of Service [QoS] parameters
- H04W28/0975—Quality of Service [QoS] parameters for reducing delays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W28/00—Network traffic management; Network resource management
- H04W28/02—Traffic management, e.g. flow control or congestion control
- H04W28/08—Load balancing or load distribution
- H04W28/09—Management thereof
- H04W28/0958—Management thereof based on metrics or performance parameters
- H04W28/0967—Quality of Service [QoS] parameters
- H04W28/0983—Quality of Service [QoS] parameters for optimizing bandwidth or throughput
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W84/00—Network topologies
- H04W84/02—Hierarchically pre-organised networks, e.g. paging networks, cellular networks, WLAN [Wireless Local Area Network] or WLL [Wireless Local Loop]
- H04W84/04—Large scale networks; Deep hierarchical networks
- H04W84/08—Trunked mobile radio systems
Landscapes
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Quality & Reliability (AREA)
- Mobile Radio Communication Systems (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
The invention belongs to the field of wireless communication, and particularly relates to a multi-user computing and unloading method based on a 5G private network shunt; the method comprises the following steps: constructing a multi-user computing unloading frame based on a 5G private network shunt; constructing an objective function of the combined unloading total time delay and the resource allocation balance degree according to the multi-user computing unloading framework; solving an objective function of the combined unloading total time delay and the resource allocation balance degree to obtain an objective function optimal solution; determining an unloading decision by a user according to the optimal solution of the objective function; the user performs task unloading according to the determined unloading decision; the invention has good expansibility and safety, effectively utilizes the bandwidth and throughput of each UE device, enhances the network data processing capability, and improves the flexibility and usability of the network; the method can effectively reduce the average unloading time delay and the average transmission time delay of multiple users, and balance the workload of each mobile edge terminal device or each edge server, thereby having good economic benefit.
Description
Technical Field
The invention belongs to the field of wireless communication, and particularly relates to a multi-user computing and unloading method based on a 5G private network shunt.
Background
In recent years, mobile user devices (mobile phones, tablet computers, etc.) are becoming an important tool for learning, entertainment, social networking and business development, with a rapid increase in mobile data traffic, and this increase is expected to continue in the future. As mobile data traffic increases rapidly, the amount of computation and data exchange between mobile user equipment and the cloud is also increasing dramatically. To address this situation, mobile edge computing (Mobile Edge Computing, MEC) technology has been introduced in the industry to supplement and enrich cloud computing. Because the mobile edge computing technology directly deploys a mobile edge computing server (Mobile Edge Computing Server, MECS) on the base station, the mobile user equipment needing computing offloading does not have to select a remote cloud server, but can adopt a nearby policy, namely offloading computing load to the edge server nearby, or directly offloading the computing load to a terminal node with other idle resource capability in the local, so that the mobile user can accept a short-range cloud computing service in the coverage area of a wireless access network, and the user enjoys high-quality network experience, namely, the mobile edge computing extends computing resources and storage resources from a core wide area network to an edge network.
The problems encountered by mobile user equipment currently offloading local computing tasks to MECS are: the data transfer of the offloading process results in a non-negligible delay between the mobile device and the MECS, and it remains difficult to co-process how the multi-user different resource requirements are handled in case the multi-mobile user needs to compute the offloading. In the related art, most methods predict the processing time of each task on each candidate MECS using linear regression, while offloading the mobile user's computing tasks to the appropriate MECS based on the previously observed states of the MECS. However, when the number of mobile users needing to be calculated and offloaded is large and there are a plurality of MECS that can accept the offloaded service, how to solve the problems of multi-user total delay minimization and multi-MECS load balancing at the same time still lacks a practical solution at present.
Disclosure of Invention
In view of this, the present invention provides a multi-user computing and unloading method based on a 5G private network splitter, which includes: .
S1: constructing a multi-user computing unloading frame based on a 5G private network shunt; the 5G private network splitter comprises user plane function equipment UPF and a small mobile edge computing platform miniMEP;
s2: constructing an objective function of the combined unloading total time delay and the resource allocation balance degree according to the multi-user computing unloading framework;
s3: solving an objective function of the combined unloading total time delay and the resource allocation balance degree to obtain an objective function optimal solution; determining an unloading decision by a user according to the optimal solution of the objective function;
s4: and the user performs task unloading according to the determined unloading decision.
Preferably, constructing the multi-user computing offload framework based on the 5G private network splitter comprises constructing a computing offload total delay framework and constructing a computing resource allocation balance framework.
Further, constructing a computing offload total delay framework includes:
miniMEP is used as a mobile edge computing cooperative server, and the resource available state of each idle service node is recorded; the available states of the computing resources and the storage resources of the service node j are represented by a joint resource pair (p j,cj), wherein p j represents the computing resources of the service node, and c j represents the storage resources of the service node; defining an intensive computing task of the mobile user k as T k, and representing the requirement of the mobile user k on resources by using a joint resource pair (z k,ek), wherein the computing resource provided by a service node for the task T k is z k, and the storage resource provided by the service node for the task T k is e k;
every t time, mobile user k sends its own joint resource demand pair (z k,ek) to miniMEP;
miniMEP calculates the total time delay for offloading each mobile user based on the remaining computing resources of each service node, the remaining storage resources, and the combined resource demand pair for the mobile user (z k,ek).
Further, the total time delay formula of the unloading is calculated as follows:
Wherein f 1 (x, y) represents an objective function of total time delay of unloading, K represents the total number of users, M represents the total number of service nodes, d kj represents a computing resource allowance of the mobile edge computing collaboration server, z i represents a computing resource applied by a user, e i represents a storage resource applied by a user, J represents a service node set, and x kj represents an unloading judgment parameter.
Further, the computing resource allocation balance framework includes:
Defining a concentration degree parameter u for representing the occupation of the resources of each service node by the unloading decision, and defining a vector A= (a 1,a2,a3...aM) for representing the cumulative load of the resources of each five-Fu node; where a j represents the normalized cumulative workload of service node j;
and calculating the resource allocation balance degree according to the concentration degree parameter and the vector of the accumulated load of the resources of each service node.
Further, each service node resource cumulative load a j is defined as:
Where a j represents the normalized cumulative workload of service node j, K 1 represents the degree of importance for the computing resource usage, K 2 represents the degree of importance for the storage resource usage, K represents the total number of users, and x kj represents the offload decision parameter.
Further, the formula for calculating the balance degree of resource allocation is as follows:
Where f 2 (x, y) represents an objective function of the resource allocation balance, M represents the total number of service nodes, y j represents a busy decision parameter, a j represents an element in a vector of the cumulative load of the resources of each service node, The average value of the cumulative load of the resources of each service node is represented, and v represents the mean square error of the cumulative load of the resources of each service node.
Preferably, the objective function of the combined total time delay and the resource allocation balance degree is:
min[f(x,y)]=min[w1f1(x,y)+w2f2(x,y)]
C1:0≤w1≤1,0≤w2≤1;
C2:w1+w2=1
C3:
C4:
Wherein f 1 (x, y) represents an objective function of total time delay of unloading, f 2 (x, y) represents an objective function of balance degree of resource allocation, w 1 represents a weight value allocated to the total time delay of unloading, w 2 represents a weight value allocated to the resource by the balance degree, K represents a total number of users, M represents a total number of service nodes, z k represents a computing resource provided by the service node for task T k, x kj represents an unloading determination parameter, p j represents a computing resource of the service node, y j represents a busy determination parameter, e k represents a storage resource provided by the service node for task T k, and c j represents a storage resource of the service node.
Preferably, the process of solving the objective function of the combined unloading total delay and the resource allocation balance degree comprises the following steps: solving an objective function by adopting a weighted resource optimization algorithm, wherein the step of solving comprises the following steps:
S1: determining the weight w 1 of the objective function f 1 (x, y) and the weight w 2 of the function f 2 (x, y);
S2: preprocessing an objective function and inequality constraint conditions according to the weight w 1 and the weight w 2 to obtain a coefficient matrix x of each variable in the objective function;
s3: traversing elements in the coefficient matrix x, and taking the elements in the coefficient matrix as intcon values in sequence;
s4: initializing a coefficient matrix A of the inequality constraint condition and a constraint vector b of the inequality constraint condition; initializing a constraint vector beq of the equality constraint coefficient matrix Aeq and the equality constraint; the optimization interval lb of the variable x takes an all-zero vector, and ub takes an all-1 vector;
s5: and (3) according to the coefficient matrix A of the initialized inequality constraint condition, the constraint vector b of the initialized inequality constraint condition, the coefficient matrix Aeq of the initialized equality constraint condition, the constraint vectors beq of the initialized equality constraint condition and the values of intcon, invoking intlinprog () function to implement linear programming on the objective function f (x, y), wherein the finally obtained x is the optimal solution xopt.
Further, the formula for solving the objective function by adopting the weighted resource optimization algorithm is as follows:
[x,f(x,y)]=intlinprog(f,intcon,A,b,Aeq,beq,lb,ub)
A×x≤b
Aeq×x=beq
lb≤x≤ub
Wherein x represents a coefficient matrix of variables in the objective function, f (x, y) is the objective function, intlinprog () is a function for performing linear programming on the objective function, intcon represents a position of an integer decision variable in x, a represents a coefficient matrix of an inequality constraint condition, b represents a constraint vector of the inequality constraint condition, aeq represents an equality constraint condition coefficient matrix, beq represents an equality constraint condition constraint vector, lb represents a constraint interval lower limit of the variable x, and ub represents a constraint interval upper limit of the variable x.
The beneficial effects of the invention are as follows: the invention provides a multi-user computing and unloading method based on UPF equipment, mini MEP and other 5G private network shunts, which has good expansibility and safety, ensures high availability of industrial Internet equipment through a cluster formed by a plurality of edge terminal equipment with idle resources, effectively utilizes the bandwidth and throughput of each UE equipment through load balancing, strengthens network data processing capacity, and improves the flexibility and availability of the network; the method can effectively reduce the average unloading time delay and the average transmission time delay of multiple users, and balance the workload of each mobile edge terminal device or each edge server, thereby having good economic benefit.
Drawings
FIG. 1 is a block diagram of a system frame in accordance with the present invention;
FIG. 2 is a flow chart of a multi-user computing offload method incorporating UPFs and miniMEP in accordance with the present invention;
FIG. 3 is a flow chart of constructing a linear optimization objective function in the present invention;
FIG. 4 is a flowchart of the weighted resource optimization algorithm of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The invention provides a multi-user computing and unloading method based on a 5G private network splitter, wherein the structure of a system framework of the invention is shown in figure 1, and mainly comprises an integral framework, a user plane function device UPF (User Plane Function), a small mobile edge computing platform miniMEP (mini Mobile Edge Platform) and the position of a local terminal device in a suitable 5G scene; the base station bound with miniUPF is used for acquiring the data information of the industrial Internet equipment transmitted by the CPE and the task information to be unloaded by the local equipment, and transmitting the data information to the mini UPF for data screening processing; the miniUPF equipment is used for receiving and acquiring data information transmitted by the base station, screening and judging, receiving the data packets conforming to the rules into the factory edge cloud, collecting the load condition of each CPE by the mini MEP, carrying out corresponding processing, and forwarding the data packets not conforming to the load condition to the core network, thereby reducing the time delay of accessing the local network which must pass through the core network. The mini MEP is used as an edge collaboration server for collaboration of mobile edge terminals and servers, gathers the offloading tasks and related information sent from the current user terminal equipment, judges which edge terminal equipment and servers in the local network have sufficient computing and storage resources, and judges which edge terminal equipment and service are best to offload to after the computation of an objective function, thereby solving the problem that the current terminal equipment has insufficient computing task capacity.
The overall flow for realizing user unloading according to UPF and miniMEP in the invention is shown in figure 2, in particular, as shown in figure 3, the multi-user computing unloading method based on the 5G private network splitter in the invention comprises the following steps:
s1: constructing a multi-user computing unloading frame based on a 5G private network shunt; the 5G private network splitter includes a UPF device and a small mobile edge computing platform miniMEP;
S2: constructing an objective function of the combined unloading total time delay and the resource allocation balance degree according to the multi-user computing unloading framework; the unloading total time delay comprises unloading time delay and transmission time delay;
s3: solving an objective function of the combined unloading total time delay and the resource allocation balance degree to obtain an objective function optimal solution; determining an unloading decision by a user according to the optimal solution of the objective function;
s4: and the user performs task unloading according to the determined unloading decision.
Constructing a multi-user computing and unloading framework based on a 5G private network shunt comprises constructing a computing and unloading total time delay framework and constructing a computing resource allocation balance degree framework; the calculation unloading framework based on the 5G private network splitter can enable user data to directly enter a local enterprise network through UPF (unified packet flow) without passing through a core network, so that transmission delay is reduced, and the safety of the whole unloading process is improved; constructing a computing offload aggregate latency framework includes: miniMEP is used as a mobile edge computing cooperative server, and the resource available state of each idle service node is recorded; the available states of the computing resources and the storage resources of the service node j are represented by a joint resource pair (p j,cj), wherein p j represents the computing resources of the service node, and c j represents the storage resources of the service node; defining an intensive computing task of the mobile user k as T k, and representing the requirement of the mobile user k on resources by using a joint resource pair (z k,ek), wherein the computing resource provided by a service node for the task T k is z k, and the storage resource provided by the service node for the task T k is e k; the service node comprises a mobile edge terminal and an edge server; every t time, at the beginning of each t time period, mobile user k sends its own joint resource demand pair (z k,ek) to miniMEP; miniMEP calculates the total offload delay for each mobile user based on the remaining computing resources of each serving node, the remaining storage resources, and the joint resource demand pair of the mobile user (z k,ek).
The computing resource margin d kj of the mobile edge computing collaboration server is:
Where z i denotes the computing resources applied by the user, e i denotes the storage resources applied by the user, and J denotes the set of service nodes.
In order to represent the service and served relationship between the mobile user and the mobile edge terminal and edge server, a matrix x=x i1×M is introduced to characterize the overall offloading decision, the matrix is in the row j column, the element X kj of the matrix is defined as an offloading decision parameter, and it is represented whether the mobile user K offloads the computing task to the service node j, where K e K, j e M. K is the total number of mobile users, M is the total number of service nodes, and the unloading judgment parameter x kj is assigned as follows:
It is assumed that a mobile terminal can only offload computing tasks to at most one edge terminal node or edge server, so that at most one element in each row of matrix X has a value of 1, and the remaining elements in this row are all 0. Further, a vector y=y i1×M is introduced to identify whether the service node j provides offload services for the mobile user, its element Y i (j e M) is defined as a busy decision parameter, which indicates whether the service node j provides offload services for the mobile user, and the busy decision parameter is assigned as:
Constructing a computing resource allocation balance framework includes:
to evaluate the load balance of the offloading decision, two parameters are defined: one is a concentration degree parameter u characterizing that the offloading scheme occupies a service node, and the other is a vector a= (a 1,a2,a3...aM) characterizing the cumulative load of edge terminal nodes or edge server resources. The element a j (j epsilon M) and the parameter u of A are normalized parameters, the numerical value is in the range of [0,1], and the definition of the concentration degree parameter u is as follows:
Element a j in vector a= (a 1,a2,a3...aM) of cumulative load of each service node resource is defined as:
Where a j (j∈m) represents the normalized cumulative workload of the service node j (comprehensively considering the computing resource and the storage resource usage load), the normalized resource cumulative loads a 1,a2,a3...aM of all M service nodes form a resource cumulative load vector a= (a 1,a2,a3...aM);k1 represents the importance of the computing resource usage, k 2 represents the importance of the storage resource usage, the weights k 1 and k 2 satisfy 0.ltoreq.k 1,k2.ltoreq.1, and k 1+k2 =1, and preferably, the computing resource and the storage resource usage of the edge terminal node or the edge server are the same, that is, k 1=k2 =0.5.
The constructing of the objective function of the combined unloading total time delay and the resource allocation balance degree comprises the following steps:
Firstly, minimizing the total computing and unloading time delay of a mobile user, improving the mobile user experience, secondly balancing the resource use load on a service node, and avoiding the resource on the service node from being consumed too early by an unbalanced computing and unloading scheme, so that the optimization of computing and unloading decision is the optimization of two targets of the total computing and unloading time delay and the resource allocation balance degree, the optimization is converted into the linear optimization problem based on integers, and an objective function of the total computing and unloading time delay and the resource allocation balance degree is constructed:
min[f(x,y)]=min[w1f1(x,y)+w2f2(x,y)]
C1:0≤w1≤1,0≤w2≤1;
C2:w1+w2=1
C3:
C4:
Where f 1 (x, y) represents the objective function of the total offload delay, f 2 (x, y) represents the objective function of the resource allocation balance, w 1 represents the weight value allocated to the total offload delay, w 2 represents the weight value allocated to the resource utilization balance, The average value of the cumulative load of the resources of each service node is represented, and v represents the mean square error of the cumulative load of the resources of each service node.
The method for solving the objective function of the combined unloading total time delay and the resource allocation balance degree adopts a weighted resource optimization algorithm to solve the objective function, and comprises the following steps:
For a fully distributed computing offload scheme, f 2 (x, y) gets the smallest possible value of 0 when all edge terminals or edge servers have exactly the same proportion of resource usage load; as the difference of the usage load proportion of different edge terminals or edge servers becomes larger, and the distribution of the edge terminals or edge servers occupied by the unloading calculation tends to be concentrated, the value of f 2 (x, y) also increases, the asymptotically reaches 1, and the resource usage unbalance degree is highest at the moment. The constraint on the resource load of each edge terminal or edge server is simple, i.e. the sum of its resources allocated to the mobile users must not exceed its total resource upper limit. The functions f 1(x,y)、f2 (x, y) are aggregated together by using a linear weighting method to obtain an objective function f (x, y) for representing the whole optimization problem, and when f (x, y) obtains the minimum value, the total unloading time delay of the calculation unloading scheme and the resource allocation balance degree of the service node are indicated to be totally optimal; as shown in fig. 4, the step of solving includes:
S1: determining the weight w 1 of the objective function f 1 (x, y) and the weight w 2 of the function f 2 (x, y);
S2: preprocessing an objective function and inequality constraint conditions according to the weight w 1 and the weight w 2 to obtain a coefficient matrix x of each variable in the objective function;
s3: traversing elements in the coefficient matrix x, and taking the elements in the coefficient matrix as intcon values in sequence;
s4: initializing a coefficient matrix A of the inequality constraint condition and a constraint vector b of the inequality constraint condition; initializing a constraint vector beq of the equality constraint coefficient matrix Aeq and the equality constraint; the optimization interval lb of the variable x takes an all-zero vector, and ub takes an all-1 vector;
s5: and (3) according to the coefficient matrix A of the initialized inequality constraint condition, the constraint vector b of the initialized inequality constraint condition, the coefficient matrix Aeq of the initialized equality constraint condition, the constraint vectors beq of the initialized equality constraint condition and the values of intcon, invoking intlinprog () function to implement linear programming on the objective function f (x, y), wherein the finally obtained x is the optimal solution xopt.
The formula for solving the objective function by adopting the weighted resource optimization algorithm is as follows:
[x,f(x,y)]=intlinprog(f,intcon,A,b,Aeq,beq,lb,ub)
A×x≤b
Aeq×x=beq
lb≤x≤ub
Wherein x represents a coefficient matrix of variables in the objective function, f (x, y) is the objective function, intlinprog () is a function for performing linear programming on the objective function, intcon represents a position of an integer decision variable in x, a represents a coefficient matrix of an inequality constraint condition, b represents a constraint vector of the inequality constraint condition, aeq represents an equality constraint condition coefficient matrix, beq represents an equality constraint condition constraint vector, lb represents a constraint interval lower limit of the variable x, and ub represents a constraint interval upper limit of the variable x.
The optimal solution xopt is the optimal solution of the objective function combining the total unloading time delay and the resource allocation balance degree, and then the system guides each mobile terminal to implement unloading according to the unloading scheme corresponding to the xopt.
The invention provides a multi-user computing and unloading method based on UPF equipment, mini MEP and other 5G private network shunts, which has good expansibility and safety, ensures high availability of industrial Internet equipment through a cluster formed by a plurality of edge terminal equipment with idle resources, effectively utilizes the bandwidth and throughput of each UE equipment through load balancing, strengthens network data processing capacity, and improves the flexibility and availability of the network; the method can effectively reduce the average unloading time delay and the average transmission time delay of multiple users, and balance the workload of each mobile edge terminal device or each edge server, thereby having good economic benefit.
While the foregoing is directed to embodiments, aspects and advantages of the present invention, other and further details of the invention may be had by the foregoing description, it will be understood that the foregoing embodiments are merely exemplary of the invention, and that any changes, substitutions, alterations, etc. which may be made herein without departing from the spirit and principles of the invention.
Claims (2)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202111519999.7A CN114173379B (en) | 2021-12-13 | 2021-12-13 | A multi-user computing offloading method based on 5G private network splitter |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202111519999.7A CN114173379B (en) | 2021-12-13 | 2021-12-13 | A multi-user computing offloading method based on 5G private network splitter |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN114173379A CN114173379A (en) | 2022-03-11 |
| CN114173379B true CN114173379B (en) | 2024-11-12 |
Family
ID=80486330
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202111519999.7A Active CN114173379B (en) | 2021-12-13 | 2021-12-13 | A multi-user computing offloading method based on 5G private network splitter |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN114173379B (en) |
Families Citing this family (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11843953B1 (en) | 2022-08-02 | 2023-12-12 | Digital Global Systems, Inc. | System, method, and apparatus for providing optimized network resources |
| US12156037B2 (en) | 2022-08-02 | 2024-11-26 | Digital Global Systems, Inc. | System, method, and apparatus for providing optimized network resources |
| US12250558B2 (en) | 2022-08-02 | 2025-03-11 | Digital Global Systems, Inc. | System, method, and apparatus for providing optimized network resources |
| US11570627B1 (en) | 2022-08-02 | 2023-01-31 | Digital Global Systems, Inc. | System, method, and apparatus for providing optimized network resources |
| US11751064B1 (en) | 2022-08-02 | 2023-09-05 | Digital Global Systems, Inc. | System, method, and apparatus for providing optimized network resources |
| US12262212B2 (en) | 2022-08-02 | 2025-03-25 | Digital Global Systems, Inc. | System, method, and apparatus for providing optimized network resources |
| US11659400B1 (en) | 2022-08-02 | 2023-05-23 | Digital Global Systems, Inc. | System, method, and apparatus for providing optimized network resources |
| US12483898B2 (en) | 2022-08-02 | 2025-11-25 | Digital Global Systems, Inc. | System, method, and apparatus for providing optimized network resources |
| US12294865B2 (en) | 2022-08-02 | 2025-05-06 | Digital Global Systems, Inc. | System, method, and apparatus for providing optimized network resources |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN108174421A (en) * | 2018-03-05 | 2018-06-15 | 重庆邮电大学 | A MEC-assisted data offloading method in a 5G network |
| CN111586762A (en) * | 2020-04-29 | 2020-08-25 | 重庆邮电大学 | A joint optimization method for task offloading and resource allocation based on edge collaboration |
Family Cites Families (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11172490B2 (en) * | 2019-11-25 | 2021-11-09 | University-Industry Cooperation Group Of Kyung Hee University | Apparatus for constructing energy-efficient communication and computation resource allocation framework for multiple communication service and method thereof |
| CN111641973B (en) * | 2020-05-29 | 2022-04-01 | 重庆邮电大学 | Load balancing method based on fog node cooperation in fog computing network |
-
2021
- 2021-12-13 CN CN202111519999.7A patent/CN114173379B/en active Active
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN108174421A (en) * | 2018-03-05 | 2018-06-15 | 重庆邮电大学 | A MEC-assisted data offloading method in a 5G network |
| CN111586762A (en) * | 2020-04-29 | 2020-08-25 | 重庆邮电大学 | A joint optimization method for task offloading and resource allocation based on edge collaboration |
Also Published As
| Publication number | Publication date |
|---|---|
| CN114173379A (en) | 2022-03-11 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN114173379B (en) | A multi-user computing offloading method based on 5G private network splitter | |
| Apostolopoulos et al. | Risk-aware data offloading in multi-server multi-access edge computing environment | |
| CN108901046B (en) | Collaborative task offloading algorithm and system design scheme for mobile edge computing | |
| CN109684075B (en) | Method for unloading computing tasks based on edge computing and cloud computing cooperation | |
| CN110098969B (en) | Fog computing task unloading method for Internet of things | |
| CN108920279B (en) | Mobile edge computing task unloading method under multi-user scene | |
| CN110087318B (en) | Task unloading and resource allocation joint optimization method based on 5G mobile edge calculation | |
| Sun et al. | Autonomous resource slicing for virtualized vehicular networks with D2D communications based on deep reinforcement learning | |
| CN109857546B (en) | Multi-server mobile edge computing unloading method and device based on Lyapunov optimization | |
| Jia et al. | Cloudlet load balancing in wireless metropolitan area networks | |
| CN110234127B (en) | SDN-based fog network task unloading method | |
| Lai et al. | TSSM: Time-sharing switch migration to balance loads of distributed SDN controllers | |
| CN111131421B (en) | A method for interconnection between industrial Internet field big data and cloud information | |
| CN111586696A (en) | A Resource Allocation and Offloading Decision Method Based on Multi-Agent Architecture Reinforcement Learning | |
| CN111182570A (en) | User association and edge computing unloading method for improving utility of operator | |
| CN112040512A (en) | A fair-based fog computing task offloading method and system | |
| CN108112037A (en) | Load-balancing method with cooperative communication network is calculated based on mist | |
| Tam et al. | Intelligent massive traffic handling scheme in 5G bottleneck backhaul networks | |
| CN111836284A (en) | Energy consumption optimization calculation and unloading method and system based on mobile edge computing | |
| CN103442412A (en) | Network selection method based on weighted optimal bipartite graph matching | |
| Ma et al. | Online NFV-enabled multicasting in mobile edge cloud networks | |
| Zhang et al. | Adaptive digital twin server deployment for dynamic edge networks in IoT system | |
| Shahrbabaki et al. | A novel sdn-enabled edge computing load balancing scheme for iot video analytics | |
| CN111835849A (en) | Method and device for enhancing service capability of access network | |
| CN112084034B (en) | A MCT scheduling method based on edge platform layer adjustment coefficient |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |