[go: up one dir, main page]

US9439081B1 - Systems and methods for network performance forecasting - Google Patents

Systems and methods for network performance forecasting Download PDF

Info

Publication number
US9439081B1
US9439081B1 US14/172,686 US201414172686A US9439081B1 US 9439081 B1 US9439081 B1 US 9439081B1 US 201414172686 A US201414172686 A US 201414172686A US 9439081 B1 US9439081 B1 US 9439081B1
Authority
US
United States
Prior art keywords
kpi
network
data
kpis
computer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US14/172,686
Inventor
Matthew Brian Knebl
Amr Albanna
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ookla LLC
Original Assignee
Further LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Further LLC filed Critical Further LLC
Priority to US14/172,686 priority Critical patent/US9439081B1/en
Assigned to Further LLC reassignment Further LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ALBANNA, AMR, KNEBL, MATTHEW BRIAN
Application granted granted Critical
Publication of US9439081B1 publication Critical patent/US9439081B1/en
Assigned to OOKLA, LLC reassignment OOKLA, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Further LLC
Assigned to MUFG UNION BANK, N.A., AS COLLATERAL AGENT reassignment MUFG UNION BANK, N.A., AS COLLATERAL AGENT SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: EVERYDAY HEALTH, INC., KEEPITSAFE, INC., OOKLA, LLC, RETAILMENOT, INC., SPICEWORKS, INC., THREATTRACK SECURITY, INC.
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W16/00Network planning, e.g. coverage or traffic planning tools; Network deployment, e.g. resource partitioning or cells structures
    • H04W16/18Network planning tools
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N99/005
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W16/00Network planning, e.g. coverage or traffic planning tools; Network deployment, e.g. resource partitioning or cells structures
    • H04W16/22Traffic simulation tools or models

Definitions

  • a cellular network or mobile network can include a wireless network distributed over land areas called cells, each served by at least one fixed-location transceiver known as a cell site or base station.
  • each cell uses a different set of frequencies or codes from neighboring cells to avoid interference and provide acceptable performance within each cell.
  • these cells provide radio coverage over a wide geographic area. This joining of cells enables a large number of portable transceivers (e.g., mobile phones, pagers, etc.) to communicate with each other and with transceivers and telephones anywhere in the network, via base stations, even if some of the transceivers are moving through more than one cell during transmission.
  • portable transceivers e.g., mobile phones, pagers, etc.
  • telecommunications providers have deployed voice and data cellular networks over most of the inhabited land area of the Earth. This wide deployment allows mobile phones and mobile computing devices to be connected to the public switched telephone network and public Internet. Private cellular networks can be used for research or for large organizations and fleets, such as dispatch for local public safety agencies.
  • the method comprises receiving historical performance data and baseline data for a cell in a network.
  • the historical performance data may comprise a plurality of key performance indicators (KPIs).
  • KPIs key performance indicators
  • the method further comprises receiving, from a user, a selection of a first KPI in the plurality of KPIs.
  • the method further comprises applying a machine learning model to the first KPI and at least one KPI in the plurality of KPIs.
  • the method further comprises generating at least one model parameter based on application of the machine learning model.
  • the method further comprises applying the baseline data to the at least one model parameter.
  • the method further comprises calculating a predicted value of the first KPI for the cell in the network based on the application of the baseline data to the at least one model parameter. At least said calculating may be performed by a computer system comprising computer hardware.
  • the system further comprises a computing system comprising one or more computing devices, the computing system in communication with the computer data repository.
  • the computing system may be programmed to implement a historical performance data collection unit configured to receive the historical performance data, the historical performance data comprising a plurality of key performance indicators (KPIs).
  • KPIs key performance indicators
  • the computing system may be further programmed to implement a network forecast determination unit configured to receive a selection of a first KPI in the plurality of KPIs.
  • the network forecast determination unit may be further configured to apply a machine learning model to at least the first KPI to produce a model parameter and calculate a predicted value of the first KPI for the cell in the network based on the model parameter.
  • a computer storage system comprising a non-transitory storage device, said computer storage system having stored thereon executable program instructions that direct a computer system to at least access historical performance data associated with a telecommunications network, the historical performance data comprising a plurality of key performance indicators (KPIs).
  • KPIs key performance indicators
  • the computer storage system further has stored thereon executable program instructions that direct a computer system to at least electronically output a graphical user interface for presentation to a user, the graphical user interface comprising functionality for the user to select one or more of the KPIs.
  • the computer storage system further has stored thereon executable program instructions that direct a computer system to at least receive, from the graphical user interface, a user selection of a first KPI of the plurality of KPIs.
  • the computer storage system further has stored thereon executable program instructions that direct a computer system to at least identify a numerical relationship between the first KPI and one or more other KPIs in the plurality of KPIs.
  • the computer storage system further has stored thereon executable program instructions that direct a computer system to at least calculate a predicted value of one or more of the plurality of KPIs based on the numerical relationship.
  • the computer storage system further has stored thereon executable program instructions that direct a computer system to at least output a representation of the predicted value of the one or more of the plurality of KPIs for presentation to the user.
  • FIG. 2 illustrates an example of a more detailed block diagram of the network forecasting server of FIG. 1 .
  • FIG. 3 illustrates an example set of statistical correlation analysis results in a chart.
  • FIG. 4 illustrates an example analysis of two KPIs.
  • FIG. 5 illustrates an example matrix generated by the network forecast determination unit.
  • FIG. 6 illustrates an example graph that compares predicted performance of a KPI and actual performance of a KPI on a cell.
  • FIG. 7 illustrates an example GUI generated by the GUI unit.
  • FIG. 8 illustrates another example GUI generated by the GUI unit.
  • FIG. 10 illustrates another example GUI generated by the GUI unit.
  • FIG. 11 illustrates a flowchart of an embodiment of a method for forecasting network performance.
  • cell capacity and performance can vary dramatically from cell to cell depending on various factors.
  • factors may include the location of users (e.g., near, far, in buildings, etc.), traffic level and mix (e.g., voice, data, etc.), location of neighboring cells, noise from own cells and/or users, noise from neighboring cells and/or users, external interferers, and/or other factors.
  • KPIs key performance indicators
  • the accuracy of typical coverage and quality predictions is degraded by the lack of precise information about the location of users, vegetation in the coverage area, building materials used in the coverage area, and/or the like.
  • network operators can identify simple, low cost ways to squeeze more capacity from existing network equipment. For example, network operators often highlight large numbers of planned new cell site builds, radio adds, and/or spectrum adds that can be cancelled or deferred. Conversely, areas that will degrade in the near future that have no relief planned can be identified and such degradation issues can be mitigated before customers experience service problems. Spectrum carving and device mix scenarios can be simulated in order to optimize timing of network adjustments and rollouts of new technologies. Networks often have excess capacity with overloaded cells located next to underutilized cells that can be proactively rebalanced through simple, low cost network changes, such as antenna azimuth, downtilt changes, and/or power changes.
  • operators can prioritize which new capacity sites to build (which can take 2 or more years), which sites get additional carriers (which can take 6 months or more), the location and quantity of small cells needed, and/or the spectrum required, and plan ahead to get these modifications done on time before KPIs are impacted.
  • the network forecasting tools described herein may be available via a network.
  • a user e.g., an individual, an employee of a network operator, etc.
  • the network forecasting tools hosted by a server or such like device, may allow the user to view current network traffic conditions for one or more cells in a network.
  • the network forecasting tools may also allow the user to view predicted network traffic conditions on a chosen future date for one or more cells in the network.
  • the predicted network traffic conditions include predicted values for one or more key performance indicators (KPIs), which are described in greater detail below.
  • KPIs key performance indicators
  • the server may calculate the predicted network conditions based on historical performance data, an estimated voice and/or data traffic growth rate, frequencies identified as available for use by the network operator in the future and/or the like. Traffic growth may be positive or negative.
  • the server may receive historical performance data for a geographical region that the user is viewing. The user may then identify one or more KPIs that are of interest to the user.
  • the server may use machine learning techniques to produce a set of model parameters.
  • the model parameters may represent coefficients for various KPIs.
  • the model parameters may be applied to baseline data (e.g., current or observed KPI values) to determine a predicted KPI value for the specific cell in the network. The above techniques may be repeated until the selected KPIs in some or all cells in the geographic region have been predicted.
  • the server may generate instructions based on the predicted KPIs for controlling network equipment. For example, the server may instruct network equipment to perform a power shift on a cell, adjust an antenna pattern, and/or perform like adjustments such that degradations in performance levels can be reduced or prevented.
  • systems and methods described herein can be implemented in any of a variety of electronic devices, including, for example, physical or virtual servers, cell phones, smart phones, personal digital assistants (PDAs), tablets, mini-tablets, laptops, desktops, and televisions, to name a few.
  • PDAs personal digital assistants
  • tablets mini-tablets
  • laptops desktops
  • televisions to name a few.
  • FIG. 1 illustrates a block diagram of an example network forecasting system 100 .
  • the network forecasting system 100 includes one or more user devices 110 , a network 120 , a network control unit 130 , a network forecasting server 140 , a performance data store 150 , and a site configuration data store 160 .
  • the one or more user devices 110 can each be configured to submit requests for access to network performance forecasts provided by the network forecasting server 140 . Such requests can be made via the network 120 .
  • the one or more user devices 110 may allow users (e.g., individuals, employees of telecommunication service providers or network operators, etc.) or automatic control units (e.g., self-optimizing network (SON) modules) to interact with the network forecasting information provided by the network forecasting server 140 .
  • users e.g., individuals, employees of telecommunication service providers or network operators, etc.
  • automatic control units e.g., self-optimizing network (SON) modules
  • the network control unit 130 can receive instructions to control automatically one or more pieces of network equipment.
  • the network equipment may be automatic control units (e.g., self-organizing network (SON) modules).
  • the network control unit 130 may receive the instructions from the network forecasting server 140 via the network 120 .
  • the instructions may instruct network equipment to automatically self-optimize or self-improve parameters or behavior in response to network performance forecasts generated by the network forecasting server 140 .
  • the network equipment may be instructed to change parameters, perform a power shift on a cell, adjust remotely-controlled components (e.g., an antenna pattern), and/or perform like adjustments such that degradations in performance levels can be reduced or prevented.
  • remotely-controlled components e.g., an antenna pattern
  • the network forecasting server 140 can generate network performance forecasts.
  • the network forecasting server 140 may generate the network performance forecasts based on cell-specific performance models (e.g., forecasts generated based on performance models for individual cells can be aggregated to provide forecasts for part of or all of an entire network).
  • the cell-specific performance models may be based on historical performance data of telecommunication networks operated by one or more network operators (e.g., carriers). Historical performance data may include historical key performance indicator (KPI) values and other measurements taken from a network.
  • KPI historical key performance indicator
  • historical performance data may include, but are not limited to, voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic (e.g., neighboring cell traffic), received signal strength indication (RSSI) levels (e.g., uplink RSSI levels, downlink RSSI levels, noise floor at a site, etc.), radio resource control (RRC) attempts, transmit power, number of voice users, number of data users, parameter settings, pilot power, load threshold, code utilization, power utilization, and/or the like.
  • RRC radio resource control
  • each network performance forecast predicts a cell's network performance as the cell's level of traffic varies to detect previously hidden network problems and/or to identify an amount of traffic a particular cell can be expected to carry before the cell's network performance KPIs degrade below acceptable thresholds.
  • a network performance forecast may be represented by one or more predicted KPI values (e.g., KPI values expected in the future).
  • Example KPIs may include, but are not limited to, voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic (e.g., neighboring cell traffic), RSSI levels (e.g., uplink RSSI levels, downlink RSSI levels, etc.), RRC attempts, transmit power, number of voice users, number of data users, parameter settings, pilot power, load threshold, code utilization, power utilization, and/or the like.
  • the network performance forecasts can be used to improve daily network troubleshooting and optimization. For example, users of the one or more user devices 110 can view the network performance forecasts to identify cells that are experiencing external interference (e.g., because the cells have a higher uplink noise level than can be explained by normal network traffic).
  • the network forecasting server 140 can generate instructions based on the network performance forecasts that are used to control the operation of network equipment.
  • the network forecasting server 140 can also generate a graphical representation of the network performance forecasts.
  • the network forecasting server 140 may generate a graphical user interface (GUI) that is viewable via a browser or other client application running on the one or more user devices 110 .
  • GUI graphical user interface
  • the network forecasting server 140 may be a computing device.
  • the network forecasting server 140 may include one or more processors to execute one or more instructions, memory, and communication devices to transmit and receive data over the network 120 .
  • the network forecasting server 140 is implemented as one or more backend servers capable of communicating over a network.
  • the network forecasting server 140 is implemented by one or more virtual machines in a hosted computing environment, such as in a cloud computing environment.
  • the hosted computing environment may include one or more rapidly provisioned and released computing resources, which computing resources may include computing, networking and/or storage devices.
  • a hosted computing environment may also be referred to as a cloud computing environment.
  • the network forecasting server 140 may be represented as a user computing device capable of communicating over a network, such as a laptop or tablet computer, personal computer, personal digital assistant (PDA), hybrid PDA/mobile phone, mobile phone, global positioning system (GPS) device, or the like.
  • the network forecasting server 140 may be in communication with one or a plurality of user devices 110 .
  • the network forecasting server 140 may be in communication with the performance data store 150 via the network 120 .
  • the performance data store 150 stores historical performance data of telecommunication networks operated by one or more network operators.
  • the network forecasting server 140 may access historical performance data stored in the performance data store 150 in generating network performance forecasts.
  • the performance data store 150 is a single data store. In other embodiments, the performance data store 150 includes multiple physical storage devices distributed over many different locations.
  • the site configuration data store 160 stores site configuration data for network equipment in telecommunication networks operated by one or more network operators.
  • site configuration data may include, but is not limited to, cell site latitude, longitude, antenna type, azimuth (e.g., direction), frequency, radio network controller (RNC), and/or the like.
  • the site configuration data store 160 is a single data store. In other embodiments, the site configuration data store 160 includes multiple physical storage devices distributed over many different locations.
  • the site configuration data may be used to depict the network (e.g., pie wedges as described below) in a user interface (e.g., the example user interfaces in FIGS. 7-10 ) and/or may be used when determining network performance forecasts (e.g., may be used as inputs in a machine learning model).
  • the network 120 may be a wired network, a wireless network, or a combination of the two.
  • the network 120 may be a personal area network, a local area network (LAN), a wide area network (WAN), or combinations of the same.
  • Protocols and components for communicating via any of the other aforementioned types of communication networks, such as the TCP/IP protocols, can be used in the network 120 .
  • FIG. 2 illustrates an example of a more detailed block diagram of the network forecasting server 140 of FIG. 1 .
  • the network forecasting server 140 may include a historical performance data collection unit 210 , a network forecast determination unit 220 , and a graphical user interface (GUI) unit 230 .
  • GUI graphical user interface
  • the historical performance data collection unit 210 can collect historical performance data stored in the performance data store 150 .
  • the historical performance data collection unit 210 may collect historical performance data for a set period of time (e.g., one week, one month, one year, etc.). The historical performance data collected may depend on the user requesting permission to access the network performance forecasts.
  • the historical performance data collection unit 210 may collect historical performance data associated with a first carrier if the user requesting permission to access the network performance forecasts is associated with the first carrier.
  • the historical performance data collection unit 210 collects the historical performance data at the same time or nearly the same time as the network forecasting server 140 receives a request for network performance forecasts.
  • the historical performance data collection unit 210 collects the historical performance data at a time before the network forecasting server 140 receives a request for network performance forecasts.
  • the network forecast determination unit 220 can generate network performance forecasts. Each network performance forecast may be specific to a cell in a cellular network or may encompass a plurality of cells in the cellular network.
  • the network forecast determination unit 220 may be configured to generate a plurality of network performance forecasts to cover a geographic area selected by a user.
  • the network forecast determination unit 220 receives a selection of one or more KPIs. For example, a user operating a user device 110 may select one or more KPIs that are of interest (e.g., a user may be interested in the impact of rising traffic on the dropped call rate).
  • the network forecast determination unit 220 identifies relationships between KPIs before beginning the process of generating network performance forecasts. For example, the relationships between KPIs may be unknown (e.g., network performance forecasts are to be generated for a wireless standard or network for the first time). In other embodiments, the network forecast determination unit 220 begins generating network performance forecasts once one or more KPIs are selected. For example, the relationships between KPIs may have been determined already at a previous time or received from another device.
  • the network forecast determination unit 220 may identify relationships between KPIs using machine-learning techniques on one or more KPIs.
  • Machine-learning techniques may include statistical correlation analyses, regression analyses (e.g., multiple regression analyses, linear regression analyses, logistic regression analyses, etc.), neural networks, combinations of the same, or the like.
  • machine learning is performed on at least one selected KPI and at least one unselected KPI.
  • machine-learning is performed on only the selected KPIs.
  • the network forecast determination unit 220 may use a statistical correlation analysis that compares each analyzed KPI with some or all other analyzed KPIs.
  • the term “correlation,” in addition to having its ordinary meaning, can refer to calculation of a specific correlation value or coefficient or more generally to a numerical comparison (e.g., of KPIs) that does not necessarily require calculation of a mathematically-defined correlation coefficient.
  • the result of the statistical correlation analysis may indicate the sensitivity of each analyzed KPI to the other analyzed KPIs.
  • the network forecast determination unit 220 may analyze the result of the statistical correlation analysis to determine which KPIs are most sensitive to other KPIs.
  • the network forecast determination unit 220 may focus the analysis of the results on the one or more selected KPIs.
  • the statistical analysis that may be performed by the network forecast determination unit 220 may also include other statistical parameters or operations, such as a mean, median, mode, variance, standard deviation, regression, or other numerical calculation based on one or more KPIs.
  • the network forecast determination unit 220 determines a relationship between two or more KPIs based on the result of the statistical correlation analysis. For example, the data set of each selected KPI may be compared with the data sets for one or more remaining KPIs individually (e.g., a pair of KPIs may be compared at a time). A KPI data set may include KPI values collected over a period of time (e.g., hourly, daily, etc.). The network forecast determination unit 220 may mathematically determine whether the relationship between KPIs is linear, exponential, logarithmic, or has some other type of relationship.
  • the network forecast determination unit 220 generates one or more charts that include a result of the KPI comparisons.
  • FIG. 3 illustrates an example chart 300 that depicts a set of example statistical correlation analysis results.
  • the chart 300 indicates, for example, that data traffic and uplink RSSI KPIs have a positive correlation and data traffic and voice access success rate KPIs have a negative correlation.
  • the chart 300 may be provided to an analyst (e.g., an individual trained to identify relationships between KPIs based on machine learning results), not shown, for further analysis.
  • the results of the analysis performed by the analyst may be relayed to the one or more user devices 110 via the network forecasting server 140 .
  • FIG. 4 illustrates an example chart 400 depicting analysis of two example KPIs.
  • uplink RSSI 410 is plotted against data access rate 420 .
  • the uplink RSSI 410 and the data access rate 420 have a linear relationship in the depicted example.
  • the network forecast determination unit 220 uses the identified relationships when generating the network performance forecasts.
  • Machine learning may also be used to generate the network performance forecasts.
  • the network forecast determination unit 220 may use a multiple regression analysis on one or more KPIs.
  • the network forecast determination unit 220 may perform a multiple regression analysis with KPIs used as independent and dependent variables.
  • the relationships identified between KPIs are used as a guide to determine which KPIs are used as independent variables and which KPIs are used as dependent variables (e.g., KPIs that have a strong correlation may be used as independent variables, etc.).
  • each selected KPI may be used as a dependent variable and some or all of the remaining KPIs may be used as independent variables.
  • the independent variables may be isolated from semi-independent and dependent variables as this may play a useful role in the selection of historical performance data and KPIs.
  • the network forecast determination unit 220 may improve the machine learning model by using interaction terms or other corrective factors (e.g., adjusting the source data to exponential or other non-linear methods to capture the inter-KPI relationship observed when mathematically analyzing or charting the data).
  • the quality of the machine learning model may be measured using a coefficient of determination (R 2 ), standard error, the Pearson correlation coefficient, or other measurements.
  • the network forecast determination unit 220 sets filter criteria to manipulate the historical performance data.
  • the filter criteria may be used to manipulate the historical performance data before any such data is used to generate the network performance forecasts.
  • the filter criteria may be set based on a baseline range of measurements taken in a network. If a portion of the historical performance data falls outside of the baseline range, the portion may be removed or adjusted.
  • the filter criteria may vary based on the KPI, technology used by the carrier, and/or the carrier requested network performance forecasts.
  • individual cell busy hour, market busy hour, 24 hour, or other timeframes may be selected to vary the loading performance of a cell based on voice erlangs and traffic megabytes.
  • the network forecast determination unit 220 may generate nested machine learning models. In an embodiment, the network forecast determination unit 220 recursively applies machine learning. For example, the network forecast determination unit 220 may apply machine learning to generate a first output. Machine learning may be applied again to generate additional outputs. Machine learning then may be applied again, with the first and/or additional outputs being used as coefficients for a dependent or independent variable. In addition, the network forecast determination unit 220 may correlate noise level measurements and power measurements with at least some performance measurements and at least some KPIs to better predict performance measurements or semi-independent KPIs to the noise level. In an embodiment, the network forecast determination unit 220 correlates the noise level measurement and the power measurement to the final performance KPIs.
  • the network forecast determination unit 220 may revise the machine learning model used to generate the network performance forecasts. Revisions may occur via the addition of independent variables (e.g., KPIs, parameter settings, etc.) or the removal of independent variables.
  • the network forecast determination unit 220 filters and/or weights the machine learning model input data to exclude incorrect data or data which is degrading the model. Once any filtering or weighting is complete, the machine learning is performed and key model parameters are recorded for the respective cell.
  • key model parameters e.g., when a multiple regression analysis is performed
  • the network forecast determination unit 220 may generate a matrix for each cell in the geographic area selected by the user.
  • FIG. 5 illustrates an example matrix 500 generated by the network forecast determination unit 220 .
  • the matrix may include example machine learning model information (e.g., the key model parameters), which can be used to predict each important or desired KPI.
  • the network forecast determination unit 220 applies the matrix coefficients to baseline data for the respective cell to determine one or more predicted KPIs (e.g., an estimated value for a KPI selected by the user).
  • the baseline data for a cell may include measured performance data for on-air cells and/or estimated data for new cells not yet on-air.
  • the baseline data may be stored in the performance data store 150 , stored in another data store (not shown), and/or provided by the user.
  • the network forecast determination unit 220 may calculate a predicted KPI as the intercept of the model plus the product of each independent variable coefficient and the cell's historical baseline data.
  • the network forecast determination unit 220 can increase or decrease the traffic baseline data and can apply the matrix coefficients to the new traffic baseline data.
  • An example predicted dropped call rate for one cell using a multiple regression analysis model is provided below in equation (1) (with coefficients derived from the model, which may vary in other implementations):
  • the network forecast determination unit 220 performs a verification test on one or more KPIs in a cell (e.g., a verification test is performed on each cell's machine learning model).
  • the verification test may ensure or attempt to ensure that as traffic increases, the performance KPIs degrade. If the verification test indicates that a cell's model is not responding as expected, the cell may be labeled as a bad cell.
  • a cell's model may fail the verification test as a result of missing counters, missing statistics, outages, or other reasons. Data for the bad cell may be replaced with new data and new models may be generated using the new data. If the new models fail the verification test, then the network forecast determination unit 220 may assign a default model to the cell.
  • the default model may be generated from RNC, base station controller (BSC), mobility management entity (MME), and/or area models that depict similar cells in similar morphology and/or topology.
  • the network forecast determination unit 220 may send instructions derived from the predicted KPI values to the network control unit 130 .
  • the network control unit 130 may instruct a SON module to direct automatic changes to the network, such as parameter or power changes or adjustments to remotely-controlled components (e.g., antennas).
  • the network forecast determination unit 220 may also send the predicted KPI values to the one or more user devices 110 and/or the GUI unit 230 .
  • the predicted KPI values are provided to the one or more user devices 110 in a table format (e.g., via an exportable file).
  • the predicted KPI values are output visually in a GUI generated by the GUI unit 230 and accessible via a browser or other client application running on the one of more user devices 110 .
  • Example GUIs generated by the GUI unit 230 are described in greater detail below with respect to FIGS. 7-10 .
  • FIGS. 7 through 10 depict example user interfaces or GUIs that may be output by the systems or processes described above.
  • Each of the user interfaces shown includes one or more user interface controls that can be selected by a user, for example, using a browser or other application software.
  • the user interface controls shown are merely illustrative examples and can be varied in other embodiments. For instance, buttons, dropdown boxes, select boxes, text boxes, check boxes, slider controls, and other user interface controls shown, may be substituted with other types of user interface controls that provide the same or similar functionality. Further, user interface controls may be combined or divided into other sets of user interface controls such that similar functionality or the same functionality may be provided with different looking user interfaces.
  • each of the user interface controls may be selected by a user using one or more input options, such as a mouse, touch screen input, or keyboard input, among other user interface input options.
  • FIG. 7 illustrates an example GUI 700 that may be generated by the GUI unit 230 .
  • the GUI 700 includes several interactive features that allow a user to view network performance forecasts for a current date or a date in the future.
  • a user can select one or more RNCs using dropdown box 702 , view one or more map views using options 704 (e.g., road view, aerial view, bird's eye view, etc.), view sector details (e.g., sector information, predicted KPI values, etc.) in window 706 by hovering over or selecting a sector in the GUI 700 , change which spectrum frequencies (or carriers or bands) are viewable using button 712 , hide or view labels 708 depicting planned capacity (e.g., a planned future carrier count in a cell of a sector) using button 716 , adjust annual voice and/or data traffic growth rate via box 718 , adjust a date of the desired network performance forecast using slider 710 (e.g., the slider 710 may step forward and/or back
  • a user can display one or more carriers simultaneously, choose to view KPIs in 24 hour or market busy hour timeframes, and/or export data (e.g., predicted KPIs for a chosen future date and growth rate, first date at which each cell will fall below a threshold that indicates an acceptable KPI performance level, cells or sectors that are the worst offenders, etc.).
  • each cell in a sector is represented by color-coded pie wedges to indicate the KPI performance level (e.g., for the selected KPI(s)). For example, green may indicate an acceptable and high KPI performance level, yellow may indicate an acceptable and medium KPI performance level, and red may indicate an unacceptable and low KPI performance level.
  • each cell may include a label 708 that includes a number depicting planned capacity in the respective cell.
  • a user may also carve spectrum, which removes one or more frequencies and distributes the traffic from those frequencies to the remaining frequencies and shows the predicted performance on the remaining frequencies.
  • FIG. 8 illustrates another example GUI 800 generated by the GUI unit 230 .
  • a sector such as sector 804 , includes one or more cells in a pie-shaped configuration. Each cell may be divided into concentric segments or rings that each represent the performance level of different frequencies. For example, three frequencies (e.g., channels 637 , 2062 , and 2087 ) are displayed in each cell. Cell 802 is displayed as having an acceptable and high KPI performance level.
  • the annual voice traffic growth rate is set at 10% and the annual data traffic growth rate is set at 100%.
  • Slider 810 is set at a current date (e.g., Nov. 1, 2013).
  • FIG. 9 illustrates another example GUI 900 generated by the GUI unit 230 .
  • the slider 810 is stepped forward one year (e.g., to Nov. 1, 2014).
  • the predicted KPI values in each cell are adjusted based on the analysis performed by the network forecast determination unit 220 described above and the growth rates selected by the user.
  • the predicted KPI values are calculated as the slider 810 is adjusted.
  • some or all possible predicted KPI values are calculated before the GUI 900 is presented to the user.
  • a frequency (e.g., channel 2062 ) of the cell 802 is displayed as having an acceptable and medium KPI performance level once slider 810 is adjusted from the position as illustrated in FIG. 8 .
  • FIG. 10 illustrates another example GUI 1000 generated by the GUI unit 230 .
  • the slider 810 remains in the same position as illustrated in FIG. 9 .
  • the spectrum carve feature has been selected such that one frequency is removed (e.g., channel 637 ) and the traffic from the removed frequency is distributed to the remaining frequencies equally, nearly equally, or based on expected loads and/or traffic patterns.
  • sector 1006 includes three cells. The inner concentric ring of each cell in sector 1006 is shown as being transparent (e.g., rather than a color indicating performance levels), indicating that the frequency band associated with that concentric ring has been removed.
  • a frequency (e.g., channel 2062 ) of the cell 802 is now displayed as having an unacceptable and low KPI performance level.
  • removal of spectrum does not necessarily mean that the performance level of a cell will degrade.
  • a cell may be able to maintain performance levels even with the removal of spectrum.
  • channel 637 is also carved from the cell 1008 and distributed to the remaining frequencies. Yet, the performance levels of the remaining frequencies are unchanged despite the removal of spectrum.
  • FIG. 11 illustrates a flowchart of an example method 1100 for forecasting network performance.
  • the method 1100 can be performed by the network forecasting server 140 discussed above with respect to FIGS. 1 and 2 or any other computing device.
  • the method 1100 may include fewer and/or additional blocks and the blocks may be performed in an order different than illustrated.
  • historical performance data and baseline data are received.
  • the historical performance data includes a plurality of KPIs.
  • a selection of a first KPI in the plurality of KPIs is received.
  • the selection may be received from a user operating a user device, such as a user device 110 .
  • a machine learning model is applied.
  • the machine learning model is applied to the first KPI and at least one KPI in the plurality of KPIs.
  • the machine learning model is a multiple regression analysis model, although the machine learning model may be any of the machine learning models described above.
  • a model parameter is generated based on the application of the machine learning model.
  • a model parameter may be a coefficient produced by the machine learning model, where the coefficient is of an independent variable used in the machine learning model.
  • the baseline data is applied to the at least one model parameter.
  • a predicted value of the first KPI is calculated based on the application of the baseline data to the at least one model parameter.
  • the method comprises receiving historical performance data and baseline data for a cell in a network.
  • the historical performance data may comprise a plurality of key performance indicators (KPIs).
  • KPIs key performance indicators
  • the method further comprises receiving, from a user, a selection of a first KPI in the plurality of KPIs.
  • the method further comprises applying a machine learning model to the first KPI and at least one KPI in the plurality of KPIs.
  • the method further comprises generating at least one model parameter based on application of the machine learning model.
  • the method further comprises applying the baseline data to the at least one model parameter.
  • the method further comprises calculating a predicted value of the first KPI for the cell in the network based on the application of the baseline data to the at least one model parameter. At least said calculating may be performed by a computer system comprising computer hardware.
  • the method of the preceding paragraph can have any sub-combination of the following features: verifying accuracy of results of the application of the machine learning model; or wherein the plurality of KPIs comprises at least one of voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic, uplink RSSI levels, downlink RSSI levels, RRC attempts, transmit power, number of voice users, number of data users, parameter settings, pilot power, code utilization, power utilization, or load threshold.
  • the system further comprises a computing system comprising one or more computing devices, the computing system in communication with the computer data repository.
  • the computing system may be programmed to implement a historical performance data collection unit configured to receive the historical performance data, the historical performance data comprising a plurality of key performance indicators (KPIs).
  • KPIs key performance indicators
  • the computing system may be further programmed to implement a network forecast determination unit configured to receive a selection of a first KPI in the plurality of KPIs.
  • the network forecast determination unit may be further configured to apply a machine learning model to at least the first KPI to produce a model parameter and calculate a predicted value of the first KPI for the cell in the network based on the model parameter.
  • the system of the preceding paragraph can have any sub-combination of the following features: where the network forecast determination unit is further configured to generate instructions based on the predicted value of the first KPI; a network control unit configured to instruct a self-organizing network module to adjust a parameter based on the generated instructions; where the network forecast determination unit is further configured to verify accuracy of results of the application of the machine learning model; or where the plurality of KPIs comprises at least one of voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic, uplink RSSI levels, downlink RSSI levels, RRC attempts, transmit power, number of voice users, number of data users, parameter settings, pilot power, code utilization, power utilization, or load threshold.
  • the computer system may, in some cases, include multiple distinct computers or computing devices (e.g., physical servers, workstations, storage arrays, etc.) that communicate and interoperate over a network to perform the described functions.
  • Each such computing device typically includes a hardware processor (or multiple hardware processors) that executes program instructions or modules stored in a memory or other computer-readable storage medium.
  • Each such processor includes digital logic circuitry.
  • the various functions disclosed herein may be embodied in such program instructions, although some or all of the disclosed functions may alternatively be implemented in application-specific circuitry (e.g., ASICs or FPGAs) of the computer system. Where the computer system includes multiple computing devices, these devices may, but need not, be co-located.
  • the results of the disclosed methods and tasks may be persistently stored by transforming physical storage devices, such as solid state memory chips and/or magnetic disks, into a different state.
  • the functional components 110 , 120 , 130 , and 140 shown in FIG. 1 may be implemented by a programmed computer system that comprises one or more physical computers or computing devices. Different components 110 , 120 , 130 , and 140 may, but need not, be implemented on or by different physical machines.
  • the data repositories 150 and 160 shown in FIG. 1 may be implemented as databases, flat files, and/or any other type of computer-based storage system that uses persistent data storage devices to store data.
  • Disjunctive language such as the phrase “at least one of X, Y or Z,” unless specifically stated otherwise, is otherwise understood with the context as used in general to present that an item, term, etc., may be either X, Y or Z, or any combination thereof (e.g., X, Y and/or Z). Thus, such disjunctive language is not generally intended to, and should not, imply that certain embodiments require at least one of X, at least one of Y or at least one of Z to each be present.
  • a device configured to are intended to include one or more recited devices. Such one or more recited devices can also be collectively configured to carry out the stated recitations.
  • a processor configured to carry out recitations A, B and C can include a first processor configured to carry out recitation A working in conjunction with a second processor configured to carry out recitations B and C.

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Telephonic Communication Services (AREA)
  • Mobile Radio Communication Systems (AREA)

Abstract

Untapped capacity and opportunities for immediate performance improvement can be brought to light in wireless networks through the use of new predictive analytics tools and processes. By knowing the specific breaking points in the network well in advance, network adjustments can be planned and implemented in time to preserve a good customer experience. To successfully manage rapidly rising traffic, network operators can adopt a performance-based approach to capacity planning and optimization. Predictive analytics tools and processes may allow a user to view current network conditions for one or more cells in a network. The tools and processes may also allow the user to view predicted network conditions on a chosen future date for one or more cells in the network.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This application claims priority to U.S. Provisional Patent Application No. 61/849,748, entitled “NETWORK PERFORMANCE FORECASTING” and filed on Feb. 4, 2013, the entire contents of which disclosure is hereby incorporated by reference.
BACKGROUND
A cellular network or mobile network can include a wireless network distributed over land areas called cells, each served by at least one fixed-location transceiver known as a cell site or base station. In a cellular network, each cell uses a different set of frequencies or codes from neighboring cells to avoid interference and provide acceptable performance within each cell. When joined together, these cells provide radio coverage over a wide geographic area. This joining of cells enables a large number of portable transceivers (e.g., mobile phones, pagers, etc.) to communicate with each other and with transceivers and telephones anywhere in the network, via base stations, even if some of the transceivers are moving through more than one cell during transmission.
Major telecommunications providers have deployed voice and data cellular networks over most of the inhabited land area of the Earth. This wide deployment allows mobile phones and mobile computing devices to be connected to the public switched telephone network and public Internet. Private cellular networks can be used for research or for large organizations and fleets, such as dispatch for local public safety agencies.
SUMMARY
One aspect of the disclosure provides a computer-implemented method of forecasting wireless network performance. The method comprises receiving historical performance data and baseline data for a cell in a network. The historical performance data may comprise a plurality of key performance indicators (KPIs). The method further comprises receiving, from a user, a selection of a first KPI in the plurality of KPIs. The method further comprises applying a machine learning model to the first KPI and at least one KPI in the plurality of KPIs. The method further comprises generating at least one model parameter based on application of the machine learning model. The method further comprises applying the baseline data to the at least one model parameter. The method further comprises calculating a predicted value of the first KPI for the cell in the network based on the application of the baseline data to the at least one model parameter. At least said calculating may be performed by a computer system comprising computer hardware.
Another aspect of the disclosure provides a system comprising a computer data repository that stores historical performance data for a cell in a network. The system further comprises a computing system comprising one or more computing devices, the computing system in communication with the computer data repository. The computing system may be programmed to implement a historical performance data collection unit configured to receive the historical performance data, the historical performance data comprising a plurality of key performance indicators (KPIs). The computing system may be further programmed to implement a network forecast determination unit configured to receive a selection of a first KPI in the plurality of KPIs. The network forecast determination unit may be further configured to apply a machine learning model to at least the first KPI to produce a model parameter and calculate a predicted value of the first KPI for the cell in the network based on the model parameter.
Another aspect of the disclosure provides a computer storage system comprising a non-transitory storage device, said computer storage system having stored thereon executable program instructions that direct a computer system to at least access historical performance data associated with a telecommunications network, the historical performance data comprising a plurality of key performance indicators (KPIs). The computer storage system further has stored thereon executable program instructions that direct a computer system to at least electronically output a graphical user interface for presentation to a user, the graphical user interface comprising functionality for the user to select one or more of the KPIs. The computer storage system further has stored thereon executable program instructions that direct a computer system to at least receive, from the graphical user interface, a user selection of a first KPI of the plurality of KPIs. The computer storage system further has stored thereon executable program instructions that direct a computer system to at least identify a numerical relationship between the first KPI and one or more other KPIs in the plurality of KPIs. The computer storage system further has stored thereon executable program instructions that direct a computer system to at least calculate a predicted value of one or more of the plurality of KPIs based on the numerical relationship. The computer storage system further has stored thereon executable program instructions that direct a computer system to at least output a representation of the predicted value of the one or more of the plurality of KPIs for presentation to the user.
Certain aspects, advantages and novel features are described herein. It is to be understood that not necessarily all such advantages may be achieved in accordance with any particular embodiment disclosed herein. Thus, the features disclosed herein may be embodied or carried out in a manner that achieves or selects one advantage or group of advantages as taught herein without necessarily achieving other advantages as may be taught or suggested herein.
BRIEF DESCRIPTION OF THE DRAWINGS
Throughout the drawings, reference numbers can be re-used to indicate correspondence between referenced elements. The drawings are provided to illustrate embodiments described herein and not to limit the scope thereof.
FIG. 1 illustrates a block diagram of an example network forecasting system.
FIG. 2 illustrates an example of a more detailed block diagram of the network forecasting server of FIG. 1.
FIG. 3 illustrates an example set of statistical correlation analysis results in a chart.
FIG. 4 illustrates an example analysis of two KPIs.
FIG. 5 illustrates an example matrix generated by the network forecast determination unit.
FIG. 6 illustrates an example graph that compares predicted performance of a KPI and actual performance of a KPI on a cell.
FIG. 7 illustrates an example GUI generated by the GUI unit.
FIG. 8 illustrates another example GUI generated by the GUI unit.
FIG. 9 illustrates another example GUI generated by the GUI unit.
FIG. 10 illustrates another example GUI generated by the GUI unit.
FIG. 11 illustrates a flowchart of an embodiment of a method for forecasting network performance.
DETAILED DESCRIPTION
Introduction
Approximately $100 billion is spent worldwide each year to keep pace with rising traffic on telecommunications networks. However, much of this spending is misspent on new cell sites, spectrum addition, spectrum migration, and/or network equipment that is added too early, too late, or in the wrong place. Telecommunications networks are still being planned and optimized in large part using legacy planning tools and methods, and also gut feel and intuition. Legacy tools for capacity planning often treat all base stations as having equal capacity and/or use simple, fixed thresholds to determine when capacity needs to be added (e.g., if traffic is forecasted to exceed X erlangs or Y megabytes, then a carrier is added). For older technologies (e.g., 2G wireless telephone technology like TDMA, GSM, EDGE, etc.), this approach was sufficient. With newer network technologies (e.g., wireless telephone technologies like CDMA2000, EVDO, UMTS, HSPA+, LTE, WiMAX, etc.), though, cell capacity and performance can vary dramatically from cell to cell depending on various factors. For example, such factors may include the location of users (e.g., near, far, in buildings, etc.), traffic level and mix (e.g., voice, data, etc.), location of neighboring cells, noise from own cells and/or users, noise from neighboring cells and/or users, external interferers, and/or other factors.
Existing third party tools (e.g., propagation prediction and Automatic Cell Planning tools) can predict a crude, generic quality metric over an expected coverage area in a cellular network. However, the tools give no clear correlation between the quality metrics and the key performance indicators (KPIs) measured by network operators and experienced directly by every customer. For example, KPIs may include a dropped call rate, an access failure rate, throughput, latency, and/or the like. The accuracy of typical coverage and quality predictions is degraded by the lack of precise information about the location of users, vegetation in the coverage area, building materials used in the coverage area, and/or the like.
Attempts to overcome this problem by randomly distributing users through simulations (e.g., Monte Carlo simulations) and categorizing areas into discrete morphology types improves accuracy. However, such techniques may not accurately predict performance as experienced by customers. Operators also typically have little visibility into other real-world capacity reducers, such as imbalanced carriers, cells, and/or areas. If this data is available, it is often in table form and a great amount of effort is needed to understand, for example, whether an overloaded site is situated next to an underutilized site. Thus, tools that fill this gap between theoretical predictions and the actual customer experience may be beneficial.
Until recently there has been no accurate way to capture each telecommunication network cell's unique relation between traffic and performance, or to know the level of traffic at which a particular cell's performance will begin to degrade. However, through data mining and smart manipulation of historical performance and network configuration data, each and every carrier on the network can be characterized in terms of its relation between traffic and network performance. For example, a network operator can step any number of months into the future (e.g., 6 months, 7 months, 8 months, 9 months, etc.) and view the predicted network performance at future traffic levels. Then, network changes can be made (manually or automatically) before the customer notices any degradation due to increased traffic (e.g., a proactive approach, rather than the existing reactive approach, to maintaining good network performance). Network operators can now receive advanced notice on what needs to be addressed today or in the future to maintain or improve network performance, even while traffic is rising.
Once each cell's unique breaking point is known, network operators can identify simple, low cost ways to squeeze more capacity from existing network equipment. For example, network operators often highlight large numbers of planned new cell site builds, radio adds, and/or spectrum adds that can be cancelled or deferred. Conversely, areas that will degrade in the near future that have no relief planned can be identified and such degradation issues can be mitigated before customers experience service problems. Spectrum carving and device mix scenarios can be simulated in order to optimize timing of network adjustments and rollouts of new technologies. Networks often have excess capacity with overloaded cells located next to underutilized cells that can be proactively rebalanced through simple, low cost network changes, such as antenna azimuth, downtilt changes, and/or power changes. By analyzing the forecasted network performance, operators can prioritize which new capacity sites to build (which can take 2 or more years), which sites get additional carriers (which can take 6 months or more), the location and quantity of small cells needed, and/or the spectrum required, and plan ahead to get these modifications done on time before KPIs are impacted.
Untapped capacity and opportunities for immediate performance improvement can be brought to light in wireless networks (including cellular networks and other wireless networks) through the use of new predictive analytics tools and processes. By knowing the specific breaking points in the network well in advance, network adjustments can be planned and implemented in time to preserve a good customer experience. To successfully manage rapidly rising traffic, network operators can adopt a performance-based approach to capacity planning and optimization.
The network forecasting tools described herein may be available via a network. For example, a user (e.g., an individual, an employee of a network operator, etc.) may access the network forecasting tools via a computing device that has access to a local area network (LAN), a wide area network (WAN), the Internet, combinations of the same, or the like. The network forecasting tools, hosted by a server or such like device, may allow the user to view current network traffic conditions for one or more cells in a network. The network forecasting tools may also allow the user to view predicted network traffic conditions on a chosen future date for one or more cells in the network. In an embodiment, the predicted network traffic conditions include predicted values for one or more key performance indicators (KPIs), which are described in greater detail below.
As described herein, the server may calculate the predicted network conditions based on historical performance data, an estimated voice and/or data traffic growth rate, frequencies identified as available for use by the network operator in the future and/or the like. Traffic growth may be positive or negative. For example, the server may receive historical performance data for a geographical region that the user is viewing. The user may then identify one or more KPIs that are of interest to the user. The server may use machine learning techniques to produce a set of model parameters. As an example, the model parameters may represent coefficients for various KPIs. The model parameters may be applied to baseline data (e.g., current or observed KPI values) to determine a predicted KPI value for the specific cell in the network. The above techniques may be repeated until the selected KPIs in some or all cells in the geographic region have been predicted.
In an embodiment, the predicted KPIs may be displayed in a graphical user interface (GUI). For example, the predicted KPIs may be displayed over a graphical representation of the geographic region (e.g., on a road or satellite map). Thus, the user may be able to visually identify areas in which performance levels are expected to degrade to unacceptable levels and areas in which performance levels are expected to remain at acceptable levels, and plan accordingly.
In further embodiments, the server may generate instructions based on the predicted KPIs for controlling network equipment. For example, the server may instruct network equipment to perform a power shift on a cell, adjust an antenna pattern, and/or perform like adjustments such that degradations in performance levels can be reduced or prevented.
Further, the systems and methods described herein can be implemented in any of a variety of electronic devices, including, for example, physical or virtual servers, cell phones, smart phones, personal digital assistants (PDAs), tablets, mini-tablets, laptops, desktops, and televisions, to name a few.
For purposes of summarizing this disclosure, certain aspects, advantages and novel features of several embodiments have been described herein. It is to be understood that not necessarily all such advantages can be achieved in accordance with any particular embodiment of the embodiments disclosed herein. Thus, the embodiments disclosed herein can be embodied or carried out in a manner that achieves one advantage or group of advantages as taught herein without necessarily achieving other advantages as taught or suggested herein.
Network Forecasting System Overview
FIG. 1 illustrates a block diagram of an example network forecasting system 100. As illustrated in FIG. 1, the network forecasting system 100 includes one or more user devices 110, a network 120, a network control unit 130, a network forecasting server 140, a performance data store 150, and a site configuration data store 160.
The one or more user devices 110 can each be configured to submit requests for access to network performance forecasts provided by the network forecasting server 140. Such requests can be made via the network 120. For example, the one or more user devices 110 may allow users (e.g., individuals, employees of telecommunication service providers or network operators, etc.) or automatic control units (e.g., self-optimizing network (SON) modules) to interact with the network forecasting information provided by the network forecasting server 140.
In an embodiment, the one or more user devices 110 may be computing devices. For example, one or more of the user devices 110 may be an electronic device, such as a cell phone, a smart phone, a tablet, a laptop, a personal digital assistant (PDA), a computer, a desktop, a workstation, a digital media player, a server, a terminal, a kiosk, or the like. The one or more user devices 110 may include a microphone, a speaker, a wireless module, a camera, and/or a display.
The network control unit 130 can receive instructions to control automatically one or more pieces of network equipment. For example, the network equipment may be automatic control units (e.g., self-organizing network (SON) modules). The network control unit 130 may receive the instructions from the network forecasting server 140 via the network 120. The instructions may instruct network equipment to automatically self-optimize or self-improve parameters or behavior in response to network performance forecasts generated by the network forecasting server 140. For example, the network equipment may be instructed to change parameters, perform a power shift on a cell, adjust remotely-controlled components (e.g., an antenna pattern), and/or perform like adjustments such that degradations in performance levels can be reduced or prevented.
The network forecasting server 140 can generate network performance forecasts. For example, the network forecasting server 140 may generate the network performance forecasts based on cell-specific performance models (e.g., forecasts generated based on performance models for individual cells can be aggregated to provide forecasts for part of or all of an entire network). The cell-specific performance models may be based on historical performance data of telecommunication networks operated by one or more network operators (e.g., carriers). Historical performance data may include historical key performance indicator (KPI) values and other measurements taken from a network. For example, historical performance data may include, but are not limited to, voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic (e.g., neighboring cell traffic), received signal strength indication (RSSI) levels (e.g., uplink RSSI levels, downlink RSSI levels, noise floor at a site, etc.), radio resource control (RRC) attempts, transmit power, number of voice users, number of data users, parameter settings, pilot power, load threshold, code utilization, power utilization, and/or the like.
As described in greater detail below, each network performance forecast predicts a cell's network performance as the cell's level of traffic varies to detect previously hidden network problems and/or to identify an amount of traffic a particular cell can be expected to carry before the cell's network performance KPIs degrade below acceptable thresholds. A network performance forecast may be represented by one or more predicted KPI values (e.g., KPI values expected in the future). Example KPIs may include, but are not limited to, voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic (e.g., neighboring cell traffic), RSSI levels (e.g., uplink RSSI levels, downlink RSSI levels, etc.), RRC attempts, transmit power, number of voice users, number of data users, parameter settings, pilot power, load threshold, code utilization, power utilization, and/or the like. Furthermore, the network performance forecasts can be used to improve daily network troubleshooting and optimization. For example, users of the one or more user devices 110 can view the network performance forecasts to identify cells that are experiencing external interference (e.g., because the cells have a higher uplink noise level than can be explained by normal network traffic). As another example, the network forecasting server 140 can generate instructions based on the network performance forecasts that are used to control the operation of network equipment.
In an embodiment, the network forecasting server 140 can also generate a graphical representation of the network performance forecasts. For example, the network forecasting server 140 may generate a graphical user interface (GUI) that is viewable via a browser or other client application running on the one or more user devices 110.
The network forecasting server 140 may be a computing device. For example, the network forecasting server 140 may include one or more processors to execute one or more instructions, memory, and communication devices to transmit and receive data over the network 120. In some embodiments, the network forecasting server 140 is implemented as one or more backend servers capable of communicating over a network. In other embodiments, the network forecasting server 140 is implemented by one or more virtual machines in a hosted computing environment, such as in a cloud computing environment. The hosted computing environment may include one or more rapidly provisioned and released computing resources, which computing resources may include computing, networking and/or storage devices. A hosted computing environment may also be referred to as a cloud computing environment. In still other embodiments, the network forecasting server 140 may be represented as a user computing device capable of communicating over a network, such as a laptop or tablet computer, personal computer, personal digital assistant (PDA), hybrid PDA/mobile phone, mobile phone, global positioning system (GPS) device, or the like. The network forecasting server 140 may be in communication with one or a plurality of user devices 110.
The network forecasting server 140 may be in communication with the performance data store 150 via the network 120. In an embodiment, the performance data store 150 stores historical performance data of telecommunication networks operated by one or more network operators. The network forecasting server 140 may access historical performance data stored in the performance data store 150 in generating network performance forecasts. In some embodiments, the performance data store 150 is a single data store. In other embodiments, the performance data store 150 includes multiple physical storage devices distributed over many different locations.
In an embodiment, the site configuration data store 160 stores site configuration data for network equipment in telecommunication networks operated by one or more network operators. For example, site configuration data may include, but is not limited to, cell site latitude, longitude, antenna type, azimuth (e.g., direction), frequency, radio network controller (RNC), and/or the like. In some embodiments, the site configuration data store 160 is a single data store. In other embodiments, the site configuration data store 160 includes multiple physical storage devices distributed over many different locations. The site configuration data may be used to depict the network (e.g., pie wedges as described below) in a user interface (e.g., the example user interfaces in FIGS. 7-10) and/or may be used when determining network performance forecasts (e.g., may be used as inputs in a machine learning model).
The network 120 may be a wired network, a wireless network, or a combination of the two. For example, the network 120 may be a personal area network, a local area network (LAN), a wide area network (WAN), or combinations of the same. Protocols and components for communicating via any of the other aforementioned types of communication networks, such as the TCP/IP protocols, can be used in the network 120.
Network Performance Forecasts
FIG. 2 illustrates an example of a more detailed block diagram of the network forecasting server 140 of FIG. 1. As illustrated in FIG. 2, the network forecasting server 140 may include a historical performance data collection unit 210, a network forecast determination unit 220, and a graphical user interface (GUI) unit 230.
In an embodiment, the historical performance data collection unit 210 can collect historical performance data stored in the performance data store 150. The historical performance data collection unit 210 may collect historical performance data for a set period of time (e.g., one week, one month, one year, etc.). The historical performance data collected may depend on the user requesting permission to access the network performance forecasts. For example, the historical performance data collection unit 210 may collect historical performance data associated with a first carrier if the user requesting permission to access the network performance forecasts is associated with the first carrier. In some embodiments, the historical performance data collection unit 210 collects the historical performance data at the same time or nearly the same time as the network forecasting server 140 receives a request for network performance forecasts. In other embodiments, the historical performance data collection unit 210 collects the historical performance data at a time before the network forecasting server 140 receives a request for network performance forecasts.
The network forecast determination unit 220 can generate network performance forecasts. Each network performance forecast may be specific to a cell in a cellular network or may encompass a plurality of cells in the cellular network. The network forecast determination unit 220 may be configured to generate a plurality of network performance forecasts to cover a geographic area selected by a user. In an embodiment, the network forecast determination unit 220 receives a selection of one or more KPIs. For example, a user operating a user device 110 may select one or more KPIs that are of interest (e.g., a user may be interested in the impact of rising traffic on the dropped call rate).
In some embodiments, once one or more KPIs are selected, the network forecast determination unit 220 identifies relationships between KPIs before beginning the process of generating network performance forecasts. For example, the relationships between KPIs may be unknown (e.g., network performance forecasts are to be generated for a wireless standard or network for the first time). In other embodiments, the network forecast determination unit 220 begins generating network performance forecasts once one or more KPIs are selected. For example, the relationships between KPIs may have been determined already at a previous time or received from another device.
The network forecast determination unit 220 may identify relationships between KPIs using machine-learning techniques on one or more KPIs. Machine-learning techniques may include statistical correlation analyses, regression analyses (e.g., multiple regression analyses, linear regression analyses, logistic regression analyses, etc.), neural networks, combinations of the same, or the like. In some embodiments, machine learning is performed on at least one selected KPI and at least one unselected KPI. In other embodiments, machine-learning is performed on only the selected KPIs. As an example, the network forecast determination unit 220 may use a statistical correlation analysis that compares each analyzed KPI with some or all other analyzed KPIs. As used herein, the term “correlation,” in addition to having its ordinary meaning, can refer to calculation of a specific correlation value or coefficient or more generally to a numerical comparison (e.g., of KPIs) that does not necessarily require calculation of a mathematically-defined correlation coefficient. The result of the statistical correlation analysis may indicate the sensitivity of each analyzed KPI to the other analyzed KPIs. For example, the network forecast determination unit 220 may analyze the result of the statistical correlation analysis to determine which KPIs are most sensitive to other KPIs. The network forecast determination unit 220 may focus the analysis of the results on the one or more selected KPIs. In other embodiments, the statistical analysis that may be performed by the network forecast determination unit 220 may also include other statistical parameters or operations, such as a mean, median, mode, variance, standard deviation, regression, or other numerical calculation based on one or more KPIs.
In an embodiment, the network forecast determination unit 220 determines a relationship between two or more KPIs based on the result of the statistical correlation analysis. For example, the data set of each selected KPI may be compared with the data sets for one or more remaining KPIs individually (e.g., a pair of KPIs may be compared at a time). A KPI data set may include KPI values collected over a period of time (e.g., hourly, daily, etc.). The network forecast determination unit 220 may mathematically determine whether the relationship between KPIs is linear, exponential, logarithmic, or has some other type of relationship.
In some embodiments, the network forecast determination unit 220 generates one or more charts that include a result of the KPI comparisons. FIG. 3 illustrates an example chart 300 that depicts a set of example statistical correlation analysis results. The chart 300 indicates, for example, that data traffic and uplink RSSI KPIs have a positive correlation and data traffic and voice access success rate KPIs have a negative correlation. The chart 300 may be provided to an analyst (e.g., an individual trained to identify relationships between KPIs based on machine learning results), not shown, for further analysis. The results of the analysis performed by the analyst may be relayed to the one or more user devices 110 via the network forecasting server 140.
FIG. 4 illustrates an example chart 400 depicting analysis of two example KPIs. As illustrated in FIG. 4, uplink RSSI 410 is plotted against data access rate 420. The uplink RSSI 410 and the data access rate 420 have a linear relationship in the depicted example.
In an embodiment, the network forecast determination unit 220 uses the identified relationships when generating the network performance forecasts. Machine learning may also be used to generate the network performance forecasts. For example, the network forecast determination unit 220 may use a multiple regression analysis on one or more KPIs.
For example, the network forecast determination unit 220 may perform a multiple regression analysis with KPIs used as independent and dependent variables. In some embodiments, the relationships identified between KPIs are used as a guide to determine which KPIs are used as independent variables and which KPIs are used as dependent variables (e.g., KPIs that have a strong correlation may be used as independent variables, etc.). As an example, each selected KPI may be used as a dependent variable and some or all of the remaining KPIs may be used as independent variables. The independent variables may be isolated from semi-independent and dependent variables as this may play a useful role in the selection of historical performance data and KPIs. The network forecast determination unit 220 may improve the machine learning model by using interaction terms or other corrective factors (e.g., adjusting the source data to exponential or other non-linear methods to capture the inter-KPI relationship observed when mathematically analyzing or charting the data). The quality of the machine learning model may be measured using a coefficient of determination (R2), standard error, the Pearson correlation coefficient, or other measurements.
In an embodiment, the network forecast determination unit 220 sets filter criteria to manipulate the historical performance data. The filter criteria may be used to manipulate the historical performance data before any such data is used to generate the network performance forecasts. For example, the filter criteria may be set based on a baseline range of measurements taken in a network. If a portion of the historical performance data falls outside of the baseline range, the portion may be removed or adjusted. The filter criteria may vary based on the KPI, technology used by the carrier, and/or the carrier requested network performance forecasts. In addition, individual cell busy hour, market busy hour, 24 hour, or other timeframes may be selected to vary the loading performance of a cell based on voice erlangs and traffic megabytes.
The network forecast determination unit 220 may generate nested machine learning models. In an embodiment, the network forecast determination unit 220 recursively applies machine learning. For example, the network forecast determination unit 220 may apply machine learning to generate a first output. Machine learning may be applied again to generate additional outputs. Machine learning then may be applied again, with the first and/or additional outputs being used as coefficients for a dependent or independent variable. In addition, the network forecast determination unit 220 may correlate noise level measurements and power measurements with at least some performance measurements and at least some KPIs to better predict performance measurements or semi-independent KPIs to the noise level. In an embodiment, the network forecast determination unit 220 correlates the noise level measurement and the power measurement to the final performance KPIs.
The network forecast determination unit 220 may revise the machine learning model used to generate the network performance forecasts. Revisions may occur via the addition of independent variables (e.g., KPIs, parameter settings, etc.) or the removal of independent variables. In an embodiment, the network forecast determination unit 220 filters and/or weights the machine learning model input data to exclude incorrect data or data which is degrading the model. Once any filtering or weighting is complete, the machine learning is performed and key model parameters are recorded for the respective cell. For example, key model parameters (e.g., when a multiple regression analysis is performed) may include the intercept and/or coefficients for all independent variables.
Once the network forecast determination unit 220 performs the machine learning, the network forecast determination unit 220 may generate a matrix for each cell in the geographic area selected by the user. FIG. 5 illustrates an example matrix 500 generated by the network forecast determination unit 220. The matrix may include example machine learning model information (e.g., the key model parameters), which can be used to predict each important or desired KPI. In an embodiment, the network forecast determination unit 220 applies the matrix coefficients to baseline data for the respective cell to determine one or more predicted KPIs (e.g., an estimated value for a KPI selected by the user). The baseline data for a cell may include measured performance data for on-air cells and/or estimated data for new cells not yet on-air. The baseline data may be stored in the performance data store 150, stored in another data store (not shown), and/or provided by the user. For example, in a multiple regression analysis model applied to an on-air cell, the network forecast determination unit 220 may calculate a predicted KPI as the intercept of the model plus the product of each independent variable coefficient and the cell's historical baseline data. As a further example, to calculate the predicted KPI as traffic is varied, the network forecast determination unit 220 can increase or decrease the traffic baseline data and can apply the matrix coefficients to the new traffic baseline data. An example predicted dropped call rate for one cell using a multiple regression analysis model is provided below in equation (1) (with coefficients derived from the model, which may vary in other implementations):
Dropped Call Rate = 370.6 - 4.691 ( total traffic ) + 1.302 ( UL RSSI ) - 0.05191 ( total traffic × UL RSSI ) - 0.5418 ( power ) + 0.2986 ( load threshold ) + 1.361 ( offset ) ( 1 )
In an embodiment, the network forecast determination unit 220 performs a verification test on one or more KPIs in a cell (e.g., a verification test is performed on each cell's machine learning model). The verification test may ensure or attempt to ensure that as traffic increases, the performance KPIs degrade. If the verification test indicates that a cell's model is not responding as expected, the cell may be labeled as a bad cell. A cell's model may fail the verification test as a result of missing counters, missing statistics, outages, or other reasons. Data for the bad cell may be replaced with new data and new models may be generated using the new data. If the new models fail the verification test, then the network forecast determination unit 220 may assign a default model to the cell. The default model may be generated from RNC, base station controller (BSC), mobility management entity (MME), and/or area models that depict similar cells in similar morphology and/or topology.
FIG. 6 illustrates an example graph 600 that compares predicted performance 620 of a KPI and actual performance 610 of a KPI on a cell with respect to time. As illustrated in FIG. 6, a dramatic rise in traffic occurred around 12:00, causing congestion. Graph 600 indicates that the predicted performance 620 may be very close to the actual performance 610 (e.g., R2=0.996, standard error=2.7%).
The network forecast determination unit 220 may send instructions derived from the predicted KPI values to the network control unit 130. The network control unit 130 may instruct a SON module to direct automatic changes to the network, such as parameter or power changes or adjustments to remotely-controlled components (e.g., antennas).
The network forecast determination unit 220 may also send the predicted KPI values to the one or more user devices 110 and/or the GUI unit 230. In some embodiments, the predicted KPI values are provided to the one or more user devices 110 in a table format (e.g., via an exportable file). In other embodiments, the predicted KPI values are output visually in a GUI generated by the GUI unit 230 and accessible via a browser or other client application running on the one of more user devices 110. Example GUIs generated by the GUI unit 230 are described in greater detail below with respect to FIGS. 7-10.
Example GUIs
FIGS. 7 through 10 depict example user interfaces or GUIs that may be output by the systems or processes described above. Each of the user interfaces shown includes one or more user interface controls that can be selected by a user, for example, using a browser or other application software. The user interface controls shown are merely illustrative examples and can be varied in other embodiments. For instance, buttons, dropdown boxes, select boxes, text boxes, check boxes, slider controls, and other user interface controls shown, may be substituted with other types of user interface controls that provide the same or similar functionality. Further, user interface controls may be combined or divided into other sets of user interface controls such that similar functionality or the same functionality may be provided with different looking user interfaces. Moreover, each of the user interface controls may be selected by a user using one or more input options, such as a mouse, touch screen input, or keyboard input, among other user interface input options.
FIG. 7 illustrates an example GUI 700 that may be generated by the GUI unit 230. As illustrated in FIG. 7, the GUI 700 includes several interactive features that allow a user to view network performance forecasts for a current date or a date in the future. For example, a user can select one or more RNCs using dropdown box 702, view one or more map views using options 704 (e.g., road view, aerial view, bird's eye view, etc.), view sector details (e.g., sector information, predicted KPI values, etc.) in window 706 by hovering over or selecting a sector in the GUI 700, change which spectrum frequencies (or carriers or bands) are viewable using button 712, hide or view labels 708 depicting planned capacity (e.g., a planned future carrier count in a cell of a sector) using button 716, adjust annual voice and/or data traffic growth rate via box 718, adjust a date of the desired network performance forecast using slider 710 (e.g., the slider 710 may step forward and/or backward in daily increments), back button 726, or forward button 727, select play button 728 to view KPIs change over time, select user growth button 714 to view the network operator's forecasted traffic, show additional KPIs simultaneously by selecting an available KPI in box 724 (e.g., circuit switch (CS) retainability percentage, packet switched (PS) accessibility percentage, PS retainability percentage, throughput, etc.), and/or remove KPIs by selecting a KPI in box 722 (e.g., CS accessibility percentage).
In addition, a user can display one or more carriers simultaneously, choose to view KPIs in 24 hour or market busy hour timeframes, and/or export data (e.g., predicted KPIs for a chosen future date and growth rate, first date at which each cell will fall below a threshold that indicates an acceptable KPI performance level, cells or sectors that are the worst offenders, etc.). In some embodiments, each cell in a sector is represented by color-coded pie wedges to indicate the KPI performance level (e.g., for the selected KPI(s)). For example, green may indicate an acceptable and high KPI performance level, yellow may indicate an acceptable and medium KPI performance level, and red may indicate an unacceptable and low KPI performance level. As described above, each cell may include a label 708 that includes a number depicting planned capacity in the respective cell. A user may also carve spectrum, which removes one or more frequencies and distributes the traffic from those frequencies to the remaining frequencies and shows the predicted performance on the remaining frequencies.
FIG. 8 illustrates another example GUI 800 generated by the GUI unit 230. As illustrated in FIG. 8, a sector, such as sector 804, includes one or more cells in a pie-shaped configuration. Each cell may be divided into concentric segments or rings that each represent the performance level of different frequencies. For example, three frequencies (e.g., channels 637, 2062, and 2087) are displayed in each cell. Cell 802 is displayed as having an acceptable and high KPI performance level. In addition, in box 818, the annual voice traffic growth rate is set at 10% and the annual data traffic growth rate is set at 100%. Slider 810 is set at a current date (e.g., Nov. 1, 2013).
FIG. 9 illustrates another example GUI 900 generated by the GUI unit 230. As illustrated in FIG. 9, the slider 810 is stepped forward one year (e.g., to Nov. 1, 2014). Accordingly, the predicted KPI values in each cell are adjusted based on the analysis performed by the network forecast determination unit 220 described above and the growth rates selected by the user. In some embodiments, the predicted KPI values are calculated as the slider 810 is adjusted. In other embodiments, some or all possible predicted KPI values are calculated before the GUI 900 is presented to the user. As an example, a frequency (e.g., channel 2062) of the cell 802 is displayed as having an acceptable and medium KPI performance level once slider 810 is adjusted from the position as illustrated in FIG. 8.
FIG. 10 illustrates another example GUI 1000 generated by the GUI unit 230. As illustrated in FIG. 10, the slider 810 remains in the same position as illustrated in FIG. 9. However, the spectrum carve feature has been selected such that one frequency is removed (e.g., channel 637) and the traffic from the removed frequency is distributed to the remaining frequencies equally, nearly equally, or based on expected loads and/or traffic patterns. For example, sector 1006 includes three cells. The inner concentric ring of each cell in sector 1006 is shown as being transparent (e.g., rather than a color indicating performance levels), indicating that the frequency band associated with that concentric ring has been removed. Due to the removal of spectrum, an expected load on the cell 802, and/or other factors, a frequency (e.g., channel 2062) of the cell 802 is now displayed as having an unacceptable and low KPI performance level. However, removal of spectrum does not necessarily mean that the performance level of a cell will degrade. Depending on an expected load on a cell and/or other factors, a cell may be able to maintain performance levels even with the removal of spectrum. For example, channel 637 is also carved from the cell 1008 and distributed to the remaining frequencies. Yet, the performance levels of the remaining frequencies are unchanged despite the removal of spectrum.
Flowchart
FIG. 11 illustrates a flowchart of an example method 1100 for forecasting network performance. In an embodiment, the method 1100 can be performed by the network forecasting server 140 discussed above with respect to FIGS. 1 and 2 or any other computing device. Depending on the embodiment, the method 1100 may include fewer and/or additional blocks and the blocks may be performed in an order different than illustrated.
In block 1102, historical performance data and baseline data are received. In an embodiment, the historical performance data includes a plurality of KPIs.
In block 1104, a selection of a first KPI in the plurality of KPIs is received. In an embodiment, the selection may be received from a user operating a user device, such as a user device 110.
In block 1106, a machine learning model is applied. In an embodiment, the machine learning model is applied to the first KPI and at least one KPI in the plurality of KPIs. In a further embodiment, the machine learning model is a multiple regression analysis model, although the machine learning model may be any of the machine learning models described above.
In block 1108, at least one model parameter is generated based on the application of the machine learning model. In an embodiment, a model parameter may be a coefficient produced by the machine learning model, where the coefficient is of an independent variable used in the machine learning model.
In block 1110, the baseline data is applied to the at least one model parameter. In block 1112, a predicted value of the first KPI is calculated based on the application of the baseline data to the at least one model parameter.
ADDITIONAL EMBODIMENTS
One aspect of the disclosure provides a computer-implemented method of forecasting wireless network performance. The method comprises receiving historical performance data and baseline data for a cell in a network. The historical performance data may comprise a plurality of key performance indicators (KPIs). The method further comprises receiving, from a user, a selection of a first KPI in the plurality of KPIs. The method further comprises applying a machine learning model to the first KPI and at least one KPI in the plurality of KPIs. The method further comprises generating at least one model parameter based on application of the machine learning model. The method further comprises applying the baseline data to the at least one model parameter. The method further comprises calculating a predicted value of the first KPI for the cell in the network based on the application of the baseline data to the at least one model parameter. At least said calculating may be performed by a computer system comprising computer hardware.
The method of the preceding paragraph can have any sub-combination of the following features: verifying accuracy of results of the application of the machine learning model; or wherein the plurality of KPIs comprises at least one of voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic, uplink RSSI levels, downlink RSSI levels, RRC attempts, transmit power, number of voice users, number of data users, parameter settings, pilot power, code utilization, power utilization, or load threshold.
Another aspect of the disclosure provides a system comprising a computer data repository that stores historical performance data for a cell in a network. The system further comprises a computing system comprising one or more computing devices, the computing system in communication with the computer data repository. The computing system may be programmed to implement a historical performance data collection unit configured to receive the historical performance data, the historical performance data comprising a plurality of key performance indicators (KPIs). The computing system may be further programmed to implement a network forecast determination unit configured to receive a selection of a first KPI in the plurality of KPIs. The network forecast determination unit may be further configured to apply a machine learning model to at least the first KPI to produce a model parameter and calculate a predicted value of the first KPI for the cell in the network based on the model parameter.
The system of the preceding paragraph can have any sub-combination of the following features: where the network forecast determination unit is further configured to generate instructions based on the predicted value of the first KPI; a network control unit configured to instruct a self-organizing network module to adjust a parameter based on the generated instructions; where the network forecast determination unit is further configured to verify accuracy of results of the application of the machine learning model; or where the plurality of KPIs comprises at least one of voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic, uplink RSSI levels, downlink RSSI levels, RRC attempts, transmit power, number of voice users, number of data users, parameter settings, pilot power, code utilization, power utilization, or load threshold.
Terminology
Some or all of the methods and tasks described above may be performed and fully automated by a computer system. The computer system may, in some cases, include multiple distinct computers or computing devices (e.g., physical servers, workstations, storage arrays, etc.) that communicate and interoperate over a network to perform the described functions. Each such computing device typically includes a hardware processor (or multiple hardware processors) that executes program instructions or modules stored in a memory or other computer-readable storage medium. Each such processor includes digital logic circuitry. The various functions disclosed herein may be embodied in such program instructions, although some or all of the disclosed functions may alternatively be implemented in application-specific circuitry (e.g., ASICs or FPGAs) of the computer system. Where the computer system includes multiple computing devices, these devices may, but need not, be co-located. The results of the disclosed methods and tasks may be persistently stored by transforming physical storage devices, such as solid state memory chips and/or magnetic disks, into a different state.
For example, the functional components 110, 120, 130, and 140 shown in FIG. 1 may be implemented by a programmed computer system that comprises one or more physical computers or computing devices. Different components 110, 120, 130, and 140 may, but need not, be implemented on or by different physical machines. The data repositories 150 and 160 shown in FIG. 1 may be implemented as databases, flat files, and/or any other type of computer-based storage system that uses persistent data storage devices to store data.
Conditional language used herein, such as, among others, “can,” “could,” “might,” “may,” “e.g.,” and the like, unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without author input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment. The terms “comprising,” “including,” “having,” and the like are synonymous and are used inclusively, in an open-ended fashion, and do not exclude additional elements, features, acts, operations, and so forth. Also, the term “or” is used in its inclusive sense (and not in its exclusive sense) so that when used, for example, to connect a list of elements, the term “or” means one, some, or all of the elements in the list.
Disjunctive language such as the phrase “at least one of X, Y or Z,” unless specifically stated otherwise, is otherwise understood with the context as used in general to present that an item, term, etc., may be either X, Y or Z, or any combination thereof (e.g., X, Y and/or Z). Thus, such disjunctive language is not generally intended to, and should not, imply that certain embodiments require at least one of X, at least one of Y or at least one of Z to each be present.
Unless otherwise explicitly stated, articles such as ‘a’ or ‘an’ should generally be interpreted to include one or more described items. Accordingly, phrases such as “a device configured to” are intended to include one or more recited devices. Such one or more recited devices can also be collectively configured to carry out the stated recitations. For example, “a processor configured to carry out recitations A, B and C” can include a first processor configured to carry out recitation A working in conjunction with a second processor configured to carry out recitations B and C.
The foregoing embodiments are intended to be non-limiting. Other embodiments, including embodiments that do not provide all of the benefits and features set forth herein, are also within the scope of protection. Accordingly, the scope of protection is defined only by reference to the appended claims.

Claims (21)

What is claimed is:
1. A computer-implemented method of forecasting wireless network performance, the method comprising:
receiving historical performance data and baseline data for a cell in a network, wherein the historical performance data comprises a plurality of key performance indicators (KPIs);
receiving, from a user, a selection of a first KPI in the plurality of KPIs;
applying a machine learning model to the first KPI and a second KPI in the plurality of KPIs to identify a relationship between the first KPI and the second KPI;
applying the machine learning model to the first KPI and a third KPI in the plurality of KPIs to identify a second relationship between the first KPI and the third KPI;
determining that a correlation between the first KPI and the third KPI is greater than a threshold correlation value based on the identified second relationship;
generating a first model parameter based on the identified relationship between the first KPI and the second KPI;
generating a second model parameter based on the identified second relationship between the first KPI and the third KPI in response to the determination that the correlation between the first KPI and the third KPI is greater than the threshold correlation value;
applying the baseline data to the first model parameter and the second model parameter; and
calculating a predicted value of the first KPI for the cell in the network based on the application of the baseline data to the first model parameter and the second model parameter;
wherein at least said calculating is performed by a computer system comprising computer hardware.
2. The computer-implemented method of claim 1, further comprising:
receiving a selection of a date in the future;
receiving a selection of an estimated traffic growth rate;
adjusting the baseline data based on the selected date and the estimated traffic growth rate;
applying the adjusted baseline data to the first model parameter; and
calculating a second predicted value of the first KPI for the cell in the network based on the application of the adjusted baseline data to the first model parameter.
3. The computer-implemented method of claim 1, wherein the machine learning model comprises a statistical correlation analysis model.
4. The computer-implemented method of claim 1, further comprising:
filtering the historical performance data based on a subset of the historical performance data identified as being incorrect; and
applying the machine learning model based on the filtered historical performance data.
5. The computer-implemented method of claim 1, further comprising transmitting data to an application running on a general purpose computer, the data including information for displaying the predicted value of the first KPI in a graphical user interface.
6. The computer-implemented method of claim 5, wherein the data includes information for displaying the predicted value of the first KPI over a graphical representation of a geographic location.
7. The computer-implemented method of claim 1, wherein the machine learning model comprises a multiple regression analysis model.
8. The computer-implemented method of claim 1, wherein the plurality of KPIs comprises at least one of voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic, uplink received signal strength indication (RSSI) levels, downlink RSSI levels, radio resource control (RRC) attempts, transmit power, number of voice users, number of data users, or parameter settings.
9. The computer-implemented method of claim 1, further comprising:
receiving, from the user, an indication to remove a first frequency from the cell, wherein the cell comprises the first frequency, a second frequency, and a third frequency, and wherein the predicted value of the first KPI for the cell comprises a first predicted value of the first KPI for the first frequency, a second predicted value of the second KPI for the second frequency, and a third predicted value of the first KPI for the third frequency;
distributing network traffic associated with the first frequency to the second frequency and the third frequency;
determining an updated second predicted value of the first KPI and an updated third predicted value of the first KPI based on the distribution of the network traffic and based on the application of the baseline data to the first model parameter and the second model parameter.
10. The computer-implemented method of claim 1, wherein the second KPI comprises one of data traffic, voice traffic, or number of users, wherein the performance data comprises a first value for the second KPI at a first time, wherein the baseline data comprises a second value for the second KPI, wherein the second value corresponds to a second time different from the first time, and wherein the computer-implemented method further comprises:
receiving, from the user, a selection of a third value for the second KPI, wherein the third value corresponds to the second time, and wherein the third value is double the first value;
modifying the baseline data such that the modified baseline data comprises the third value for the second KPI; and
calculating the predicted value of the first KPI for the cell in the network based on at least a product of the first model parameter and the third value.
11. A system comprising:
a computer data repository that stores historical performance data for a cell in a network;
a second computer data repository that stores baseline data for the cell in the network;
a computing system comprising one or more computing devices, the computing system in communication with the computer data repository, the computing system programmed to implement:
a historical performance data collection unit configured to receive the historical performance data and the baseline data, the historical performance data comprising a plurality of key performance indicators (KPIs); and
a network forecast determination unit configured to receive a selection of a first KPI in the plurality of KPIs, the network forecast determination unit further configured to:
apply a machine learning model to at least the first KPI and a second KPI in the plurality of KPIs to identify a relationship between the first KPI and the second KPI,
apply the machine learning model to the first KPI and a third KPI in the plurality of KPIs to identify a second relationship between the first KPI and the third KPI,
determine that a correlation between the first KPI and the third KPI is greater than a threshold correlation value based on the identified second relationship,
generate a first model parameter based on the identified relationship between the first KPI and the second KPI,
generate a second model parameter based on the identified second relationship between the first KPI and the third KPI in response to the determination that the correlation between the first KPI and the third KPI is greater than the threshold correlation value,
apply the baseline data to the first model parameter and the second model parameter, and
calculate a predicted value of the first KPI for the cell in the network based on the application of the baseline data to the first model parameter and the second model parameter.
12. The system of claim 11, wherein the network forecast determination unit is further configured to:
receive a selection of a date in the future;
receive a selection of an estimated traffic growth rate;
adjust the baseline data based on the selected date and the estimated traffic growth rate;
apply the adjusted baseline data to the first model parameter; and
calculate a second predicted value of the first KPI for the cell in the network based on the application of the adjusted baseline data to the first model parameter.
13. The system of claim 11, wherein the computing system is further programmed to implement a graphical user interface unit configured to transmit data to an application running on a general purpose computer, the data including information for displaying the predicted value of the first KPI in a graphical user interface.
14. The system of claim 13, wherein the data includes information for displaying the predicted value of the first KPI over a graphical representation of a geographic location.
15. The system of claim 11, wherein the machine learning model comprises a multiple regression analysis model.
16. The system of claim 11, wherein the plurality of KPIs comprises at least one of voice dropped call rates, data dropped call rates, successful call establishment rate, successful session establishment rate, throughput, voice traffic, data traffic, neighbor traffic, uplink received signal strength indication (RSSI) levels, downlink RSSI levels, radio resource control (RRC) attempts, transmit power, number of voice users, number of data users, or parameter settings.
17. A computer storage system comprising a non-transitory storage device, said computer storage system having stored thereon executable program instructions that direct a computer system to at least:
access historical performance data and baseline data for a cell in a telecommunications network, the historical performance data comprising a plurality of key performance indicators (KPIs);
electronically output a graphical user interface for presentation to a user, the graphical user interface comprising functionality for the user to select one or more of the KPIs;
receive, from the graphical user interface, a user selection of a first KPI of the plurality of KPIs;
apply a machine learning model to the first KPI and a second KPI in the plurality of KPIs to identify a relationship between the first KPI and the second KPI;
apply the machine learning model to the first KPI and a third KPI in the plurality of KPIs to identify a second relationship between the first KPI and the third KPI;
determine that a correlation between the first KPI and the third KPI is greater than a threshold correlation value based on the identified second relationship;
generate a first model parameter based on the identified relationship between the first KPI and the second KPI;
generate a second model parameter based on the identified second relationship between the first KPI and the third KPI in response to the determination that the correlation between the first KPI and the third KPI is greater than the threshold correlation value;
apply the baseline data to the first model parameter and the second model parameter;
calculate a predicted value of the first KPI for the cell in the telecommunications network based on the application of the baseline data to the first model parameter and the second model parameter; and
output a representation of the predicted value of the first KPI for presentation to the user.
18. The computer storage system of claim 17, wherein the predicted value of the first KPI is calculated for a first date.
19. The computer storage system of claim 18, wherein the computer storage system further has stored thereon executable program instructions that direct the computer system to at least calculate a second predicted value of the first KPI for a second date.
20. The computer storage system of claim 19, wherein the graphical user interface comprises functionality for the user to adjust a date.
21. The computer storage system of claim 20, wherein the computer storage system further has stored thereon executable program instructions that direct the computer system to at least:
receive, from the graphical user interface, an adjustment from the first date to the second date; and
output a representation of the second predicted value of the first KPI for presentation to the user.
US14/172,686 2013-02-04 2014-02-04 Systems and methods for network performance forecasting Active 2034-10-05 US9439081B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/172,686 US9439081B1 (en) 2013-02-04 2014-02-04 Systems and methods for network performance forecasting

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201361849748P 2013-02-04 2013-02-04
US14/172,686 US9439081B1 (en) 2013-02-04 2014-02-04 Systems and methods for network performance forecasting

Publications (1)

Publication Number Publication Date
US9439081B1 true US9439081B1 (en) 2016-09-06

Family

ID=56878108

Family Applications (1)

Application Number Title Priority Date Filing Date
US14/172,686 Active 2034-10-05 US9439081B1 (en) 2013-02-04 2014-02-04 Systems and methods for network performance forecasting

Country Status (1)

Country Link
US (1) US9439081B1 (en)

Cited By (66)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160323804A1 (en) * 2014-01-09 2016-11-03 Nec Corporation Node device
US20170164334A1 (en) * 2015-12-08 2017-06-08 At&T Intellectual Property I, Lp. Method and apparatus for transmitting a control signal to a self organizing network controller
US20170351689A1 (en) * 2016-06-06 2017-12-07 Gauthaman Vasudevan System and method for automated key-performance-indicator discovery
EP3382938A1 (en) * 2017-03-31 2018-10-03 Telefonica Digital España, S.L.U. A computer implemented method, a system and computer programs to forecast the performance of a network
US20190028909A1 (en) * 2017-07-20 2019-01-24 Cisco Technology, Inc. Adaptive health status scoring for network assurance
US10200877B1 (en) * 2015-05-14 2019-02-05 Roger Ray Skidmore Systems and methods for telecommunications network design, improvement, expansion, and deployment
EP3438899A1 (en) * 2017-07-31 2019-02-06 Accenture Global Solutions Limited Using machine learning to make network management decisions
US10210189B2 (en) * 2017-01-06 2019-02-19 International Business Machines Corporation Root cause analysis of performance problems
US10291498B1 (en) * 2017-11-14 2019-05-14 Sprint Communications Company L.P. Mobile communication device diagnostic client and error remediation sharing
EP3496015A1 (en) * 2017-12-07 2019-06-12 Accenture Global Solutions Limited Data transformation of performance statistics and ticket information for network devices for use in machine learning models
CN109983798A (en) * 2016-12-08 2019-07-05 华为技术有限公司 The prediction of performance indicator in cellular network
CN110098966A (en) * 2019-05-13 2019-08-06 中国联合网络通信集团有限公司 Wireless network capacitance analysis method and equipment
US10484255B2 (en) * 2017-06-19 2019-11-19 Cisco Technology, Inc. Trustworthiness index computation in a network assurance system based on data source health monitoring
US20190387422A1 (en) * 2017-11-21 2019-12-19 At&T Intellectual Property I, L.P. Network reconfiguration using genetic algorithm-based predictive models
US10555191B1 (en) 2019-08-01 2020-02-04 T-Mobile Usa, Inc. Optimum network performance improvement solutions selection systems and methods
WO2020033424A1 (en) * 2018-08-06 2020-02-13 Intel Corporation Management data analytical kpis for 5g network traffic and resource
US20200057933A1 (en) * 2018-08-14 2020-02-20 Ca, Inc. Neural network predicting communications network infrastructure outages based on forecasted performance metrics
WO2020070739A1 (en) * 2018-10-02 2020-04-09 Cellwize Wireless Technologies Ltd. Method of controlling traffic in a cellular network and system thereof
CN111133795A (en) * 2017-09-26 2020-05-08 株式会社Ntt都科摩 Cell selection or access method, user terminal, maintenance method and base station
WO2020111748A1 (en) 2018-11-26 2020-06-04 Samsung Electronics Co., Ltd. Methods and apparatus for coverage prediction and network optimization in 5g new radio networks
CN111372259A (en) * 2018-12-26 2020-07-03 华为技术有限公司 Performance evaluation method and device of wireless system
US10764184B2 (en) * 2018-07-17 2020-09-01 Facebook, Inc. Detecting communication network insights of alerts
WO2020218956A1 (en) * 2019-04-24 2020-10-29 Telefonaktiebolaget Lm Ericsson (Publ) First network node, and method performed thereby, for handling a performance of a communications network
US10887778B2 (en) 2017-12-22 2021-01-05 At&T Intellectual Property I, L.P. Proactively adjusting network infrastructure in response to reporting of real-time network performance
US10884844B2 (en) * 2015-11-06 2021-01-05 HomeAway.com, Inc. Data stream processor and method to counteract anomalies in data streams transiting a distributed computing system
US20210012047A1 (en) * 2017-08-10 2021-01-14 Allstate Insurance Company Multi-Platform Model Processing and Execution Management Engine
CN112232559A (en) * 2020-10-12 2021-01-15 国网江西省电力有限公司信息通信分公司 A kind of short-term forecasting method and device of electric power area load
US10904114B2 (en) 2019-01-31 2021-01-26 Cisco Technology, Inc. KPI trajectory-driven outlier detection in a network assurance service
CN112425122A (en) * 2018-07-17 2021-02-26 脸谱公司 Communication network insight to detect alerts
US10944641B1 (en) * 2019-11-01 2021-03-09 Cisco Technology, Inc. Systems and methods for application traffic simulation using captured flows
US11037251B2 (en) * 2017-03-01 2021-06-15 Microsoft Technology Licensing, Llc Understanding business insights and deep-dive using artificial intelligence
US11064382B1 (en) 2020-05-07 2021-07-13 T-Mobile Usa, Inc. Management of telecommunications network congestion on roadways
US20210224699A1 (en) * 2020-01-21 2021-07-22 Hcl Technologies Limited Building time series based prediction / forecast model for a telecommunication network
KR20210092305A (en) * 2019-05-22 2021-07-23 후아웨이 테크놀러지 컴퍼니 리미티드 Network simulation platform creation method, network simulation method and corresponding device
CN113453261A (en) * 2021-06-25 2021-09-28 维沃移动通信有限公司 Abnormal cell identification method and device and electronic equipment
US11153765B1 (en) 2020-05-15 2021-10-19 T-Mobile Usa, Inc. Capacity planning of telecommunications network by detecting anomalies in site behavior
US20220014424A1 (en) * 2020-07-09 2022-01-13 Telefonaktiebolaget Lm Ericsson (Publ) Cell accessibility prediction and actuation
CN114124733A (en) * 2020-08-27 2022-03-01 中国电信股份有限公司 Service flow prediction method and device
CN114189882A (en) * 2020-09-14 2022-03-15 中国移动通信有限公司研究院 Information transmission method, device, related equipment and storage medium
US11283889B2 (en) * 2013-03-14 2022-03-22 Comcast Cable Communications, Llc Systems and methods for abandonment detection and mitigation
US20220131625A1 (en) * 2019-02-04 2022-04-28 Nokia Solutions And Networks Oy Selecting Uplink Transmission Band in Wireless Network
US11343683B2 (en) 2020-04-22 2022-05-24 T-Mobile Usa, Inc. Identification and prioritization of optimum capacity solutions in a telecommunications network
US11350289B2 (en) 2020-05-14 2022-05-31 T-Mobile Usa, Inc. Identification of indoor and outdoor traffic usage of customers of a telecommunications network
US11375463B2 (en) * 2020-11-30 2022-06-28 Silicon Laboratories Inc. System, apparatus and method for dynamic transmit power determination
WO2022144207A1 (en) 2020-12-28 2022-07-07 Telecom Italia S.P.A. Cellular communication system featuring son functionality
US11405296B1 (en) * 2020-06-12 2022-08-02 Amazon Technologies, Inc. Automated validation of network matrices
US11412390B2 (en) * 2019-11-29 2022-08-09 Jio Platforms Limited System and method of automatic outdoor small cell planning
US20220279452A1 (en) * 2021-03-01 2022-09-01 At&T Intellectual Property I, L.P. Method and system for controlling downlink transmit power
CN115209443A (en) * 2021-04-12 2022-10-18 中国移动通信集团设计院有限公司 Wireless network capacity expansion analysis method and device and electronic equipment
WO2022250454A1 (en) * 2021-05-27 2022-12-01 Samsung Electronics Co., Ltd. Method and system for managing robust header compression (rohc) in a wireless network
US11558271B2 (en) * 2019-09-04 2023-01-17 Cisco Technology, Inc. System and method of comparing time periods before and after a network temporal event
IT202100032966A1 (en) 2021-12-29 2023-06-29 Telecom Italia Spa CELLULAR COMMUNICATION SYSTEM WITH ADVANCED SON FUNCTIONALITY
US20230209367A1 (en) * 2021-12-29 2023-06-29 T-Mobile Innovations Llc Telecommunications network predictions based on machine learning using aggregated network key performance indicators
WO2023132851A1 (en) * 2022-01-05 2023-07-13 Rakuten Mobile, Inc. Unified coverage system
WO2023130787A1 (en) * 2022-01-05 2023-07-13 Huawei Technologies Co., Ltd. Method and apparatus for managing network traffic via uncertainty
US11800398B2 (en) 2021-10-27 2023-10-24 T-Mobile Usa, Inc. Predicting an attribute of an immature wireless telecommunication network, such as a 5G network
US11870559B1 (en) 2021-12-15 2024-01-09 Juniper Networks, Inc. Identifying and correlating metrics associated with unhealthy key performance indicators
US11979311B2 (en) 2021-12-10 2024-05-07 Cisco Technology, Inc. User-assisted training data denoising for predictive systems
US20240171479A1 (en) * 2022-11-22 2024-05-23 At&T Intellectual Property I, L.P. Methods, systems, and devices for scalable and layered architecture for real-time key performance indicator (kpi) prediction in mobile networks
WO2024188472A1 (en) * 2023-03-16 2024-09-19 Telefonaktiebolaget Lm Ericsson (Publ) Ml model based hw-aware power management in communication networks
WO2025004090A1 (en) * 2023-06-29 2025-01-02 Jio Platforms Limited System and method for automated change impact analysis
US12192790B2 (en) 2021-10-27 2025-01-07 T-Mobile Usa, Inc. Predicting an attribute of a wireless telecommunication network, such as a 5G network
WO2025013078A1 (en) * 2023-07-13 2025-01-16 Jio Platforms Limited System and method for forecasting expansion of a network
US12204399B1 (en) * 2024-06-27 2025-01-21 Paul Hightower Revise animation control program (TAC)
US12212988B2 (en) 2022-08-09 2025-01-28 T-Mobile Usa, Inc. Identifying a performance issue associated with a 5G wireless telecommunication network
WO2025032608A1 (en) * 2023-08-05 2025-02-13 Jio Platforms Limited System and method for managing operations in a network

Citations (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6754487B1 (en) 2000-02-28 2004-06-22 Telecom Network Optimization, Inc. Radio network test analysis system
US6996374B1 (en) * 2002-07-30 2006-02-07 Cellco Partnership Sector capacity prediction
WO2006060358A2 (en) 2004-12-03 2006-06-08 Interdigital Technology Corporation Method for improving wireless network performance in a multi-cell communication network
US20070192065A1 (en) * 2006-02-14 2007-08-16 Sun Microsystems, Inc. Embedded performance forecasting of network devices
US20080109731A1 (en) * 2006-06-16 2008-05-08 Groundhog Technologies Inc. Management system and method for wireless communication network and associated graphic user interface
US20080144905A1 (en) * 2006-09-29 2008-06-19 Tony Tallman Novel display adjustment features
WO2008107020A1 (en) 2007-03-08 2008-09-12 Telefonaktiebolaget L M Ericsson (Publ) An arrangement and a method relating to performance monitoring
WO2008121062A1 (en) 2007-03-29 2008-10-09 Telefonaktiebolaget Lm Ericsson (Publ) Method and apparatus for evaluating services in communication networks
WO2008131473A1 (en) 2007-04-30 2008-11-06 Ubo Wireless Pty Limited Wireless broadband network management
WO2009004661A2 (en) 2007-07-02 2009-01-08 Salvatore Lucifora Method and computer system for performance analysis and for dimensioning a gsm access network
US20110149782A1 (en) * 2007-12-05 2011-06-23 Cellco Partnership D/B/A Verizon Wireless Methodology to analyze sector capacity in data-only mobile-wireless network
EP2372937A2 (en) 2010-03-29 2011-10-05 Vodafone Group PLC Improving network performance in a cellular telecommunications network
US20120303413A1 (en) * 2011-05-27 2012-11-29 Vpisystems Inc. Methods and systems for network traffic forecast and analysis
EP2557830A1 (en) 2011-08-11 2013-02-13 Deutsche Telekom AG Method for predicting a MIMO throughput of an LTE downlink in a radio network planning tool
WO2013026469A1 (en) 2011-08-22 2013-02-28 Nokia Siemens Networks Oy Method and device for processing a communication network
US20130143561A1 (en) * 2011-11-17 2013-06-06 Intucell Ltd. Method and system for optimizing cellular networks operation
US8538787B2 (en) 2007-06-18 2013-09-17 International Business Machines Corporation Implementing key performance indicators in a service model
WO2013144950A1 (en) 2012-03-25 2013-10-03 Intucell Ltd. System and method for optimizing performance of a communication network
US20130281100A1 (en) 2010-12-30 2013-10-24 Telecom Italia S.P.A. Method for the prediction of coverage areas of a cellular network
US20130281089A1 (en) 2012-04-23 2013-10-24 Accelera Mobile Broadband, Inc. Interference management and network performance optimization in small cells
US20130291034A1 (en) * 2012-04-27 2013-10-31 General Instrument Corporation Network Monitoring with Estimation of Network Path to Network Element Location
US20130331109A1 (en) 2012-06-08 2013-12-12 T-Mobile Usa, Inc. Cell Cluster Network Performance Optimizations
US20140113600A1 (en) * 2010-09-28 2014-04-24 The Ohio State University Predictive network system and method
US20140111517A1 (en) * 2012-10-22 2014-04-24 United States Cellular Corporation Detecting and processing anomalous parameter data points by a mobile wireless data network forecasting system
US20140205207A1 (en) * 2013-01-21 2014-07-24 Apple Inc. Techniques for presenting user adjustments to a digital image

Patent Citations (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6754487B1 (en) 2000-02-28 2004-06-22 Telecom Network Optimization, Inc. Radio network test analysis system
US6996374B1 (en) * 2002-07-30 2006-02-07 Cellco Partnership Sector capacity prediction
WO2006060358A2 (en) 2004-12-03 2006-06-08 Interdigital Technology Corporation Method for improving wireless network performance in a multi-cell communication network
US20070192065A1 (en) * 2006-02-14 2007-08-16 Sun Microsystems, Inc. Embedded performance forecasting of network devices
US20080109731A1 (en) * 2006-06-16 2008-05-08 Groundhog Technologies Inc. Management system and method for wireless communication network and associated graphic user interface
US20080144905A1 (en) * 2006-09-29 2008-06-19 Tony Tallman Novel display adjustment features
WO2008107020A1 (en) 2007-03-08 2008-09-12 Telefonaktiebolaget L M Ericsson (Publ) An arrangement and a method relating to performance monitoring
WO2008121062A1 (en) 2007-03-29 2008-10-09 Telefonaktiebolaget Lm Ericsson (Publ) Method and apparatus for evaluating services in communication networks
WO2008131473A1 (en) 2007-04-30 2008-11-06 Ubo Wireless Pty Limited Wireless broadband network management
US8538787B2 (en) 2007-06-18 2013-09-17 International Business Machines Corporation Implementing key performance indicators in a service model
WO2009004661A2 (en) 2007-07-02 2009-01-08 Salvatore Lucifora Method and computer system for performance analysis and for dimensioning a gsm access network
US20110149782A1 (en) * 2007-12-05 2011-06-23 Cellco Partnership D/B/A Verizon Wireless Methodology to analyze sector capacity in data-only mobile-wireless network
EP2372937A2 (en) 2010-03-29 2011-10-05 Vodafone Group PLC Improving network performance in a cellular telecommunications network
US20140113600A1 (en) * 2010-09-28 2014-04-24 The Ohio State University Predictive network system and method
US20130281100A1 (en) 2010-12-30 2013-10-24 Telecom Italia S.P.A. Method for the prediction of coverage areas of a cellular network
US20120303413A1 (en) * 2011-05-27 2012-11-29 Vpisystems Inc. Methods and systems for network traffic forecast and analysis
EP2557830A1 (en) 2011-08-11 2013-02-13 Deutsche Telekom AG Method for predicting a MIMO throughput of an LTE downlink in a radio network planning tool
WO2013026469A1 (en) 2011-08-22 2013-02-28 Nokia Siemens Networks Oy Method and device for processing a communication network
US20130143561A1 (en) * 2011-11-17 2013-06-06 Intucell Ltd. Method and system for optimizing cellular networks operation
WO2013144950A1 (en) 2012-03-25 2013-10-03 Intucell Ltd. System and method for optimizing performance of a communication network
US20130281089A1 (en) 2012-04-23 2013-10-24 Accelera Mobile Broadband, Inc. Interference management and network performance optimization in small cells
US20130291034A1 (en) * 2012-04-27 2013-10-31 General Instrument Corporation Network Monitoring with Estimation of Network Path to Network Element Location
US20130331109A1 (en) 2012-06-08 2013-12-12 T-Mobile Usa, Inc. Cell Cluster Network Performance Optimizations
US20140111517A1 (en) * 2012-10-22 2014-04-24 United States Cellular Corporation Detecting and processing anomalous parameter data points by a mobile wireless data network forecasting system
US20140205207A1 (en) * 2013-01-21 2014-07-24 Apple Inc. Techniques for presenting user adjustments to a digital image

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Eduardo Yusta Padilla et al., "Method and system for predicting the channel usage", EP 2 750 432 A1, Date of filing: Dec. 28, 2012, Date of publication: Jul. 2, 2014. (pertinent date prohibits use as art). *

Cited By (108)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11283889B2 (en) * 2013-03-14 2022-03-22 Comcast Cable Communications, Llc Systems and methods for abandonment detection and mitigation
US10225786B2 (en) * 2014-01-09 2019-03-05 Nec Corporation Delay tolerant network (DTN) and ad-hoc node device
US20160323804A1 (en) * 2014-01-09 2016-11-03 Nec Corporation Node device
US10200877B1 (en) * 2015-05-14 2019-02-05 Roger Ray Skidmore Systems and methods for telecommunications network design, improvement, expansion, and deployment
US10884844B2 (en) * 2015-11-06 2021-01-05 HomeAway.com, Inc. Data stream processor and method to counteract anomalies in data streams transiting a distributed computing system
US20170164334A1 (en) * 2015-12-08 2017-06-08 At&T Intellectual Property I, Lp. Method and apparatus for transmitting a control signal to a self organizing network controller
US10834616B2 (en) * 2015-12-08 2020-11-10 At&T Intellectual Property I, L.P. Method and apparatus for transmitting a control signal to a self organizing network controller
US20190191320A1 (en) * 2015-12-08 2019-06-20 At&T Intellectual Property I, L.P. Method and apparatus for transmitting a control signal to a self organizing network controller
US10219261B2 (en) * 2015-12-08 2019-02-26 At&T Mobility Ii Llc Method and apparatus for transmitting a control signal to a self organizing network controller
US11416533B2 (en) * 2016-06-06 2022-08-16 Avlino Inc. System and method for automated key-performance-indicator discovery
US20170351689A1 (en) * 2016-06-06 2017-12-07 Gauthaman Vasudevan System and method for automated key-performance-indicator discovery
CN109983798A (en) * 2016-12-08 2019-07-05 华为技术有限公司 The prediction of performance indicator in cellular network
EP3539316A4 (en) * 2016-12-08 2019-09-18 Huawei Technologies Co., Ltd. PREDICTING PERFORMANCE INDICATORS IN CELLULAR NETWORKS
US10552390B2 (en) 2017-01-06 2020-02-04 International Business Machines Corporation Root cause analysis of performance problems
US10210189B2 (en) * 2017-01-06 2019-02-19 International Business Machines Corporation Root cause analysis of performance problems
US11037251B2 (en) * 2017-03-01 2021-06-15 Microsoft Technology Licensing, Llc Understanding business insights and deep-dive using artificial intelligence
EP3382938A1 (en) * 2017-03-31 2018-10-03 Telefonica Digital España, S.L.U. A computer implemented method, a system and computer programs to forecast the performance of a network
US10484255B2 (en) * 2017-06-19 2019-11-19 Cisco Technology, Inc. Trustworthiness index computation in a network assurance system based on data source health monitoring
US20190028909A1 (en) * 2017-07-20 2019-01-24 Cisco Technology, Inc. Adaptive health status scoring for network assurance
US10735273B2 (en) 2017-07-31 2020-08-04 Accenture Global Solutions Limited Using machine learning to make network management decisions
EP3438899A1 (en) * 2017-07-31 2019-02-06 Accenture Global Solutions Limited Using machine learning to make network management decisions
US20210012047A1 (en) * 2017-08-10 2021-01-14 Allstate Insurance Company Multi-Platform Model Processing and Execution Management Engine
US12190026B2 (en) * 2017-08-10 2025-01-07 Allstate Insurance Company Multi-platform model processing and execution management engine
CN111133795A (en) * 2017-09-26 2020-05-08 株式会社Ntt都科摩 Cell selection or access method, user terminal, maintenance method and base station
US10291498B1 (en) * 2017-11-14 2019-05-14 Sprint Communications Company L.P. Mobile communication device diagnostic client and error remediation sharing
US20190387422A1 (en) * 2017-11-21 2019-12-19 At&T Intellectual Property I, L.P. Network reconfiguration using genetic algorithm-based predictive models
US10945145B2 (en) * 2017-11-21 2021-03-09 At&T Intellectual Property I, L.P. Network reconfiguration using genetic algorithm-based predictive models
US10693740B2 (en) 2017-12-07 2020-06-23 Accenture Global Solutions Limited Data transformation of performance statistics and ticket information for network devices for use in machine learning models
EP3496015A1 (en) * 2017-12-07 2019-06-12 Accenture Global Solutions Limited Data transformation of performance statistics and ticket information for network devices for use in machine learning models
US20240340663A1 (en) * 2017-12-22 2024-10-10 At&T Intellectual Property I, L.P. Proactively adjusting network infrastructure in response to reporting of real-time network performance
US10887778B2 (en) 2017-12-22 2021-01-05 At&T Intellectual Property I, L.P. Proactively adjusting network infrastructure in response to reporting of real-time network performance
US11477668B2 (en) 2017-12-22 2022-10-18 At&T Intellectual Property I, L.P. Proactively adjusting network infrastructure in response to reporting of real-time network performance
US12015936B2 (en) 2017-12-22 2024-06-18 At&T Intellectual Property I, L.P. Proactively adjusting network infrastructure in response to reporting of real-time network performance
US11258709B2 (en) * 2018-07-17 2022-02-22 Meta Platforms, Inc. Detecting communication network insights of alerts
CN112425122A (en) * 2018-07-17 2021-02-26 脸谱公司 Communication network insight to detect alerts
US10764184B2 (en) * 2018-07-17 2020-09-01 Facebook, Inc. Detecting communication network insights of alerts
WO2020033424A1 (en) * 2018-08-06 2020-02-13 Intel Corporation Management data analytical kpis for 5g network traffic and resource
CN112534778A (en) * 2018-08-06 2021-03-19 苹果公司 Management data analysis KPI for 5G network traffic and resources
US20210160147A1 (en) * 2018-08-06 2021-05-27 Apple Inc. Management data analytical kpis for 5g network traffic and resource
US11729067B2 (en) * 2018-08-06 2023-08-15 Apple Inc. Management data analytical KPIS for 5G network traffic and resource
CN112534778B (en) * 2018-08-06 2023-08-22 苹果公司 Management data analytics KPIs for 5G network traffic and resources
US20200057933A1 (en) * 2018-08-14 2020-02-20 Ca, Inc. Neural network predicting communications network infrastructure outages based on forecasted performance metrics
US11228924B2 (en) 2018-10-02 2022-01-18 Cellwize Wireless Technologies Ltd. Method of controlling traffic in a cellular network and system thereof
EP4142345A1 (en) * 2018-10-02 2023-03-01 Cellwize Wireless Technologies Ltd. Method of controlling traffic in a cellular network and system thereof
WO2020070739A1 (en) * 2018-10-02 2020-04-09 Cellwize Wireless Technologies Ltd. Method of controlling traffic in a cellular network and system thereof
EP3868143A4 (en) * 2018-11-26 2021-12-29 Samsung Electronics Co., Ltd. Methods and apparatus for coverage prediction and network optimization in 5g new radio networks
US11259191B2 (en) 2018-11-26 2022-02-22 Samsung Electronics Co., Ltd. Methods and apparatus for coverage prediction and network optimization in 5G new radio networks
WO2020111748A1 (en) 2018-11-26 2020-06-04 Samsung Electronics Co., Ltd. Methods and apparatus for coverage prediction and network optimization in 5g new radio networks
CN111372259A (en) * 2018-12-26 2020-07-03 华为技术有限公司 Performance evaluation method and device of wireless system
US10904114B2 (en) 2019-01-31 2021-01-26 Cisco Technology, Inc. KPI trajectory-driven outlier detection in a network assurance service
US20220131625A1 (en) * 2019-02-04 2022-04-28 Nokia Solutions And Networks Oy Selecting Uplink Transmission Band in Wireless Network
US12052065B2 (en) * 2019-02-04 2024-07-30 Nokia Solutions And Networks Oy Selecting uplink transmission band in wireless network
WO2020218956A1 (en) * 2019-04-24 2020-10-29 Telefonaktiebolaget Lm Ericsson (Publ) First network node, and method performed thereby, for handling a performance of a communications network
CN110098966A (en) * 2019-05-13 2019-08-06 中国联合网络通信集团有限公司 Wireless network capacitance analysis method and equipment
US20210306881A1 (en) * 2019-05-22 2021-09-30 Huawei Technologies Co., Ltd. Method for creating network simulation platform, network simulation method, and corresponding apparatuses
US11856424B2 (en) * 2019-05-22 2023-12-26 Huawei Technologies Co., Ltd. Method for creating network simulation platform, network simulation method, and corresponding apparatuses
EP3879758A4 (en) * 2019-05-22 2022-03-09 Huawei Technologies Co., Ltd. METHOD FOR CREATING A NETWORK SIMULATION PLATFORM AND NETWORK SIMULATION METHOD AND RELATED DEVICE
KR20210092305A (en) * 2019-05-22 2021-07-23 후아웨이 테크놀러지 컴퍼니 리미티드 Network simulation platform creation method, network simulation method and corresponding device
US11146974B2 (en) 2019-08-01 2021-10-12 T-Mobile Usa, Inc. Optimum network performance improvement solutions selection systems and methods
US10555191B1 (en) 2019-08-01 2020-02-04 T-Mobile Usa, Inc. Optimum network performance improvement solutions selection systems and methods
US11558271B2 (en) * 2019-09-04 2023-01-17 Cisco Technology, Inc. System and method of comparing time periods before and after a network temporal event
US10944641B1 (en) * 2019-11-01 2021-03-09 Cisco Technology, Inc. Systems and methods for application traffic simulation using captured flows
US11412390B2 (en) * 2019-11-29 2022-08-09 Jio Platforms Limited System and method of automatic outdoor small cell planning
US11729631B2 (en) 2019-11-29 2023-08-15 Jio Platforms Limited System and method of automatic outdoor small cell planning
US11900282B2 (en) * 2020-01-21 2024-02-13 Hcl Technologies Limited Building time series based prediction / forecast model for a telecommunication network
US20210224699A1 (en) * 2020-01-21 2021-07-22 Hcl Technologies Limited Building time series based prediction / forecast model for a telecommunication network
US11792662B2 (en) * 2020-04-22 2023-10-17 T-Mobile Usa, Inc. Identification and prioritization of optimum capacity solutions in a telecommunications network
US20220256364A1 (en) * 2020-04-22 2022-08-11 T-Mobile Usa, Inc. Identification and prioritization of optimum capacity solutions in a telecommunications network
US11343683B2 (en) 2020-04-22 2022-05-24 T-Mobile Usa, Inc. Identification and prioritization of optimum capacity solutions in a telecommunications network
US11418993B2 (en) 2020-05-07 2022-08-16 T-Mobile Usa, Inc. Management of telecommunications network congestion on roadways
US11064382B1 (en) 2020-05-07 2021-07-13 T-Mobile Usa, Inc. Management of telecommunications network congestion on roadways
US11350289B2 (en) 2020-05-14 2022-05-31 T-Mobile Usa, Inc. Identification of indoor and outdoor traffic usage of customers of a telecommunications network
US11678200B2 (en) 2020-05-14 2023-06-13 T-Mobile Usa, Inc. Identification of indoor and outdoor traffic usage of customers of a telecommunications network
US12200506B2 (en) 2020-05-14 2025-01-14 T-Mobile Usa, Inc. Identification of indoor and outdoor traffic usage of customers of a telecommunications network
US11153765B1 (en) 2020-05-15 2021-10-19 T-Mobile Usa, Inc. Capacity planning of telecommunications network by detecting anomalies in site behavior
US11405296B1 (en) * 2020-06-12 2022-08-02 Amazon Technologies, Inc. Automated validation of network matrices
US11271797B2 (en) * 2020-07-09 2022-03-08 Telefonaktiebolaget Lm Ericsson (Publ) Cell accessibility prediction and actuation
US20220014424A1 (en) * 2020-07-09 2022-01-13 Telefonaktiebolaget Lm Ericsson (Publ) Cell accessibility prediction and actuation
CN114124733B (en) * 2020-08-27 2024-05-14 中国电信股份有限公司 Service flow prediction method and device
CN114124733A (en) * 2020-08-27 2022-03-01 中国电信股份有限公司 Service flow prediction method and device
CN114189882A (en) * 2020-09-14 2022-03-15 中国移动通信有限公司研究院 Information transmission method, device, related equipment and storage medium
CN112232559A (en) * 2020-10-12 2021-01-15 国网江西省电力有限公司信息通信分公司 A kind of short-term forecasting method and device of electric power area load
CN112232559B (en) * 2020-10-12 2023-06-13 国网江西省电力有限公司信息通信分公司 A short-term forecasting method and device for power regional load
US11375463B2 (en) * 2020-11-30 2022-06-28 Silicon Laboratories Inc. System, apparatus and method for dynamic transmit power determination
WO2022144207A1 (en) 2020-12-28 2022-07-07 Telecom Italia S.P.A. Cellular communication system featuring son functionality
US11582070B2 (en) * 2021-03-01 2023-02-14 At&T Intellectual Property I, L.P. Method and system for controlling downlink transmit power
US20220279452A1 (en) * 2021-03-01 2022-09-01 At&T Intellectual Property I, L.P. Method and system for controlling downlink transmit power
CN115209443A (en) * 2021-04-12 2022-10-18 中国移动通信集团设计院有限公司 Wireless network capacity expansion analysis method and device and electronic equipment
WO2022250454A1 (en) * 2021-05-27 2022-12-01 Samsung Electronics Co., Ltd. Method and system for managing robust header compression (rohc) in a wireless network
CN113453261A (en) * 2021-06-25 2021-09-28 维沃移动通信有限公司 Abnormal cell identification method and device and electronic equipment
US12133109B2 (en) 2021-10-27 2024-10-29 T-Mobile Usa, Inc. Predicting an attribute of an immature wireless telecommunication network, such as a 5G network
US12192790B2 (en) 2021-10-27 2025-01-07 T-Mobile Usa, Inc. Predicting an attribute of a wireless telecommunication network, such as a 5G network
US11800398B2 (en) 2021-10-27 2023-10-24 T-Mobile Usa, Inc. Predicting an attribute of an immature wireless telecommunication network, such as a 5G network
US11979311B2 (en) 2021-12-10 2024-05-07 Cisco Technology, Inc. User-assisted training data denoising for predictive systems
US11870559B1 (en) 2021-12-15 2024-01-09 Juniper Networks, Inc. Identifying and correlating metrics associated with unhealthy key performance indicators
US12212478B1 (en) 2021-12-15 2025-01-28 Juniper Networks, Inc. Identifying and correlating metrics associated with unhealthy key performance indicators
WO2023126185A1 (en) 2021-12-29 2023-07-06 Telecom Italia S.P.A. Cellular communication system featuring advanced son functionality
US20230209367A1 (en) * 2021-12-29 2023-06-29 T-Mobile Innovations Llc Telecommunications network predictions based on machine learning using aggregated network key performance indicators
IT202100032966A1 (en) 2021-12-29 2023-06-29 Telecom Italia Spa CELLULAR COMMUNICATION SYSTEM WITH ADVANCED SON FUNCTIONALITY
WO2023130787A1 (en) * 2022-01-05 2023-07-13 Huawei Technologies Co., Ltd. Method and apparatus for managing network traffic via uncertainty
WO2023132851A1 (en) * 2022-01-05 2023-07-13 Rakuten Mobile, Inc. Unified coverage system
US12212988B2 (en) 2022-08-09 2025-01-28 T-Mobile Usa, Inc. Identifying a performance issue associated with a 5G wireless telecommunication network
US20240171479A1 (en) * 2022-11-22 2024-05-23 At&T Intellectual Property I, L.P. Methods, systems, and devices for scalable and layered architecture for real-time key performance indicator (kpi) prediction in mobile networks
WO2024188472A1 (en) * 2023-03-16 2024-09-19 Telefonaktiebolaget Lm Ericsson (Publ) Ml model based hw-aware power management in communication networks
WO2025004090A1 (en) * 2023-06-29 2025-01-02 Jio Platforms Limited System and method for automated change impact analysis
WO2025013078A1 (en) * 2023-07-13 2025-01-16 Jio Platforms Limited System and method for forecasting expansion of a network
WO2025032608A1 (en) * 2023-08-05 2025-02-13 Jio Platforms Limited System and method for managing operations in a network
US12204399B1 (en) * 2024-06-27 2025-01-21 Paul Hightower Revise animation control program (TAC)

Similar Documents

Publication Publication Date Title
US9439081B1 (en) Systems and methods for network performance forecasting
US10039016B1 (en) Machine-learning-based RF optimization
US9204319B2 (en) Estimating long term evolution network capacity and performance
JP6411435B2 (en) Method and system for maintaining or optimizing a cellular network
US10477426B1 (en) Identifying a cell site as a target for utilizing 5th generation (5G) network technologies and upgrading the cell site to implement the 5G network technologies
US10440503B2 (en) Machine learning-based geolocation and hotspot area identification
US9848337B2 (en) Metro cell planning
US7929459B2 (en) Method and apparatus for automatically determining the manner in which to allocate available capital to achieve a desired level of network quality performance
WO2017016472A1 (en) Predicting network performance
US20130324076A1 (en) Method & system for cellular network load balance
EP2750432A1 (en) Method and system for predicting the channel usage
EP2934037B1 (en) Technique for Evaluation of a Parameter Adjustment in a Mobile Communications Network
US20220256364A1 (en) Identification and prioritization of optimum capacity solutions in a telecommunications network
US10645604B2 (en) Intelligent optimization of cells in a mobile network
US11811598B2 (en) Systems and methods for modifying device operation based on datasets generated using machine learning techniques
US10327165B2 (en) Methods and systems for monitoring mobile networks
EP2681967B1 (en) Resource managing method, resource management device and apparatus for supporting operation of a radio communication network
WO2017108106A1 (en) Method and network node for identifiying specific area of wireless communication system
Emeruwa Analysis of standalone dedicated control channel congestion for 4g networks in Yenagoa–Southern Nigeria
KR20240040445A (en) Method for optimizing wireless network and management server using the same

Legal Events

Date Code Title Description
AS Assignment

Owner name: FURTHER LLC, PENNSYLVANIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KNEBL, MATTHEW BRIAN;ALBANNA, AMR;REEL/FRAME:032872/0500

Effective date: 20140502

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: OOKLA, LLC, NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:FURTHER LLC;REEL/FRAME:050749/0145

Effective date: 20191001

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

AS Assignment

Owner name: MUFG UNION BANK, N.A., AS COLLATERAL AGENT, ARIZONA

Free format text: SECURITY INTEREST;ASSIGNORS:EVERYDAY HEALTH, INC.;KEEPITSAFE, INC.;OOKLA, LLC;AND OTHERS;REEL/FRAME:056969/0755

Effective date: 20210407

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8