WO1995034981A2 - Enhancement of network operation and performance - Google Patents
Enhancement of network operation and performance Download PDFInfo
- Publication number
- WO1995034981A2 WO1995034981A2 PCT/SE1995/000704 SE9500704W WO9534981A2 WO 1995034981 A2 WO1995034981 A2 WO 1995034981A2 SE 9500704 W SE9500704 W SE 9500704W WO 9534981 A2 WO9534981 A2 WO 9534981A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- accordance
- logical
- logical networks
- objective function
- network
- Prior art date
Links
- 238000000034 method Methods 0.000 claims abstract description 118
- 230000005540 biological transmission Effects 0.000 claims abstract description 38
- 238000005457 optimization Methods 0.000 claims abstract description 38
- 238000000638 solvent extraction Methods 0.000 claims abstract description 27
- 230000000903 blocking effect Effects 0.000 claims description 24
- 238000004891 communication Methods 0.000 claims description 15
- 238000012886 linear function Methods 0.000 claims description 11
- 230000008569 process Effects 0.000 claims description 11
- 230000004044 response Effects 0.000 claims description 5
- 239000000872 buffer Substances 0.000 claims description 2
- 238000005315 distribution function Methods 0.000 claims description 2
- 230000002441 reversible effect Effects 0.000 claims 1
- 230000006870 function Effects 0.000 description 53
- 230000009194 climbing Effects 0.000 description 18
- 238000013461 design Methods 0.000 description 13
- 238000010586 diagram Methods 0.000 description 12
- 230000014509 gene expression Effects 0.000 description 7
- 238000013459 approach Methods 0.000 description 6
- 239000013256 coordination polymer Substances 0.000 description 6
- 238000007726 management method Methods 0.000 description 6
- 230000008859 change Effects 0.000 description 5
- 238000000926 separation method Methods 0.000 description 5
- 238000013528 artificial neural network Methods 0.000 description 4
- 238000005192 partition Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 230000010354 integration Effects 0.000 description 3
- 238000013468 resource allocation Methods 0.000 description 3
- 230000001360 synchronised effect Effects 0.000 description 3
- 238000012546 transfer Methods 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000000593 degrading effect Effects 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/32—Specific management aspects for broadband networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/08—Configuration management of networks or network elements
- H04L41/0896—Bandwidth or capacity management, i.e. automatically increasing or decreasing capacities
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/14—Network analysis or design
- H04L41/142—Network analysis or design using statistical or mathematical methods
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04J—MULTIPLEX COMMUNICATION
- H04J2203/00—Aspects of optical multiplex systems other than those covered by H04J14/05 and H04J14/07
- H04J2203/0001—Provisions for broadband connections in integrated services digital network using frames of the Optical Transport Network [OTN] or using synchronous transfer mode [STM], e.g. SONET, SDH
- H04J2203/0057—Operations, administration and maintenance [OAM]
- H04J2203/0058—Network management, e.g. Intelligent nets
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04J—MULTIPLEX COMMUNICATION
- H04J2203/00—Aspects of optical multiplex systems other than those covered by H04J14/05 and H04J14/07
- H04J2203/0001—Provisions for broadband connections in integrated services digital network using frames of the Optical Transport Network [OTN] or using synchronous transfer mode [STM], e.g. SONET, SDH
- H04J2203/0064—Admission Control
- H04J2203/0067—Resource management and allocation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04J—MULTIPLEX COMMUNICATION
- H04J2203/00—Aspects of optical multiplex systems other than those covered by H04J14/05 and H04J14/07
- H04J2203/0001—Provisions for broadband connections in integrated services digital network using frames of the Optical Transport Network [OTN] or using synchronous transfer mode [STM], e.g. SONET, SDH
- H04J2203/0064—Admission Control
- H04J2203/0067—Resource management and allocation
- H04J2203/0071—Monitoring
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04J—MULTIPLEX COMMUNICATION
- H04J2203/00—Aspects of optical multiplex systems other than those covered by H04J14/05 and H04J14/07
- H04J2203/0001—Provisions for broadband connections in integrated services digital network using frames of the Optical Transport Network [OTN] or using synchronous transfer mode [STM], e.g. SONET, SDH
- H04J2203/0073—Services, e.g. multimedia, GOS, QOS
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/14—Network analysis or design
- H04L41/145—Network analysis or design involving simulating, designing, planning or modelling of a network
Definitions
- the present invention relates to telecommunication networks and in particular to overall network performance.
- a main characteristic of a modern telecommunication network is its ability to provide different services.
- One efficient way of providing said services is to logically separate the resources of a physical network - resource separation (see Fig. 1).
- a physical network PN On top of a physical network PN there is established a number of logical networks LN, also referred to as logical or virtual subnetworks, each of which comprises nodes N and logical links LL interconnecting the nodes.
- Each logical network forms a logical view of parts of the physical network or of the complete physical network.
- a first logical network LN1 comprises one view of parts of the physical network and a second logical network LN2 comprises another view, different from that of the first logical network.
- the logical links of the various logical networks share the capacities of physical links present in said physical network.
- a physical network comprises switches S (physical nodes) or equivalents, physical links interconnecting said switches, and various auxiliary devices.
- a physical link utilizes transmission equipment, such as fiber optic conductors, coaxial cables or radio links.
- transmission equipment such as fiber optic conductors, coaxial cables or radio links.
- physical links are grouped into trunk groups TG which extend between said switches.
- access points to the physical network to which access points access units, such as telephone sets and computer modems, are connected.
- Each physical link has limited transmission capacity.
- FIG. 2 is a simple schematic drawing explaining the relationship between physical links, logical links and also routes.
- a simple underlying physical network with physical switches S and trunk groups TG, i.e. physical links, interconnecting the switches is illustrated.
- On top of this physical network a number of logical networks are established, only one of which is shown in the drawing.
- the logical networks can be established by a network manager, a network operator or other organization.
- the logical networks can be established by a network manager, a network operator or other organization.
- Swedish Patent Application 9403035-0 incorporated herein by reference, there is described a method of creating and configuring logical networks.
- the single logical network shown comprises logical nodes N1, N2, N3 corresponding to physical switches S1, S2 and S3 respectively.
- the logical network comprises logical links LL interconnecting the logical nodes N1-N3.
- a physical link is logically subdivided into one or more logical links, each logical link having an individual traffic capacity referred to as logical link capacity. It is important to note that each logical link may use more than one physical link or trunk group.
- a routing table which is used to route a connection from node to node in the particular logical network starting from the node associated with the terminal that originates the connection and ending at the node associated with the terminal which terminates said connection. Said nodes together form an origin-destination pair.
- a node pair with two routes is also illustrated. One of the routes is a direct route DR while the other one is an alternative route AR.
- the links and the routes may be interpreted as being bidirectional.
- a route is a subset of logical links which belong to the same logical network, i.e. a route have to live in a single logical network. Note that it can be an arbitrary subset that is not necessarily a path in the graph theoretic sense. Nevertheless, for practical purposes, routes are typically conceived as simple paths.
- the conception of a route is used to define the way a connection follows between nodes in a logical network.
- a node pair in a logical network, the nodes of which are associated with access points, is called an origin-destination (O-D) pair.
- O-D origin-destination
- all node pairs in a logical network are not O-D pairs, but instead some nodes in a logical network may be intermediate nodes to which no access points are associated.
- a logical link is a subset of physical links.
- Bearer services are STM 64 (Synchronous Transmission Mode with standard 64 kbit/s), STM 2 Mb (Synchronous Transmission Mode with 2 Mbit/s) and ATM (Asynchronous Transfer Mode).
- PSTN Public Switched Telephone Network
- B- ISDN Broadband Integrated Services Digital Network
- a method for adaptive link capacity control, and the integration of call admission control and link capacity control, by using distributed neural networks is disclosed in the article entitled "Integration of ATM Call Admission Control and Link Capacity Control by Distributed Neural Networks" by A. Hiramatsu in IEEE Journal on Selected Areas in Communications, vol. 9, no. 7 (1991).
- neural networks are trained to estimate the call loss rate given the link capacity and observed traffic.
- an objective function of the link capacity assignment optimization problem constituted by the maximum call loss rate in the network, is optimized by a simple random optimization method according to the estimated call loss rate.
- Hiramatsu only considers the optimization problem on the level of logical links.
- the concept of logical networks is not at all incorporated in the approach of Hiramatsu.
- the optimization method the Matyas random optimization method, is a simple method generally leading to a suboptimal solution.
- only one bit-rate class is considered in the model.
- the method described in the above U.S. Patent assumes that the resource allocation problem can be adequately described by a linear programming model.
- a linear programming model In its application to resource allocation, such a model consists of a number of linear expressions representing the quantitative relationships between the various possible allocations, their constraints and their costs, i.e. the objective function is a linear function of the allocated resources.
- the method according to the above U.S. patent does not take teletraffic models, such as e.g. Erlang's B formula, describing the statistical fluctuation of traffic, into consideration. Consequently, the linear programming model described in the above U.S. patent is quite unsatisfactory.
- the partitioning problem that the present invention considers gives rise to an objective function that depends on the allocated resources in an indirectly defined non-linear way. The dependence is defined through a complicated non-linear system of equations that follows from a teletraffic model.
- logical networks On top of a physical network a number of logical networks are established in which logical links, used by routes, share the same physical transmission and switching resources. There are several reasons for logically separating physical resources. Logical resource separation for offering different Grade of Service classes, virtual leased networks with guaranteed resources and peak rate allocated virtual paths are some examples of interesting features in the design, dimensioning and management of physical networks. However, it is still necessary to decide how to distribute or partition said physical network resources among the logical networks. In addition, the distribution of offered traffic load among routes interconnecting the nodes of node pairs in the logical networks will also affect the overall network performance.
- a set of logical networks is established on top of a physical network, which physical network comprises physical transmission and switching resources.
- the logical networks comprise nodes and logical links extending between the nodes so as to form the logical networks.
- the logical links are used by routes interconnecting the nodes of node pairs in the logical networks.
- an objective function which is closely related to the operation and overall performance of the resource separated physical network is optimized with respect to at least one set of decision variables, given physical network parameters and the requirements of each logical network. Examples of objective functions are the carried traffic in the complete network, the link utilization and the network revenue or some other function representing resource utilization or network performance.
- Two sets of decision variables are the logical link capacities, and the load sharing variables controlling the distribution of offered traffic load among routes.
- Each set of decision variables is related to a separate aspect of the invention. If the objective function have been optimized with respect to the logical link capacities then the physical transmission resources of the physical network are allocated among the logical links of the various logical networks in accordance with the optimization. On the other hand, if the objective function have been optimized with respect to the load sharing variables, then the offered traffic load is distributed, for each individual node pair in each one of the logical networks, among the routes interconnecting the nodes of the individual node pair, in accordance with the optimization.
- optimizing with respect to both the logical link capacities and the load sharing variables relates to yet another aspect of the present invention.
- the physical transmission resources of the physical network are allocated among the logical links of the various logical networks and the offered traffic load is distributed, for each individual node pair in each one of the logical networks, among the routes interconnecting the nodes of the individual node pair, in accordance with the optimization.
- Figure 1 shows a physical network, on top of which a number of logical networks are established, and an operation and support system (OSS) which controls the operation of the overall network,
- OSS operation and support system
- Figure 2 is a schematic drawing explaining the relationship between physical links and switches, logical links and nodes, and also routes,
- Figure 3 is a schematic drawing of a B-ISDN network from the viewpoint of the Stratified Reference Model
- Figure 4 is a schematic flow diagram illustrating a method in accordance with a general inventive concept of the present invention
- Figure 5 is a flow diagram illustrating, in more detail, a method in accordance with a first preferred embodiment of the invention
- Figure 6 is a schematic flow diagram illustrating how the method in accordance with a first preferred embodiment of the present invention flexibly adapt the overall network system to changing traffic conditions, but also to facility failures and demands for new logical network topologies,
- Figure 7 is a flow diagram illustrating a method in accordance with a second preferred embodiment of the invention.
- Figure 8 is a schematic flow diagram illustrating how the method in accordance with a second preferred embodiment of the present invention flexibly adapt the overall network system to changing traffic conditions, but also to facility failures and demands for new logical network topologies,
- Figure 9 is a flow diagram illustrating a method in accordance with a third preferred embodiment of the invention.
- Figure 10 presents experimental results illustrating how much gain is achieved by the proposed invention in comparison to initial values obtained from a convex optimization method (COM).
- COM convex optimization method
- VP Virtual Path
- a physical network e.g. a large telecommunication network, with physical resources is considered.
- Fig. 1 there is illustrated a physical network PN on top of which a set of logical networks LN1, LN2, ..., LNX (assuming there are X logical networks) is established.
- Each logical network comprises nodes N and logical links LL interconnecting the nodes.
- the topology of these logical or virtual networks will in general differ from the topology of the underlying physical network.
- the network system is preferably controlled by an operation and support system OSS.
- An operation and support system OSS usually comprises a processor system PS, terminals T and a control program module CPM with a number of control programs CP along with other auxiliary devices.
- the architecture of the processor system is usually that of a multiprocessor system with several processors working in parallel. It is also possible to use a hierarchical processor structure with a number of regional processors and a central processor.
- the switches themselves can be equipped with their own processor units in a not completely distributed system, where the control of certain functions are centralized.
- the processor system may consist of a single processor, often a large capacity processor.
- a database DB preferably an interactive database, comprising e.g.
- OSS a description of the physical network, traffic information and other useful data about the telecommunication system, is connected to the OSS.
- Special data links through which a network manager/operator controls the switches, connect the OSS with those switches which form part of the network system.
- the OSS contains e.g. functions for monitoring and controlling the physical network and the traffic.
- the network manager establishes a number of logical networks on top of the physical network by associating different parts of the traffic with different parts of the transmission and switching resources of the physical network. This can e.g. be realized by controlling the port assignment of the switches and cross connect devices of the physical network, or by call admission control procedures.
- the process of establishing logical networks means that the topology of each one of the logical networks is defined. In other words, the structure of the nodes and logical links in each logical network is determined.
- traffxc classes are arranged into groups in such a way that those with similar demands to bandwidth are handled together in a separate logical network.
- all traffic types requiring more than a given amount of bandwidth can be integrated in one logical network, and those traffic types that require less bandwidth than this given amount can be integrated in another logical network.
- the two traffic groups are handled separately in different logical subnetworks. In particular, this is advantageous for an ATM network carrying a wide variety of traffic types.
- each individual traffic type is handled in a separate logical network.
- the present invention is applied in the B-ISDN (Broadband Integrated Services Digital Network) environment.
- B-ISDN Broadband Integrated Services Digital Network
- a fully developed B-ISDN network will have a very complex structure with a number of overlaid networks.
- One conceptual model suitable of describing overlaid networks is the Stratified Reference Model as described in "The stratified Reference Model: An Open Architecture to B-ISDN" by T. Hadoung, B. Stavenow, J. Dejean, ISS'90, Sweden.
- Fig. 3 a schematic drawing of a B-ISDN network from the viewpoint of the Stratified Reference Model is illustrated (the protocol viewpoint to the left and the network viewpoint to the right). Accordingly, the B-ISDN will consist of the following strata.
- SDH Serial Digital Hierarchy
- ATM Asynchronous Transfer Mode
- the large set of possible applications uses the cross connect stratum as an infrastructure.
- it is the infrastructure network modelling the cross connect stratum in a B-ISDN overlaid network that is considered. In general, this infrastructure network is referred to as a physical network.
- Fig. 4 shows a schematic flow diagram illustrating a method in accordance with a general inventive concept of the present invention.
- a set of logical networks is established on top of a physical network comprising physical transmission and switching resources, said logical networks comprising nodes and logical links extending between the nodes so as to define the topology of said logical networks.
- the logical networks are completely separated from each other.
- the logical links are used by routes interconnecting the nodes of node pairs in the logical networks.
- a predefined objective function closely related to the operation and performance of the physical network, which physical network is viewed as the set of logical networks, is optimized with respect to decision variables.
- the decision variables in accordance with the optimization are used to control the operation of the overall network system.
- the physical transmission resources i.e. the transmission capacities of the physical links
- a natural objective is to partition the physical transmission resources so as to optimize the operation of the complete physical network, viewed as the set of logical networks, according to a given predefined objective function.
- the logical networks share the same given physical transmission and switching resources, which means that the operation of the physical network has to be optimized with respect to all the logical networks, i.e. the complete set of logical networks, at the same time.
- the cross connect stratum can be realized by either SDH or ATM. If the cross connect stratum is based on SDH and the infrastructure network is realizing e.g. different quality of service classes by resource separation, the partitioning can only be performed in integer portions of the STM modules of the SDH structure. On the other hand, if the cross connect is realized by ATM virtual paths then no integrality restriction exists and the partitioning can be performed in any real portions. Therefore, whether the cross connect stratum is based on SDH or ATM will have important implications for the partitioning of the physical network resources.
- the SDH cross connect solution gives rise to a model that is discrete with regard to the logical link capacities, while the ATM cross connect solution gives rise to a continuous model.
- the continuous model requires that the ATM switches support partitioning on the individual input and output ports. For example, this is realized by multiple logical buffers at the output ports.
- an infrastructure network modelling the ATM cross connect stratum is considered while in an alternative embodiment an infrastructure modelling the SDH cross connect is considered, as can be seen in Fig. 1.
- partitioning as opposed to complete sharing, is a reduction of the full flexibility of ATM. This is however not the case if the partitioning is considered on a general level.
- the complete sharing schemes e.g. priority queuing, Virtual Spacing etc. tell us how to realize resource sharing on the cell level, while the partitioning approach seeks for the call scale characteristics, e.g. how to assign rates to various logical links, that is then to be realized on the cell level. In this sense the complete partitioning approach complements, rather than excludes, the complete sharing approaches.
- the objective function according to the present invention is preferably defined as the total carried traffic, although other objective functions can be used. Examples of other objective functions are the link utilization in the complete network, the network revenue or some other function representing resource utilization or network performance.
- the optimization associated with resource partitioning is to calculate, given a description of the physical network, the topology of the logical networks, the traffic types, the routes in each of the logical networks and also the offered traffic to each route or to each node pair in each logical network, the logical link capacities for the corresponding logical networks so as to maximize the total carried traffic or the network revenue.
- C is defined above, and C phys refer to the vector of given physical link capacities. In addition it is required that C ⁇ 0.
- each route carries only a single type of traffic. This means that if several traffic types are to be carried, they are represented by parallel routes.
- v denotes logical networks
- p denotes node pairs (O-D pairs)
- i (sometimes also g) denotes traffic types.
- a ijr which are equal to 1 when route r uses logical link j and carries traffic type i, otherwise said variables are 0.
- a ijr is not to be interpreted as the amount of bandwidth that route r requires on logical link j.
- a ij will denote the amount of bandwidth (capacity) that a call belonging to traffic type i requires on logical link j.
- all routes that carry a given traffic type i require the same amount of bandwidth on link j. Since the bandwidth requirement is associated with the traffic type, this is not seen as a restriction.
- the bandwidth requirement of calls on a given route is allowed to vary along the logical links of the route. In fact, this is needed if the concept of effective or equivalent bandwidth is adopted and the involved logical links have different capacities.
- R the total set of routes over all logical networks, that is,
- R U v U p U i R ( ⁇ ,p,i) (1)
- R ( ⁇ , p,i ) is the set of routes in logical network ⁇ realizing communication between node pair p regarding traffic type i. It is important to understand that a route is not associated with more than one logical network. Each logical network is assumed to operate under fixed non-alternate routing.
- ⁇ r be the Poissonian call arrival rate to route r
- l/ ⁇ r be the average holding time of calls on route r
- v r ⁇ r / ⁇ r be the offered traffic to route r.
- v ( ⁇ ,p,i) be the aggregated offered traffic of type i to node pair p in logical network v .
- the offered traffic for each route in each logical network is given while in another preferred embodiment of the invention the above aggregated offered traffic is given for all logical networks, node pairs and traffic types. In the latter case, the load is e.g. distributed on shortest paths.
- w r be the revenue coefficient parameter for route r, meaning that one unit of carried traffic on route r is associated with revenue w r .
- the revenue coefficients can easily be incorporated into the total carried traffic function so as to obtain the network revenue function.
- the network revenue is the objective function.
- the total carried traffic is the main objective function and the network revenue is an extension which is a weighted version of the total carried traffic.
- the objective function can be expressed as the total network revenue summed up over all routes in all the logical networks: where L r is the end-to-end blocking probability for traffic on route r.
- this route blocking probability is defined as the probability of the event that at least one logical link is blocked along the route.
- the objective function to be optimized is inherently difficult to deal with, since it requires knowledge of the carried traffic and thereby route blocking probabilities which can only be computed in an exact way for very small networks.
- the objective function depends on the allocated resources in an indirectly defined non-linear way. The dependence is defined through a complicated non-linear system of equations that follows from a teletraffic model, as will be described below.
- the objective of the optimization task associated with the partitioning of physical network resources is to maximize the total network revenue, as defined above, subject to the physical constraints SC ⁇ C phys , C ⁇ 0. In accordance with a first preferred embodiment of the present invention, this is achieved by computing the partial derivatives of the network revenue with respect to the logical link capacities and subsequently using them in a gradient method.
- B ik denotes the blocking probability for traffic type i on logical link k.
- ⁇ ik be the offered bandwidth demand from traffic type i to logical link k when blocking elsewhere is taken into account.
- E ik which, given the logical link offered classwise bandwidth demand ⁇ 1k , ..., ⁇ Ik and logical link capacity C k , returns the blocking probability on logical link k regarding traffic type i.
- any blocking function is allowed that is jointly smooth in all the variables.
- the partial derivatives of the total network revenue are calculated for the multirate case, as calls on different routes are allowed to have different bandwidth requirements.
- the formulas are presented in the following and it is useful to have the usual notion of a differential like dW as a small change in W.
- the revenue derivative with respect to logical link capacity C k can be formulated as: A where
- the blocking function is Erlang's B formula that is defined for integer capacity values but has a simple analytic extension to any non-negative real capacity value.
- the present invention considers the multirate case. It is possible to use Kaufman and Robert's recursive blocking formula from the stochastic knapsack problem. Unfortunately it is quite complicated to find explicitly a smooth extension to real capacity values. Therefore, in a preferred embodiment of the present invention, in order to enhance computational feasibility, a normal approximation is used for the blocking function (see Appendix A for details):
- the gradient method is an iterative method capable of solving both linear and non-linear problems. It is based on the fact that the gradient of a multivariable function, i.e. the vector of partial derivatives for the function, points, at each point, in the direction in which the function changes (increases or decreases) at the highest rate. In addition, the optimal step size in this direction is determined by a line search. Hence, the method is appropriately designed for climbing towards a maximum of a multivariable function.
- the partial derivatives for the network revenue with respect to the logical link capacities are applied in a gradient based hill climbing procedure in order to maximize the total network revenue as defined by (2).
- W(C 1 , C 2 , ..., C J ) denote the gradient vector of the total network revenue with respect to the logical link capacities.
- an initial design point for the logical link capacities associated with the various logical networks is selected.
- the logical link capacities are iteratively calculated by an alternating sequence of calculating the optimal ascend or step direction using the revenue gradient vector W(C 1 , C 2 , ..., C J ) and performing a one-dimensional line search to find an optimal point.
- Each step in the actual hill climbing must be in consistency with the feasibility region defined by the physical constraints.
- the iteration process is terminated when convergence is achieved with a required level of accuracy.
- the physical link capacities are then allocated among the logical links of the various logical networks in accordance with the finally calculated logical link capacities.
- a method according to a first preferred embodiment will be described in more detail with reference to the schematic flow diagram of Fig. 5.
- a set of logical networks is established on top of a physical network by associating different parts of the traffic with different parts of the physical transmission and switching resources.
- initial values for the logical link capacities C j initial (for all j), which can be seen as an initial design point, are selected.
- the fixed point equations defined by (3)-(6) are solved by successive substitutions, thereby computing a set of blocking probabilities B ik to be used in subsequent steps.
- the set of linear equations that is defined by ( 9 ) and (10) has to be solved.
- the solution will yield the set of auxiliary parameters c ik that is needed for calculating the revenue derivatives.
- the present logical link capacities, the blocking probabilities and the auxiliary parameters are known.
- the actual calculation of the revenue derivatives with respect to the present values of the logical link capacities takes place.
- the gradient vector for the network revenue is known and the optimal ascend or step direction at the present design point can be determined. If the present design point is situated at the boundary of the feasibility region and the calculated gradient vector points out from the feasibility region, then the direction of the next step in the hill climbing is calculated by projecting the gradient vector so that it follows the boundary. In fact, if identity projections (i.e.
- next step in the actual hill climbing is always taken in the direction of the projection of the gradient of the network revenue to the feasibility region.
- a penalty function procedure can be utilized to determine the direction of the next step.
- a one-dimensional line search is performed along the ascend or step direction to optimize the size of the step to be taken in the actual hill climbing.
- every step, determined by the direction and the step size, in the hill climbing has to be consistent with the feasible region.
- the partitioning can only be performed in integer portions of the STM modules of the SDH structure, as mentioned above.
- the real capacity values obtained from the method according to the first preferred embodiment of the invention are preferably rounded into integer values such that the physical constraints are satisfied. In one embodiment of the invention this is realized by independently repeated random rounding trials.
- the method according to the first preferred embodiment of the invention is preferably performed by one or more control programs CP of the control program module of the operation and support system OSS. These control programs, in turn, are executed by one or more of the processors in the processor system PS described above.
- the operation and support system OSS collects the required information from the network system and uses this information together with the database DB information as input to the respective control programs CP. Furthermore, the OSS controls the network switches through the data links so as to partition the physical link capacities among the logical links of the logical networks.
- the network manager can flexibly adapt the overall network system to changing traffic conditions, such as changes in offered traffic, but also to facility failures and new demands on the logical network topology from e.g. business users, as is illustrated in the schematic flow diagram of Fig. 6.
- the partitioning is optimal.
- the topology of one or more logical networks have to be changed for some reason ( facility failure or demands for new topologies) or additional logical networks are requested, then the complete set of steps according to the first preferred embodiment of the present invention has to be performed in order to optimize the overall network configuration. If no changes regarding the topology of the logical networks is necessary, but e.g.
- the optimization step and the allocation step are repeated in response to changing traffic conditions so as to change the logical link capacities of the various logical networks in a flexible and optimal way.
- This is realized by the switches and cross connect devices of the physical network in a very short period of time.
- the realization of the present invention renders the operation of the complete physical network both safe and flexible.
- the method in accordance with the present invention regards the logical link capacities as fixed in each iteration step.
- the logical link capacities are not fixed parameters since it is desired to optimize them.
- the strictly non-linear objective function is approximated as a linear function valid in a neighborhood of the fixed point, but seen from the whole procedure the objective function is nonlinear since the objective function itself is changing in an indirectly non-linear way in each iteration step.
- the initial point is selected as the solution of a convexification of the problem.
- the convex optimization method yields a global optimum, it is obtained in a relatively rough model.
- a gradient based hill climbing is used to come from the initial point obtained from the COM method to an improved value in the more refined model of the present invention.
- a device as described in our Patent Application 9500838-9 can be used to obtain an initial design point for the logical link capacities.
- the device comprises two artificial neural networks interworking to compute a set of logical link capacities representing a global optimum.
- the global optimum is obtained in a model which is relatively rougher than the model of the present invention.
- the physical constraint C ⁇ O can be altered to C ⁇ C constant , where C constant is a constant capacity vector representing a minimum guaranteed resource for each individual logical link.
- C constant is a constant capacity vector representing a minimum guaranteed resource for each individual logical link.
- SC ⁇ C phys can not be violated.
- the aggregated offered traffic of type i to node pair p in logical network ⁇ is given for all logical networks, node pairs and traffic types.
- the total network revenue depends also on the offered traffic to each route, v r , and hence the overall network performance is affected by the distribution of offered traffic among the routes which can realize the communication even within a single logical network. It may appear natural to distribute the offered traffic load uniformly among parallel routes, but in general this distribution is far from optimal.
- the distribution of the offered traffic load of type i to node pair p in logical network ⁇ , for all i, p and v, is determined together with the partitioning of the physical link capacities among the logical links of the various logical networks, such that the overall network performance is optimized.
- the offered traffic load among the routes carrying traffic between the nodes of the individual node pair overload situations are avoided and in general load balancing is achieved.
- load sharing The distribution of the offered traffic between the possible routes is termed load sharing, and the parameters according to which it takes place is called load sharing coefficients.
- the optimization task according to a second preferred embodiment of the invention is to determine the logical link capacities of the various logical networks and the typewise load sharing coefficients for the node pairs in each one of the logical networks so that the total expected network revenue is maximized, while the physical constraints are satisfied.
- this can be formulated as follows: , (
- the revenue derivative with respect to offered traffic along route r can be expressed as: where the set of auxiliary parameters c ik is defined by the system of linear equations given by (9) and (10). Together with the partial derivatives for the network revenue with respect to the logical link capacities, as given above, the revenue derivatives with respect to the route offered traffic values are applied in a gradient based hill climbing procedure in order to maximize the network revenue function.
- the physical constraints define a convex feasibility region, which has to be considered in the actual hill climbing.
- a method in accordance with a second preferred embodiment of the invention is illustrated.
- a set of logical networks is established on top of a physical network.
- an initial design point for the logical link capacities C j inltial and the route offered traffic values v r initial in each one of the logical networks is selected.
- the offered load is distributed along shortest paths as a good initial choice for the route offered traffic values.
- logical link capacities and route offered traffic values are iteratively calculated by an alternating sequence of calculating the optimal ascend or step direction using the revenue gradient vector with respect to both the logical link capacities and the route offered traffic and performing a one-dimensional line search to find an optimal point.
- the partial derivatives with respect to the present logical link capacities and the present route offered traffic values, constituting the revenue gradient vector in this particular embodiment, have to be calculated in each iteration step.
- the iteration process is terminated when convergence is achieved with a required level of accuracy.
- the gradient based hill climbing procedure is similar to the one in which only the revenue derivatives with respect to the logical link capacities were considered.
- the logical link capacities for the various logical networks and the route offered traffic values in each logical network which maximizes the revenue function are known. Consequently, the optimal distribution of the offered traffic load of type i to node pair p in logical network v is known for all i, p and v, and the corresponding load sharing coefficients are calculated in a straightforward way.
- the physical link capacities are then allocated among the logical links of the various logical networks in accordance with the finally calculated logical link capacities.
- the traffic load is apportioned, for each individual node pair in each one of the logical networks, among the routes which can realize communication between the nodes of the individual node pair, in accordance with the finally calculated set of load sharing coefficients.
- the second preferred embodiment is preferably realized by one or more control programs CP which are executed by the processor system PS incorporated in the operation and support system OSS described above.
- the operation and support system OSS collects the required information from the network system and uses this information together with the database DB information as input to the respective control programs CP. Furthermore, the OSS controls the overall network system through the data links.
- the apportioning is preferably realized by routing decision means. For example, assume that there are two different routes between a node pair.
- the load sharing coefficients for the first and second route are 0,6 and 0,4 respectively.
- a random number between 0 and 1 is generated by random number generator means. If the random number is less than 0,6 then the first route is to be used, and if the random number is greater or equal to 0,6 then the second route is to be used.
- the second preferred embodiment allows the network manager to flexibly adapt the overall network system to changing traffic conditions, such as changes in offered traffic, but also to facility failures and new demands on the logical network topology from e.g. business users.
- This is illustrated in the schematic flow diagram of Fig. 8.
- the complete set of steps according to the second preferred embodiment of the present invention has to be performed in order to optimize the overall network configuration. If no changes regarding the topology of the logical networks are necessary, but the traffic varies, then only the optimizing, allocating and apportioning steps of the second preferred embodiment of the present invention have to be carried out. That is, the optimizing step, the allocating step and the apportioning step are repeated in response to changing traffic conditions so as to change the logical link capacities and the route offered traffic values of the various logical networks in a flexible and optimal way.
- the typewise load sharing coefficients for each node pair in each one of the logical networks are determined so that the total expected network revenue is maximized, while the physical constraints are satisfied.
- this is formulated in the following way:
- Fig. 9 there is illustrated a schematic flow diagram of a third preferred embodiment of the present invention.
- initial values for the route offered traffic in each one of the logical networks are selected as an initial design point.
- the offered load is distributed along shortest paths as an initial choice.
- route offered traffic values are iteratively calculated by an alternating sequence of calculating the optimal ascend or step direction using the revenue gradient vector with respect to the route offered traffic (using expression (16)) and performing a one-dimensional line search to find an optimal point by optimizing the step size. Since the route offered traffic is changed in the actual hill climbing, the fixed point equations (3)-(6) have to be solved in each iteration.
- the set of linear equations (9)-(10) has to be solved in each iteration in order to calculate the partial derivatives of the network revenue with respect to the present route offered traffic values.
- Each step along the ascend or step direction in the actual hill climbing must be in consistency with the feasibility region defined by the physical constraints. This is achieved by a projection procedure similar to the one in the first and second preferred embodiments of the invention. The iteration process is terminated when convergence is achieved with a required level of accuracy. In all other regards the gradient based hill climbing procedure is similar to the one in the second preferred embodiment.
- the route offered traffic values for each logical network which maximizes the network revenue function are known.
- the optimal distribution of the offered traffic load of type i to node pair p in logical network v is known for all i, p and v, and the corresponding load sharing coefficients are calculated in a straightforward way. Then, the offered traffic load is apportioned, for each individual node pair in each one of the logical networks, among the routes which can realize communication between the nodes of the individual node pair, in accordance with the finally calculated load sharing coefficients.
- the third preferred embodiment is preferably realized by one or more control programs CP which are executed by the processor system PS incorporated in the operation and support system OSS.
- the apportioning of offered traffic load is performed by routing decision means using random number generator means.
- the present invention was tried on a very simple network. It was a four-node ring network on top of which two logical networks with one traffic class in each were established. These traffic classes were different, so there were altogether two traffic classes. Two parameters were varied: the ratio of the single-call bandwidth demand in the two traffic classes (bandwidth ratio - BR) and the ratio of the offered traffic load in the two classes (offered traffic ratio - OTR).
- the revenue ratio - RR i.e. the revenue achieved by the COM method divided by the revenue achieved by the second preferred embodiment of the present invention.
- the result is shown in Fig. 10. If the varied ratios were below 10, the present invention improved the result insignificantly.
- the traffic scenario gets more inhomogeneous, the additional gain obtained by the present invention becomes more and more substantial. Consequently, if the proposed invention is applied to large telecommunication networks, such as an ATM network, with several logical networks carrying a number of widely different traffic types it is very likely that the invention will considerably improve the result, as is shown by the overall tendency of the experiments.
Abstract
Description
Claims
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA002192794A CA2192794C (en) | 1994-06-13 | 1995-06-12 | Enhancement of network operation and performance |
AU27587/95A AU692884B2 (en) | 1994-06-13 | 1995-06-12 | Enhancement of network operation and performance |
EP95922843A EP0765552B1 (en) | 1994-06-13 | 1995-06-12 | Enhancement of network operation and performance |
DE69533064T DE69533064D1 (en) | 1994-06-13 | 1995-06-12 | IMPROVE NETWORK OPERATION AND PERFORMANCE |
JP8502038A JPH10506243A (en) | 1994-06-13 | 1995-06-12 | Network operation and performance enhancement |
US08/765,106 US6069894A (en) | 1995-06-12 | 1995-06-12 | Enhancement of network operation and performance |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SE9402059A SE9402059D0 (en) | 1994-06-13 | 1994-06-13 | Methods and apparatus for telecommunications |
SE9402059-1 | 1994-06-13 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO1995034981A2 true WO1995034981A2 (en) | 1995-12-21 |
WO1995034981A3 WO1995034981A3 (en) | 1996-02-08 |
Family
ID=20394357
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/SE1995/000704 WO1995034981A2 (en) | 1994-06-13 | 1995-06-12 | Enhancement of network operation and performance |
PCT/SE1995/000703 WO1995034973A2 (en) | 1994-06-13 | 1995-06-12 | A method and device for partitioning physical netword resources |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/SE1995/000703 WO1995034973A2 (en) | 1994-06-13 | 1995-06-12 | A method and device for partitioning physical netword resources |
Country Status (9)
Country | Link |
---|---|
US (1) | US6104699A (en) |
EP (2) | EP0765552B1 (en) |
JP (2) | JPH10506243A (en) |
CN (2) | CN1080501C (en) |
AU (2) | AU692884B2 (en) |
CA (1) | CA2192793A1 (en) |
DE (2) | DE69534216D1 (en) |
SE (1) | SE9402059D0 (en) |
WO (2) | WO1995034981A2 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2317533A (en) * | 1996-07-29 | 1998-03-25 | Northern Telecom Ltd | Communications network |
EP1098546A2 (en) * | 1999-11-04 | 2001-05-09 | Lucent Technologies Inc. | Methods and apparatus for derivative based optimization of wireless network performance |
WO2001093504A1 (en) * | 2000-05-31 | 2001-12-06 | Nokia Corporation | Partitioning of a communications network |
US7006466B2 (en) * | 2001-03-09 | 2006-02-28 | Lucent Technologies Inc. | Dynamic rate control methods and apparatus for scheduling data transmissions in a communication network |
EP2651054A1 (en) | 2004-10-22 | 2013-10-16 | Cisco Technology, Inc. | Fibre channel over ethernet |
Families Citing this family (197)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB9606708D0 (en) * | 1996-03-29 | 1996-06-05 | Plessey Telecomm | Bandwidth bidding |
US5844886A (en) * | 1996-12-30 | 1998-12-01 | Telefonaktiebolaget Lm Ericsson (Publ.) | System and method for network optimization using code blocking |
IL120449A0 (en) | 1997-03-13 | 1997-07-13 | Ben Ami Raphael | Apparatus and method for expanding communication networks |
EP1013131A2 (en) * | 1997-03-13 | 2000-06-28 | Urizen Ltd. | Apparatus and method for expanding communication networks |
GB2332334A (en) | 1997-12-10 | 1999-06-16 | Northern Telecom Ltd | Trail management system |
US6434619B1 (en) * | 1998-04-29 | 2002-08-13 | Alcatel Canada Inc. | Internet-enabled service management system and method |
JP3609256B2 (en) * | 1998-05-19 | 2005-01-12 | 株式会社日立製作所 | Network management device, node device, and network management system |
US6999421B1 (en) * | 1998-10-26 | 2006-02-14 | Fujitsu Limited | Adjustable connection admission control method and device for packet-based switch |
US6381237B1 (en) | 1998-10-30 | 2002-04-30 | Nortel Networks Limited | Trail explorer and method for exploring trails in a communication network |
US6442507B1 (en) | 1998-12-29 | 2002-08-27 | Wireless Communications, Inc. | System for creating a computer model and measurement database of a wireless communication network |
US6493679B1 (en) * | 1999-05-26 | 2002-12-10 | Wireless Valley Communications, Inc. | Method and system for managing a real time bill of materials |
US6317599B1 (en) * | 1999-05-26 | 2001-11-13 | Wireless Valley Communications, Inc. | Method and system for automated optimization of antenna positioning in 3-D |
US6850946B1 (en) * | 1999-05-26 | 2005-02-01 | Wireless Valley Communications, Inc. | Method and system for a building database manipulator |
US7243054B2 (en) | 1999-07-14 | 2007-07-10 | Wireless Valley Communications, Inc. | Method and system for displaying network performance, cost, maintenance, and infrastructure wiring diagram |
US6499006B1 (en) * | 1999-07-14 | 2002-12-24 | Wireless Valley Communications, Inc. | System for the three-dimensional display of wireless communication system performance |
US6986137B1 (en) * | 1999-09-28 | 2006-01-10 | International Business Machines Corporation | Method, system and program products for managing logical processors of a computing environment |
US6519660B1 (en) * | 1999-09-28 | 2003-02-11 | International Business Machines Corporation | Method, system and program products for determining I/O configuration entropy |
US6810422B1 (en) | 2000-01-14 | 2004-10-26 | Lockheed Martin Tactical Defense Systems | System and method for probabilistic quality of communication service determination |
FI20001312A (en) * | 2000-05-31 | 2001-12-01 | Nokia Networks Oy | Formation of a telecommunications network |
US6971063B1 (en) | 2000-07-28 | 2005-11-29 | Wireless Valley Communications Inc. | System, method, and apparatus for portable design, deployment, test, and optimization of a communication network |
US6912203B1 (en) * | 2000-07-31 | 2005-06-28 | Cisco Technology, Inc. | Method and apparatus for estimating delay and jitter between many network routers using measurements between a preferred set of routers |
US6625454B1 (en) | 2000-08-04 | 2003-09-23 | Wireless Valley Communications, Inc. | Method and system for designing or deploying a communications network which considers frequency dependent effects |
US7096173B1 (en) | 2000-08-04 | 2006-08-22 | Motorola, Inc. | Method and system for designing or deploying a communications network which allows simultaneous selection of multiple components |
US7680644B2 (en) * | 2000-08-04 | 2010-03-16 | Wireless Valley Communications, Inc. | Method and system, with component kits, for designing or deploying a communications network which considers frequency dependent effects |
US7246045B1 (en) | 2000-08-04 | 2007-07-17 | Wireless Valley Communication, Inc. | System and method for efficiently visualizing and comparing communication network system performance |
US7085697B1 (en) | 2000-08-04 | 2006-08-01 | Motorola, Inc. | Method and system for designing or deploying a communications network which considers component attributes |
US7055107B1 (en) | 2000-09-22 | 2006-05-30 | Wireless Valley Communications, Inc. | Method and system for automated selection of optimal communication network equipment model, position, and configuration |
US6973622B1 (en) * | 2000-09-25 | 2005-12-06 | Wireless Valley Communications, Inc. | System and method for design, tracking, measurement, prediction and optimization of data communication networks |
BR0116693A (en) * | 2000-12-18 | 2004-08-17 | Wireless Valley Comm Inc | Computer system and method for collecting data from a spatially distributed object group or network and method for visualizing a spatially distributed object group or network |
US7133410B2 (en) * | 2001-02-12 | 2006-11-07 | Tellabs Operations, Inc. | Method and system for designing ring-based telecommunications networks |
US7164883B2 (en) * | 2001-02-14 | 2007-01-16 | Motorola. Inc. | Method and system for modeling and managing terrain, buildings, and infrastructure |
DE10116835A1 (en) * | 2001-04-04 | 2002-10-17 | Alcatel Sa | Fast restoration mechanism e.g. for determining restoration capacity in transmission network, involves determining mesh in network with mesh closed sequence of bi-directional links traversing each node part of mesh exactly once |
US7031253B1 (en) * | 2001-06-01 | 2006-04-18 | Cisco Technology, Inc. | Method and apparatus for computing a path through specified elements in a network |
US6880002B2 (en) * | 2001-09-05 | 2005-04-12 | Surgient, Inc. | Virtualized logical server cloud providing non-deterministic allocation of logical attributes of logical servers to physical resources |
CA2411806A1 (en) * | 2001-11-16 | 2003-05-16 | Telecommunications Research Laboratory | Wide-area content-based routing architecture |
US7574496B2 (en) | 2001-11-30 | 2009-08-11 | Surgient, Inc. | Virtual server cloud interfacing |
KR100428721B1 (en) * | 2001-12-04 | 2004-04-28 | 주식회사 케이티 | Method of Generation and Contolling Logical Port in ATM Exchange System |
US7574323B2 (en) * | 2001-12-17 | 2009-08-11 | Wireless Valley Communications, Inc. | Textual and graphical demarcation of location, and interpretation of measurements |
US7339897B2 (en) * | 2002-02-22 | 2008-03-04 | Telefonaktiebolaget Lm Ericsson (Publ) | Cross-layer integrated collision free path routing |
US6990666B2 (en) * | 2002-03-18 | 2006-01-24 | Surgient Inc. | Near on-line server |
US7257584B2 (en) | 2002-03-18 | 2007-08-14 | Surgient, Inc. | Server file management |
US7197553B2 (en) * | 2002-04-19 | 2007-03-27 | Nortel Networks Limited | Network system having a virtual-service-module |
US7804785B2 (en) * | 2002-04-19 | 2010-09-28 | Avaya Inc. | Network system having an instructional sequence for performing packet processing and optimizing the packet processing |
US7246178B2 (en) * | 2002-05-07 | 2007-07-17 | Nortel Networks Limited | Methods and systems for changing a topology of a network |
US7386628B1 (en) | 2002-05-08 | 2008-06-10 | Nortel Networks Limited | Methods and systems for processing network data packets |
US8463947B2 (en) * | 2002-08-28 | 2013-06-11 | Tellabs Operations, Inc. | Method of finding rings for optimal routing of digital information |
US7346709B2 (en) * | 2002-08-28 | 2008-03-18 | Tellabs Operations, Inc. | Methods for assigning rings in a network |
US7319675B1 (en) * | 2002-09-04 | 2008-01-15 | At&T Mobility Ii Llc | Systems and methods for calculating call blocking for alternate call routing schemes |
CN100459534C (en) * | 2002-10-07 | 2009-02-04 | 日本电信电话株式会社 | Layer network node and network constituted throuth said nodes, the node and layer network thereof |
US7295119B2 (en) | 2003-01-22 | 2007-11-13 | Wireless Valley Communications, Inc. | System and method for indicating the presence or physical location of persons or devices in a site specific representation of a physical environment |
US7295960B2 (en) * | 2003-01-22 | 2007-11-13 | Wireless Valley Communications, Inc. | System and method for automated placement or configuration of equipment for obtaining desired network performance objectives |
US20040259555A1 (en) * | 2003-04-23 | 2004-12-23 | Rappaport Theodore S. | System and method for predicting network performance and position location using multiple table lookups |
US20040259554A1 (en) * | 2003-04-23 | 2004-12-23 | Rappaport Theodore S. | System and method for ray tracing using reception surfaces |
US7287186B2 (en) | 2003-06-02 | 2007-10-23 | Surgient Inc. | Shared nothing virtual cluster |
CN100370736C (en) * | 2003-09-02 | 2008-02-20 | 华为技术有限公司 | Method for managing resources based on planes |
US7643484B2 (en) * | 2003-09-26 | 2010-01-05 | Surgient, Inc. | Network abstraction and isolation layer rules-based federation and masquerading |
US7769004B2 (en) | 2003-09-26 | 2010-08-03 | Surgient, Inc. | Network abstraction and isolation layer for masquerading machine identity of a computer |
US7382738B2 (en) * | 2003-11-24 | 2008-06-03 | Nortel Networks Limited | Method and apparatus for computing metric information for abstracted network links |
US7801125B2 (en) | 2004-10-22 | 2010-09-21 | Cisco Technology, Inc. | Forwarding table reduction and multipath network forwarding |
US7969971B2 (en) * | 2004-10-22 | 2011-06-28 | Cisco Technology, Inc. | Ethernet extension for the data center |
US8238347B2 (en) | 2004-10-22 | 2012-08-07 | Cisco Technology, Inc. | Fibre channel over ethernet |
US7830793B2 (en) * | 2004-10-22 | 2010-11-09 | Cisco Technology, Inc. | Network device architecture for consolidating input/output and reducing latency |
US7602720B2 (en) * | 2004-10-22 | 2009-10-13 | Cisco Technology, Inc. | Active queue management methods and devices |
US20070070898A1 (en) * | 2005-09-29 | 2007-03-29 | Khrais Nidal N | Channel resource allocation based upon call blocking probabilities |
US7961621B2 (en) | 2005-10-11 | 2011-06-14 | Cisco Technology, Inc. | Methods and devices for backward congestion notification |
CN100544300C (en) * | 2005-11-02 | 2009-09-23 | 华为技术有限公司 | Realize the method for signalling network interconnection |
US8078728B1 (en) | 2006-03-31 | 2011-12-13 | Quest Software, Inc. | Capacity pooling for application reservation and delivery |
US8924524B2 (en) | 2009-07-27 | 2014-12-30 | Vmware, Inc. | Automated network configuration of virtual machines in a virtual lab data environment |
US8619771B2 (en) | 2009-09-30 | 2013-12-31 | Vmware, Inc. | Private allocated networks over shared communications infrastructure |
US8892706B1 (en) | 2010-06-21 | 2014-11-18 | Vmware, Inc. | Private ethernet overlay networks over a shared ethernet in a virtual environment |
CN101001395B (en) * | 2006-12-30 | 2010-10-27 | 华为技术有限公司 | Method and device of selecting call route |
US8259720B2 (en) | 2007-02-02 | 2012-09-04 | Cisco Technology, Inc. | Triple-tier anycast addressing |
US8149710B2 (en) | 2007-07-05 | 2012-04-03 | Cisco Technology, Inc. | Flexible and hierarchical dynamic buffer allocation |
US8121038B2 (en) | 2007-08-21 | 2012-02-21 | Cisco Technology, Inc. | Backward congestion notification |
US9083609B2 (en) | 2007-09-26 | 2015-07-14 | Nicira, Inc. | Network operating system for managing and securing networks |
US7860012B2 (en) * | 2007-12-18 | 2010-12-28 | Michael Asher | Employing parallel processing for routing calls |
US8194674B1 (en) | 2007-12-20 | 2012-06-05 | Quest Software, Inc. | System and method for aggregating communications and for translating between overlapping internal network addresses and unique external network addresses |
US8195774B2 (en) | 2008-05-23 | 2012-06-05 | Vmware, Inc. | Distributed virtual switch for virtualized computer systems |
WO2010115060A2 (en) | 2009-04-01 | 2010-10-07 | Nicira Networks | Method and apparatus for implementing and managing virtual switches |
US9680750B2 (en) | 2010-07-06 | 2017-06-13 | Nicira, Inc. | Use of tunnels to hide network addresses |
US8964528B2 (en) | 2010-07-06 | 2015-02-24 | Nicira, Inc. | Method and apparatus for robust packet distribution among hierarchical managed switching elements |
US8817621B2 (en) | 2010-07-06 | 2014-08-26 | Nicira, Inc. | Network virtualization apparatus |
US9525647B2 (en) | 2010-07-06 | 2016-12-20 | Nicira, Inc. | Network control apparatus and method for creating and modifying logical switching elements |
US10103939B2 (en) | 2010-07-06 | 2018-10-16 | Nicira, Inc. | Network control apparatus and method for populating logical datapath sets |
WO2012141241A1 (en) * | 2011-04-13 | 2012-10-18 | 日本電気株式会社 | Network, data transfer node, communication method, and program |
US9043452B2 (en) | 2011-05-04 | 2015-05-26 | Nicira, Inc. | Network control apparatus and method for port isolation |
US9288081B2 (en) | 2011-08-17 | 2016-03-15 | Nicira, Inc. | Connecting unmanaged segmented networks by managing interconnection switching elements |
US8913611B2 (en) | 2011-11-15 | 2014-12-16 | Nicira, Inc. | Connection identifier assignment and source network address translation |
CN106850878B (en) | 2011-08-17 | 2020-07-14 | Nicira股份有限公司 | Logical L3 routing |
US9137107B2 (en) | 2011-10-25 | 2015-09-15 | Nicira, Inc. | Physical controllers for converting universal flows |
US9203701B2 (en) | 2011-10-25 | 2015-12-01 | Nicira, Inc. | Network virtualization apparatus and method with scheduling capabilities |
US9178833B2 (en) | 2011-10-25 | 2015-11-03 | Nicira, Inc. | Chassis controller |
US9288104B2 (en) | 2011-10-25 | 2016-03-15 | Nicira, Inc. | Chassis controllers for converting universal flows |
AU2013249154B2 (en) | 2012-04-18 | 2015-12-10 | Nicira, Inc. | Exchange of network state information between forwarding elements |
US9231892B2 (en) | 2012-07-09 | 2016-01-05 | Vmware, Inc. | Distributed virtual switch configuration and state management |
US9471385B1 (en) | 2012-08-16 | 2016-10-18 | Open Invention Network Llc | Resource overprovisioning in a virtual machine environment |
US9432215B2 (en) | 2013-05-21 | 2016-08-30 | Nicira, Inc. | Hierarchical network managers |
US9571386B2 (en) | 2013-07-08 | 2017-02-14 | Nicira, Inc. | Hybrid packet processing |
US10218564B2 (en) | 2013-07-08 | 2019-02-26 | Nicira, Inc. | Unified replication mechanism for fault-tolerance of state |
US9602312B2 (en) | 2013-07-08 | 2017-03-21 | Nicira, Inc. | Storing network state at a network controller |
US9282019B2 (en) | 2013-07-12 | 2016-03-08 | Nicira, Inc. | Tracing logical network packets through physical network |
US9344349B2 (en) | 2013-07-12 | 2016-05-17 | Nicira, Inc. | Tracing network packets by a cluster of network controllers |
US9407580B2 (en) | 2013-07-12 | 2016-08-02 | Nicira, Inc. | Maintaining data stored with a packet |
US9952885B2 (en) | 2013-08-14 | 2018-04-24 | Nicira, Inc. | Generation of configuration files for a DHCP module executing within a virtualized container |
US9887960B2 (en) | 2013-08-14 | 2018-02-06 | Nicira, Inc. | Providing services for logical networks |
US9973382B2 (en) | 2013-08-15 | 2018-05-15 | Nicira, Inc. | Hitless upgrade for network control applications |
US9432204B2 (en) | 2013-08-24 | 2016-08-30 | Nicira, Inc. | Distributed multicast by endpoints |
US9577845B2 (en) | 2013-09-04 | 2017-02-21 | Nicira, Inc. | Multiple active L3 gateways for logical networks |
US9503371B2 (en) | 2013-09-04 | 2016-11-22 | Nicira, Inc. | High availability L3 gateways for logical networks |
US9674087B2 (en) | 2013-09-15 | 2017-06-06 | Nicira, Inc. | Performing a multi-stage lookup to classify packets |
US9602398B2 (en) | 2013-09-15 | 2017-03-21 | Nicira, Inc. | Dynamically generating flows with wildcard fields |
US9596126B2 (en) | 2013-10-10 | 2017-03-14 | Nicira, Inc. | Controller side method of generating and updating a controller assignment list |
US9575782B2 (en) | 2013-10-13 | 2017-02-21 | Nicira, Inc. | ARP for logical router |
US10063458B2 (en) | 2013-10-13 | 2018-08-28 | Nicira, Inc. | Asymmetric connection with external networks |
US10193771B2 (en) | 2013-12-09 | 2019-01-29 | Nicira, Inc. | Detecting and handling elephant flows |
US9967199B2 (en) | 2013-12-09 | 2018-05-08 | Nicira, Inc. | Inspecting operations of a machine to detect elephant flows |
US9569368B2 (en) | 2013-12-13 | 2017-02-14 | Nicira, Inc. | Installing and managing flows in a flow table cache |
US9996467B2 (en) | 2013-12-13 | 2018-06-12 | Nicira, Inc. | Dynamically adjusting the number of flows allowed in a flow table cache |
US9602392B2 (en) | 2013-12-18 | 2017-03-21 | Nicira, Inc. | Connectivity segment coloring |
US9602385B2 (en) | 2013-12-18 | 2017-03-21 | Nicira, Inc. | Connectivity segment selection |
US9419855B2 (en) | 2014-03-14 | 2016-08-16 | Nicira, Inc. | Static routes for logical routers |
US9313129B2 (en) | 2014-03-14 | 2016-04-12 | Nicira, Inc. | Logical router processing by network controller |
US9590901B2 (en) | 2014-03-14 | 2017-03-07 | Nicira, Inc. | Route advertisement by managed gateways |
US9225597B2 (en) | 2014-03-14 | 2015-12-29 | Nicira, Inc. | Managed gateways peering with external router to attract ingress packets |
US9503321B2 (en) | 2014-03-21 | 2016-11-22 | Nicira, Inc. | Dynamic routing for logical routers |
US9647883B2 (en) | 2014-03-21 | 2017-05-09 | Nicria, Inc. | Multiple levels of logical routers |
US9413644B2 (en) | 2014-03-27 | 2016-08-09 | Nicira, Inc. | Ingress ECMP in virtual distributed routing environment |
US9893988B2 (en) | 2014-03-27 | 2018-02-13 | Nicira, Inc. | Address resolution using multiple designated instances of a logical router |
US10193806B2 (en) | 2014-03-31 | 2019-01-29 | Nicira, Inc. | Performing a finishing operation to improve the quality of a resulting hash |
US9385954B2 (en) | 2014-03-31 | 2016-07-05 | Nicira, Inc. | Hashing techniques for use in a network environment |
US9794079B2 (en) | 2014-03-31 | 2017-10-17 | Nicira, Inc. | Replicating broadcast, unknown-unicast, and multicast traffic in overlay logical networks bridged with physical networks |
US9985896B2 (en) | 2014-03-31 | 2018-05-29 | Nicira, Inc. | Caching of service decisions |
US9602422B2 (en) | 2014-05-05 | 2017-03-21 | Nicira, Inc. | Implementing fixed points in network state updates using generation numbers |
US9742881B2 (en) | 2014-06-30 | 2017-08-22 | Nicira, Inc. | Network virtualization using just-in-time distributed capability for classification encoding |
US9553803B2 (en) | 2014-06-30 | 2017-01-24 | Nicira, Inc. | Periodical generation of network measurement data |
US9379956B2 (en) | 2014-06-30 | 2016-06-28 | Nicira, Inc. | Identifying a network topology between two endpoints |
US9858100B2 (en) | 2014-08-22 | 2018-01-02 | Nicira, Inc. | Method and system of provisioning logical networks on a host machine |
US9768980B2 (en) | 2014-09-30 | 2017-09-19 | Nicira, Inc. | Virtual distributed bridging |
US10511458B2 (en) | 2014-09-30 | 2019-12-17 | Nicira, Inc. | Virtual distributed bridging |
US11178051B2 (en) | 2014-09-30 | 2021-11-16 | Vmware, Inc. | Packet key parser for flow-based forwarding elements |
US10250443B2 (en) | 2014-09-30 | 2019-04-02 | Nicira, Inc. | Using physical location to modify behavior of a distributed virtual network element |
US10020960B2 (en) | 2014-09-30 | 2018-07-10 | Nicira, Inc. | Virtual distributed bridging |
US10469342B2 (en) | 2014-10-10 | 2019-11-05 | Nicira, Inc. | Logical network traffic analysis |
US9787605B2 (en) | 2015-01-30 | 2017-10-10 | Nicira, Inc. | Logical router with multiple routing components |
US10038628B2 (en) | 2015-04-04 | 2018-07-31 | Nicira, Inc. | Route server mode for dynamic routing between logical and physical networks |
US9923760B2 (en) | 2015-04-06 | 2018-03-20 | Nicira, Inc. | Reduction of churn in a network control system |
US10225184B2 (en) | 2015-06-30 | 2019-03-05 | Nicira, Inc. | Redirecting traffic in a virtual distributed router environment |
CN105162716A (en) * | 2015-07-28 | 2015-12-16 | 上海华为技术有限公司 | Flow control method and apparatus under NFV configuration |
US10129142B2 (en) | 2015-08-11 | 2018-11-13 | Nicira, Inc. | Route configuration for logical router |
US10075363B2 (en) | 2015-08-31 | 2018-09-11 | Nicira, Inc. | Authorization for advertised routes among logical routers |
US10204122B2 (en) | 2015-09-30 | 2019-02-12 | Nicira, Inc. | Implementing an interface between tuple and message-driven control entities |
US10095535B2 (en) | 2015-10-31 | 2018-10-09 | Nicira, Inc. | Static route types for logical routers |
US10333849B2 (en) | 2016-04-28 | 2019-06-25 | Nicira, Inc. | Automatic configuration of logical routers on edge nodes |
US10841273B2 (en) | 2016-04-29 | 2020-11-17 | Nicira, Inc. | Implementing logical DHCP servers in logical networks |
US11019167B2 (en) | 2016-04-29 | 2021-05-25 | Nicira, Inc. | Management of update queues for network controller |
US10484515B2 (en) | 2016-04-29 | 2019-11-19 | Nicira, Inc. | Implementing logical metadata proxy servers in logical networks |
US10091161B2 (en) | 2016-04-30 | 2018-10-02 | Nicira, Inc. | Assignment of router ID for logical routers |
US10205651B2 (en) | 2016-05-13 | 2019-02-12 | 128 Technology, Inc. | Apparatus and method of selecting next hops for a session |
US10153973B2 (en) | 2016-06-29 | 2018-12-11 | Nicira, Inc. | Installation of routing tables for logical router in route server mode |
US10560320B2 (en) | 2016-06-29 | 2020-02-11 | Nicira, Inc. | Ranking of gateways in cluster |
US10454758B2 (en) | 2016-08-31 | 2019-10-22 | Nicira, Inc. | Edge node cluster network redundancy and fast convergence using an underlay anycast VTEP IP |
US10341236B2 (en) | 2016-09-30 | 2019-07-02 | Nicira, Inc. | Anycast edge service gateways |
US10237123B2 (en) | 2016-12-21 | 2019-03-19 | Nicira, Inc. | Dynamic recovery from a split-brain failure in edge nodes |
US10742746B2 (en) | 2016-12-21 | 2020-08-11 | Nicira, Inc. | Bypassing a load balancer in a return path of network traffic |
US10212071B2 (en) | 2016-12-21 | 2019-02-19 | Nicira, Inc. | Bypassing a load balancer in a return path of network traffic |
US10616045B2 (en) | 2016-12-22 | 2020-04-07 | Nicira, Inc. | Migration of centralized routing components of logical router |
US10805239B2 (en) | 2017-03-07 | 2020-10-13 | Nicira, Inc. | Visualization of path between logical network endpoints |
US10681000B2 (en) | 2017-06-30 | 2020-06-09 | Nicira, Inc. | Assignment of unique physical network addresses for logical network addresses |
US10637800B2 (en) | 2017-06-30 | 2020-04-28 | Nicira, Inc | Replacement of logical network addresses with physical network addresses |
US10608887B2 (en) | 2017-10-06 | 2020-03-31 | Nicira, Inc. | Using packet tracing tool to automatically execute packet capture operations |
US10374827B2 (en) | 2017-11-14 | 2019-08-06 | Nicira, Inc. | Identifier that maps to different networks at different datacenters |
US10511459B2 (en) | 2017-11-14 | 2019-12-17 | Nicira, Inc. | Selection of managed forwarding element for bridge spanning multiple datacenters |
US10999220B2 (en) | 2018-07-05 | 2021-05-04 | Vmware, Inc. | Context aware middlebox services at datacenter edge |
US11184327B2 (en) | 2018-07-05 | 2021-11-23 | Vmware, Inc. | Context aware middlebox services at datacenter edges |
CN108958940A (en) * | 2018-07-09 | 2018-12-07 | 苏州浪潮智能软件有限公司 | A kind of computer processing method and device |
US10931560B2 (en) | 2018-11-23 | 2021-02-23 | Vmware, Inc. | Using route type to determine routing protocol behavior |
US10735541B2 (en) | 2018-11-30 | 2020-08-04 | Vmware, Inc. | Distributed inline proxy |
US10797998B2 (en) | 2018-12-05 | 2020-10-06 | Vmware, Inc. | Route server for distributed routers using hierarchical routing protocol |
US10938788B2 (en) | 2018-12-12 | 2021-03-02 | Vmware, Inc. | Static routes for policy-based VPN |
US10778457B1 (en) | 2019-06-18 | 2020-09-15 | Vmware, Inc. | Traffic replication in overlay networks spanning multiple sites |
US11159343B2 (en) | 2019-08-30 | 2021-10-26 | Vmware, Inc. | Configuring traffic optimization using distributed edge services |
US11641305B2 (en) | 2019-12-16 | 2023-05-02 | Vmware, Inc. | Network diagnosis in software-defined networking (SDN) environments |
US11283699B2 (en) | 2020-01-17 | 2022-03-22 | Vmware, Inc. | Practical overlay network latency measurement in datacenter |
US11606294B2 (en) | 2020-07-16 | 2023-03-14 | Vmware, Inc. | Host computer configured to facilitate distributed SNAT service |
US11616755B2 (en) | 2020-07-16 | 2023-03-28 | Vmware, Inc. | Facilitating distributed SNAT service |
US11611613B2 (en) | 2020-07-24 | 2023-03-21 | Vmware, Inc. | Policy-based forwarding to a load balancer of a load balancing cluster |
US11451413B2 (en) | 2020-07-28 | 2022-09-20 | Vmware, Inc. | Method for advertising availability of distributed gateway service and machines at host computer |
US11902050B2 (en) | 2020-07-28 | 2024-02-13 | VMware LLC | Method for providing distributed gateway service at host computer |
US11558426B2 (en) | 2020-07-29 | 2023-01-17 | Vmware, Inc. | Connection tracking for container cluster |
US11196628B1 (en) | 2020-07-29 | 2021-12-07 | Vmware, Inc. | Monitoring container clusters |
US11570090B2 (en) | 2020-07-29 | 2023-01-31 | Vmware, Inc. | Flow tracing operation in container cluster |
US11736436B2 (en) | 2020-12-31 | 2023-08-22 | Vmware, Inc. | Identifying routes with indirect addressing in a datacenter |
US11336533B1 (en) | 2021-01-08 | 2022-05-17 | Vmware, Inc. | Network visualization of correlations between logical elements and associated physical elements |
US11784922B2 (en) | 2021-07-03 | 2023-10-10 | Vmware, Inc. | Scalable overlay multicast routing in multi-tier edge gateways |
US11687210B2 (en) | 2021-07-05 | 2023-06-27 | Vmware, Inc. | Criteria-based expansion of group nodes in a network topology visualization |
US11711278B2 (en) | 2021-07-24 | 2023-07-25 | Vmware, Inc. | Visualization of flow trace operation across multiple sites |
US11855862B2 (en) | 2021-09-17 | 2023-12-26 | Vmware, Inc. | Tagging packets for monitoring and analysis |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05207068A (en) * | 1992-01-27 | 1993-08-13 | Nec Corp | Packet network design system |
EP0631413A2 (en) * | 1993-06-28 | 1994-12-28 | AT&T Corp. | Method for shortest path routing |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4669113A (en) * | 1985-04-26 | 1987-05-26 | At&T Company | Integrated network controller for a dynamic nonhierarchical routing switching network |
US4713806A (en) * | 1986-03-14 | 1987-12-15 | American Telephone And Telegraph Company, At&T Bell Laboratories | Communication system control arrangement |
US5297137A (en) * | 1991-01-30 | 1994-03-22 | International Business Machines Corporation | Process for routing data packets around a multi-node communications network |
GB2253970A (en) * | 1991-03-15 | 1992-09-23 | Plessey Telecomm | Data network management |
US5381404A (en) * | 1992-07-14 | 1995-01-10 | Mita Industrial Co., Ltd. | Packet-switching communication network and method of design |
US5345444A (en) * | 1992-09-30 | 1994-09-06 | At&T Bell Laboratories | Chuted, growable packet switching arrangement |
US5289303A (en) * | 1992-09-30 | 1994-02-22 | At&T Bell Laboratories | Chuted, optical packet distribution network |
JPH0793645B2 (en) * | 1993-01-11 | 1995-10-09 | 日本電気株式会社 | Signal connection controller |
JP3672341B2 (en) * | 1993-07-21 | 2005-07-20 | 富士通株式会社 | Communication network separation design method and its management method |
SE9500838L (en) * | 1994-06-13 | 1995-12-14 | Ellemtel Utvecklings Ab | Device and method for allocating resources of a physical network |
US5872918A (en) * | 1995-07-14 | 1999-02-16 | Telefonaktiebolaget Lm Erisson (Publ) | System and method for optimal virtual path capacity dimensioning with broadband traffic |
US5764740A (en) * | 1995-07-14 | 1998-06-09 | Telefonaktiebolaget Lm Ericsson | System and method for optimal logical network capacity dimensioning with broadband traffic |
-
1994
- 1994-06-13 SE SE9402059A patent/SE9402059D0/en unknown
-
1995
- 1995-06-12 AU AU27587/95A patent/AU692884B2/en not_active Ceased
- 1995-06-12 CN CN95194459A patent/CN1080501C/en not_active Expired - Fee Related
- 1995-06-12 DE DE69534216T patent/DE69534216D1/en not_active Expired - Fee Related
- 1995-06-12 JP JP8502038A patent/JPH10506243A/en active Pending
- 1995-06-12 DE DE69533064T patent/DE69533064D1/en not_active Expired - Fee Related
- 1995-06-12 CN CN95194460.6A patent/CN1104120C/en not_active Expired - Fee Related
- 1995-06-12 WO PCT/SE1995/000704 patent/WO1995034981A2/en active IP Right Grant
- 1995-06-12 US US08/765,159 patent/US6104699A/en not_active Expired - Lifetime
- 1995-06-12 EP EP95922843A patent/EP0765552B1/en not_active Expired - Lifetime
- 1995-06-12 JP JP8502037A patent/JPH10504426A/en active Pending
- 1995-06-12 WO PCT/SE1995/000703 patent/WO1995034973A2/en active IP Right Grant
- 1995-06-12 EP EP95922842A patent/EP0765554B1/en not_active Expired - Lifetime
- 1995-06-12 CA CA002192793A patent/CA2192793A1/en not_active Abandoned
- 1995-06-12 AU AU27586/95A patent/AU688917B2/en not_active Ceased
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05207068A (en) * | 1992-01-27 | 1993-08-13 | Nec Corp | Packet network design system |
EP0631413A2 (en) * | 1993-06-28 | 1994-12-28 | AT&T Corp. | Method for shortest path routing |
Non-Patent Citations (2)
Title |
---|
IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, Volume 2, April 1990, (Atlanta), G. GOPAL et al., "Dynamic Network Configuration Management Claims X -47-51,57-61. Claims A 52-56, 62-63". * |
PATENT ABSTRACTS OF JAPAN, Vol. 17, No. 642, E-1466; & JP,A,05 207 068 (NEC CORP), 13 August 1993. * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2317533A (en) * | 1996-07-29 | 1998-03-25 | Northern Telecom Ltd | Communications network |
US5940008A (en) * | 1996-07-29 | 1999-08-17 | Northern Telecom Limited | Communications switching network |
EP1098546A2 (en) * | 1999-11-04 | 2001-05-09 | Lucent Technologies Inc. | Methods and apparatus for derivative based optimization of wireless network performance |
EP1098546A3 (en) * | 1999-11-04 | 2001-10-31 | Lucent Technologies Inc. | Methods and apparatus for derivative based optimization of wireless network performance |
US6611500B1 (en) * | 1999-11-04 | 2003-08-26 | Lucent Technologies, Inc. | Methods and apparatus for derivative-based optimization of wireless network performance |
WO2001093504A1 (en) * | 2000-05-31 | 2001-12-06 | Nokia Corporation | Partitioning of a communications network |
US7257627B2 (en) | 2000-05-31 | 2007-08-14 | Nokia Corporation | Partitioning of a communications network |
US7006466B2 (en) * | 2001-03-09 | 2006-02-28 | Lucent Technologies Inc. | Dynamic rate control methods and apparatus for scheduling data transmissions in a communication network |
EP2651054A1 (en) | 2004-10-22 | 2013-10-16 | Cisco Technology, Inc. | Fibre channel over ethernet |
Also Published As
Publication number | Publication date |
---|---|
JPH10506243A (en) | 1998-06-16 |
SE9402059D0 (en) | 1994-06-13 |
DE69534216D1 (en) | 2005-06-23 |
EP0765552A2 (en) | 1997-04-02 |
AU692884B2 (en) | 1998-06-18 |
US6104699A (en) | 2000-08-15 |
EP0765554B1 (en) | 2005-05-18 |
AU2758795A (en) | 1996-01-05 |
WO1995034973A2 (en) | 1995-12-21 |
CN1104120C (en) | 2003-03-26 |
CN1155360A (en) | 1997-07-23 |
AU688917B2 (en) | 1998-03-19 |
EP0765552B1 (en) | 2004-05-19 |
WO1995034973A3 (en) | 1996-02-01 |
AU2758695A (en) | 1996-01-05 |
JPH10504426A (en) | 1998-04-28 |
DE69533064D1 (en) | 2004-06-24 |
EP0765554A2 (en) | 1997-04-02 |
CA2192793A1 (en) | 1995-12-21 |
WO1995034981A3 (en) | 1996-02-08 |
CN1080501C (en) | 2002-03-06 |
CN1154772A (en) | 1997-07-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6069894A (en) | Enhancement of network operation and performance | |
EP0765552B1 (en) | Enhancement of network operation and performance | |
US6304639B1 (en) | System and methods for controlling virtual paths within a network based on entropy rate function | |
US5727051A (en) | System and method for adaptive routing on a virtual path broadband network | |
US5764740A (en) | System and method for optimal logical network capacity dimensioning with broadband traffic | |
Guerin et al. | Equivalent capacity and its application to bandwidth allocation in high-speed networks | |
US6331986B1 (en) | Method for resource allocation and routing in multi-service virtual private networks | |
US7266122B1 (en) | System and method for allocating bandwidth in a communications environment | |
US6097722A (en) | Bandwidth management processes and systems for asynchronous transfer mode networks using variable virtual paths | |
US5953338A (en) | Dynamic control processes and systems for asynchronous transfer mode networks | |
EP0753979A1 (en) | Routing method and system for a high speed packet switching network | |
WO1998026612A9 (en) | Dynamic control processes and systems for asynchronous transfer mode networks | |
WO1998029984A9 (en) | Bandwidth management system using variable virtual paths | |
US20030138253A1 (en) | Dynamic wavelength management method in OBS networks | |
US6301267B1 (en) | Smart switch | |
US6798747B1 (en) | System and method for time slot assignment in a fiber optic network simulation plan | |
Arvidsson | High level B-ISDN/ATM traffic management in real time | |
US20030179717A1 (en) | Methods and apparatus for controlling multi-layer communications networks | |
CA2192794C (en) | Enhancement of network operation and performance | |
Plante et al. | A typology for multi-technology, multi-service broadband network synthesis | |
Randhawa et al. | Traffic measurement based bandwidth dimensioning of broadband networks | |
Medhi et al. | Dimensioning and computational results for wide-area broadband networks with two-level dynamic routing | |
AU750013B2 (en) | Explicit rate marking for flow control in ATM networks | |
Valadas | Dimensioning and Resource Management of ATM Networks | |
Chiotis et al. | The effective bandwidth approach: configuration of ATM virtual connections |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 95194460.6 Country of ref document: CN |
|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): AM AT AU BB BG BR BY CA CH CN CZ DE DK EE ES FI GB GE HU JP KE KG KP KR KZ LK LR LT LU LV MD MG MN MW MX NO NZ PL PT RO RU SD SE SG SI SK TJ TT UA US UZ VN |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): KE MW SD SZ UG AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN ML MR NE SN TD TG |
|
AK | Designated states |
Kind code of ref document: A3 Designated state(s): AM AT AU BB BG BR BY CA CH CN CZ DE DK EE ES FI GB GE HU JP KE KG KP KR KZ LK LR LT LU LV MD MG MN MW MX NO NZ PL PT RO RU SD SE SG SI SK TJ TT UA US UZ VN |
|
AL | Designated countries for regional patents |
Kind code of ref document: A3 Designated state(s): KE MW SD SZ UG AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN ML MR NE SN TD TG |
|
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 1995922843 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2192794 Country of ref document: CA |
|
WWP | Wipo information: published in national office |
Ref document number: 1995922843 Country of ref document: EP |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 08765106 Country of ref document: US |
|
WWG | Wipo information: grant in national office |
Ref document number: 1995922843 Country of ref document: EP |