1 Introduction

Wireless Multiple-Input–Multiple-Output (MIMO) communication systems [1,2,3,4] involve multiple antennas at both the transmitter and receiver to enhance data throughput, reliability, and coverage. Different data streams are transmitted and received simultaneously, exploiting the spatial diversity of the wireless channel.

Joint Transmission (JT), commonly named Coordinated MultiPoint (CoMP) in the mobile access network context, is a multiuser MIMO system which aims to improve users’ service quality by pre-compensating the Inter-Cell Interference (ICI) components, i.e., increasing the Signal-to-Interference-and-Noise-Ratio (SINR), through a beamforming technique. This consists of splitting information symbols among different base stations (BSs) to maximize the transmission rate (TR) of served users, making the channel’s attenuation and phase-shifting effects negligible. However, attenuation could imply a preliminary signal amplification unsustainable by each BS, thus necessarily requiring the definition, for each symbol associated with a user to be served, of a power scaling transmission factor \(p_{s} < 1\), which is also responsible for TR reduction.

An optimization problem, looking for the optimal subset of users to be served in each time slot, can be identified to make the most of the mechanism. From a practical perspective, it consists of finding the subset of users having the highest TR after defining the highest \(p_{s}\) values not exceeding the BSs maximum power. This subset can be theoretically identified by evaluating all those possible iteratively. For each of these, the optimal \(p_{s}\) values for the considered users must be found (e.g., exploiting a convex optimizer), and then the total TR can be computed. However, this brute-force approach can be computationally expensive and becomes unfeasible for many potential users because of subsets factorial scaling. It must be always reminded that in JT-based systems users change in time and the optimal subset in a given time frame must be detected in the shortest time possible. For this reason, it is important to define strategies for the solutions space exploration characterized by a good trade-off between the quality of the solution and the execution time.

Quantum computation [5] can offer advantages in tackling search and optimization problems. Thanks to quantum superposition and entanglement, it virtually performs a simultaneous evaluation of multiple possibilities, data, and combinations, thus significantly accelerating search processes. These features hold promise for addressing optimization problems, where quantum algorithms and procedures can find solutions with higher quality or in a lower time than classical methods. Reminding that quantum computing has already provided interesting results when solving problems concerning wireless communications [6,7,8,9], the investigation of the potentialities of this paradigm for users scheduling in JT-based MIMO systems can be considered reasonable.

This work is going to propose new approaches—one based on quantum-complaint heuristic (QUBO-assisted), the other on a simplification of the \(p_{s}\) exploration (Naïve)—and to compare them with the brute-force one in terms of total execution time and quality of the obtained solution. The analysis was performed with ad-hoc developed software prototypes, verifying the effectiveness of the proposed methods. A longer-term perspective of this work can be the definition of a distributed computing system for monitoring the JT mechanism in a quite small time window based on application-specific hardware.

The article is structured as follows. Sect. 2 introduces JT mechanism in wireless MIMO systems, the fundamentals of QUBO formulation are described in Sects. 3, 4 presents the proposed quantum-compliant strategies for JT scheduling optimization, Sect. 5 discusses results, and Sect. 6 concludes the article with some future perspectives.

2 Joint transmission theoretical foundations

In this section, the mathematical formalism and the working principles of JT systems are introduced. It emphasizes that it follows the mathematical formalism described in [10], which is based on a phasor-domain representation for sinusoidal signals and normalizes electric powers to an implicit resistance (\(P_{\text {normalized}} = \vert V \vert ^{2}\)).

2.1 MIMO model

In a wireless MIMO system with \(N_{\text {BS}}\) base stations (BSs) (composed of a single antenna) and \(S \le N_{\text {BS}}\) receiving users, the \(n^{\text {th}}\) BS transmits a symbol \(x_n\), which is received by a user s as \(h_{sn} x_n\), where \(h_{sn}\) is a channel coefficient related to path-loss attenuation and random phase shift of the transmitted signal. The overall signal received by user s is given by:

$$\begin{aligned} y_s = \sum _{n=1}^{N_{\text {BS}}} h_{sn} x_n + \theta _{s}, \end{aligned}$$
(1)

where \(\theta _{s}\) is the intrinsic thermal noise. Therefore, the MIMO system transfer function can be represented in matrix form as:

$$\begin{aligned} Y= H \cdot X + \theta \,, \end{aligned}$$
(2)

where \( H\in \mathbb {C}^{S \times N_{\text {BS}}}\) is the channel matrix, X and Y are transmitted and received signal vectors and \(\theta \) is the intrinsic thermal noise (TN) vector. Assuming that each BS transmits the symbol of one user, which implies \(N_{\text {BS}} = S\) and \(x_{s} = b_{s}\), the user s receives the signal:

$$\begin{aligned} y_{s} = \sum _{n=1}^{N_{\text {BS}}} h_{sn} x_{n} + \theta _s = \overset{{{\textbf {useful}}}}{h_{ss} x_{s}} + \overset{{{\textbf {interference}}}}{\sum _{n=1, n\ne s}^{N_{\text {BS}}} h_{sn} x_{n}} + \overset{{{\textbf {noise}}}}{\theta _{s}} \,. \end{aligned}$$
(3)

It is characterized by two unwanted components: the interference of the signals transmitted by other BSs and the intrinsic thermal noise,

$$\begin{aligned} \text {SINR} {\mathop {=}\limits ^{\Delta }} \frac{P_{\text {useful}}}{P_{\text {interference}} + P_{\text {noise}}} = \frac{|h_{ss} x_{s}|^2}{\sum _{n=1, n\ne s}^{N_{\text {BS}}} |h_{sn} x_{n}|^2 +\sigma _{\theta }^2} \,, \end{aligned}$$
(4)

where \(\sigma _{\theta }^2\) is the TN power. SINR is also used to determine the transmission rate (TR) over a fixed bandwidth B:

$$\begin{aligned} r {\mathop {=}\limits ^{\Delta }} B \log _2{(1+\text {SINR})} \,. \end{aligned}$$
(5)

Maximizing SINR is crucial for maximizing the transmission rate because bandwidth is typically constant.

Fig. 1
figure 1

Conceptual scheme of a MIMO system with Joint transmission. N BSs are exploited to transmit S symbols to S users. The symbol \(b_s\) to be transmitted to user s is associated with the base station n according to the element \(w_{ns}\) of the beamforming matrix, while the signal of the BS n is provided to user s according to the element \(h_{sn}\) of the channel matrix. The received signal \(y_s\) of each user depends on the linear combination of the BS signals and on the thermal noise \(n_s\) (Color figure online)

2.2 Join transmission mechanism

In JT-based networks, the goal is to remove interference through compensation by distributing user symbols (\(b_{s}\))— equal to one in this work—to all base stations (BSs) through a technique called beamforming (Fig. 1). Therefore, each BS transmits a linear combination of user symbols:

$$\begin{aligned} x_{n} = \sum _{s=1}^{S} w_{ns}b_{s} \,, \end{aligned}$$
(6)

thus making the received signals equal to:

$$\begin{aligned} Y= H \cdot X + \theta = H \cdot ( W \cdot b )+ \theta \,, \end{aligned}$$
(7)

where \(W \in \mathbb {C}^{N_{\text {BS}} \times S}\) is the beamforming matrix engineered to compensate for interference and H depends on the environment (\(H \times W = \mathbb {I}_{\text {dim}\{S\}} \Leftrightarrow W = H^{\dagger }\), where \(\mathbb {I}_{\text {dim}\{S\}}\) is the identity matrix and \(H^{\dagger }\) is Moore-Penrose inverse matrix of H).

Beamforming also implies signal amplification of the transmitted signal against channel attenuation, potentially exceeding the maximum supported power (\(\sum _{s=1}^{S} |w_{ns} b_{s}|^2 \ge P_{\text {max}_{\text {BS}_n}}\)), so a power partition factor (\(p_{s}\)) is introduced for each symbol. In this way, user s receives the following signal:

$$\begin{aligned} y_{s} = h_{ss} w_{ss} \sqrt{p_{s}} b_{s} + \theta _{s} \,, \end{aligned}$$
(8)

and the TR can be written, according to the Shannon–Hartley theorem, as:

$$\begin{aligned} r {\mathop {=}\limits ^{\Delta }} B \log _2{\biggr (1+ \frac{p_s |b_{s}|^2}{\sigma _{\theta }^2}\biggl )} \,. \end{aligned}$$
(9)

For optimizing the performance of a mobile network, the goal is to maximize the overall TR of the network:

$$\begin{aligned} \text {find} \qquad \max \left\{ r_{S_{\text {served}}} \right\} = \max \biggl \{ B \sum _{s \in S_{\text {served}}} \log _2{\biggr (1+ \frac{p_s |b_{s}|^2}{\sigma _{\theta }^2}\biggl )} \biggr \} \,. \end{aligned}$$
(10)

In practice, there are more potential users (\(N_{\text {UE}}\)) than BSs (\( N_{\text {BS}}\)), and finding the best user subset involves iterating through all possible subsets and optimizing \(p_{s}\) for each user in each subset. Considering the increasing monotone metric in Eq. 9, the optimal \(p_s\) values of a given user subset can be identified with a convex optimizer.

The hardness of the task comes from the necessity of solving \(N_{\text {subset}} = \left( {\begin{array}{c}N_{\text {UE}}\\ S\end{array}}\right) = \frac{N_{\text {UE}}!}{S! (N_{\text {UE}}-S)!}\) convex optimization problems and compare the obtained results for the selection of the best subset of users (brute-force method). Therefore, the complexity scales factorially with the number of users, thus making brute-force exploration unfeasible.

In this article, alternative methods are proposed to overcome this computational limitation, as quickly identifying the optimal user subset is essential in JT-based systems where users change over time.

3 Quadratic unconstrained binary optimization

QUBO modeling and quantum computational paradigms for QUBO problem solving are briefly introduced in the following. More details are available at [11,12,13].

3.1 Mathematical model

Quadratic Unconstrained Binary Optimization (QUBO) is a mathematical formulation capable of describing many real-world problems [11]. The term Optimization suggests the goal of this formulation: to minimize an objective function. As the term Binary implies, it employs unipolar binary variables (0/1). The Quadratic term refers to the highest power applied to variables, while Unconstrained indicates that the constraints on variables are not introduced in a standard way. The objective function can be written as:

$$\begin{aligned} \text {Obj}(c,a_{i}, b_{ij}, x_{i}) = c+ \sum _{i} a_{i} \cdot x_{i} + \sum _{i<j} b_{ij} \cdot x_{i}x_{j} \,, \end{aligned}$$
(11)

where \(x_{i} \in [0,1]\) is a binary variable, \(x_{i}x_{j}\) is a coupler of two variables, \(a_{i}\) is a single-variable weight, \(b_{ij}\) is a strength which controls the influence of variables i and j and c is an offset, which can be neglected during the optimization.

In reality, QUBO formulation can take into account the variable constraints, introducing quadratic penalties to the objective function

$$\begin{aligned} \text {minimize/maximize} \quad y = f({\textbf {x}}) + \lambda g({\textbf {x}}) \,, \end{aligned}$$
(12)

where \(\lambda \) is a positive penalty parameter to be multiplied by the constraint function \(g({\textbf {x}})\). This trick allows the evaluation of constraints during the solver execution, but correctly sizing \(\lambda \) is critical and can affect the solution quality. In fact, a too-low value makes the constraint function negligible, while a too-high value flattens the objective function. According to [11], \(\lambda g({\textbf {x}})\) should be a fraction (\(\in [75\%; 150\%]\)) of \(f({\textbf {x}})\).

It is possible to prove that the objective function of a QUBO problem is equivalent to the Ising Hamiltonian, used in statistical mechanics for the description of ferromagnetism in terms of atomic spins:

$$\begin{aligned} \mathcal {H} = \sum _{i} h_{i} \cdot s_{i} + \sum _{i<j} J_{ij} \cdot s_{i}s_{j} \,. \end{aligned}$$
(13)

Ising model involves bipolar binary variables (\(s_{i} \in \{-1,+1\}\), associated with spin orientations) instead of unipolar ones, \(h_{i}\) is the single-spin external field, whose sign is related to the preferred orientation of a spin with respect to that of an applied external field, and \(J_{ij}\) magnitude and sign quantify the interaction between neighbor spin pairs and their preferred alignment (ferromagnetic or anti-ferromagnetic), respectively. It is possible to move from Ising to QUBO representation (and vice-versa) with the following relations:

$$\begin{aligned} \begin{aligned} x_{i}&= \frac{1+ s_{i}}{2} \,, \\ s_{i}&= 2x_{i}-1 \,. \end{aligned} \end{aligned}$$
(14)

Even though the conversion changes the values of the single-variable and two-variable coefficients, the optimal solution does not change. A QUBO solution variable equal to 0(1) has a corresponding Ising solution variable equal to -1(+1).

3.2 Solving QUBO models with quantum computers

QUBO and Ising are the currently most feasible formulations for exploiting quantum-related solvers. Quantum computing is a computational paradigm which exploits quantum mechanics for executing calculations. It is possible to define quantum algorithms capable of solving specific problems with a computational complexity lower than their corresponding classical counterparts. This is possible thanks to quantum principles [5] like:

  • Superposition, allowing the implementation of a virtual parallel computation;

  • Entanglement, an intrinsic interaction between quantum systems exploitable to accelerate the convergence to the problem’s solution;

  • Tunneling, thanks to which quantum systems and particles can pass through high and narrow barriers instead of jum** them, as done by classic methods, such as hill climbing and simulated annealing.

The two most consolidated quantum computation paradigms are the general-purpose quantum circuit model and the special-purpose quantum annealing. In the first case, operations are executed in terms of quantum gates [5], a sort of extension to qubit (the unit of information of quantum computing) of logic gates employed in classical computing.

In quantum annealing [14], an objective function is mapped onto the energy profile of a quantum physical system, so its minimum-energy state can be associated with the problem’s optimal solution. This can be achieved with quantum properties, like superposition, tunneling, and adiabatic evolution, differently from the classical simulated annealing, where the barrier is jumped instead of tunneled.

Fig. 2
figure 2

Exploration capabilities and limits of quantum annealing (left) and simulated annealing (right). The first can efficiently explore the rough part of the energy landscape, while cannot easily overcome the wide energy barrier. On the contrary, simulated annealing exploits thermal fluctuations to overcome the wide energy barrier but shows more difficulties with high and narrow peaks (Color figure online)

The barrier-overcoming mechanism is strictly related to the fact that the effectiveness of an optimization algorithm on a specific problem is strongly related to the characteristic of its energy profile [15]. For example, simulated annealing and other approaches based on local search can overcome wide and smooth barriers more easily than high and narrow ones, while, on the contrary, quantum annealing performs well with high and narrow peaks and is expected to be less effective in exploring vast and flat regions [16, 17].

The energy profiles associated with real-world problems are usually heterogeneous, as reported in Fig. 2. This implies that the probability of success of each solver depends on the compatibility of its exploration mechanism with the problem’s energy profile region and that the definition of alternative QUBO models for the same problem could imply so different profiles that their best solver could change.

Current quantum annealers natively encode and solve Ising problems, so a QUBO problem can be solved by translating it into an Ising one —according to (14)—and handling the partial connectivity of the backend with minor embedding techniques [18]. Moreover, these are generally affected by errors that can be associated with decoherence phenomena; for this reason, quantum-inspired techniques, such as digital annealing [19], simulated quantum annealing [13] and simulated adiabatic bifurcation [20], have been proposed to emulate the quantum annealing behavior on classical hardware, thus avoiding errors associated with qubits decoherence.

The same issues concerning qubit connectivity and decoherence are observed on quantum computers based on the circuit model. Moreover, the currently best procedures exploiting these quantum computers to solve QUBO/Ising problems require the assistance of classical computers [21,22,23]. In the context of this article, quantum annealing is employed as quantum QUBO solver, since real hardware based on this paradigm [24] is nowadays capable of processing more QUBO variables than the corresponding one based on the circuit model [25].

4 Proposed solutions

This work proposes alternative approaches for reducing the computational complexity and the time required for selecting the optimal subset of JT-served users. The first proposal (Sect. 4.1), called Naïve, simplifies the \(p_{s}\) exploration for each subset, accepting sub-optimal solutions. The second approach performs a quantum-compliant preliminary fast-approximated user selection, which will be examined with the brute-force method. In this strategy, the user set reduction is done by exploiting a QUBO model (Sect. 4.2).

Before discussing the two proposed solutions, it is important to clarify that they operate at the single-BS-triplet level and in a specific sub-channel. As detailed in Sect. 5.1, in a specific sub-channel users could receive signals from more than three BSs belonging to different triplets. For this reason, in a specific sub-channel, users are preliminarily separated in such a way that each of them is associated with a single BS triplet. In particular, each user is associated with the triplet whose BSs are all capable of providing a signal \(h_{sn}x_n \ne 0\) (Equation (1)) and the highest receivable power, according to TIM S.p.A. MIMO simulator. If the previous condition is not satisfied, the same procedure is applied by decreasing the highest number of BSs providing non-null received signals.

The described procedure can be repeated for all sub-channels of the analyzed MIMO system. Assuming that the optimization BS-triplet-level problems of each sub-channel can be analyzed independently on those of the other sub-channels, tasks can be also parallelized.

4.1 Naïve

The problem’s computational complexity can be reduced, assuming all the S users in each subset have the same \(p_{s}\). In this way, the \(p_s\), allowing each BS to transmit at its maximum power, can be computed independently of the others; then, the power constraint is satisfied for all the \(N_{\text {BS}}\) available BSs by choosing the lowest \(p_s\).

An algebraic example with three BSs and two users is reported in the following;

$$\begin{aligned} \begin{aligned} \begin{bmatrix} \vert w_{11} \vert ^2 &{} \vert w_{12} \vert ^2 \\ \vert w_{21} \vert ^2 &{} \vert w_{22} \vert ^2 \\ \vert w_{31} \vert ^2 &{} \vert w_{32} \vert ^2 \end{bmatrix} \begin{bmatrix} p_1 \\ p_2 \end{bmatrix}&= \begin{bmatrix} \vert w_{11} \vert ^2 p_1 + \vert w_{12} \vert ^2 p_2 \\ \vert w_{21} \vert ^2 p_1 + \vert w_{22} \vert ^2 p_2 \\ \vert w_{31} \vert ^2 p_1 + \vert w_{32} \vert ^2 p_2 \end{bmatrix} = \\&{\mathop {=}\limits ^{\text {if }p_1 = p_2}} \begin{bmatrix} (\vert w_{11} \vert ^2 + \vert w_{12} \vert ^2) p \\ (\vert w_{21} \vert ^2 + \vert w_{22} \vert ^2) p \\ (\vert w_{31} \vert ^2 + \vert w_{32} \vert ^2) p \end{bmatrix} \le \begin{bmatrix} P_{\max _{\text {BS}}} \\ P_{\max _{\text {BS}}} \\ P_{\max _{\text {BS}}} \end{bmatrix} \\ p&= \min \left\{ \begin{bmatrix} P_{\max _{\text {BS}}}/(\vert w_{11} \vert ^2 + \vert w_{12} \vert ^2) \\ P_{\max _{\text {BS}}}/(\vert w_{21} \vert ^2 + \vert w_{22} \vert ^2) \\ P_{\max _{\text {BS}}}/(\vert w_{31} \vert ^2 + \vert w_{32} \vert ^2) \end{bmatrix} \right\} \,. \end{aligned} \end{aligned}$$
(15)

The total time required to find a solution is reduced because the convex optimizer is not involved. However, the obtained solution is sub-optimal because only one BS transmits at the maximum power, so exploiting all the antennas at their best is impossible.

Fig. 3
figure 3

Block scheme of the proposed quantum-compliant approach, where users are reduced by solving a QUBO problem and a classical computer evaluates all the reduced users’ set combinations, identifying the optimal power partition factor and the best users to serve (Color figure online)

4.2 Quantum-compliant users set reduction

The second approach divides the problem into two parts:

  • Reduction of the user set with a QUBO optimization problem to solve in a quantum or classical way;

  • Application of the convex classical solver for optimizing the power partition factors on each combination of users in the reduced set, to find one with the highest total transmission rate.

Therefore, the QUBO solver allows a first screening of the users, based on their characteristics and the problem constraints, to analyze with the accurate but more complex brute-force approach.

This strategy, which is graphically represented in Fig. 3, and whose pseudocode is reported in Algorithm 1 significantly reduces the computational complexity and the execution time, without affecting significantly the quality of the final solution. As shown in the aforementioned figure, a feedback mechanism could be inserted to improve the quality of the user reduction performed by the QUBO solver.

From a theoretical point of view, a hybrid approach, where the quantum-compliant method exploits the Naïve method in its second step, can be considered. However, this is not reported in this article because it is expected to require more execution time than the Naïve-only approach and to provide solutions with lower transmission rate than the canonical quantum-compliant method involving the convex optimizer.

figure a
Fig. 4
figure 4

Structure of the binary variables chosen, which is, fundamentally, a \(N_{\text {UE}} \times N_{\text {BS}} \) matrix of unipolar binary variables—\(x_{sn}\) is equal to one (colored red) if the \(n^{\text {th}}\) BS serves the \(s^{\text {th}}\) user in the optimal solution, and graphical representation of the principle of the knapsack problem exploited for the QUBO formulation (Color figure online)

4.2.1 First QUBO formulation

The data structure of the proposed QUBO formulation is fundamentally a matrix \(N_{\text {UE}} \times N_{\text {BS}}\) of unipolar binary variables (as shown in Fig. 4a), one for each user–BS pair, in which rows are devoted to users, while columns to BSs. The \(x_{sn}\) variable assumes value one if the \(n^{\text {th}}\) BS serves the \(s^{\text {th}}\) user.

Considering that the optimal subset of users has to be identified such that the total TR is maximized without exceeding the BSs maximum power, a similarity to the well-known knapsack problem (shown in Fig. 4b) [11] can be found. Indeed, it aims to define, for a set of objects X, each of which is labeled as \(x_i\) and is characterized by a weight \(w_i\) and a preference score \(p_i\), the best subset to be put into a bag, guaranteeing that the total weight does not exceed a threshold W (\(0 < \sum _{i=1}^{\dim (X)} w_i x_i \le W\)) while maximizing the total preference score \(P = \sum _{i=1}^{\dim (X)} p_{i} x_{i}\).

For each BS of the group employed for the JT, i.e., for each column, the problem has to satisfy knapsack-like conditions:

  • Maximization of the TR, which plays the role of the total preference scores;

  • Verifying that the BS transmitted power does not exceed the maximum one, which is equivalent to assuring that the objects’ total weight is lower than a certain threshold.

In the QUBO formulation, for optimizing the overall TR of the network, it was chosen to optimize the sum SNR:

$$\begin{aligned} \text {find} \qquad \max \biggl \{ \sum _{s \in S_{\text {served}}} \frac{p_s |b_{s}|^2}{\sigma _{\theta }^2}\biggr \} \, . \end{aligned}$$
(16)

This choice was made because, under the Shannon law assumption, TR has a logarithmic dependence from SNR, which introduces a nonlinearity that is not compatible with QUBO formulation. Optimizing the sum of SNR of the network can be a simpler and more tractable approach. Moreover, by improving the SNR of each link in the network, the overall quality of the communication channels is expected to be improved.

Even if optimizing the sum of SNR of a mobile network may not directly guarantee the highest TR sum, it can be a reasonable approximation, since the optimization of the SNR sum usually leads to the optimization of the overall system’s TR.

Moreover, since the BSs have to serve the same users, the complete QUBO formulation requires more than the combinations of \(N_{\text {BS}}\) parallel knapsack problems. In particular, all variables in a row must be all equal (as shown in Fig. 5a), thus ensuring that a user is served by all the BSs or not, and the sum of binary variables on a column must be equal to S (as shown in Fig. 5b), so that the BSs serve a total number of users equal to S.

Fig. 5
figure 5

Constraints (Color figure online)

Fig. 6
figure 6

Graphical representation of the SNR optimization and a possible final solution (Color figure online)

To summarize, the cost function is composed of four contributions:

  1. 1.

    Maximize SNR:

    $$\begin{aligned} f_{\text {SNR}}({\textbf {x}}) = - \sum _{s} \frac{p_{s}}{\sigma _{\theta }^{2}} \sum _{n} x_{sn} \,, \end{aligned}$$
    (17)

    where \(p_s\) is the power scaling factor of the \(s^{\text {th}}\) user, \(\sigma _{\theta }^{2}\) is the thermal noise power, and the minus sign is required to move the maximization problem to a minimization one. This term tends to force all \(x_{sn}\) to one. As observable in Fig. 6a, Boolean variables belonging to the same row/user have the same \(p_s\). Moreover, different purple intensities refer to the different \(p_s\), and consequently SNR, of users;

  2. 2.

    Power constraint, i.e., the BS maximum power must be not exceeded:

    $$\begin{aligned} f_{\text {pow}}({\textbf {x}}) = \sum _{n}\biggl [\sum _{s} |w_{sn}|^2 p_{s} x_{sn} + \sum _{i \in A} \alpha _{i} a_{i} - P_{\text {max}}\biggr ]^2 \,, \end{aligned}$$
    (18)

    where \(|w_{sn}|^2 p_{s}\) is the power required to the \(n^{\text {th}}\) BS to transmit the symbol \(b_s\) to the \(s^{\text {th}}\) user, \(P_{\text {max}}\) is the maximum power available for each base station, \(a_{i}\) is the \(i^{\text {th}}\) auxiliary variable required for expressing the inequality constraint (as presented in [11, 26] the number of auxiliary variables required depends from \(P_{\text {max}}\) value), and \(\alpha _i\) is the weight of \(a_{i}\);

  3. 3.

    Row constraint, i.e., all BSs have to serve the same users:

    $$\begin{aligned} f_{\text {row}}({\textbf {x}}) = \sum _{s} \left[ \left( \sum _{n} x_{sn}\right) - N_{\text {BS}}\right] ^2 \,; \end{aligned}$$
    (19)
  4. 4.

    Column constraint, i.e., the number of users served by each BS has to be equal to S:

    $$\begin{aligned} f_{\text {column}}({\textbf {x}}) = \sum _{n} \left[ \left( \sum _{s} x_{sn} \right) - S)\right] ^2 \,. \end{aligned}$$
    (20)

Therefore, the final objective function is:

$$\begin{aligned} \begin{aligned} f({\textbf {x}})&= \lambda _{1} \overset{{{\textbf {column constraint}}}}{\sum _{n} \biggl [ \biggl ( \sum _{s} x_{sn} \biggr ) - S)\biggr ]^2} - \lambda _{2} \overset{{{\textbf {maximize SNR}}}}{\sum _{s} \frac{p_{s}}{\sigma _{\theta }^{2}} \sum _{n} x_{sn}}+\\&+ \lambda _{3} \overset{{{\textbf {row constraint}}}}{ \sum _{s} \biggl [ \biggl (\sum _{n} x_{sn}\biggr ) - N_{\text {BS}}\biggr ]^2 } + \lambda _{4} \overset{{{\textbf {power}}}}{f_{{pow}}({\textbf {x}})} , \end{aligned} \end{aligned}$$
(21)

where \(\lambda _{1 \cdots 4}\) are the weights of the objective function terms to be properly sized, as mentioned in Paragraph 3. The principles behind the QUBO formulation are graphically resumed in Fig. 7.

As mentioned, \(f_{\text {pow}}\) allows the expression of the inequality contribution in QUBO, translating them as equality constraints by adding auxiliary variables whose number depends on the \(P_{\text {max}}\) values, as explained in [11, 26]. Python libraries for QUBO problem formulation, such as qubovert [27], can automatically do this transformation. However, introducing the auxiliary variables could dramatically increase the QUBO model and, consequently, the time required for its resolution, thus also reducing the advantage of employing the proposed solution. Therefore, this is the strongest limit of this QUBO formulation.

Unfortunately, the final solution could be similar to the one reported in Fig. 6b, where the elements of each row have different values. A cause could be the not-perfect sizing of penalty parameters. However, the proposed solver can activate a majority voting mechanism, performing a sort of error mitigation: when more than half of the elements on a row are equal to one, the corresponding user is selected for the second step of the algorithm.

Another problem to be managed is when the reduced subset size is smaller than S. In this case, the solver modifies the weights \(\lambda _2\) and \(\lambda _4\) and tries to solve the QUBO problem again. This procedure, substantially equivalent to the so-called sequential scaling of QUBO penalty functions [28], can be repeated a couple of times, and if the number of users is still insufficient, the Naïve approach (Sect. 4.1) is employed. It is also involved if the convex solver in the second step gives unfeasible negative \(p_{s}\).

Fig. 7
figure 7

Overview of QUBO model for users selection: graphical description of the knapsack-like problem associated with BS maximum power (left), QUBO variables data structure with row and column constraints (middle), an expected QUBO result, i.e., a subset of users, and an example of a problem cost function (right) (Color figure online)

4.2.2 Normalized QUBO formulation

The rapid increase of the QUBO problem dimension due to the insertion of auxiliary variables for representing the inequality constraints could reduce the efficiency of the proposed method, so evaluating a strategy for reducing the auxiliary variables number is crucial. First, it is necessary to remember that these are inserted to transform the inequality constraint into an equality one:

$$\begin{aligned} \sum _{s} |w_{sn}|^2 p_s x_{sn} + k = P_{\text {max}} \,, \end{aligned}$$
(22)

where the multi-level variable k can be described by exploiting binary expansions so that the number of inserted variables increases as \(P_{\text {max}}\) increases [26].

The first proposal for solving the presented problem is to perform a normalization of the power factors:

$$\begin{aligned} \sum _{s} \frac{|w_{sn}|^2 p_s x_{sn}}{P_{\text {max}}} + k = \frac{P_{\text {max}}}{P_{\text {max}}} \,. \end{aligned}$$
(23)

In order to maintain all coherence, the constraint parameters must also be scaled by the same quantity.

The QUBO normalization significantly reduces the number of required variables (about one order of magnitude, as shown in Fig. 8), thus positively impacting on solving time.

Fig. 8
figure 8

Number of binary variables required as a function of the number of users for the first, normalized and compacted QUBO formulations (Color figure online)

Compacted QUBO formulation In order to further reduce the QUBO complexity, an approximation for the power constraint can be performed, taking into account that QUBO simply performs the first selection and the \(p_{s}\) will be optimized in the next step, considering the power constraint again. The proposed approximation applies an equality constraint instead of an inequality one. In particular, it is imposed that:

$$\begin{aligned} \sum _{s} |w_{sn}|^2 p_s x_{sn} = c \cdot P_{\text {max}} \,, \end{aligned}$$
(24)

where c is a reduction factor, e.g., 0.8, so that a much higher quadratic penalty is imposed on a solution that is further from the imposed value (as shown in Fig. 9). Solutions with a lower penalty are around the forced value \(c \cdot P_{\text {max}}\). Considering the proportionality between the SNR and the transmitted power, solutions with a transmitted power close to the maximum should be characterized by a higher SNR. This approach avoids the introduction of auxiliary variables, thus obtaining a QUBO model with only \(N_{\text {UE}} \times N_{\text {BS}}\) variables (a reduction of two orders of magnitude with respect to the First formulation, as shown in Fig. 8).

Other QUBO formulations analyzed

Starting from the compact QUBO version, some constraints, which seem redundant considering the role played by the others, can be removed. In particular, three alternatives were considered:

  • Compacted QUBO without column constraint (QUBO NCC): the removal of the column constraints was considered since the power limitations should be sufficient to select a number of users close to the wanted one.

  • Compacted QUBO without power constraint (QUBO NPC): considering the observations in Sect. 4.2 and the presence of column constraints for limiting the number of users, the removal of the maximum power constraint can be tried.

  • Compacted QUBO without SNR optimization (QUBO NSNR): since the current QUBO works with random \(p_{s}\) values not coming from previous experience, the objective function SNR contribution was tried to be removed.

Fig. 9
figure 9

Quadratic penalty imposed by an equality constraint (Color figure online)

Fig. 10
figure 10

Overall block scheme of the quantum-compliant solver for Joint transmission allocation problem, considering a single sub-channel, 19 BS triplets, and users grouped for triplet: the environment considered by the TIM S.p.A. simulator, with a focus on a BSs collaborating triplet, for which a dedicated JT analysis can be evaluated (left), block scheme of the proposed quantum-compliant approach, where users are reduced by solving a QUBO problem and a classical computer evaluates all the reduced user set combinations, identifying the optimal power partition factor and the best users to serve (middle), graphical description of the JT MIMO system, with the found optimal users served by BSs (right) (Color figure online)

5 Results

This section reports the most significant results for the proposed approaches for solving the JT optimization problem.

5.1 Setup overview

The tests conducted in this study involved the implementation of proposed strategies using Python and the qubovert [27] library for Quadratic unconstrained binary optimization (QUBO) formulation. The best power scaling (\(p_{s}\)) values were determined using the Splitting Conic Solver (SCS) [29] of the CVXPy [30] library. QUBO problems were solved using different solvers, including the D-Wave Neal simulated annealer (SA) [31],D-Wave Leap Quantum Annealer(QA) [24], accessible through a remote connection, and digital annealer (DA) solver [19], which was obtained by modifying from the D-Wave Neal SA code for fair comparison.

The SA tests were performed on a single-process Intel(R) Xeon(R) Gold 6134 CPU @ 3.20 GHz opta-core, Model 85, with a memory of about 103 GB [32], while tests comparing different QUBO solvers were carried out on a single-process Intel(R) Core(TM) i7-1165G7 @ 2.80 GHz, with a memory of 8 GB.

In all cases, independently on the employed QUBO solver, the majority-voting-based error mitigation technique was employed for the identification of the reduced user set for each BS triplet.

The tests exploit synthetic datasets and data from the TIM S.p.A. MATLAB-based MIMO simulator. The first scenario considers random sub-channel matrices for varying numbers of users (\(N_{\text {UE}}\)), while the second sub-channel matrices from the TIM S.p.A. simulator, which simulated a complex environment (shown in the leftmost part of Fig. 10). with 57 base stations, 570 users, and 52 sub-channels. The system-level simulator for mobile networks evaluates network-level performance indicators, such as cell capacity, user throughput distribution, radio resource utilization, and interference levels, considering a multiplicity of users and BSs. The propagation channel is computed by following the indication provided by 3GPP in the technical report TR 38.901 [33], describing the modeling of the radio channel in the radio bands used by 5 G technologies. The selected scenario is Urban Macro (UMa), composed of a hexagonal grid of 19 macro BS sites, with three sectors (i.e., cells) per site and an inter-site distance (ISD) of 500 m. BSs are assumed at 25 m, and a given number of pedestrian users, moving at a speed of 3 km/h, are randomly dropped in each cell (10 per cell in the considered scenario). As per TR 38.901, different propagation laws are used to compute path-loss and shadow for Line of Sight (LOS) and Non-Line of Sight users, with the probability of being in LOS, computed according to the technical report. Following the clustered delay line (CDL) approach, fast fading effects are included, considering a CDL-B channel profile.

All parameters were computed by considering each BS triplet independently from the others, through the method for associating each user with a single BS triplet introduced in Sect. 4.

The H matrix of each sub-channel was computed as the element-by-element product of the path-loss matrix in linear units L and the complex dephasing matrix P associated with the phase shift of each signal during its transmission from a BS to a user. Noise power was always considered constant and equal to \(-109 \text {dBm}\), for staying in a worst-case scenario.

5.2 Figures of merit

The following figures of merit were considered to evaluate the proposed approaches’ effectiveness:

  • The number of users combination to evaluate obtained as:

    $$\begin{aligned} n_{c,\text {init}} = \left( {\begin{array}{c}N_{\text {UE}}\\ S\end{array}}\right) = \frac{N_{\text {UE}}!}{S! (N_{\text {UE}}-S)!} \end{aligned}$$
    (25)

    in the brute-force case and as:

    $$\begin{aligned} n_c = \left( {\begin{array}{c}N_{\text {UE}_{\text {reduced}}}\\ S\end{array}}\right) = \frac{N_{\text {UE}_{\text {reduced}}}!}{S! (N_{\text {UE}_{\text {reduced}}}-S)!} \end{aligned}$$
    (26)

    in the QUBO-assisted case;

  • The percentage of selected users from the original set:

    $$\begin{aligned} p_{\text {sel}} = \frac{N_{\text {UE}_{\text {reduced}}}}{N_{\text {UE}}}\cdot 100 \,; \end{aligned}$$
    (27)
  • The complexity of the obtained QUBO formulation in terms of the number of binary variables exploited (\(v_{q}\));

  • The relative percentage transmission rate error:

    $$\begin{aligned} \epsilon _{r} = \frac{Tr_{\text {brute-force}} - Tr_{\text {QUBO\_assisted}}}{Tr_{\text {brute-force}}} \cdot 100 \,; \end{aligned}$$
    (28)
  • The time required for solving the QUBO problem (\(t_q\));

  • The time required for solving the QUBO problem with the digital annealer normalized for the possible degree of parallelization (\(t_{q_n} = \frac{t_q}{v_{q}}\));

  • The time required for the overall optimization (\(t_t\)).

The first two figures of merit (\(n_c\) and \(p_{\text {sel}}\)) evaluate the effectiveness of different QUBO formulations in reducing the user set and selecting optimal subsets. Moreover, the comparison between \(n_{c,\text {init}}\) and \(n_c\) quantifies the reduction in the number of convex solver calls required when employing the QUBO-assisted approach. Furthermore, the proposed QUBO formulations were compared in terms of binary variables amount \(v_{q}\) and time required for their resolution (QUBO-only) \(t_q\).

Furthermore, the DA solver could be reasonably compared with the others by introducing a post-processing normalized time \(t_{q_n}\). In fact, since this approach gives a computational advantage only involving parallelization (on its hand achievable with the required amount \(v_{q}\) only with dedicated hardware and not in software), the performed post-processing operation permits to estimate, with a software functionally-equivalent model, the QUBO solving time expected with the desired degree of parallel computation.

Finally, the relative percentage TR error \(\epsilon _{r}\) and the overall execution time \(t_t\) compare the proposed approaches and the brute-force one in terms of both results quality and quickness.

5.3 Comparison of different strategies

Table 1 reports some results obtained with the synthetic dataset, changing the number of involved users \(N_{\text {UE}}\), with the original brute-force approach and the QUBO-assisted strategy, considering each proposed QUBO formulation solved with simulated annealing. This solver was chosen because it can be launched locally without requiring time-limited accesses to remote machines and can currently handle more binary variables than a quantum annealer, thus allowing the validation of the proposed QUBO model. For each \(N_{\text {UE}}\), five different problems solved ten times were considered for averaging the results. The QUBO-assisted approach significantly reduces execution time while maintaining an acceptable relative error (\(\epsilon _{r}\)). This advantage becomes more pronounced with an increasing number of users (as shown in Fig. 11). The variable reduction of QUBO formulation allows an appreciable reduction of the time required for solving it, which results in a reduction of the total execution time, even if less evident because, in the QUBO-assisted approach, \(t_q\) has a limited weight on \(t_t\).

The complexity advantage of the proposed QUBO-assisted strategy is highlighted in Fig. 12a, reporting the number of saved convex solver calls as a function of the number of users. Additionally, Fig. 12b compares the number of convex optimizations required by the brute-force and the proposed QUBO-assisted approaches, varying the number of users. It is evident that the complexity reduction increases with the problem dimension, consistently with the time advantage demonstrated in Fig. 11.

Table 2 shows the results obtained by solving two synthetically generated problems for each \(N_{\text {UE}}\) ten times with the original brute-force approach and with the proposed Naïve and QUBO-assisted strategies, considering the first three formulations of Sect. 4.2. It is possible to notice that the Naïve approach allows a more relevant reduction of the execution time than the QUBO-assisted one but at the expense of the quality of the results (Fig. 13). Moreover, it is possible to ascertain that the total execution time \(t_t\) is usually increasing with the number of users for all the QUBO solvers and that \(p_{\text {sel}}\) usually tends to decrease with higher \(N_{\text {UE}}\). The apparent exceptions in \(p_{\text {sel}}\) trends are expected to be overcome by increasing the number of repetitions for averaging the results and by improving the choice of QUBO penalty weights with a feedback mechanism. These trends also allow one to justify the \(t_t\) values obtained for Normalized QUBO and Compacted QUBO, which are, with \(N_{\text {UE}} = [35;40]\), lower than those obtained with \(N_{\text {UE}}\le 30\). In fact, the contribution of the QUBO solving time \(t_q\) in the calculation of \(t_t\) is, in these cases, more negligible than in the other QUBO-assisted strategies, so the total execution time mainly depends on the brute-force exploration time, which is lower for \(N_{\text {UE}} = [35;40]\) because of a lower \(n_c\).

Fig. 11
figure 11

Graphical view of the results of Table 1 (Color figure online)

Fig. 12
figure 12

Reduction of the computational complexity achieved with the proposed QUBO-assisted user selection (Color figure online)

Table 1 Results obtained with the synthetic dataset, changing the number of involved users \(N_{\text {UE}}\), with the original brute-force approach and the QUBO-assisted strategy, considering each obtained QUBO formulation solved with simulated annealing
Fig. 13
figure 13

Graphical view of the results of Table 2

Table 2 Results obtained by solving two randomly generated problems for each \(N_{\text {UE}}\) ten times with the original brute-force approach, the proposed Naïve and the QUBO-assisted strategies (QUBO solved with simulated annealing), considering the first three formulations of Sect. 4.2
Table 3 Results obtained by solving the problems with data retrieved from the TIM S.p.A. simulator, with all the analyzed approaches
Fig. 14
figure 14

TIM S.p.A. MATLAB-based MIMO simulator data. BSs are separated in 19 triplets and simulations can be performed on each of the 52 sub-channels

Table 4 QUBO-assisted results obtained with a classical (SA), a quantum (QA) and a quantum-inspired (DA) solvers
Fig. 15
figure 15

Graphical view of the results of Table 4

Table 3 presents the results obtained with data retrieved from the TIM S.p.A. simulator with all the analyzed approaches, with QUBO problems solved with simulated annealing. In particular, the results of the triplet of BSs 1-6-20 (\(N_{\text {UE}}= 30\)) of Fig. 10 for different sub-channels are reported. According to the performed tests, the same overall trends in the results can be observed for other triplets and sub-channels. Also in this case, the Naïve approach gives the lowest execution time but the worst results, since the obtained transmission rates are about 40% lower than those obtained with the brute-force exploration. On the other hand, apart from NCC, QUBO models allow a reduction of the selected users for brute-force exploration of more than 25%, thus also reducing \(t_t\). From the transmission rate perspective, \(\epsilon _r\) never exceeds 12.4%, which is about three times lower than the value obtained with the Naïve approach. In general, the QUBO NSNR seems to be the best approach, since it achieves the best trade-off between low execution time (in terms of \(p_{\text {sel}}\) or \(t_t\)) and transmission rate (\(\epsilon _r\), never exceeding 6%) (Fig. 14).

Finally, Table 4 proves how the chosen QUBO solver influences the solution time and the quality of the results (Fig. 15). Results are associated with tests performed by solving one synthetically generated problem for each \(N_{\text {UE}}\) ten times. The synthetic data were chosen because they permit to granularly change the number of users and the application of QA also for the First QUBO formulation. It is possible to recognize the flexibility of QUBO formulation, which can be solved by a classical optimizer (SA), a quantum one (QA) and a quantum-inspired one (DA). In evaluating QA results, it is necessary to consider that the solving time also includes the one required for minor embedding (which increases with the number of involved variables), for remote connection and any waiting one spent in the execution queue (the effective annealing time was set to 100\({\upmu \textrm{s}}\)). Furthermore, the First QUBO formulation was solved by QA only until the number of users was lower than eight because current devices cannot embed larger problems. The best solver (in terms of the quality of the results) depends on the QUBO formulation considered. This is coherent with the expectation because, as already discussed when Fig. 2 of Sect. 3.2 has been commented, modifying the QUBO formulation changes the energy profile of the problem and, consequently, the effectiveness of classical and quantum explorations [13, 15]. According to the performed tests, SA is most frequently characterized by the lowest \(\epsilon _r\) with the QUBO formulations with all constraints. This could be traced back to the fact that the presence of auxiliary variables or the current choice of \(\lambda \) values for the constraints penalty functions imply a problem’s profile characterized by wide barriers, not necessarily high. On the other hand, the reduction of variables obtained with normalization and the removal of some constraints imply an overall quadratic profile with narrower barriers, which are more suitable for QA and DA solvers. In particular, QA seems to be the best solver for QUBO NCC formulation, while DA is the best for NPC and NSNR formulations.

It is also interesting to observe that QA and DA achieve, for low \(N_{\text {UE}}\) values, results in terms of \(\epsilon _r\) very close to SA with QUBO formulations without constraints removals, usually with a lower \(p_{\text {sel}}\), thus having a smaller reduced user set and a consequent brute-force exploration with less iterations. For this reason, in a performance-complexity trade-off evaluation, the quantum and quantum-inspired methods can be considered competitive with respect to a classical QUBO solver.

6 Conclusions

This work proposed new strategies for optimizing the JT mechanism, in terms of TR, according to the potential users and the constraints on BS transmitting power. The results show that the QUBO-assisted mechanism achieves a better trade-off between execution time and the quality of the solution, compared to the brute-force and Naïve approaches. Indeed, the first gives the best solution but its execution time scales factorially, while the other gives the lowest solving time but gives results with poor quality. At the same time, QUBO-assisted solvers guarantee a significant reduction of the execution time without excessively degrading the performance. Coherently with the expectation, the Compacted, NPC and NSNR QUBO formulations, having the lowest number of binary variables, give the lowest execution time with performance comparable to First and Normalized formulations. Moreover, the performance obtained employing SA, QA, and DA solvers was compared, showing that the best solver strictly depends on QUBO formulation.

QUBO-assisted method can be undoutedly improved, exploiting alternative techniques for \(\lambda \) tuning. For example, scaled-sequential scaling and binary-search scaling [28] can be considered when the identified candidate users set is too low. Moreover, a feedback mechanism should be implemented (green arrow in Fig. 10) for adjusting the penalty weights of the different contributions in the QUBO formulation, exploiting the previous experience and performing the QUBO optimization with the \(p_s\) values computed in the previous time slot, instead of generating them randomly. This approach can be considered reasonable assuming that the users’ spatial motion should be limited during the time-slots of evaluation of the optimization problem, thus not dramatically changing their contribution in the channel matrix H. Finally, it is interesting to remind that the state of art presents some methods for a preliminary \(\lambda \) estimation [34,35,36,37], thus allowing reasonable initial weights for the constraints of the optimization problem.

The performed analysis with software prototypes lays the fundamentals, in the long term perspective, for a distributed computing system which can optimize in a quite small time window the JT mechanism. In particular, the idea is to have a hardware (Field-Programmable Gate Array or an Application-Specific Integrated Circuit [38]) solver in each BS or in each group of interacting BSs, which allows the optimization of the group of collaborating transmitting antennas. The design of a specialized architecture should guarantee the reactivity of the system, acquire information from sensors and work in a relatively small window of time. The flexibility of QUBO allows for the use of classical, quantum-emulated, or quantum-inspired solvers. Remote connections to quantum annealers are also anticipated as scaling improvements are expected.

Improvements are not strictly related to the definition of the QUBO model, but also concern benchmarking. In fact, attempts to compare the proposed QUBO-assisted method with alternative optimization formulations, such as linear programming [39, 40], and the solvers with other classical approaches, such as CPLEX and Gurobi, can be done.

The proposed final distributed optimization mechanism should significantly improve the communication quality thanks to JT, thus guaranteeing a more comfortable social interaction between users of wireless mobile networks.