Parametric devices based on four-wave mixing in fibers perform many signal-processing functions required by optical communication systems. In these devices, strong pumps drive weak signal and idler sidebands, which can have one or two polarization components, and one or many frequency components. The evolution of these components (modes) is governed by a system of coupled-mode equations. Schmidt decompositions of the associated transfer matrices determine the natural input and output mode vectors of such systems, and facilitate the optimization of device performance. In this paper, the basic properties of Schmidt decompositions are derived from first principles and are illustrated by two simple examples (one- and two-mode parametric amplification). In a forthcoming paper, several nontrivial examples relevant to current research (including four-mode parametric amplification) will be discussed.
© 2013 Optical Society of America
Parametric devices based on four-wave mixing (FWM) in fibers can amplify, frequency convert, phase conjugate, regenerate and sample optical signals in classical communication systems [1–6]. They can also generate photon pairs for quantum information experiments [7–10]. Three different types of FWM are illustrated in Fig. 1. Modulation interaction (MI) is the degenerate process in which two photons from the same pump are destroyed, and signal and idler (sideband) photons are created (2πp → πs + πi, where πj represents a photon with frequency ωj). Inverse MI is the degenerate process in which two photons from different pumps are destroyed and two signal photons are created (πp + πq → 2πs). Phase conjugation (PC) is the nondegenerate process in which two different pump photons are destroyed and two different sideband photons are created (πp + πq → πs + πi). The polarization properties of these processes are reviewed in [11–14].
Parametric interactions of weak sidebands, driven by strong pumps, are governed by coupled-mode equations (CMEs) of the formEq. (1) is linear in the amplitude vector and its conjugate, the (explicit or implicit) solution of Eq. (1) can be written in the input–output (IO) form
For the aforementioned one- and two-mode interactions (scalar MI and PC of continuous waves), it is easy to solve the CMEs and interpret the IO relations. However, for multiple-mode interactions, the CMEs and IO relations are complicated and two related questions arise: Under what conditions can one solve the CMEs explicitly and how should one interpret the IO relations physically?
Recall that every complex matrix M has the singular value (Schmidt) decomposition M = VDU†, where U and V are unitary and D is diagonal [15, 16]. The columns of U are the eigenvectors of M†M, the columns of V are the eigenvectors of MM†, and the entries of D are the (common) non-negative eigenvalues of M†M and MM†. In the context of parametric interactions, the laws of Hamiltonian mechanics impose constraints on the transfer matrices M and N, which ensure that they have the simultaneous (related) decompositions M = VDμU† and N = VDνUt, where Dμ = diag(μj), Dν = diag(νj) and their entries (Schmidt coefficients) satisfy the auxiliary equations [17, 18]. Hence, Eq. (2) can be rewritten in the formEq. (3) that the columns of U define input (Schmidt) mode vectors, the columns of V define output vectors, and the mode amplitudes x̄j(0) and x̄j(z), which are the components of X relative to the input and output Schmidt bases, respectively, satisfy the one-mode squeezing equations Equations (3) and (4) are remarkable. They imply that every parametric process, no matter how complicated, can be decomposed into independent squeezing processes, the properties of which are known (in-phase signal quadratures are stretched, whereas out-of-phase quadratures are squeezed) [7, 19, 20]. This decomposition is the answer to the second of the aforementioned questions. The first question is harder to answer, but in a forthcoming paper it will be shown that the (known) form of the decomposition sometimes provides guidance that facilitates the solution of the CMEs.
The simultaneous Schmidt decomposition is more than an elegant mathematical result. Determining the Schmidt modes and coefficients of a parametric process, and the device based upon it, facilitates the optimization of device performance. For example, in photon pair generation by pulsed pumps, the Schmidt modes are the (temporal) wave-packets of the signal and idler photons, and the squares of the Schmidt coefficients are the probabilities with which photon pairs are produced [21, 22]. If one designs the system so that only one coefficient is nonzero, the output state is pure, as required for a variety of quantum information experiments. In photon frequency conversion by pulsed pumps, the Schmidt modes are the natural input and output wave-packets of the signal and idler photons, and the squares of the Schmidt coefficients are the conversion probabilities [23, 24]. One can optimize two-photon interference (or single-photon conversion) experiments by designing the system so that at least one squared coefficient is 0.5 (or 1.0). It is well known that one-mode squeezing and stretching processes dilate the coherent and incoherent parts of input signals by the same amounts. The former processes are of interest in quantum optics, because the out-of-phase quadratures have smaller fluctuations than vacuum quadratures [19, 20], whereas the latter are of interest in optical communications, because they can amplify in-phase signals without degrading their signal-to-noise ratios [25, 26]. To minimize the effects of noise in a communication system, one should encode information in the Schmidt modes of the system (superpositions of frequency or polarization components), not the physical modes (individual components). Encoding information in this way also maximizes the information capacity of the system [27, 28].
This paper is organized as follows: In Sec. 2, one- and two-mode parametric amplification are analyzed in detail. The transfer matrices for these processes are determined explicitly by solving the CMEs analytically. These matrices are shown to have several interesting and useful properties, which are not accidental. In Sec. 3, the Schmidt decomposition theorem is proved constructively (from first principles) and the aforementioned properties are established. Previous derivations of these results were based on the laws of quantum optics [17, 18], in which context they are standard [21–24]. In contrast, the present derivation is based solely on the laws of classical mechanics, which are less fundamental, but sufficient. Readers who are interested in Schmidt decompositions and their applications can learn about them easily, unburdened by the complexities of quantum optics. In Sec. 4 the specific results derived in Sec. 3 are related to general properties of Hamiltonian systems. Finally, in Sec. 5 the main results of this paper are summarized.
Aspects of Schmidt decompositions were also discussed in [29, 30]. In these articles, the mathematical properties of Schmidt decompositions, and the adjoint decompositions to which they are related, were described in detail, and some simple examples (including one- and two-mode amplification) were mentioned briefly. In this paper, the physical consequences of Schmidt decompositions are emphasized, the aforementioned simple examples are discussed in detail and key results are explained in the context of Hamiltonian dynamics. In a forthcoming paper, new solutions of the CMEs are obtained and used to discuss several nontrivial examples of current interest (including four-mode parametric amplification).
2. Simple examples of Schmidt decompositions
In this section, two simple parametric processes are considered, for which the Schmidt decompositions are easy to determine. These examples provide useful checks and illustrations of the general results derived in the next section.
2.1. One-mode amplification
Consider a one-mode parametric process (inverse MI), which is governed by the equation31, 32]. Equation (5) depends linearly on x and x*, so its solution can be written in the input–output (IO) form (7) satisfy the auxiliary equation
Because the transfer functions usually have different phases, some analysis is required to determine the consequences of Eq. (6). Let μ = |μ|eiϕμ and ν = |ν|eiϕν, and define the sum and difference phases ϕs = (ϕμ + ϕν)/2 and ϕd = (ϕν − ϕμ)/2, respectively. Then Eq. (6) can be rewritten in the formEq. (10) add constructively: The signal is said to be in-phase and is amplified (stretched) by the factor |μ| + |ν|. Conversely, if ϕx = ϕd + π/2, the terms on the right side of Eq. (10) add destructively: The signal is said to be out-of-phase and is attenuated (squeezed) by the factor |μ|+ |ν| = 1/(|μ|−|ν|). If one were to measure the phase of the input signal relative to the aforementioned reference phase, one would say that the real quadrature is amplified and the imaginary quadrature is attenuated. Notice that Eq. (10) has the canonical form of Eq. (3).
It is instructive to formalize the method of solution. Equation (5) and its conjugate can be written in the augmented matrix formEq. (11) can be written in the IO form Eqs. (12) and (14). First, tr(Ly) = 0, so det(Ty) = 1, and second, . Because Eqs. (11) and (13) describe two copies of the same process (the original and its conjugate), the transfer matrix can be written in the form Eqs. (9).
It was stated in Sec. 1 that the Schmidt vectors of Ty are the eigenvectors of and , and the Schmidt coefficients are the square roots of the (common) eigenvalues of these matrices. One can determine these eigensystems explicitly, or simply verify thatEq. (10). All three matrices in Eq. (17) depend on z. The evolution equation (11) governs x and x* simultaneously, so it is natural that the associated transfer matrix describes stretching and squeezing simultaneously. Specifically, Eq. (17) shows that the stretching condition is 2ϕx = 2ϕd, whereas the squeezing condition is 2ϕx = 2ϕd + π, and the associated Schmidt coefficients are reciprocals. The input Schmidt vectors, which are the natural inputs for one-mode amplification, correspond to in-phase and out-of-phase signals. These results are consistent with the discussion that follows Eq. (10).
Equations (15) and (16) show that there is a simple relation between the transfer matrix and its inverse. The replacements μ → μ* and ν → −ν (which do not affect the moduli of the transfer functions) are equivalent to ϕμ → −ϕμ and ϕν → ϕν + π and, ultimately, to eiϕs → ieiϕd and eiϕd → ieiϕs. By making these replacements in decomposition (17), one obtains the inverse decomposition(17), nor the backward decomposition (18), is unique.
2.2. Two-mode amplification
Now consider a two-mode parametric process (MI or PC), which is governed by the CMEs1, 2]. We will refer to mode 1 as the signal and mode 2 as the idler. The solutions of Eqs. (19) can be written in the IO forms Eqs. (7), with the mismatch coefficient δ replaced by (δ1 + δ2)/2, and the phase factor e(z) = exp[i(δ1 − δ2)z/2] = exp(iϕδ). The two-mode transfer functions satisfy the auxiliary equations (20) reduce to solution (6).
In the first of Eqs. (20), the transfer functions have the common factor e(z), which affects the output signal phase, but does not affect the interference conditions. Hence, if ϕ1 + ϕ2 = ϕν − ϕμ = 2ϕd, the terms in the first of Eqs. (20) add constructively: The sidebands are said to be in-phase and (if their input amplitudes are equal) are stretched by the factor |μ| + |ν|. Conversely, if ϕ1 + ϕ2 = ϕν − ϕμ + π, the terms in the first of Eqs. (20) add destructively: The sidebands are said to be out-of-phase and (if their input amplitudes are equal) are squeezed by the factor |μ| + |ν| = 1/(|μ| − |ν|). The same interference conditions apply to the second of Eqs. (20).
Equations (19) can be written in the standard matrix formEq. (23) is closed (involves only x1 and ), so no augmentation is necessary. (The equation for is simply the conjugate of the stated equation and contains no extra information.) The solution of Eq. (23) can be written in the IO form Eqs. (24) and (26)] differ only slightly from those of Ly and Ty [Eqs. (12) and (14)]. Because tr(Lx) = δ1 − δ2 ≠ 0, det(Tx) = exp[i(δ1 − δ2)z] ≠ 1. Nonetheless, Tx(−z) = T−1(z).
It follows from Eqs. (20) that the transfer matrixEqs. (22).Eq. (9), and ϕδ was defined after Eq. (20). Suppose that the input sideband phases are measured relative to the reference phase ϕd. Then decomposition (29) implies that the combination is stretched, whereas the combination is squeezed. For stretching, the optimal phase condition is ϕ1 + ϕ2 = 0, whereas for squeezing, the optimal phase condition is ϕ1 + ϕ2 = π. These results are consistent with the results stated after Eq. (22). It follows from Eq. (18) that the backward transfer matrix has the Schmidt decomposition Eq. (30) from Eq. (29) by making the replacements eiϕs → ieiϕd, eiϕd → ieiϕs and eiϕδ → e−iϕδ.Eqs. (33) and (34) that allows M to be diagonal and N to be off-diagonal. Decompositions (33) and (34) are not quite in canonical form, because the elements of the diagonal matrices (μ and ν) are complex. However, by generalizing the derivation of Eq. (10), one obtains the Schmidt decompositions (29), (30), (35) and (36) are not unique.
At this point, it is instructive to introduce the superposition modesEqs. (19), one obtains the superposition-mode equations Eq. (4) or (6). Because the coupling coefficients in Eqs. (38) differ by a factor of −1, the input phases required for stretching differ by π/2, as do the input phases required for squeezing. This method of analysis fails for the general case in which δd ≠ 0. However, the concept of a superposition mode remains useful.
Decompositions (35) and (36), which are valid for arbitrary values of δd, show that it is appropriate to measure the input sideband phases relative to the (common) reference phase ϕd, as did decomposition (29). With this convention, the real quadrature of the sum mode and the imaginary quadrature of the difference mode are stretched. It follows from Eq. (4), in which the transfer functions are non-negative by construction, that the imaginary sum quadrature and the real difference quadrature are squeezed.
In summary, the Schmidt decompositions (17), (29), (35) and (36) reveal automatically the input-phase conditions required for stretching and squeezing . The decompositions of the augmented and standard transfer matrices feature stretched and squeezed modes, whereas the decompositions of the canonical matrices feature only stretched modes (from which the squeezed modes can be deduced). These decompositions also lead to the concept of superposition modes: The Schmidt modes of a system are the natural superposition modes of that system.
3. Basic theory of Schmidt decompositions
In the preceding section, the Schmidt decompositions of two simple parametric processes were determined explicitly, and were shown to have interesting and useful properties. In this section, the basic properties of such decompositions are established for arbitrary parametric processes. The evolution of a conservative system of m coupled modes is governed by the HamiltonianEq. (39) to be real, J must be Hermitian. The sum of the second and third terms is real by construction, and one can always write these terms in such a way that K is symmetric. For reference, J is specified by (up to) m2 real parameters, whereas K is specified by m(m + 1) real parameters. By applying the (complex) Hamilton equation (39), one obtains the CME Appendix.) Equation (41) depends linearly on X and X*, so its solution can be written in the IO form Equation (41) is just Eq. (1), with A = iJ and B = iK, and Eq. (42) is just Eq. (2), repeated for convenience.] For the special case in which m = 1, J = δ and K = γ, and Eq. (41) reduces to Eq. (5).
Alternatively, one can write Eq. (41) and its conjugate as the augmented matrix equationEq. (43) can be written in the IO form Eq. (43) describes two copies of the same process (the original and its conjugate), the m × m blocks of Ty are the transfer matrices M and N, which appeared in Eq. (42), and their conjugates [see Eq. (15)]. Clearly, Eqs. (43)–(46) are generalizations of Eqs. (11)–(14).
In general, each component of X is coupled to every component of X and X*. However, there are many important systems in which a subset of the components of X (denoted by X1 and called the signal vector) is coupled to itself and a different subset of X* (denoted by and called the idler vector). Such systems, which include the MI- and PC-based systems described in the introduction, are governed by the Hamiltonian(47), one obtains the CMEs Eqs. (49) can be written in the IO forms Eqs. (49) reduce to Eqs. (19).
Alternatively, Eqs. (49) can be rewritten as the standard matrix equationEq. (51) can be written in the IO form Eq. (27)]. Clearly, Eqs. (51)–(54) are generalizations of Eqs. (23)–(26).
Although the special system described by Eq. (51) is of lower order that the general system described by Eq. (43), its mathematical structure is more complicated, because the diagonal blocks of Lx are not necessarily equal and the off-diagonal blocks are not necessarily symmetric, as are the corresponding blocks of Ly [Eqs. (44) and (52)]. Hence, we will derive the properties of the special system and deduce the corresponding properties of the general system.Equation (56) has several important (mathematical and physical) consequences.
In two-mode amplification, the Manley-Rowe-Weiss (MRW) variable c = |x1|2 − |x2|2 is conserved [33, 34]. The classical (quantal) interpretation of this result is that the difference between the action (photon) fluxes of the signal and idler is conserved (sideband photons are produced in pairs). In the present context, the MRW variable . By combining this definition with Eq. (51), one finds thatEq. (56) is both a necessary and sufficient condition for the conservation of action (photon) flux.
Starting from Eq. (56), one can prove by induction thatEqs. (54) and (58), one finds that Equation (54) implies that Eq. (59) is equivalent to the equations and . The first equation provides a useful formula for the inverse transfer matrix (if the forward matrix is known, so also is the backward matrix), whereas the second provides a link to the theory of Hamiltonian systems (which will be described in Sec. 4).
The Schmidt decomposition theorem  states that the transfer matrix Tx = VDU†, where U and V are unitary, and D is positive and diagonal. (Positivity is required because Tx is invertible.) The columns of U (input Schmidt vectors) are the eigenvectors of , the columns of V (output Schmidt vectors) are the eigenvectors of and the entries of D (Schmidt coefficients) are the square roots of the (common) eigenvalues of and . The stated decomposition is not unique: One can multiply the columns of U and V by the same set of phase factors, and one can permute (reorder) the columns of U and V in the same way, without invalidating the decomposition. Notice that the input vectors of are the eigenvectors of and the output vectors of are the eigenvectors of . Hence, the input (output) vectors of are the output (input) vectors of Tx and the Schmidt coefficients of equal those of Tx.
Before determining the Schmidt decomposition in its entirety, we pause to prove an important result about the Schmidt coefficients. For any matrix A,Equation (62) implies that and have the same (positive) eigenvalues, and the identity implies that Eqs. (17) and (29).
We now return to the Schmidt decomposition. Equations (59) and (60) impose several constraints on the block matrices M11, N12, N21 and M22, which allow the decomposition to be determined. The former equation implies thatEq. (65), one finds that Eqs. (66) reduce to Eqs. (9) and (22) in the appropriate limits. The latter equation implies that Eqs. (67) and (68), one finds that Equation (69) can be rewritten as Because the unitary decomposition of a Hermitian matrix is unique (apart from phase shifts and ordering), and . By continuing in this way, one finds that and , V12 = V11 = V1 and , and and . Hence, D11 = D22 = Dμ and D12 = D21 = Dν, where . The equations associated with the off-diagonal terms in Eqs. (67) and (68) are satisfied identically. By assembling the preceding results, one finds that [35, 36] Eq. (73) does not define the overall phases of the blocks uniquely. If one were to replace Uj and Vj by Ujeiϕj and Vjeiϕj, respectively, the diagonal blocks of the transfer matrix would be unaltered, whereas the off-diagonal blocks would be multiplied by ei(ϕ1+ϕ2) and e−i(ϕ1+ϕ2). One can exploit this non-uniqueness to write some decompositions in particularly simple ways.
For each of the n sets of Schmidt modes in decomposition (73), there are two Schmidt coefficients (|μ| and |ν|) and four phase combinations (ϕv1 − ϕu1, ϕv1 + ϕu2, − ϕv2 − ϕu1 and −ϕv2 + ϕu2). However, only one of the coefficients is independent (|ν|) and only three of the combinations are independent. If one defines the reference phase (ϕv1 − ϕv2 − ϕu1 + ϕu2)/2, then the first and fourth combinations have the relative phases ±(ϕv1 + ϕv2 − ϕu1 − ϕu2)/2, whereas the second and third combinations have the relative phases ±(ϕv1 + ϕv2 + ϕu1 + ϕu2)/2. Thus, if the Schmidt modes are known, only 4n real parameters are required to specify the transfer matrix. [The other 4n(n − 1) parameters in the coefficient matrix specify the Schmidt modes.] For the special case in which J1 = J2 and K = Kt, the signal and idler equations are identical, so U1 = U2, V1 = V2 and the reference phase is 0. Thus, if the Schmidt modes are known, only 3n real parameters are required to specify the transfer matrix. [The other 2n(n − 1) parameters in the coefficient matrix specify the Schmidt modes.] These results are consistent with Eqs. (15) and (27), which apply to cases in which n = 1.
According to the Schmidt decomposition theorem, the Schmidt coefficients and input Schmidt modes are determined by the eigenvalues and eigenvectors of the Hermitian matrixEq. (64). Hence, if E± are the eigenvectors of or , then SE± = E∓ are the eigenvectors of or . It follows from Eqs. (75) and (77) that the transfer matrix (73) has the Schmidt decomposition (78) implies that the (input) combinations are stretched, whereas the combinations are squeezed. Notice that Eq. (78) reduces to Eqs. (17) and (29) in the appropriate limits.
If the forward transfer matrix can be written in the form (73), the backward transfer matrix can be written in the formEq. (79) from Eq. (73) by interchanging the input and output modes, and changing the sign of Dν. However, Dν is non-negative by construction, so this empirical rule is not canonical.
It is instructive to consider the decomposition of the backward matrix, which can be calculated in (at least) three ways. First, Tx = VDU†, so the laws of matrix algebra require that . Hence,(80), which is consistent with Eq. (79), the input and output modes are interchanged, as are the associated stretching and squeezing factors.
Second, the identity requires that(81), which is also consistent with Eq. (79), the input and output modes are interchanged, as are the associated stretched and squeezed modes, but the stretching and squeezing factors are unaltered.(80) works by permuting the Schmidt coefficients so that stretched modes are squeezed and vice versa. It is also easy to verify that (81) works by permuting the Schmidt vectors so that squeezed modes are stretched and vice versa. These actions are equivalent ways to obtain the same result: In the (common) inversion formula , the P matrices can act to the middle, or to the outsides.
Third, the identity implies that there are simple relations between the forward and backward Schmidt coefficients and vectors. Equations (73) and (79) do not determine these relations uniquely. However, by applying the transformationsEq. (73), one does obtain Eq. (79). Transformations (85) allow Dμ and Dν to remain positive. By applying them twice, one finds that Uj(z) → iVj(−z) → i2Uj(z) and Vj(z) → iUj(−z) → i2Vj(z). These results are acceptable in the context of a Schmidt decomposition, because the signs (phases) of the Schmidt modes are not unique. By applying transformations (85) to the forward decomposition (78), one obtains a backward decomposition that is similar to decomposition (81): The first unitary matrix is multiplied by i and the second is multiplied by −i, so the decompositions are equivalent. Notice that transformations (85) are consistent with Eqs. (17) and (18), and Eqs. (29) and (30). In the former case u1 = u2 = eiϕd and v1 = v2 = eiϕs, whereas in the latter case u1 = u2 = eiϕd, v1 = eiϕs+iϕδ and v2 = eiϕs−iϕδ.Eq. (4), in which the transfer functions are non-negative by construction, that the imaginary sum quadratures and the real difference quadratures are squeezed. These results are valid for the general case in which J1 ≠ J2 and K ≠ Kt. Notice that Eqs. (87) and (88) reduce to Eqs. (35) and (36) in the appropriate limit.
4. Unifying principles
The specific results derived in the preceding section are closely connected to the general properties of Hamiltonian dynamical systems [37,38]. To make these connections, one must rewrite the CMEs (49) in the simplest form possible. By using the fact that J2 is Hermitian, one can rewrite Hamiltonian (47) in the alternative form(89), one can reproduce the aforementioned CMEs. Equations (89) and (90) are equivalent to the Hamiltonian Eq. (55). Notice that G is Hermitian. (Consequently, if L = SG, then SL = L†S.) Equations (91) and (92) are said to be in canonical form.
Suppose that X′ = TX, where T is an arbitrary transformation (change-of-variables) matrix. Then, in component form,Equation (94) shows that T is symplectic if and only if (95) can be rewritten in the matrix form (96) is satisfied, then T−1 = ST†S and (T†)−1 = STS. For reference, the set of (nonsingular) matrices that satisfy condition (96) form a group with respect to multiplication.
Now suppose that T(z) is the transfer matrix for the system, which satisfies the evolution equationEquation (99) implies that Equation (99) also implies that T−1 = ST†S and (T†)−1 = STS.
The transfer matrix must also satisfy the symplectic condition (96), because the Hamilton equation (92) retains its form as X evolves. One can prove this statement directly. Alternatively, by multiplying the identity T†ST = S by S(T†)−1 on the left and T−1S on the right, one can show that S = (TS)S(ST†) = TST†. Hence, the transfer matrix satisfies the symplectic condition, which is equivalent to the MRW condition (99). Notice that the proofs of the preceding results were based on the assumption that T is a linear transformation, but not on the assumption that G is a constant: The results remain valid when G is a function of z.
The key results of the preceding section are consequences of the identity T−1 = ST†S. For example, if (T†T)E = λE, where λ ≠ 0, then(73) owes its form to the constraints imposed on the blocks of the transfer matrix [Eqs. (67) and (68)], which are just (ST†S)T = I and T(ST†S) = I. Decompositions (78), (79), (81), (87) and (88) all follow directly from decomposition (73). Furthermore, if T = VDU†, then Equations (101) and (102) show that the stretched modes for the forward transformation are the squeezed modes for the backward transformation. These relationships guarantee that the combined transformation is the identity transformation.
The preceding results show that the Schmidt decomposition of the transfer matrix owes its (extremely useful) form to the symplectic properties of the associated evolution equation, which is Hamiltonian. There are many physical processes for which knowledge of the underlying mathematical (algebraic) structure facilitates the derivation of important physical results. Consequently, it is worthwhile to review some definitions and make some specific connections.
Nonsingular matrices form a variety of groups with respect to multiplication. GL(m,C) is the general linear group, whose members are m × m complex matrices. SL(m,C) is the special linear group of degree m, whose members are unimodular (have determinant 1).
U(m) is the unitary group, whose members are m × m unitary matrices (which are complex by definition). The actions of these matrices preserve the quadratic form X†X, where X is an arbitrary m × 1 vector. SU(m) is the special unitary group of degree m, whose members have determinant 1. It is sometimes called the unimodular unitary group. These groups occur in models of conservative phenomena, in which X†X is the total power (or energy). For example, U(2) and SU(2) underly polarization rotation, beam splitting, directional coupling and (stable) frequency conversion.
U(n,n) is the pseudo-unitary group, whose members are 2n × 2n complex matrices. The actions of these matrices preserve the quadratic form X†SX, where X is an arbitrary 2n × 1 vector. (This group has a subgroup of diagonal matrices that are unitary, but most of its members are nonunitary.) In the context of parametric amplification, X†SX is the MRW variable and Tx is a member of U(n,n). SU(n,n) is the special pseudo-unitary group of degree 2n, whose members have determinant 1. (This group also has a unitary subgroup.) In the aforementioned context, Ty is a member of SU(n,n). Ty is also a member of the symplectic group Sp(2n), whose members satisfy condition (96) and have determinant 1. (In the Appendix, it is shown that this definition of the symplectic group is equivalent to the standard definition, which involves a different auxiliary matrix.) The mathematical properties of continuous groups are described in [39, 40] and several simple examples (including those mentioned above) are described in [41–44].
Parametric devices based on four-wave mixing in fibers perform many signal-processing functions required by optical communication systems. In these devices, strong pumps drive weak signal and idler sidebands, which can have one or two polarization components, and one or many frequency components. The evolutions of these components (modes) are governed by systems of coupled-mode equations [Eq. (1)], the solutions of which are specified by transfer matrices [Eq. (2)]. Schmidt decompositions of these transfer matrices [Eq. (3)] determine the natural input and output mode vectors of such systems, and the effects of the systems on the associated mode amplitudes [Eqs. (4)].
In Sec. 2, two simple examples were considered: one- and two-mode parametric amplification. The transfer matrices for these processes were determined explicitly, as were their Schmidt decompositions. In the first process, different quadratures of the same mode are stretched and squeezed (dilated), whereas in the second process, different combinations (superpositions) of the signal and idler modes are dilated. These superpositions are the Schmidt modes. For every mode (quadrature) that is stretched, there is another mode (quadrature) that is squeezed by the same amount. Furthermore, simple relations exist between the transfer matrices for the forward and backward processes [Eqs. (27) and (28)], and their Schmidt decompositions [Eqs. (29) and (30)]. These properties turn out to be generic.
In Sec. 3, the properties of Schmidt decompositions were derived from first principles, for parametric processes that involve 2n modes (n is arbitrary). The coefficient matrix that appears in the coupled-mode equation (51) has a symmetry property [Eq. (56)] that constrains the associated transfer matrix [Eq. (59)]. This constraint links the decompositions of the signal and idler blocks, and allows the decomposition of the forward transfer matrix to be determined [Eqs. (73) and (78)]. The transfer matrix involves 4n mode vectors, for the input and output signal and idler, which are equivalent to 2n input and output Schmidt vectors (combinations of the signal and idler vectors). In addition to these vectors, the transfer matrix involves (up to) 4n real parameters (n Schmidt coefficients and 3n phase factors). This number of parameters is much smaller than the number required to specify the aforementioned coefficient matrix, which is of order n2. If the forward matrix and decomposition are known, so also are the backward matrix and decomposition [Eqs. (79) and (80)]: One obtains the latter entities from the former by interchanging the input and output vectors, and interchanging the stretching and squeezing factors.
In Sec. 4, the specific properties established for parametric processes were shown to be general properties of Hamiltonian dynamical systems. The symplectic identity [Eq. (96)] was shown to be equivalent to the Manley-Rowe-Weiss identity [Eq. (99)]. Either identity (together with the laws of matrix algebra) is sufficient to establish the main properties of Schmidt decompositions: For every Schmidt mode that is stretched in a forward (or backward) transformation, there is another, closely related, mode that is squeezed by the same amount. Furthermore, the forward and backward transformations have the same Schmidt coefficients, and the input squeezed (stretched) modes for the backward transformation are the output stretched (squeezed) modes for the forward transformation. These relationships guarantee that the combined transformation is the identity transformation.
Not only do Schmidt decompositions provide physical insight into complicated parametric processes, they also provide the mathematical means to optimize the performance of devices based on these processes. In a forthcoming paper, several nontrivial examples relevant to current research (including four-mode parametric amplification) will be discussed in detail.
Appendix: Real and complex Hamiltonian systems
Consider a real dynamical system with Hamiltonian H(q, p), where q and p are conjugate variables. The Hamilton equations (for z-evolution) areEqs. (103) in the matrix form
Now consider the quadratic Hamiltonian(103) to Hamiltonian (106), one obtains the linear evolution equations (106) in the compact form (104) to Hamiltonian (108), one obtains the matrix evolution equation (107).
Suppose that X′ = TX, where T is an arbitrary transformation (change-of-variables) matrix. Then, in component form,(113) is satisfied, then T−1 = −JTtJ and (Tt)−1 = −JTJ.
Now suppose that T(z) is the transfer matrix for the system, which satisfies the evolution equation(113) and (116) are equivalent, and require that det(T) = ±1. However, G is symmetric, so tr(JG) = 0 and det(T) = 1. Hence, T is a member of Sp(2,R), the three-parameter group whose members are symplectic 2 × 2 matrices with determinant 1. [This group is isomororphic to SL(2,R) and SU(1,1).]
To establish the connection between real and complex dynamical systems, one defines the complex variablesEqs. (103), (117) and (118), one obtains the complex Hamilton equations Eqs. (119) in the matrix form Eq. (55) and the derivative of H is taken componentwise. Equation (120) is equivalent to Eq. (104).
Now consider the quadratic Hamiltonian(119) to Hamiltonian (121), one obtains the linear evolution equations Eqs. (122) is just Eq. (5), which describes one-mode squeezing. One can reconcile the real and complex descriptions of this process by setting (121) in the compact form (120) to Hamiltonian (124) componentwise, and reassembling the results, one obtains the matrix evolution equation (124) without the factor of 2, and differentiate it with respect to X†, treated as a single (vector) variable. This approach was taken in the main text.
Because Eqs. (110) and (126) are equivalent descriptions of the same system, their consequences, prime among which are the symplectic identities, must also be equivalent. To prove this result explicitly, define Xr = [p, q]t and Xc = [a, a*]t. Then Xc = UXr, where the unitary matrixEq. (113) by U on the left and U† on the right, one finds that Eq. (128) is just Eq. (96). Thus, the real and complex symplectic identities are equivalent, so the groups formed by Tr and Tc [Sp(2,R) and SU(1,1)] are isomorphic. This equivalence extends to systems of 2n variables.
References and links
1. M. E. Marhic, Fiber Optical Parametric Amplifiers, Oscillators and Related Devices (Cambridge, 2007). [CrossRef]
2. J. Hansryd, P. A. Andrekson, M. Westlund, J. Li, and P. O. Hedekvist, “Fiber-based optical parametric amplifiers and their applications,” IEEE J. Sel. Top. Quantum Electron. 8, 506–520 (2002). [CrossRef]
3. J. H. Lee, “All-optical signal processing devices based on holey fiber,” IEICE Trans. Electron. E88-C, 327–334 (2005). [CrossRef]
4. S. Radic and C. J. McKinstrie, “Optical amplification and signal processing in highly nonlinear optical fiber,” IEICE Trans. Electron. E88-C, 859–869 (2005). [CrossRef]
5. P. A. Andrekson and M. Westlund, “Nonlinear optical fiber based high resolution all-optical waveform sampling,” Laser Photon. Rev. 1, 231–248 (2007). [CrossRef]
6. S. Radic, “Parametric signal processing,” IEEE J. Sel. Top. Quantum Electron. 18, 670–680 (2012). [CrossRef]
7. R. Loudon, The Quantum Theory of Light, 3rd Ed. (Oxford, 2000).
8. M. Fiorentino, P. L. Voss, J. E. Sharping, and P. Kumar, “All-fiber photon-pair source for quantum communications,” IEEE Photon. Technol. Lett. 14, 983–985 (2002). [CrossRef]
9. M. Halder, J. Fulconis, B. Cemlyn, A. Clark, C. Xiong, W. J. Wadsworth, and J. G. Rarity, “Nonclassical 2-photon interference with separate intrinsically narrowband fibre sources,” Opt. Express 17, 4670–4676 (2009). [CrossRef] [PubMed]
11. K. Inoue, “Polarization effect on four-wave mixing efficiency in a single-mode fiber,” IEEE J. Quantum Electron. 28, 883–894 (1992). [CrossRef]
13. M. E. Marhic, K. K. Y. Wong, and L. G. Kazovsky, “Fiber optic parametric amplifiers with lineary or circularly polarized waves,” J. Opt. Soc. Am. B 20, 2425–2433 (2003). [CrossRef]
15. G. W. Stewart, “On the early history of the singular value decomposition,” SIAM Rev. 35, 551–566 (1993). [CrossRef]
16. G. J. Gbur, Mathematical Methods for Optical Physics and Engineering (Cambridge, 2011), Sec. 5.4.
18. S. L. Braunstein, “Squeezing as an irreducible resource,” Phys. Rev. A 71, 055801 (2005). [CrossRef]
19. H. P. Yuen, “Two-photon states of the radiation field,” Phys. Rev. A 13, 2226–2243 (1976). [CrossRef]
20. C. M. Caves, “Quantum limits on noise in linear amplifiers,” Phys. Rev. D 26, 1817–1839 (1982). [CrossRef]
22. W. P. Grice, A. B. U’Ren, and I. A. Walmsley, “Eliminating frequency and space-time correlations in multiphoton states,” Phys. Rev. A 64, 063815 (2001). [CrossRef]
23. M. G. Raymer, S. J. van Enk, C. J. McKinstrie, and H. J. McGuinness, “Interference of two photons of different color,” Opt. Commun. 283, 747–752 (2010). [CrossRef]
24. C. J. McKinstrie, L. Mejling, M. G. Raymer, and K. Rottwitt, “Quantum-state-preserving optical pulse reshaping and multiplexing by four-wave mixing in fibers,” Phys. Rev. A 85, 053829 (2012). [CrossRef]
26. Z. Tong, C. Lundström, P. A. Andrekson, M. Karlsson, and A. Bogris, “Ultralow noise, broadband phase-sensitive optical amplifiers and their applications,” IEEE J. Sel. Top. Quantum Electron. 18, 1016–1032 (2012). [CrossRef]
27. E. Telatar, “Capacity of multi-antenna Gaussian channels,” Euro. Trans. Telecom. 10, 585–595 (1999). [CrossRef]
28. C. J. McKinstrie and N. Alic, “Information efficiencies of parametric devices,” J. Sel. Top. Quantum Electron. 18, 794–811 (2012). [CrossRef]
29. C. J. McKinstrie, “Unitary and singular value decompositions of parametric processes in fibers,” Opt. Commun. 282, 583–593 (2009). [CrossRef]
30. D. A. Edwards, J. D. Fehribach, R. O. Moore, and C. J. McKinstrie, “An application of matrix theory to the evolution of coupled modes,” to appear in SIAM Rev.
32. K. Croussore and G. Li, “Phase and amplitude regeneration of differential phase-shift keyed signals using phase-sensitive amplification,” IEEE J. Sel. Top. Quantum Electron. 14, 648–658 (2008). [CrossRef]
33. J. M. Manley and H. E. Rowe, “Some general properties of nonlinear elements—Part I. General energy relations,” Proc. IRE 44, 904–913 (1956). [CrossRef]
34. M. T. Weiss, “Quantum derivation of energy relations analogous to those for nonlinear reactances,” Proc. IRE 45, 1012–1013 (1957).
35. A. I. Lvovsky, W. Wasilewski, and K. Banaszek, “Decomposing a pulsed optical parametric amplifier into independent squeezers,” J. Mod. Opt. 54, 721–733 (2007). [CrossRef]
36. C. J. McKinstrie, M. G. Raymer, and H. J. McGuinness, “Spatial-temporal evolution of asymmetrically-pumped phase conjugation I: General formalism,” Alcatel-Lucent ITD-09-48636Q, available upon request.
37. H. Goldstein, Classical Mechanics, 2nd Ed. (Addison-Wesley, 1980).
38. V. I. Arnold, Mathematical Methods of Classical Mechanics, 2nd Ed. (Springer, 2000).
39. D. H. Sattinger and O. L. Weaver, Lie groups and Algebras with Applications to Physics, Geometry and Mechanics (Springer, 1986). [CrossRef]
40. M. Hamermesh, Group Theory and its Application to Physical Problems (Dover, 1989).
41. H. Takenaka, “A unified formalism for polarization optics by using group theory,” Nou. Rev. Opt. 4, 37–41 (1973). [CrossRef]
42. Y. S. Kim and M. E. Noz, “Illustrative examples of the symplectic group,” Am. J. Phys. 51, 368–375 (1983). [CrossRef]
43. A. Mufti, H. A. Schmitt, and M. Sargent, “Finite-dimensional matrix representations as calculational tools in quantum optics,” Am. J. Phys. 61, 729–733 (1993). [CrossRef]