The Structures of Ring and Field

Maurice R. Kibler , in Galois Fields and Galois Rings Made Easy, 2017

1.2.7 Characteristic of a field

The definition of the characteristic of a unitary ring applies to a field since a field is a particular unitary ring. In terms of field, we have the following formulation.

1.2.7.1 Characteristic

Definition 1.20

The characteristic of a field ( K , +, ×) is the smallest positive integer p (p    2) such that

x K : p × x = 0 1 + 1 + + 1 = 0

where the sum contains p terms. If 1 + 1 + + 1 0 whatever the number of 1 in the sum is, then the field is said to be of characteristic 0.

1.2.7.2 Example: Z p , with p prime

The field Z p (p prime) is a field of characteristic p. We will see that there are other fields of characteristic p.

1.2.7.3 Example: Q , R , C and H

The field of rational numbers Q , the field of real numbers R , the field of complex numbers and the field of quaternions H are fields of characteristic 0.

1.2.7.4 Possible values of the characteristic of a field

Proposition 1.8

The characteristic of a field is either zero (for infinite fields) or a prime number (for finite fields).

In other words, if a field has a non-vanishing characteristic, then its characteristic is a prime number and the field is finite. A field of characteristic 2 is called a binary field (the field F 2 is the smallest of the binary fields).

1.2.7.5 Characteristic of two isomorphic fields

Proposition 1.9

Two isomorphic fields have the same characteristic.

1.2.7.6 Characteristic of a sub-field

Proposition 1.10

Let J be a proper sub-field of a field K . The fields J and K have the same characteristic.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9781785482359500014

Random Matrices

In Pure and Applied Mathematics, 2004

1.3.1 Level Density

As the excitation energy increases, the nuclear energy levels occur on the average at smaller and smaller intervals. In other words, level density increases with the excitation energy. The first question we might ask is how fast does this level density increase for a particular nucleus and what is the distribution of these levels with respect to spin and parity? This is an old problem treated by Bethe (1937). Even a simple model in which the nucleus is taken as a degenerate Fermi gas with equidistant single-particle levels gives an adequate result. It amounts to determining the number of partitions Λ(n) of a positive integer n into smaller positive integers v 1, v 2, …

n = v 1 + v 2 + + v , v 1 v 2 v > 0.

For large n this number, according to the Hardy-Ramanujan formula, is given by

λ ( n ) exp [ ( θ π 2 n / 3 ) 1 / 2 ] ,

where θ is equal to 1 or 2 according to whether the vi are all different or whether some of them are allowed to be equal. With a slight modification due to later work (Lang and Lecouteur, 1954; Cameron, 1956). Bethe's result gives the level density as

ρ ( E , j , π ) ( 2 j + 1 ) ( E Δ ) 5 / 4 exp [ j ( j + 1 ) / 2 σ 2 ] exp [ 2 a ( E Δ ) 1 / 2 ] ,

where E is the excitation energy, j is the spin and π is the parity. The dependence of the parameters σ, a and Δ on the neutron and proton numbers is complicated and only imperfectly understood. However, for any particular nucleus a few measurements will suffice to determine them all; the formula will then remain valid for a wide range of energy that contains thousands and even millions of levels.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/S0079816904800916

Ordinary differential equations

Brent J. Lewis , ... Andrew A. Prudil , in Advanced Mathematics for Engineering Students, 2022

Method of undetermined coefficients

Similar to Section 2.2.5 for second-order differential equations, this method gives y p for the constant coefficient equation

(2.81)

In fact, the only difference from Section 2.2.5 concerns the modification rule since now the characteristic equation of the present homogeneous equation

(2.82)

can have multiple roots (that is, greater than simple or double roots). The basic rules are summarized as follows:

(a)

Basic rule as in Section 2.2.5 with Table 2.4.

(b)

Modification rule. If the choice for y p is also a solution of the homogeneous Eq. (2.82), then multiply y p ( x ) by x k , where k is the smallest positive integer such that no term of x k y p ( x ) is a solution of Eq. (2.82).

(c)

Sum rule as in Section 2.2.5 with Table 2.4.

Example 2.3.5

(rule b) Solve y + 6 y + 12 y + 8 y = 6 e 2 x .

Solution.

(i) The characteristic equation λ 3 + 6 λ 2 + 12 λ + 8 = ( λ + 2 ) 3 = 0 has a triple root λ = 2 . Hence by Section 2.3 (Case III), y h = c 1 e 2 x + c 2 x e 2 x + c 3 x 2 e 2 x .

(ii) With y p = C e 2 x , one obtains 8 C + 24 C 24 C + 8 C = 6 (that is, there is no solution). Therefore, by rule (b) choose y p = C x 3 e 2 x . As such, y p = C ( 2 x 3 + 3 x 2 ) e 2 x , y p = C ( 4 x 3 12 x 2 + 6 x ) e 2 x , and y p = C ( 8 x 3 + 36 x 2 36 x + 6 ) e 2 x . Substituting these expressions into the differential equation gives ( 8 x 3 + 36 x 2 36 x + 6 ) C + 6 ( 4 x 3 12 x 2 + 6 x ) C + 12 ( 2 x 3 + 3 x 2 ) C + 8 x 3 C = 6 . Simplifying, one obtains C = 1 . Thus, the general solution is given by y = y h + y p = ( c 1 + c 2 x + c 3 x 2 ) e 2 x + x 3 e 2 x . [answer]

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780128236819000101

LINEAR STATE-SPACE MODELS AND SOLUTIONS OF THE STATE EQUATIONS

BISWA NATH DATTA , in Numerical Methods for Linear Control Systems, 2004

5.3.5 Evaluating an Integral with the Matrix Exponential

We have discussed methods for computing the matrix exponential. We now present a method due to Van Loan (1978) to compute integrals involving exponential matrices.

The method can be used, in particular, to compute the state-space solution (5.3.3) of the Eq. (5.3.1), and the controllability and observability Grammians, which will be discussed in the next chapter.

The method uses diagonal Padé approximation discussed in the last section.

Let

H ( Δ ) = 0 Δ e A s B d s , M ( Δ ) = 0 Δ e A T s Q H ( s ) d s , N ( Δ ) = 0 Δ e A T s Q e A s d s , W ( Δ ) = 0 Δ H ( s ) T Q H ( s ) d s

where A and B are matrices of order n and n × m, respectively, and Q is a symmetric positive semidefinite matrix.

Algorithm 5.3.3.

An Algorithm for Computing Integrals involving Matrix Exponential.

Inputs.

A—The n × n state matrix

B—An n × m matrix

Q—A symmetric positive semidefinite matrix.

Outputs. F, H, Q, M, and W which are, respectively, the approximations to e AΔ, H(Δ), N(Δ), M(Δ), and W(Δ).

Step 1.

Form the (3n + m) × (3n + m) matrix

C ˆ = ( A T l 0 0 0 A T Q 0 0 0 A B 0 0 0 0 ) .

Find the smallest positive integer j such that ( | | C ˆ Δ | | F / 2 j ) 1 2 . Set t 0 = (Δ/2 j ).

Step 2.

For some q ⩾ 1, compute

Y 0 = R q q ( C ˆ Δ 2 j ) ,

where R qq is the (q, q) Padé approximant to e z :

R q q ( z ) = Σ k = 0 z c k z k Σ k = 0 q c k ( z ) k , w h e r e c k = ( 2 q k ) ! q ! ( 2 q ) ! k ! ( q k ) ! .

Write

Y 0 = ( F 1 ( t 0 ) G 1 ( t 0 ) H 1 ( t 0 ) K 1 ( t 0 ) 0 F 2 ( t 0 ) G 2 ( t 0 ) H 2 ( t 0 ) 0 0 F 3 ( t 0 ) G 3 ( t 0 ) 0 0 0 F 4 ( t 0 ) )

and set

F 0 = F 3 ( t 0 ) M 0 = F 3 ( t 0 ) T H 2 ( t 0 ) H 0 = G 3 ( t 0 ) W 0 = [ B T F 3 ( t 0 ) T K 1 ( t 0 ) ] + ( B T F 3 ( t 0 ) T K 1 ( t 0 ) ] T . Q 0 = F 3 ( t 0 ) T G 2 ( t 0 ) .

Step 3.

For k = 0, 1,…, j − 1 do

W k + 1 = 2 W k + H k T M k + M k T H k + H k T + H k T Q k H k M k + 1 = M k + F k T [ Q k H k + M k ] Q k + 1 = Q k + F k T Q k F k H k + 1 = H k + F k H k F k + 1 = F k 2

End

Step 4.

Set FF j , HH j , QQ j , MM j , and WW j .

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780122035906500094

Stream Ciphers and Number Theory

In North-Holland Mathematical Library, 2004

3.2 Two Basic Problems from Stream Ciphers

For sequences of period N over the field GF(q), their linear and sphere complexity are closely related with the factorization of cyclotomic polynomials Q n (x) over GF(q) for all factors n of N. Proposition 3.1.1 says that Q n (x) factors into ϕ(n)/d distinct monic irreducible polynomials in GF(q) of the same degree d, where d is the least positive integer such that q d ≡ 1 (mod n). It follows that, to design sequences with both large linear and sphere complexity, we should find pairs (N, q) such that

1.

N has as few factors as possible; and

2.

for each factor n of N, d = ord n (q) should be as large as possible.

This leads to the following two basic problems in designing cryptographic sequences for certain applications.

Basic Problem 1

Find large positive integers N and small positive integers q which are powers of primes such that

1.

gcd(N, q) = 1;

2.

ord n (q) = ϕ(n) for any factor n ≠ 1 of N.

Basic Problem 2

Find large positive integers N and small positive integers q, q a power of a prime, such that

1.

gcd(N, q) = 1;

2.

N has few factors;

3.

ord n (q), a factor of ϕ(n), is as large as possible for any factor n ≠ 1 of N.

An integer q is said to be a primitive root of (or modulo) n if ord n (q) = ϕ(n). If gg' (mod N), then g' is a primitive root of N if and only if g' is a primitive root of N. So for our cryptographic purposes, we discuss here and hereafter primitive roots modulo N only in the range between 2 and N − 1. To study the two problems further, we need the following important result of Gauss whose proof can be found in most books about number theory.

Proposition 3.2.1

If p is a prime, then there exist ϕ(p − 1) primitive roots of p. The only integers having primitive roots are p e , 2p e , 1, 2 and 4, with p being an odd prime.

This proposition shows that Basic Problem 1 has a solution if and only if N = r k , or 2r k , with r being an odd prime. We shall investigate this basic problem in detail in Sections 3.4 and 3.5.

Before dealing with Basic Problem 2, we present some basic results about the order of integers modulo n. If gcd(a, n) = 1, Euler's theorem states that a ϕ(n) ≡ (mod n). This implies that ord n (a) divides ϕ(n). The order of a has a close relation to the Carmichael function λ(n), which is defined by

λ ( 1 ) = 1 , λ ( 2 ) = 1 , λ ( 4 ) = 2 , λ ( 2 r ) = 2 r 2 ( for r 3 ) λ ( p r ) = p r 1 ( p 1 ) = ϕ ( p r ) for any odd prime p and r 1 , λ ( 2 r p 1 r 1 p 2 r 2 p s r s ) = lcm ( λ ( 2 r ) , λ ( p 1 r 1 ) , , λ ( p s r s ) ) ,

where lcm denotes the least common multiple. It is not difficult to see that the order of a modulo n is at most equal to λ(n), and that λ(n) divides ϕ(n).

It seems difficult to solve Basic Problem 2 completely. However, for those N's which are a product of two distinct primes, it is possible to find the associated q's such that (N, q) is a solution of Basic Problem 2. We shall deal with this problem in Section 3.8.

Before ending this section, we make some preparations for the following two sections. Specifically, we introduce now the concept of negative order of an integer a modulo an integer N, and discuss the relation of the negative order with the order.

Definition 3.2.2

Let N and a be positive integers. If there is a positive integer m such that a m ≡ −1 (mod N), then we call the smallest such m the negative order of a modulo N (we coin the word "negord" to denote the negative order), and denote it as nord N (a).

An integer a may have a negord modulo an integer N or not. As an example, we consider N = 23. It is easily checked that 1, 2, 4, 8, 16, 9, 18, 13, 36 and 12 have no negord, but 17, 11, 22, 21, 19, 15, 7 and 14 have a negord. It is for the purpose of investigating the order that we introduce the concept of the negord.

The relation of the order and negord is stated in the following theorem.

Theorem 3.2.3

Let N be a positive integer. If an integer a, where 1 ≤ aN − 1 and gcd(a, N) = 1, has a negord modulo N, then

ord N ( a ) = 2 nord N ( a )

Proof: By definition a nord N (a) = −1 (mod N). It follows that a 2nord N (a) ≡ 1 (mod N). Hence, ord N (a) divides 2nord N (a). We now prove that ord N (a) ≥ 2nord N (a). If not so, then there are two possibilities: ord N (a) < nord N (a) and nord N (a) < ord N (a) < 2nord N (a). It is easily verified that in both cases there must exist an integer l, where 1 ≤ l < nord N (a), such that a l ≡ −1 (mod N). This is contrary to the minimality of the negord of a modulo N. Thus, ord N (a) must be equal to 2nord N (a).

A simple property of negord, which is similar to that of order, is the following conclusion.

Theorem 3.2.4

If a m ≡ −1 (mod N) for a positive integer m, then nord N (a)|m and m/nord N (a) is odd.

Proof: Let m = novd N (a)h+l, where 0 ≤ l <nord N (a). We first prove that h must be odd. From a m ≡ (a nord N (a)) h a l (mod N) we get a l ≡ (−1) h+1 (mod N). By the definition of the negord h is odd.

If l ≠ 0, then l ≥ 1. The equation a l ≡ 1 (mod N) gives that ord N (a) < nord N (a), which is contrary to Theorem 3.2.3. Therefore, l = 0. This completes the proof.

Now we give a characterization of primitive roots in terms of negord. This characterization is useful in searching for primitive roots.

Theorem 3.2.5

Let N be a positive integer > 4 which has primitive roots. Then a is a primitive root modulo N if and only nord N (a) = ϕ(N)/2.

Proof: If a is a primitive root modulo N, by Proposition 3.2.1 N must be of the form p e or 2p e , where p is an odd prime. Thus ϕ(N) must be even. Since a ϕ(N) ≡ 1 (mod N), we get

( a ϕ ( N ) / 2 + 1 ) ( a ϕ ( N ) / 2 1 ) 0 ( mod N )

This gives a ϕ(N)/2 ≡ −1 (mod N). Thus, the negord of a modulo N exists. Now by Theorem 3.2.3 we have nord N (a) = ϕ(N)/2. The remaining part then follows from Theorem 3.2.3.

This theorem shows that a necessary condition for a to be a primitive root is a ϕ(N)/2 ≡ −1 (mod N). It can be used as a criterion for primitivity. As an example, we take N = 43. Then we have 2ϕ(N)/2 = 2(N − ½) = 23 × 7 = −1 (mod N). But 2 is not a primitive root of 43. This is because nord43(2) = 7 ≠ 21.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/S0924650904800059

Cellular Automata

Jean-Paul Allouche , ... Gencho Skordev , in Encyclopedia of Physical Science and Technology (Third Edition), 2003

V Cellular Automata as Dynamical Systems

We can also consider cellular automata as dynamical systems. A topological dynamical system (X, f) consists of a compact set X, together with a continuous map f: X  X. We will restrict our attention to compact metric spaces. In the previous section, we saw that a cellular automaton on Z n can be considered as a continuous map on the (compact metric) space of configurations.

We shall discuss cellular automata in the context of classical topological dynamics, focusing particularly on the "repetitiveness" properties, such as the existence of periodic points, chain recurrence, nonwandering sets, center of the dynamical system, and so on, as well as the properties of "attracting" or "repelling" invariant sets, which are related to stability properties of the dynamical system.

V.A Periodic Points

Let (X, f) be a dynamical system. A point x 0  X is periodic if there exists a positive integer n such that f n (x 0)   = x 0. If this is the case, the orbit of x 0 under f, that is, the set {f k (x 0); k    0}, is finite. Its cardinality is called the period of the point x 0. A point x 0  X is ultimately periodic if there exist two integers k    0 and n    1 such that f n+k (x 0)   = f k (x 0). Then the orbit of x 0 is finite. The transient part of the orbit of x 0 is the set {f j(x 0); 0   j  k−1}, where k is the smallest non-negative integer such that for some positive integer n, f n+k (x 0)   = f k (x 0 ). The smallest positive integer n associated with this k is called the period of x 0.

Periodic points and their generalizations (almost periodic points and recurrent points) for dynamical systems arising from cellular automata were studied by Hedlund and others. Periodic points of cylindrical cellular automata, that is, cellular automata defined on the Cayley graph of the integers modulo m, and relations between spatial and temporal periods were studied by Martin et al. (1984).

V.B Attractors

Intuitively, an attractor of a topological dynamical system (X, f) is a compact invariant set which "attracts" all the points in some neighborhood, in the sense that iterating the map f from any one of these points gives points that converge to the attractor. Such an attractor is Lyapunov stable; that is, every orbit under the map f starting sufficiently close to the attractor remains in a neighborhood of the attractor. These orbits ultimately converge to the attractor. Formally, a set A  X is an attractor of the dynamical system (X, f) if there is an open neighborhood U of A such that f ( U ) U and

A = n = 0 f n ( U ) ,

where U is the closure of U. The open set,

B ( A ) : = n = 0 f n ( U )

is called the basin of attraction of A.

It follows from the definition that the set A is compact and the basin B(A) is also the set of points x such that all limit points in their orbits belong to A. In other words B(A) does not depend on the choice of the open set U.

Although complete descriptions of all attractors are rare in the theory of dynamical systems, the following result of Hurley (1990) holds for cellular automata.

Theorem.

A cellular automaton on Z n , considered as a dynamical system, satisfies exactly one of the following conditions:

There exists a unique minimal attractor contained in every attractor.

There is a unique minimal quasi-attractor (that is, an intersection n = 1 A n of a sequence of attractors) contained in every attractor.

There exist two disjoint attractors. In this case, the dynamical system associated with the cellular automaton has uncountably many minimal quasi-attractors.

V.C Expansiveness and Permutivity

In the remainder of this section, we shall only consider topological dynamical systems whose topology is induced by a metric or distance. This is always the case for the cellular automata under consideration.

A topological dynamical system (X, f) is expansive at a point x 0 in X if there exists a positive real constant δ, called the constant of expansiveness at x 0, with the property that for every point y not equal to x 0, there exists a non-negative integer k such that the distance between f k (y) and f k (x 0) is at least δ. The dynamical system is called expansive if it is expansive at every point x in X. The concept of μ-expansivity for a topological dynamical system (X,f,μ) with a measure μ on X is defined analogously by replacing "for every point y" with "for μ-almost every point y".

For example, the one-dimensional shift on a finite set A defined on the set of two-sided sequences on A by:

u n n Z u n + 1 n Z

is expansive. Indeed, the following theorem, due to Hedlund, shows that expansive maps are "close" to shifts.

Theorem.

If the topological dynamical system (X, f) is expansive, then it is isomorphic to a subshift; that is, there exists a one-dimensional shift σ and an imbedding ι from X to A Z such that the set ι(X) is shift invariant, and σ ∘ ι   =   ι ∘ f.

A cellular automaton is permutive if its local function depends "in an essential way" on the values of the leftmost and the rightmost neighbors. For permutive cellular automata, there is a more precise result due to Gilman (1998).

Theorem.

Any permutive cellular automaton is expansive. Furthermore, the corresponding dynamical system is isomorphic to a one-dimensional, one-sided shift; that is, the dynamical system (A N , σ), where A is a finite set, and σ is the map defined on A N by:

u n n 0 u n + 1 n 0

When n    2, n-dimensional cellular automata are never expansive.

V.D Equicontinuity or Lyapunov Stability

A topological dynamical system (X, f) is equicontinuous or Lyapunov stable at the point x 0 in X if, for every positive real number ε, there exists a neighborhood of x 0 such that for every y in this neighborhood, all iterations f i (x 0) and f i (y) are ε-close. The dynamical system is called equicontinuous if it is equicontinuous at every x  X. For a cellular automaton with values in a finite set A, the uniform Bernoulli measure μ on the configuration space is the product measure defined on A by μ({x})   =   1/# A (where # A is the number of elements in A) for each value x in A. The following theorem about equicontinuity in cellular automata is due to Gilman (1988).

Theorem.

Every cellular automaton with values in a finite set A satisfies one of the following conditions (where μ is the uniform Bernoulli measure): Either for every ε   >   0, there exists a compact shift-invariant subset Y ε of the set of configurations such that μ(Y ε)   >   1     ε and the dynamical system obtained by restricting the cellular automaton to Y ε is equicontinuous, or the cellular automaton is μ-expansive.

V.E Sensitivity and Transitivity

A topological dynamical system (X, f) is sensitive at a point x 0 in X if it is not equicontinuous at x 0. It is sensitive if it is sensitive for all x  X. If a dynamical system is sensitive, it is, in some sense, "chaotic." Another property associated with a chaotic behavior is (topological) transitivity. A topological dynamical system (X, f) is transitive or topologically mixing if for every pair of non-empty open subsets U and V of X, some iteration f k (U) of the set U intersects V. There are several different definitions of a "chaotic" dynamical system. One definition is due to Devaney. A topological dynamical system (X, f) is chaotic if it is transitive and sensitive and if the set of all periodic points is dense in X It can be proved that sensitivity follows from the other two properties. Sensitivity, transitivity, and other properties for cellular automata seen as dynamical systems have been studied.

V.F Cellular Automata as Measurable Dynamical Systems

A measurable dynamical system is a triple (X, μ, f) consisting of a set X with a probability measure μ, and a function f: X  X which is measure preserving; that is, for all Y  X, μ(Y)   =   μ(f −1(Y)). A measurable dynamical system is ergodic if every subset of X invariant under f has measure 0 or 1. The following is another theorem of Hedlund.

Theorem.

The dynamical system associated with a one-dimensional cellular automaton equipped with the Bernoulli measure is a measurable dynamical system if and only if it is surjective.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B0122274105000910

Inferring the Topology of Gene Regulatory Networks: An Algebraic Approach to Reverse Engineering

Brandilyn Stigler , Elena Dimitrova , in Mathematical Concepts and Methods in Modern Biology, 2013

3.6 Discretization

For reasons explained at the beginning of Section 3.2, we have been assuming that the experimental data we use for reverse engineering have already been discretized into a (small) finite number of states. Typically, however, experimental measurements come to us represented by computer floating point numbers and consequently data discretization is in fact part of the modeling process and can be viewed as a preprocessing step. We will use the definition of discretization presented in [35].

Definition 3.11

A discretization of a real-valued vector v = ( v 1 , , v N ) is an integer-valued vector d = ( d 1 , , d N ) with the following properties:

1.

Each element of d is in the set 0 , 1 , , D - 1 for some (usually small) positive integer D, called the degree of the discretization.

2.

For all 1 i , j N , we have d i d j if and only if v i v j .

Without loss of generality, assume that v is sorted, i.e., for all i < j , v i v j . Spanning discretizations of degree D satisfy the additional property that the smallest element of d is equal to 0 and that the largest element of d is equal to D - 1 .

There is no universal way for data discretization that works for all data sets and all purposes. Sometimes discretization is a straightforward process. For example, if a gene expression time series has a sigmoidal shape, e.g., ( 0.1 , 1.2 , 2 , 23.04 , 26 ) , it is reasonable to discretize it as ( 0 , 0 , 0 , 1 , 1 ) . More complicated expression profiles may be easy to discretize too and it is often true that the human eye is the best discretization "tool" whose abilities to discern patterns cannot be reproduced by any software. Large data sets, on the other hand, do require some level of automatization in the discretization process. Regardless of the particular situation, it is good practice to look at the data first and explore for any patterns that may help with the discretization before inputting the data into any discretization algorithm. Afterwards, the way you choose to discretize your data, which includes selecting the number of discrete states, should depend on the type and amount of data and the specific reason for discretization. Below we present several possible approaches which by no means comprise a complete list.

Binary discretizations are the simplest way of discretizing data, used, for instance, for the construction of Boolean network models for gene regulatory networks [36,37]. The expression data are discretized into only two qualitative states as either present or absent. An obvious drawback of binary discretization is that labeling real-valued data according to a present/absent scheme may cause the loss of large amounts of information.

Interval discretizations divide the interval [ v 1 , v N ] into k equally sized bins, where k is user-defined. Another simple method is quantile discretization which places N / k (possibly duplicated) values in each bin [38]. Any method based on those two approaches would suffer from problems that make it inappropriate for some data sets. Interval discretizations are very sensitive to outliers and may produce a strongly skewed range [39]. In addition, some discretization levels may not be represented at all which may cause difficulties with their interpretation as part of the state space of a discrete model.

On the other hand, quantile discretizations depend only on the ordering of the observed values of v and not on the relative spacing values. Since distance between the data points is often the only information that comes with short time series, losing it is very undesirable. A shortcoming, common for both interval and quantile, as well as for most other discretization methods, is that they require the number of discrete states, k, to be user-provided.

A number of entropy-based discretization methods deserve attention. An example of those is Hartemink's Information-Preserving Discretization (IPD) [35]. It relies on minimizing the loss of pairwise mutual information between each two real-valued vectors (variables). The mutual information between two random variables X and Y with joint distribution p ( X , Y ) and marginal distributions p ( x ) and p ( y ) is defined as

I ( X ; Y ) = x y p ( x , y ) log p ( x , y ) p ( x ) p ( y ) .

Note that if X and Y are independent, by definition of independence p ( x , y ) = p ( x ) p ( y ) , so I ( X ; Y ) = 0 . Unfortunately, when modeling regulatory networks and having as variables, for instance, mRNA, protein, and metabolite concentrations, the joint distribution function is rarely known and it is often hard to determine whether two variables are independent.

Another family of discretization techniques is based on clustering [40]. One of the most often used clustering algorithms is the k-means [41]. It is a non-hierarchical clustering procedure whose goal is to minimize dissimilarity in the elements within each cluster while maximizing this value between elements in different clusters. Many applications of the k-means clustering such as the MultiExperiment Viewer [42] start by taking a random partition of the elements into k clusters and computing their centroids. As a consequence, a different clustering may be obtained every time the algorithm is run. Another inconvenience is that the number k of clusters to be formed has to be specified in advance. Although there are methods for choosing "the best k" such as the one described in [43], they rely on some knowledge of the data properties that may not be available.

Another method is single-link clustering (SLC) with the Euclidean distance function. SLC is a divisive (top-down) hierarchical clustering that defines the distance between two clusters as the minimal distance of any two objects belonging to different clusters [40]. In the context of discretization, these objects will be the real-valued entries of the vector to be discretized, and the distance function that measures the distance between two vector entries v and w will be the one-dimensional Euclidean distance | v - w | . Top-down clustering algorithms start from the entire data set and iteratively split it until either the degree of similarity reaches a certain threshold or every group consists of one object only. For the purpose of data analysis, it is impractical to let the clustering algorithm produce clusters containing only one real value. The iteration at which the algorithm is terminated is crucial since it determines the degree of the discretization. SLC with the Euclidean distance function has one major advantage: very little starting information is needed (only distances between points). In addition, being a hierarchical clustering procedure it lends itself to adjustment in case that clusters need to be split or merged. It may result, however, in a discretization where most of the points are clustered into a single partition if they happen to be relatively close to one another. Another problem with SLC is that its direct implementation takes D, the desired number of discrete states, as an input. However, we would like to choose D as small as possible, without losing information about the system dynamics and the correlation between the variables, so that an essentially arbitrary choice is unsatisfactory.

In [44], a hybrid method for discretization of short time series of experimental data into a finite number of states was introduced. It is a modification of the SLC algorithm: it begins by discretizing a vector in the same way as SLC does but instead of providing D as part of the input, the algorithm contains termination criteria which determine the appropriate value of D. After that each discrete state is checked for information content and if it is determined that this content can be considerably increased by further discretization, then the state is separated into two states in a way that may not be consistent with SLC.

If more than one vector is to be discretized, the algorithm discretizes each vector independently. (For details on multiple vector discretization, see [44]). If the vector contains m distinct entries, a complete weighted graph on m vertices is constructed, where a vertex represents an entry and an edge weight is the Euclidean distance between its endpoints. The discretization process starts by deleting the edge(s) of highest weight until the graph gets disconnected. If there is more than one edge labeled with the current highest weight, then all of the edges with this weight are deleted. The order in which the edges are removed leads to components, in which the distance between any two vertices is smaller than the distance between any two components, a requirement of SLC. We define the distance between two components G and H to be dist ( G , H ) = min { | g - h | | g G , h H } .The output of the algorithm is a discretization of the vector, in which each cluster corresponds to a discrete state and the vector entries that belong to one component are discretized into the same state.

Example 3.12

Suppose that vector v = ( 1 , 2 , 7 , 9 , 10 , 11 ) is to be discretized. The diagram obtained by the SLC algorithm is given in Figure 3.3. The complete weighted graph in Figure 3.4 corresponds to iteration 0 of the diagram.

Figure 3.3. Diagram representing the SLC algorithm applied to the data of Example 3.12. The column on the right gives the corresponding Shannon's entropy increasing at each consecutive level.

Figure 3.4. The complete weighted graph constructed from vector entries 1, 2, 7, 9, 10, 11. Only the edge weights of the outer edges are given.

Having disconnected the graph, the next task is to determine if the obtained degree of discretization is sufficient; if not, the components need to be further disconnected in a similar manner to obtain a finer discretization. A component is further disconnected if, and only if, both (1) and (2) below hold:

1.

The minimum vertex degree of the component is less than the number of its vertices minus 1. The contrary implies that the component is a complete graph by itself, i.e., the distance between its minimum and maximum vertices is smaller than the distance between the component and any other component.

2.

One of the following three conditions is satisfied ("disconnect further" criteria):

a.

The average edge weight of the component is greater than half the average edge weight of the complete graph.

b.

The distance between its smallest and largest vertices is greater than or equal to half this distance in the complete graph. For the complete graph, the distance is the graph's highest weight.

c.

Finally, if the above two conditions fail, a third one is applied: disconnect the component if it leads to a substantial increase in the information content carried by the discretized vector.

The result of applying only the first two criteria is analogous to SLC clustering with the important property that the algorithm chooses the appropriate level to terminate. Applying the third condition, the information measure criterion may, however, result in a clustering which is inconsistent with any iteration of the SLC.

Exercise 3.18

Consider the following vector v 1 = ( 1.3 , 2.1 , 7.2 , 9.05 , 10.5 , 11.00 ) . Plot the vector entries on the real number line and propose an appropriate discretization based on your intuition. How did you decide on the number of discrete states? Would your decision change if the entry 4.6 is added?  

Exercise 3.19

Discretize vector v 1 from Problem 3.18 using the specified method.

1.

Interval.

2.

Quantile.

3.

The hybrid method presented in this section.

 

Exercise 3.20

If you know that the experimental data is very noisy, would you be willing to use a smaller or a larger number of discrete states?  

Exercise 3.21

Suppose a second vector, v 2 = ( 0.8 , 1.8 , 3.1 , 8.0 , 9.5 , 10.7 ) , is to be discretized and it is known that v 1 and v 2 are strongly correlated (assume Spearman rank correlation). How would you discretize v 2 based on your discretization of v 1 ?  

Project 3.7

Use the principle of relationship discretization that is behind Hartemink's information-preserving discretization method to discretize v 1 and v 2 .  

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B978012415780400003X

Quantum Algorithms and Methods

Ivan B. Djordjevic , in Quantum Information Processing, Quantum Computing, and Quantum Error Correction (Second Edition), 2021

5.12 Problems

1.

By using mathematical induction, prove that the following circuit can be used to implement the Deutsch–Jozsa algorithm, that is, to verify whether the mapping {0,1} n → {0,1} is constant or balanced.

2.

By using mathematical induction, prove that the following circuit can be used to calculate the FT for arbitrary n:

3.

Assume that a unitary operator U has an eigneket |u〉 with eigenvalue exp(j2πφ u ). The goal of phase estimation is to estimate the phase φ u . Describe how the quantum FT can be used for phase estimation.

4.

This problem is related to the shift-invariance property of quantum FT. Let G be a group and H be a subgroup of G. If a function f on G is constant on cosets of H, then the FT of f is invariant over cosets of H. Prove the claim.

5.

The quantum circuit to perform the Grover search for n  =   3 was shown in Fig. 5.11. Provide a quantum circuit that can be used to perform the Grover search for arbitrary n. Explain the operation principle and prove that this circuit can indeed be used for any n.

6.

Provide a quantum circuit to implement Shor's factorization algorithm for n  =   15. Describe the operation principle of this circuit.

7.

This problem is devoted to quantum discrete logarithms. Let us consider the following function: f ( x 1 , x 2 ) = a b x 1 + x 2 mod N , where all variables are integers. Let r be the smallest positive integer for which a r mod N  =   1; this integer can be determined by using the order-finding algorithm. Clearly, this function is periodic as f(x 1 + i,x 2 − ib)   = f(x 1,x 2), where i is an integer. The discrete logarithm problem can be formulated as follows: given a and c  = a b , determine b. This problem is important in cracking the RSA encryption protocol as discussed in Section 5.5. Your task is to provide a quantum algorithm that can solve this problem by using one query of a quantum block U that performs the following unitary mapping: U|x 1〉|x 2〉|y〉 → |x 1〉|x 2〉|yf(x 1,x 2)〉.

8.

In Problem 6 you were asked to provide a quantum circuit to implement Shor's factorization algorithm for n  =   15. By using Simon's algorithm, describe how to perform the same task. Provide the corresponding quantum circuit. Analyze the complexity of both algorithms.

9.

Suppose that the list of numbers x 1,…,x n is stored in quantum memory. How many memory accesses are needed to determine the smallest number in the list with a success probability of ≥1/2?

10.

Provide the quantum circuit to perform the following mapping:

| m 1 p n = 0 p 1 e j 2 π m n / p | n ,

where p is a prime.

11.

Design a quantum circuit to perform the following mapping: |x〉→ |x  + c mod 2 n 〉, where x ∈ [0,2 n–1] and c is the constant, by using quantum FT.

12.

The following circuit can be used to perform addition:

Describe the operation principle, and provide the result of addition. Can you generalize this addition problem?

13.

The following problem is related to Kitaev's algorithm, which represents an alternative way to estimate the phase. Let us observe the following quantum circuit:

where |u〉 is an eigneket of U with eigenvalue exp(j2πφ u ). Show that the result of the measurement 0 appears with a probability of p  =   cos2(πφ). Since the eigenket is insensitive to measurement the operator U can be replaced with U m , where m is an arbitrary positive integer. Show that by repeating this circuit appropriately we can obtain the arbitrary precision of p and consequently estimate the phase φ with desired precision. Compare the complexity of Kitaev's algorithm with respect to that from Problem 3.

14.

The state ket after application of Hadamard gates of the following circuit:

can be written in the following form:

| ψ = 1 N 1 / 2 x a x ( 0 ) | x , a x ( 0 ) = 1.

The application of operator GO on |ψ〉 leads to the following ket:

| ψ = 1 N 1 / 2 x a x ( 1 ) | x .

Establish the connection between a x ( 1 ) and a x ( 0 ) .

15.

This problem is related to quantum simulation. When the Hamiltonian H that can be represented as the sum of polynomial many terms H m , namely H  =   m H m , each of which can be efficiently implemented, we can efficiently simulate the evolution operator exp ( j H t ) and approximate the evolution of state | ψ ( t ) = exp ( j H t ) | ψ ( 0 ) . . If for all m,n [H m ,H n ]   =   0, then we can write exp ( j H t ) = k e j H k t . However, if [H m ,H n ] ≠ 0, then the previous equation is not valid. The following formula, known as the Trotter formula, can be used for approximations leading to quantum simulation algorithms: lim n ( e j U 1 t / n e j U 2 t / n ) = e j ( U 1 + U 2 ) t , where U 1 and U 2 are Hermitian operators. Prove the Trotter formula. Prove also the following useful approximations:

e j ( U 1 + U 2 ) Δ t = e j U 1 Δ t e j U 2 Δ t + O ( Δ t 2 ) , e j ( U 1 + U 2 ) Δ t = e j U 1 Δ t / 2 e j U 2 Δ t e j U 1 Δ t / 2 + O ( Δ t 3 ) . Finally, the following approximation, known as the Baker–Campbell–Hausdorf formula, is also useful in quantum simulation: e ( U 1 + U 2 ) Δ t = e U 1 Δ t e U 2 Δ t e [ U 1 , U 2 ] Δ t 2 / 2 + O ( Δ t 3 ) . Prove it. Consider now a single particle living in 1D potential V(x), governed by Hamiltonian: H  = p 2/(2m) + V(x). Perform the computation |ψ(t)〉   = exp(−jHt/)|ψ(0)〉 by using the foregoing approximations.

16.

Construct the quantum circuit to simulate the Hamiltonian H  = Z 1 Z 2Z n , performing the unitary transform |ψ(t)〉   = exp(−jHt/)|ψ(0)〉for arbitrary Δt.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B978012821982900006X

Fuzzy Measures of Molecular Shape and Size

PAUL G. MEZEY , in Fuzzy Logic in Chemistry, 1997

XIV FUZZY SET GENERALIZATIONS OF ZPA FOLDING-UNFOLDING CONTINUOUS SYMMETRY MEASURES BASED ON THE FUZZY FSNDSM METRIC AND FUZZY HAUSDORFF-TYPE METRICS

The fuzzy average F fav of a family of crisp or fuzzy sets F 1, F 2,…,F m has a much simpler construction 29 than the average of crisp sets used in Section XIII and that also simplifies the extension of ZPA-type folding-unfolding continuous symmetry measures to both crisp continua and fuzzy sets, using dissimilarity metrics designed for fuzzy sets and used as the actual measures of symmetry deficiency. Here we shall follow the technique 29 based on the fuzzy membership function μ F fav (x) of the fuzzy average F fav, as defined earlier by Eq. (199).

Consider a crisp or fuzzy subset A of the Euclidean space X, a (possibly approximate) symmetry element R, and the associated symmetry operator R. A fixed point of R is chosen as a reference point cX, and a local Cartesian coordinate system of origin c is specified, with coordinate axes oriented according to the usual conventions with respect to the symmetry operator R, as described for crisp sets in Section XIII.

Choose m as the smallest positive integer that satisfies the condition R m = E of Eq. (171) and take the powers R 0= E, R, R 2,…,R m − 1 of the symmetry operator R. Following the technique described in Section XIII, partition the Euclidean space X into m segments X 0, X 1,…,X m − 1, where the union of these segments generates the space X,

(236) X = j = 0 , m 1 X i

and where segments of subsequent indices are related to one another by the symmetry operation

(237) R X j = X ( j + 1 ) m o d m

The interpretation of these segments and the notation P for the convention used for the positioning of R with respect to set A and for the partitioning X 0, X 1,…, X m − 1 of the space X are the same as those used for crisp sets, described in Section XIII.

The segment A j of the crisp or fuzzy set A is defined as

(238) A j = A X j

and the "folded" version of the jth segment A j of the crisp or fuzzy set A, according to the (mj)th power R mj of the (possibly only approximate) symmetry operator R is denoted by

(239) B j = R m - j A j

For the family S A of segments A 0, A l, A 2,…,A m − 1, the corresponding folded sets A 0, R m − 1 A 1, R m − 2 A 2,…,R mj A j,…,R A m − 1 are obtained, that is, the family S B of sets B 0, B 1, B 2,…,B m − 1 is generated.

The fuzzy folded set A ffold is the fuzzy average S Bfav of these B j sets, defined by the fuzzy membership function μ S Bfav(x) as follows:

(240) μ A f t 1 d ( x ) = μ S B 4 V ( x ) = [ k = 0 , m 1 μ B k ( x ) ] / m

The "unfolding" of the fuzzy folded set A ffold is obtained using the appropriate inverse powers of symmetry operator R generating the following sets: A ffold,R 1 − m A ffold,R 2 − m A ffold,…,R jm A ffold,…,R −1 A ffold. The fuzzy union of all these sets is the folded-unfolded fuzzy set Aff, uf, R, P of crisp or fuzzy set A, generated according to symmetry element R and the actual partitioning P:

(241) A f f . uf , R , P = j = 0 , m 1 R j A f fold .

The folded-unfolded set A ff, uf, R, F of crisp or fuzzy set A is a fuzzy R set, by construction.

Fuzzy dissimilarity measures, such as the fuzzy FSNDSM metric fs(A, B), and any one of the fuzzy Hausdorff-type dissimilarity metrics, for example, f(A, B), can be applied to the pair of set A and the folded-unfolded set A ff, uf, R, P . These fuzzy dissimilarity measures generate fuzzy symmetry deficiency measures analogous to the ZPA continuous symmetry measure of discrete point sets.

If the FSNDSM metric fs(A, B) is selected, then the corresponding symmetry deficiency measure is defined as d fs(A, A ff, uf, R, P ). The measure d fs(A, A ff, uf, R, P ) describes the fuzzy FSNDSM "shape distance" between the crisp or fuzzy set A and the fully R-symmetric, fuzzy, folded-unfolded set A ff, uf, R, P of the original set A. By analogy with the case of crisp sets discussed in Section XIII, the d fs(A, A ff, uf, R, P ) symmetry deficiency measure is P-dependent, that is, it depends on the positioning P of R with respect to A and on the choice of the corresponding partitioning of the underlying Euclidean space X and that of the original crisp or fuzzy set A.

By taking the infimum for all the allowed choices of P, a symmetry deficiency measure of crisp or fuzzy set A is obtained that is independent of positioning and partitioning. The corresponding d fs(A, A ff, uf, R ) measure is defined as the infimum of d fs(A, A ff, uf, R P ) generated over all the allowed positionings and partitionings:

(242) d fs ( A , A f f , u f . R ) = inf P { d f S ( A A f f , u f , R , P ) }

Using any one of the versions of the fuzzy Hausdorff-type metrics for the dissimilarity of sets A and A ff, uf, R, P , for example, the "commitment weighted" fuzzy Hausdorff-type dissimilarity metric f(A, B), one obtains another generalization of the ZPA continuous symmetry measure of discrete point sets to crisp or fuzzy sets. The corresponding symmetry deficiency measure f(A, A ff, uf, R, P ) provides a measure for the symmetry aspect R for crisp or fuzzy set A, with reference to the given positioning P of R with respect to A and to the choice of the associated partitioning of A.

For a symmetry deficiency measure of set A, independent of positioning and partitioning, the infimum of f(A, A ff, uf, R, P ) can be taken over all the allowed positionings and partitionings P. This measure, f(A, A ff, uf, R ), is defined as

(243) f ( A , A f f . u f , R ) = inf P { f ( A , A f f , u f , R , P ) }

Following the principles of the ZPA approach, these symmetry deficiency measures are generalizations of the folding-unfolding approach, equally applicable to crisp continuum sets and fuzzy sets, for example, to entire electron density distributions of molecules and various molecular fragments representing fuzzy functional groups.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B978012598910750007X

Higher Order Equations

Martha L. Abell , James P. Braselton , in Introductory Differential Equations (Fourth Edition), 2014

Undetermined Coefficients

In the special case that the corresponding homogeneous equation has constant coefficients and the forcing function is a linear combination of functions of the form

(4.23) 1 , t , , t 2 , , e α t , t e α t , t 2 e α t , , cos β t , sin β t , t cos β t , t sin β t , t 2 cos β t , t 2 sin β t , or e α t cos β t , e α t sin β t , t e α t cos β t , t e α t sin β t , t 2 e α t cos β t , t 2 e α t sin β t , ,

then the method of undetermined coefficients can be used to determine the form of a particular solution of the nonhomogeneous equation in the same way as that discussed in Section 4.2.

Example 4.6.2

Solve y (5) + 4y‴ = 48t − 6 − 10et .

Solution

The corresponding homogeneous equation is y (5) + 4y‴ = 0, which has characteristic equation r 5 + 4r 3 = r 3(r 2 + 4) = 0 so r 1,2,3 = 0 has multiplicity three and r 4,5 = ±2i each have multiplicity one. Thus, a fundamental set for the corresponding homogeneous equation is S = 1 , t , t 2 , cos 2 t , sin 2 t and a general solution of the corresponding homogeneous equation is y h = c 1 + c 2 t + c 3 t 2 + c 4 cos 2 t + c 5 sin 2 t .

For the forcing function, f(t) = 48t − 6 − 10et we have two associated sets:

F 1 = e t and F 2 = t , 1 ,

corresponding to the terms − 10et and 48t − 6, respectively, in the forcing function. Note that no element of F 1 is a solution of the corresponding homogeneous equation. On the other hand, functions in F 2 are solutions to the corresponding homogeneous equation. So, following the outline in Section 4.2, we multiply F 2 by t n where n is the smallest positive integer so that no function in t n F 2 is a solution of the corresponding homogeneous equation. In this case, we multiply F 2 by t 3 to obtain t 3 F 2 = t 4 , t 3 .

Thus, we assume that a particular solution to the nonhomogeneous equation has the form y p = At 4 + Bt 3 + Cet . Differentiating we have,

y p = 4 A t 3 + 3 B t 2 C e t , y p = 1 2 A t 2 + 6 B t + C e t , y p = 2 4 A + C e t and y p ( 4 ) = C e t .

Substituting into the nonhomogeneous equation, simplifying the result, and equating coefficients gives us

2 4 B + 9 6 A t 5 C e t = 4 8 t 6 1 0 e t ,

so 24B = −6, 96A = 48, and − 5C = −10. Thus, A = 1/2, B = −1/4, and C = 2 so a particular solution of the nonhomogeneous equation is y p = 1 2 t 4 1 4 t 3 + 2 e t .

A general solution of the nonhomogeneous equation is then given by

y = y h + y p = c 1 + c 2 t + c 3 t 2 + c 4 cos 2 t + c 5 sin 2 t + 1 2 t 4 1 4 t 3 + 2 e t .

What is the form of a particular solution of y ( 5 ) + 4 y = t cos 2 t ?

To solve an IVP, first determine a general solution and then use the initial conditions to solve for the unknown constants in the general solution.

Example 4.6.3

Solve the IVP

4 y + 4 y + 6 5 y = e t 3 2 8 6 3 cos 2 t + 5 7 7 2 7 sin 2 t ,

y(0) = 4/3, y′(0) = −5/2, y″(0) = −173/12.

Solution

The corresponding homogeneous equation is 4y‴ + 4y″ + 65y′ = 0 with characteristic equation 4r 3 + 4r 2 + 65r = r(4r 2 + 4r + 65) = 0 so r 1 = 0 and using the quadratic formula to solve 4r 2 + 4r + 65 = 0 gives us r 2 , 3 = 1 2 ± 4 i . Thus, a fundamental set of solutions for the corresponding homogeneous equation is S = { 1 , e t 2 cos 4 t , e t 2 sin 4 t } and a general solution of the corresponding homogeneous equation is y h = c 1 + e t 2 ( c 2 cos 4 t + c 3 sin 4 t ) .

The associated set of functions for the forcing function

g ( t ) = e t 3 2 8 6 3 cos 2 t + 5 7 7 2 7 sin 2 t

is F = e t 3 cos 2 t , e t 3 sin 2 t . No function in F is a solution of the corresponding homogeneous equation so we assume that a particular solution of the nonhomogeneous equation has the form y p = A e t 3 cos 2 t + B e t 3 sin 2 t , where a and b are constants to be determined. Differentiating y p three times gives us

y p = 1 3 A + 2 B e t 3 cos 2 t + 2 A 1 3 B e t 3 sin 2 t , y p = 3 5 9 A 4 3 B e t 3 cos 2 t + 4 3 A 3 5 9 e t 3 sin 2 t , and y p = 1 0 7 2 7 A 2 2 3 B e t 3 cos 2 t + 2 2 3 A + 1 0 7 2 7 B e t 3 sin 2 t .

Substituting y p and its derivatives into the nonhomogeneous equation and simplifying the result gives us

When algebra becomes unusually cumbersome, we usually use a computer algebra system to assist in checking our calculations.

4 y p + 4 y p + 6 5 y p = 5 7 7 2 7 A + 2 8 6 3 B e t 3 cos 2 t + 2 8 6 3 A 5 7 7 2 7 B e t 3 sin 2 t = e t 3 2 8 6 3 cos 2 t + 5 7 7 2 7 sin 2 t .

Equating coefficients gives us the system

5 7 7 2 7 A + 2 8 6 3 B = 2 8 6 3 2 8 6 3 A 5 7 7 2 7 B = 5 7 7 2 7 ,

which has solution A = 0 and B = −1. Thus, y p = e t 3 sin 2 t and a general solution of the nonhomogeneous equation is y = y h + y p = c 1 + e t 2 ( c 2 cos 4 t + c 3 sin 4 t e t 3 sin 2 t (see Figure 4.11(a)).

Figure 4.11. (a) Various solutions of the nonhomogeneous equation. (b) The solution of the nonhomogeneous equation that satisfies the initial conditions y(0) = 4/3, y (0) = −5/2, and y (0) = −173/12.

To solve the IVP, we first differentiate Y twice resulting in

y = 1 2 c 2 + 4 c 3 e t 2 cos 4 t + 4 c 2 1 2 c 3 e t 2 sin 4 t 2 e t 3 cos 2 t + 1 3 e t 3 sin 2 t and y = 6 3 4 c 2 + 4 c 3 e t 2 cos 4 t + 4 c 2 6 3 4 c 3 e t 2 sin 4 t + 4 3 e t 3 cos 2 t + 3 5 9 e t 3 sin 2 t .

Evaluating at t = 0 gives us the system of equations

y ( 0 ) = c 1 + c 2 = 4 3 y ( 0 ) = 2 1 2 c 2 + 4 c 3 = 5 2 y ( 0 ) = 2 2 3 + 1 9 1 8 c 2 6 1 c 3 = 1 7 3 1 2 ,

which has solution c 1 = 1/3, c 2 = 1, and c 3 = 0 so the solution to the IVP is y = 1 3 + e t 2 cos 4 t e t 3 sin 2 t (see Figure 4.11(b)).
What is the form of a particular solution of 4y‴ + 4y″ + 65y′ = t2 + te-t/3 cos2t?

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780124172197000041