1 Limits and Continuity of Multivariate Functions
1.1 Examples of Multivariate Functions
Multivariate Function
An m m m -variable function is a mapping f : D → R f : D \to \mathbb{R} f : D → R , where D D D is R m \mathbb{R}^m R m or a subset of it, and the function value y y y and independent variables x 1 , x 2 , … , x m x_1, x_2, \dots, x_m x 1 , x 2 , … , x m satisfy y = f ( x 1 , x 2 , … , x m ) y = f(x_1, x_2, \dots, x_m) y = f ( x 1 , x 2 , … , x m ) .
Multivariate Mapping
The values of multiple variables y 1 , … , y n y_1, \dots, y_n y 1 , … , y n are determined by a set of independent variables x 1 , … , x m x_1, \dots, x_m x 1 , … , x m :
( y 1 ⋮ y n ) = ( f 1 ( x 1 , … , x m ) ⋮ f n ( x 1 , … , x m ) ) , i.e., { y 1 = f 1 ( x 1 , … , x m ) , ⋮ y n = f n ( x 1 , … , x m ) . \begin{equation*}
\begin{pmatrix}
y_1 \\
\vdots \\
y_n
\end{pmatrix}
=
\begin{pmatrix}
f_1(x_1, \dots, x_m) \\
\vdots \\
f_n(x_1, \dots, x_m)
\end{pmatrix},
\quad \text{i.e.,} \quad
\begin{cases}
y_1 = f_1(x_1, \dots, x_m), \\
\vdots \\
y_n = f_n(x_1, \dots, x_m).
\end{cases}
\end{equation*} y 1 ⋮ y n = f 1 ( x 1 , … , x m ) ⋮ f n ( x 1 , … , x m ) , i.e., ⎩ ⎨ ⎧ y 1 = f 1 ( x 1 , … , x m ) , ⋮ y n = f n ( x 1 , … , x m ) .
1.2 Distance and Limits of Sequences in R m \mathbb{R}^m R m
Distance on R m \mathbb{R}^m R m
A bivariate function d : R m × R m → R d : \mathbb{R}^m \times \mathbb{R}^m \to \mathbb{R} d : R m × R m → R is called a distance on R m \mathbb{R}^m R m if it satisfies:
For any x , y ∈ R m \mathbf{x}, \mathbf{y} \in \mathbb{R}^m x , y ∈ R m , d ( x , y ) = d ( y , x ) ≥ 0 d(\mathbf{x}, \mathbf{y}) = d(\mathbf{y}, \mathbf{x}) \ge 0 d ( x , y ) = d ( y , x ) ≥ 0 ;
For any x , y ∈ R m \mathbf{x}, \mathbf{y} \in \mathbb{R}^m x , y ∈ R m , d ( x , y ) = 0 d(\mathbf{x}, \mathbf{y}) = 0 d ( x , y ) = 0 if and only if x = y \mathbf{x} = \mathbf{y} x = y ;
For any x , y , z ∈ R m \mathbf{x}, \mathbf{y}, \mathbf{z} \in \mathbb{R}^m x , y , z ∈ R m , d ( x , z ) ≤ d ( x , y ) + d ( y , z ) d(\mathbf{x}, \mathbf{z}) \le d(\mathbf{x}, \mathbf{y}) + d(\mathbf{y}, \mathbf{z}) d ( x , z ) ≤ d ( x , y ) + d ( y , z ) .
For r > 0 r > 0 r > 0 , denote
B ( x , r ) : = { y ∈ R m ∣ d ( x , y ) < r } , \begin{equation*}
B(\mathbf{x}, r) := \{\mathbf{y} \in \mathbb{R}^m \mid d(\mathbf{x}, \mathbf{y}) < r\},
\end{equation*} B ( x , r ) := { y ∈ R m ∣ d ( x , y ) < r } ,
which is called the open ball centered at x \mathbf{x} x with radius r r r .
A distance d d d on R m \mathbb{R}^m R m is said to be translation-invariant if for any x , y , z ∈ R m \mathbf{x}, \mathbf{y}, \mathbf{z} \in \mathbb{R}^m x , y , z ∈ R m ,
d ( x + z , y + z ) = d ( x , y ) . \begin{equation*}
d(\mathbf{x} + \mathbf{z}, \mathbf{y} + \mathbf{z}) = d(\mathbf{x}, \mathbf{y}).
\end{equation*} d ( x + z , y + z ) = d ( x , y ) .
Norm on R m \mathbb{R}^m R m
A function ∥ ⋅ ∥ : R m → R \|\cdot\| : \mathbb{R}^m \to \mathbb{R} ∥ ⋅ ∥ : R m → R is called a norm on R m \mathbb{R}^m R m if it satisfies:
For any x , y ∈ R m \mathbf{x}, \mathbf{y} \in \mathbb{R}^m x , y ∈ R m , ∥ x + y ∥ ≤ ∥ x ∥ + ∥ y ∥ \|\mathbf{x} + \mathbf{y}\| \le \|\mathbf{x}\| + \|\mathbf{y}\| ∥ x + y ∥ ≤ ∥ x ∥ + ∥ y ∥ ;
For any x ∈ R m \mathbf{x} \in \mathbb{R}^m x ∈ R m , ∥ x ∥ ≥ 0 \|\mathbf{x}\| \ge 0 ∥ x ∥ ≥ 0 ; ∥ x ∥ = 0 \|\mathbf{x}\| = 0 ∥ x ∥ = 0 if and only if x = 0 \mathbf{x} = \mathbf{0} x = 0 ;
For any x ∈ R m \mathbf{x} \in \mathbb{R}^m x ∈ R m and any λ ∈ R \lambda \in \mathbb{R} λ ∈ R , ∥ λ x ∥ = ∣ λ ∣ ∥ x ∥ \|\lambda \mathbf{x}\| = |\lambda| \|\mathbf{x}\| ∥ λ x ∥ = ∣ λ ∣∥ x ∥ .
Matrix Norm
Given a norm ∥ ⋅ ∥ \|\cdot\| ∥ ⋅ ∥ in R n \mathbb{R}^n R n and a square matrix A A A of order n n n , the matrix norm of A A A is defined as:
∥ A ∥ = max ∥ v ∥ = 1 ∥ A v ∥ . \begin{align*}
\|A\| = \max_{\|\mathbf{v}\|=1} \|A\mathbf{v}\|.
\end{align*} ∥ A ∥ = ∥ v ∥ = 1 max ∥ A v ∥.
Also
∥ A ∥ = max v ≠ 0 ∥ A v ∥ ∥ v ∥ . \begin{align*}
\|A\| = \max_{\mathbf{v}\neq 0} \frac{\|A\mathbf{v}\|}{\|\mathbf{v}\|}.
\end{align*} ∥ A ∥ = v = 0 max ∥ v ∥ ∥ A v ∥ .
Theorem 1.1.
For any matrix A A A , the norm is always non-negative, and it is zero if and only if the matrix itself is the zero matrix.
∥ A ∥ ≥ 0 , and ∥ A ∥ = 0 ⟺ A = 0 \begin{align*}
\|A\| \geq 0, \quad \text{and} \quad \|A\| = 0 \iff A = \mathbf{0}
\end{align*} ∥ A ∥ ≥ 0 , and ∥ A ∥ = 0 ⟺ A = 0
Multiplying a matrix by a scalar α \alpha α scales the norm by the absolute value of that scalar.
∥ α A ∥ = ∣ α ∣ ⋅ ∥ A ∥ \begin{align*}
\|\alpha A\| = |\alpha| \cdot \|A\|
\end{align*} ∥ α A ∥ = ∣ α ∣ ⋅ ∥ A ∥
The norm of the sum of two matrices is less than or equal to the sum of their individual norms.
∥ A + B ∥ ≤ ∥ A ∥ + ∥ B ∥ \begin{align*}
\|A + B\| \leq \|A\| + \|B\|
\end{align*} ∥ A + B ∥ ≤ ∥ A ∥ + ∥ B ∥
For induced norms, the norm of a product of matrices is bounded by the product of their norms.
∥ A B ∥ ≤ ∥ A ∥ ⋅ ∥ B ∥ \begin{align*}
\|AB\| \leq \|A\| \cdot \|B\|
\end{align*} ∥ A B ∥ ≤ ∥ A ∥ ⋅ ∥ B ∥
By the very definition of the induced norm, the transformation of any vector v \mathbf{v} v by matrix A A A is bounded by the matrix norm:
∥ A v ∥ ≤ ∥ A ∥ ⋅ ∥ v ∥ \begin{align*}
\|A\mathbf{v}\| \leq \|A\| \cdot \|\mathbf{v}\|
\end{align*} ∥ A v ∥ ≤ ∥ A ∥ ⋅ ∥ v ∥
Inner Product on R m \mathbb{R}^m R m
A bivariate function ⟨ ⋅ , ⋅ ⟩ : R m × R m → R \langle \cdot, \cdot \rangle : \mathbb{R}^m \times \mathbb{R}^m \to \mathbb{R} ⟨ ⋅ , ⋅ ⟩ : R m × R m → R is called an inner product on R m \mathbb{R}^m R m if it satisfies:
For any x , y ∈ R m \mathbf{x}, \mathbf{y} \in \mathbb{R}^m x , y ∈ R m , ⟨ x , y ⟩ = ⟨ y , x ⟩ \langle \mathbf{x}, \mathbf{y} \rangle = \langle \mathbf{y}, \mathbf{x} \rangle ⟨ x , y ⟩ = ⟨ y , x ⟩ ;
For any x , y , z ∈ R m \mathbf{x}, \mathbf{y}, \mathbf{z} \in \mathbb{R}^m x , y , z ∈ R m and any α , β ∈ R \alpha, \beta \in \mathbb{R} α , β ∈ R ,
⟨ α x + β y , z ⟩ = α ⟨ x , z ⟩ + β ⟨ y , z ⟩ ; \begin{equation*}
\langle \alpha \mathbf{x} + \beta \mathbf{y}, \mathbf{z} \rangle = \alpha \langle \mathbf{x}, \mathbf{z} \rangle + \beta \langle \mathbf{y}, \mathbf{z} \rangle;
\end{equation*} ⟨ α x + β y , z ⟩ = α ⟨ x , z ⟩ + β ⟨ y , z ⟩ ;
For any x ∈ R m \mathbf{x} \in \mathbb{R}^m x ∈ R m , ⟨ x , x ⟩ ≥ 0 \langle \mathbf{x}, \mathbf{x} \rangle \ge 0 ⟨ x , x ⟩ ≥ 0 ; and ⟨ x , x ⟩ = 0 \langle \mathbf{x}, \mathbf{x} \rangle = 0 ⟨ x , x ⟩ = 0 if and only if x = 0 \mathbf{x} = \mathbf{0} x = 0 .
Definition. For any p ≥ 1 p \ge 1 p ≥ 1 ,
∥ x ∥ p : = ( ∣ x 1 ∣ p + ⋯ + ∣ x m ∣ p ) 1 p \begin{equation*}
\|\mathbf{x}\|_p := (|x^1|^p + \dots + |x^m|^p)^{\frac{1}{p}}
\end{equation*} ∥ x ∥ p := ( ∣ x 1 ∣ p + ⋯ + ∣ x m ∣ p ) p 1
defines a norm on R m \mathbb{R}^m R m ,which induces a distance on R m \mathbb{R}^m R m :
d p ( x , y ) : = ( ∣ x 1 − y 1 ∣ p + ⋯ + ∣ x m − y m ∣ p ) 1 p . \begin{equation*}
d_p(\mathbf{x}, \mathbf{y}) := (|x^1 - y^1|^p + \dots + |x^m - y^m|^p)^{\frac{1}{p}}.
\end{equation*} d p ( x , y ) := ( ∣ x 1 − y 1 ∣ p + ⋯ + ∣ x m − y m ∣ p ) p 1 .
Theorem 1.2.
∥ x ∥ ∞ : = max { ∣ x 1 ∣ , … , ∣ x m ∣ } \|\mathbf{x}\|_\infty := \max\{|x^1|, \dots, |x^m|\} ∥ x ∥ ∞ := max { ∣ x 1 ∣ , … , ∣ x m ∣ } .
∥ x ∥ ∞ ≤ ∥ x ∥ p ≤ m 1 p ∥ x ∥ ∞ \|\mathbf{x}\|_\infty \le \|\mathbf{x}\|_p \le m^{\frac{1}{p}} \|\mathbf{x}\|_\infty ∥ x ∥ ∞ ≤ ∥ x ∥ p ≤ m p 1 ∥ x ∥ ∞ .
1 m 1 q ∥ x ∥ q ≤ ∥ x ∥ p ≤ m 1 p ∥ x ∥ q \frac{1}{m^{\frac{1}{q}}} \|\mathbf{x}\|_q \le \|\mathbf{x}\|_p \le m^{\frac{1}{p}} \|\mathbf{x}\|_q m q 1 1 ∥ x ∥ q ≤ ∥ x ∥ p ≤ m p 1 ∥ x ∥ q .
Corollary 1. Any norm ∥ ⋅ ∥ \|\cdot\| ∥ ⋅ ∥ on R m \mathbb{R}^m R m is equivalent to the norm ∥ ⋅ ∥ ∞ \|\cdot\|_\infty ∥ ⋅ ∥ ∞ ,
Limits of Sequences in R m \mathbb{R}^m R m
Let { x n } n ≥ 1 \{\mathbf{x}_n\}_{n \ge 1} { x n } n ≥ 1 be a sequence in R m \mathbb{R}^m R m .
The sequence { x n } n ≥ 1 \{\mathbf{x}_n\}_{n \ge 1} { x n } n ≥ 1 is said to be bounded if there exists M > 0 M > 0 M > 0 such that for any positive integer n n n , ∥ x n ∥ ≤ M \|\mathbf{x}_n\| \le M ∥ x n ∥ ≤ M .
The sequence { x n } n ≥ 1 \{\mathbf{x}_n\}_{n \ge 1} { x n } n ≥ 1 is said to be convergent if there exists A ∈ R m A \in \mathbb{R}^m A ∈ R m such that for any ε > 0 \varepsilon > 0 ε > 0 , there exists a positive integer N ( ε ) > 0 N(\varepsilon) > 0 N ( ε ) > 0 such that for any positive integer n n n , whenever n ≥ N ( ε ) n \ge N(\varepsilon) n ≥ N ( ε ) , we have ∥ x n − A ∥ < ε \|\mathbf{x}_n - A\| < \varepsilon ∥ x n − A ∥ < ε . Such an A A A is called a limit of the sequence { x n } n ≥ 1 \{\mathbf{x}_n\}_{n \ge 1} { x n } n ≥ 1 . We denote A = lim n → + ∞ x n A = \lim\limits_{n \to +\infty} \mathbf{x}_n A = n → + ∞ lim x n .
The sequence { x n } n ≥ 1 \{\mathbf{x}_n\}_{n \ge 1} { x n } n ≥ 1 is said to be a Cauchy sequence if for any ε > 0 \varepsilon > 0 ε > 0 , there exists a positive integer N ( ε ) > 0 N(\varepsilon) > 0 N ( ε ) > 0 such that for any positive integers p , q p, q p , q , whenever p , q ≥ N ( ε ) p, q \ge N(\varepsilon) p , q ≥ N ( ε ) , we have ∥ x p − x q ∥ < ε \|\mathbf{x}_p - \mathbf{x}_q\| < \varepsilon ∥ x p − x q ∥ < ε .
Theorem 1.3. Let { x n } \{\mathbf{x}_n\} { x n } be a sequence in R m \mathbb{R}^m R m , where x n = ( x n 1 , … , x n m ) T \mathbf{x}_n = (x_n^1, \dots, x_n^m)^T x n = ( x n 1 , … , x n m ) T . Then:
{ x n } n ≥ 1 \{\mathbf{x}_n\}_{n \ge 1} { x n } n ≥ 1 is a bounded sequence if and only if for any 1 ≤ k ≤ m 1 \le k \le m 1 ≤ k ≤ m , the sequence { x n k } n ≥ 1 \{x_n^k\}_{n \ge 1} { x n k } n ≥ 1 is a bounded sequence.
{ x n } n ≥ 1 \{\mathbf{x}_n\}_{n \ge 1} { x n } n ≥ 1 is a Cauchy sequence if and only if for any 1 ≤ k ≤ m 1 \le k \le m 1 ≤ k ≤ m , the sequence { x n k } n ≥ 1 \{x_n^k\}_{n \ge 1} { x n k } n ≥ 1 is a Cauchy sequence.
lim n → + ∞ x n = A \lim\limits_{n \to +\infty} \mathbf{x}_n = A n → + ∞ lim x n = A if and only if for any 1 ≤ k ≤ m 1 \le k \le m 1 ≤ k ≤ m , lim n → + ∞ x n k = A k \lim\limits_{n \to +\infty} x_n^k = A_k n → + ∞ lim x n k = A k .
Theorem 1.4.
A sequence in R m \mathbb{R}^m R m is convergent if and only if it is a Cauchy sequence.
Any bounded sequence in R m \mathbb{R}^m R m has a convergent subsequence.
Closed Set
A set E ⊆ R m E \subseteq \mathbb{R}^m E ⊆ R m is said to be a closed set if: for any sequence { x n } n ≥ 1 \{\mathbf{x}_n\}_{n \ge 1} { x n } n ≥ 1 in E E E , if the sequence has a limit lim n → + ∞ x n \lim\limits_{n \to +\infty} \mathbf{x}_n n → + ∞ lim x n , then this limit is also in E E E .
1.3 Continuity of Multivariable Functions
Continuity of Multivariable Function
Let E ⊆ R m E \subseteq \mathbb{R}^m E ⊆ R m . A mapping f : E → R p f: E \to \mathbb{R}^p f : E → R p is said to be continuous at x 0 ∈ E \mathbf{x}_0 \in E x 0 ∈ E , if for any ϵ > 0 \epsilon > 0 ϵ > 0 , there exists a δ > 0 \delta > 0 δ > 0 such that for any x ∈ E \mathbf{x} \in E x ∈ E , as long as
∥ x − x 0 ∥ < δ \begin{align*}
\|\mathbf{x} - \mathbf{x}_0\| < \delta
\end{align*} ∥ x − x 0 ∥ < δ
the following inequality holds:
∥ f ( x ) − f ( x 0 ) ∥ ∗ < ϵ \begin{align*}
\|f(\mathbf{x}) - f(\mathbf{x}_0)\|_* < \epsilon
\end{align*} ∥ f ( x ) − f ( x 0 ) ∥ ∗ < ϵ
Here, ∥ ⋅ ∥ \|\cdot\| ∥ ⋅ ∥ and ∥ ⋅ ∥ ∗ \|\cdot\|_* ∥ ⋅ ∥ ∗ denote the norms in the domain space and the codomain space of f f f , respectively.
f f f is said to be continuous on E 0 ⊆ E E_0 \subseteq E E 0 ⊆ E , if f f f is continuous at every point x 0 ∈ E 0 \mathbf{x}_0 \in E_0 x 0 ∈ E 0 .
f f f is called a continuous mapping , if it is continuous on its entire domain E E E .
Theorem 1.5. Any linear mapping on R m \mathbb{R}^m R m is continuous.
Theorem 1.6. Any bilinear mapping B : R m × R n → R p B: \mathbb{R}^m \times \mathbb{R}^n \to \mathbb{R}^p B : R m × R n → R p , i.e., ∀ α 1 , α 2 , β 1 , β 2 ∈ R , x 1 , x 2 , x ∈ R m , y 1 , y 2 , y ∈ R n \forall \alpha_1, \alpha_2, \beta_1, \beta_2 \in \mathbb{R}, \mathbf{x}_1, \mathbf{x}_2, \mathbf{x} \in \mathbb{R}^m, \mathbf{y}_1, \mathbf{y}_2, \mathbf{y} \in \mathbb{R}^n ∀ α 1 , α 2 , β 1 , β 2 ∈ R , x 1 , x 2 , x ∈ R m , y 1 , y 2 , y ∈ R n
B ( α 1 x 1 + α 2 x 2 , y ) = α 1 B ( x 1 , y ) + α 2 B ( x 2 , y ) , B ( x , β 1 y 1 + β 2 y 2 ) = β 1 B ( x , y 1 ) + β 2 B ( x , y 2 ) , \begin{align*}
B(\alpha_1 \mathbf{x}_1 + \alpha_2 \mathbf{x}_2, \mathbf{y}) &= \alpha_1 B(\mathbf{x}_1, \mathbf{y}) + \alpha_2 B(\mathbf{x}_2, \mathbf{y}), \\
B(\mathbf{x}, \beta_1 \mathbf{y}_1 + \beta_2 \mathbf{y}_2) &= \beta_1 B(\mathbf{x}, \mathbf{y}_1) + \beta_2 B(\mathbf{x}, \mathbf{y}_2),
\end{align*} B ( α 1 x 1 + α 2 x 2 , y ) B ( x , β 1 y 1 + β 2 y 2 ) = α 1 B ( x 1 , y ) + α 2 B ( x 2 , y ) , = β 1 B ( x , y 1 ) + β 2 B ( x , y 2 ) ,
is continuous.
Corollary 1. Any multilinear mapping is a continuous mapping.
Theorem 1.7. Vector addition, scalar multiplication, and the inner product in R m \mathbb{R}^m R m , as well as real number multiplication, are all continuous mappings.
Theorem 1.8. Let E ⊆ R m E \subseteq \mathbb{R}^m E ⊆ R m , F ⊆ R p F \subseteq \mathbb{R}^p F ⊆ R p , and f : E → R p f: E \to \mathbb{R}^p f : E → R p be continuous at x 0 ∈ E \mathbf{x}_0 \in E x 0 ∈ E . If g : F → R q g: F \to \mathbb{R}^q g : F → R q is continuous at y 0 = f ( x 0 ) ∈ F \mathbf{y}_0 = f(\mathbf{x}_0) \in F y 0 = f ( x 0 ) ∈ F , then the composite mapping
g ∘ f : E ∩ f − 1 ( F ) → R q \begin{align*}
g \circ f: E \cap f^{-1}(F) \to \mathbb{R}^q
\end{align*} g ∘ f : E ∩ f − 1 ( F ) → R q
is continuous at x 0 \mathbf{x}_0 x 0 .
Theorem 1.9. The determinant is a continuous function of the matrix; therefore, small perturbations of an invertible matrix remain invertible.
Theorem 1.10. Let E ⊆ R m E \subseteq \mathbb{R}^m E ⊆ R m and f : E → R p f: E \to \mathbb{R}^p f : E → R p be a continuous mapping. Then:
For any bounded closed subset K K K of E E E , the image f ( K ) f(K) f ( K ) is also a bounded closed set.
If p = 1 p = 1 p = 1 , then for any non-empty bounded closed subset K K K of E E E , f f f attains both a maximum value and a minimum value on K K K .
Theorem 1.11. Let ⟨ ⋅ , ⋅ ⟩ \langle \cdot, \cdot \rangle ⟨ ⋅ , ⋅ ⟩ be an inner product on R n \mathbb{R}^n R n , and let A A A be a symmetric matrix of order n n n , i.e.,
⟨ A x , y ⟩ = ⟨ x , A y ⟩ , ∀ x , y ∈ R n . \begin{align*}
\langle A\mathbf{x}, \mathbf{y} \rangle = \langle \mathbf{x}, A\mathbf{y} \rangle, \quad \forall \mathbf{x}, \mathbf{y} \in \mathbb{R}^n.
\end{align*} ⟨ A x , y ⟩ = ⟨ x , A y ⟩ , ∀ x , y ∈ R n .
There exist n n n orthonormal unit vectors v 1 , … , v n \mathbf{v}_1, \dots, \mathbf{v}_n v 1 , … , v n in R n \mathbb{R}^n R n and real numbers λ 1 , … , λ n \lambda_1, \dots, \lambda_n λ 1 , … , λ n such that
A v k = λ k v k , k = 1 , … , n . \begin{align*}
A\mathbf{v}_k = \lambda_k \mathbf{v}_k, \quad k = 1, \dots, n.
\end{align*} A v k = λ k v k , k = 1 , … , n .
Theorem 1.12. Let E ⊆ R m E \subseteq \mathbb{R}^m E ⊆ R m and f : E → R p f: E \to \mathbb{R}^p f : E → R p be a continuous mapping. For any bounded closed subset K K K of E E E , f f f is uniformly continuous on K K K .
This means that for any ϵ > 0 \epsilon > 0 ϵ > 0 , there exists a δ > 0 \delta > 0 δ > 0 such that for any x ∈ K \mathbf{x} \in K x ∈ K and any y ∈ E \mathbf{y} \in E y ∈ E , as long as
∥ x − y ∥ < δ \begin{align*}
\|\mathbf{x} - \mathbf{y}\| < \delta
\end{align*} ∥ x − y ∥ < δ
the following holds:
∥ f ( x ) − f ( y ) ∥ ∗ < ϵ \begin{align*}
\|f(\mathbf{x}) - f(\mathbf{y})\|_* < \epsilon
\end{align*} ∥ f ( x ) − f ( y ) ∥ ∗ < ϵ
Here, ∥ ⋅ ∥ \|\cdot\| ∥ ⋅ ∥ and ∥ ⋅ ∥ ∗ \|\cdot\|_* ∥ ⋅ ∥ ∗ denote the norms in the domain space and the codomain space of f f f , respectively.
Path
A set E ⊆ R m E \subseteq \mathbb{R}^m E ⊆ R m is called a path-connected set if, for any x , y ∈ E \mathbf{x}, \mathbf{y} \in E x , y ∈ E , there exists a continuous mapping f : [ 0 , 1 ] → R m f : [0, 1] \to \mathbb{R}^m f : [ 0 , 1 ] → R m such that
f ( 0 ) = x , f ( 1 ) = y , and f ( t ) ∈ E for all t ∈ [ 0 , 1 ] . \begin{align*}
f(0) = \mathbf{x}, \quad f(1) = \mathbf{y}, \quad \text{and } f(t) \in E \text{ for all } t \in [0, 1].
\end{align*} f ( 0 ) = x , f ( 1 ) = y , and f ( t ) ∈ E for all t ∈ [ 0 , 1 ] .
Such a continuous mapping f f f is called a path from x \mathbf{x} x to y \mathbf{y} y in E E E .
1.4 Limits of Multivariable Functions
Limits of Multivariable Function
Let E ⊆ R m E \subseteq \mathbb{R}^m E ⊆ R m . A point x 0 ∈ R m \mathbf{x}_0 \in \mathbb{R}^m x 0 ∈ R m is called a cluster point (or accumulation point) of E E E if, for any ϵ > 0 \epsilon > 0 ϵ > 0 , there exists x ∈ E \mathbf{x} \in E x ∈ E such that
0 < d ( x , x 0 ) < ϵ \begin{align*}
0 < d(\mathbf{x}, \mathbf{x}_0) < \epsilon
\end{align*} 0 < d ( x , x 0 ) < ϵ
If x 0 \mathbf{x}_0 x 0 is both a cluster point of E ⊆ R m E \subseteq \mathbb{R}^m E ⊆ R m and a cluster point of its complement E c = R m ∖ E E^c = \mathbb{R}^m \setminus E E c = R m ∖ E (meaning every neighborhood of x 0 \mathbf{x}_0 x 0 contains points from both E E E and R m ∖ E \mathbb{R}^m \setminus E R m ∖ E ), then x 0 \mathbf{x}_0 x 0 is called a boundary point of E E E . The set of all boundary points of E E E is denoted as ∂ E \partial E ∂ E and is called the boundary set of E E E .
Let E E E be a subset of the domain of mapping f f f , and let x 0 ∈ R m \mathbf{x}_0 \in \mathbb{R}^m x 0 ∈ R m be a cluster point of E E E . We say that A ∈ R p \mathbf{A} \in \mathbb{R}^p A ∈ R p is the limit of f ( x ) f(\mathbf{x}) f ( x ) as x ∈ E \mathbf{x} \in E x ∈ E approaches x 0 \mathbf{x}_0 x 0 , denoted as
lim E ∋ x → x 0 f ( x ) = A \begin{align*}
\lim_{E \ni \mathbf{x} \to \mathbf{x}_0} f(\mathbf{x}) = \mathbf{A}
\end{align*} E ∋ x → x 0 lim f ( x ) = A
if for any ϵ > 0 \epsilon > 0 ϵ > 0 , there exists a δ > 0 \delta > 0 δ > 0 such that for any x ∈ E \mathbf{x} \in E x ∈ E , as long as 0 < d ( x , x 0 ) < δ 0 < d(\mathbf{x}, \mathbf{x}_0) < \delta 0 < d ( x , x 0 ) < δ , we have
d 2 ( f ( x ) , A ) < ϵ \begin{align*}
d_2(f(\mathbf{x}), \mathbf{A}) < \epsilon
\end{align*} d 2 ( f ( x ) , A ) < ϵ
Here, d 2 d_2 d 2 is the distance in the codomain space of f f f .
When E E E is the entire domain of mapping f f f , lim E ∋ x → x 0 f ( x ) \lim\limits_{E \ni \mathbf{x} \to \mathbf{x}_0} f(\mathbf{x}) E ∋ x → x 0 lim f ( x ) is simply written as lim x → x 0 f ( x ) \lim\limits_{\mathbf{x} \to \mathbf{x}_0} f(\mathbf{x}) x → x 0 lim f ( x ) .
Theorem 1.13. Let E E E be the domain of the composite function g ∘ f g \circ f g ∘ f , with lim E ∋ x → x 0 f ( x ) = y 0 \lim\limits_{E \ni \mathbf{x} \to \mathbf{x}_0} f(\mathbf{x}) = \mathbf{y}_0 E ∋ x → x 0 lim f ( x ) = y 0 and lim y → y 0 g ( y ) = A \lim\limits_{\mathbf{y} \to \mathbf{y}_0} g(\mathbf{y}) = A y → y 0 lim g ( y ) = A . Define
g ~ ( y ) = { g ( y ) , y ≠ y 0 ; A , y = y 0 . \begin{align*}
\tilde{g}(\mathbf{y}) = \begin{cases} g(\mathbf{y}), & \mathbf{y} \neq \mathbf{y}_0; \\ A, & \mathbf{y} = \mathbf{y}_0. \end{cases}
\end{align*} g ~ ( y ) = { g ( y ) , A , y = y 0 ; y = y 0 .
Then lim x → x 0 g ~ ( f ( x ) ) = A \lim\limits_{\mathbf{x} \to \mathbf{x}_0} \tilde{g}(f(\mathbf{x})) = A x → x 0 lim g ~ ( f ( x )) = A . In particular, if g g g is continuous at y 0 \mathbf{y}_0 y 0 , then
lim x → x 0 g ( f ( x ) ) = A . \begin{align*}
\lim_{\mathbf{x} \to \mathbf{x}_0} g(f(\mathbf{x})) = A.
\end{align*} x → x 0 lim g ( f ( x )) = A .