Geometric Analysis of Curves

While in Chapter 4 curves played the secondary role of illustrating vector-valued functions and their derivatives, in this Chapter, they will become the central object of our analysis. However, we will save a full analytical description of curves for a future book that will rely on the techniques developed in our present narrative. Instead, our primary motivation is to continue exploring geometric concepts without the use of coordinate systems, since in order to fully understand the power of coordinates, one must appreciate what is possible in their absence.
That said, our present analysis will represent somewhat of a compromise with respect to the use of coordinates. The ambient space will remain coordinate-free and we will therefore be working directly with geometric vectors rather than their components. Meanwhile, the points on the curve will be enumerated by a parameter γ\gamma, as was the case in the previous Chapter for the curve associated a vector-valued function R(γ)\mathbf{R}\left( \gamma\right) . There is no denying that the parameter γ\gamma amounts to a coordinate system on the curve. Thus, while our analysis will remain mostly geometric and will, as a result, provide great geometric clarity and insight, we will also begin to see some of the benefits of coordinate systems.
At the same time, we will observe that the analytical methods developed in this Chapter are insufficient for practical calculations, such as determining the curvature or the torsion of a specific curve. Most practical calculations can only be reasonably accomplished when all geometric objects are converted either into analytical expressions, which can be analyzed by means of Algebra and Calculus, or into numerical data which can be analyzed by computational methods. This is the role of coordinate systems.
In all of the following discussions, unless we are specifically discussing a planar curve, you should imagine a curve in a three-dimensional space, such as the "spiraling funnel" in the following figure.
(5.1)
In the previous Chapter, we learned that a curve is characterized by its unit tangent vector T\mathbf{T}. We would now like to add that a curve in a three-dimensional space is also characterized by a plane orthogonal to T\mathbf{T} known as the normal plane.
(5.2)
Thus, in three dimensions, there does not exist a unique orthogonal direction, but rather a two-dimensional orthogonal space. Below, we will discover a natural basis for this space.
In Chapter 3, we discussed the possibility of choosing a signed arc length parameterization ss on a curve. Denote by R(s)\mathbf{R}\left( s\right) the corresponding vector equation of the curve. In selecting an arc length parameterization, we have two arbitrary choices to make: the location of the origin, i.e. the point where s=0s=0 (which is distinct from the origin OO), and the orientation of the parameterization, i.e. the direction in which ss increases.
(5.3)
An important intuitive characteristic of an arc-length parameterization is the regular spacing in the sense of the distance along the curve between points corresponding to coordinates separated by constant values. This is illustrated in the figure above. Note that while the figure shows a planar curve, arc-length parameterization works perfectly well for curves in a three-dimensional space and throughout the discussion, the reader should try to visualize three-dimensional configurations.
As we established in Chapter 4 for general vector equations of the curve, the derivative R(s)\mathbf{R}^{\prime}\left( s\right) is tangential to the curve at each point. However, in the special case of an arc length parameterization, R(s)\mathbf{R}^{\prime}\left( s\right) satisfies the additional property that it is unit length, i.e.
lenR(s)=1.(5.4)\operatorname{len}\mathbf{R}^{\prime}\left( s\right) =1.\tag{5.4}
As a result, the vector R(s)\mathbf{R}^{\prime}\left( s\right) is referred to as the unit tangent. We will denote it by the symbol T\mathbf{T}, i.e.
T=R(s).(5.5)\mathbf{T}=\mathbf{R}^{\prime}\left( s\right) .\tag{5.5}
To see why R(s)\mathbf{R}^{\prime}\left( s\right) is unit length, let us examine its construction according to the limiting procedure described in Chapter 4. Consider two nearby points on the curve corresponding to the values of the arc length ss and s+hs+h. For any value of hh, the distance Δs\Delta s along the curve between the two points is, by definition, hh. When we consider a small hh, and zoom in on the section of the curve between the two points, it will appear to be essentially straight. Therefore, the length of the vector
ΔR=R(s+h)R(s)(5.6)\Delta\mathbf{R}=\mathbf{R}\left( s+h\right) -\mathbf{R}\left( s\right)\tag{5.6}
will essentially equal hh, although, in actuality, it will be slightly shorter. The following figure this insight at s=1s=1 for h=0.4h=0.4, where the length of the vector ΔR\Delta\mathbf{R} is 0.39980.3998. We chose this value of hh because it is small enough to illustrate the near-equality of Δs\Delta s and the length of ΔR\Delta\mathbf{R}, yet still large enough for us to see that the curve is not actually straight.
(5.7)
Had we chosen h=0.04h=0.04, the corresponding section of the curve would be visually indistinguishable from ΔR\Delta\mathbf{R}, whose length would be approximately 0.0399980.039998, i.e. 100100 times closer to Δs\Delta s compared to h=0.4h=0.4.
Since for small hh, the length of the vector ΔR\Delta\mathbf{R} is essentially hh, the length of the vector
R(s+h)R(s)h(5.8)\frac{\mathbf{R}\left( s+h\right) -\mathbf{R}\left( s\right) }{h}\tag{5.8}
is essentially 11 and, in the limit as h0h\rightarrow0, it is 11, i.e.
len(limh0R(s+h)R(s)h)=1.(5.9)\operatorname{len}\left( \lim_{h\rightarrow0}\frac{\mathbf{R}\left( s+h\right) -\mathbf{R}\left( s\right) }{h}\right) =1.\tag{5.9}
In other words,
lenR(s)=1.(5.4)\operatorname{len}\mathbf{R}^{\prime}\left( s\right) =1. \tag{5.4}
as we set out to prove.
Thus, referring to R(s)\mathbf{R}^{\prime}\left( s\right) as the unit tangent is justified. Note, however, that at each point there are two unit tangents since the vector T-\mathbf{T} is also one. Nevertheless, having arbitrarily selected one unit tangent out of the two available options, we can refer to it as the unit tangent to indicate that the choice has already been made. Between these two legitimate unit tangents, the derivative R(s)\mathbf{R}^{\prime}\left( s\right) corresponds to the one that points in the direction of increasing arc-length ss. In the future, we will observe this interesting lack of uniqueness for other important geometric objects, including the unit normal N\mathbf{N} characterizing planar curves.
Let us now return to a curve parameterized by an arbitrary parameter γ\gamma. Our present goal is to demonstrate that the integral
γ0γ1R(γ)R(γ)dγ,(5.10)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma,\tag{5.10}
where γ0<γ1\gamma_{0} \lt \gamma_{1}, yields the arc length of the section of the curve between the points AA and BB corresponding to γ0\gamma_{0} and γ1\gamma_{1}.
(5.11)
Note that the integrand is simply the length of R(γ)\mathbf{R}^{\prime}\left( \gamma\right) . However we prefer the expression R(γ)R(γ)\sqrt{\mathbf{R}^{\prime }\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) } since it yields itself more readily to analytical manipulations.
The integral
γ0γ1R(γ)R(γ)dγ,(5.10)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma, \tag{5.10}
is our first example of a central concept in Tensor Calculus -- that of an invariant. It is an invariant because it yields the same value for any parameterization γ\gamma, provided that the limits of integration correspond to the section of the curve between the points AA and BB and that γ0<γ1\gamma_{0} \lt \gamma_{1}. The fact that the above integral is an invariant means that its value is a characteristic of the curve itself rather than that of a particular parameterization. And more often than not, invariants, and especially those given by simple expressions, have a clear geometric meaning, although that meaning may not always be apparent. In the particular case of the above integral, it is the length of the curve between AA and B.B.
Before we prove this, we would like to point out two special parameterizations for which it is intuitively so. First, if γ\gamma represents time and the curve, therefore, corresponds to the trajectory of a material particle, then the vector R(γ)\mathbf{R}^{\prime}\left( \gamma\right) represents the velocity of the particle, while its magnitude R(γ)R(γ)\sqrt {\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) } represents speed. Thus,
γ0γ1R(γ)R(γ)dγ,(5.10)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma, \tag{5.10}
is an integral of speed over the period of time from γ0\gamma_{0} to γ1\gamma_{1}, and thus represents the total distance traveled between those times. In other words, it is the length of the curve between the points AA and BB.
Second, suppose that γ\gamma corresponds to arc length ss, and let the points AA and BB correspond to the values s0s_{0} and s1s_{1}, where s0<s1s_{0} \lt s_{1}.
(5.12)
The presumed length-of-curve integral for this parameterization reads
s0s1R(s)R(s)ds.(5.13)\int_{s_{0}}^{s_{1}}\sqrt{\mathbf{R}^{\prime}\left( s\right) \cdot \mathbf{R}^{\prime}\left( s\right) }ds.\tag{5.13}
In the previous Section, we established that the length of R(s)\mathbf{R} ^{\prime}\left( s\right) is 11. Therefore,
R(s)R(s)=1(5.14)\mathbf{R}^{\prime}\left( s\right) \cdot\mathbf{R}^{\prime}\left( s\right) =1\tag{5.14}
and we have
s0s1R(s)R(s)ds=s0s11ds=s1s0.(5.15)\int_{s_{0}}^{s_{1}}\sqrt{\mathbf{R}^{\prime}\left( s\right) \cdot \mathbf{R}^{\prime}\left( s\right) }ds=\int_{s_{0}}^{s_{1}}1ds=s_{1}-s_{0}.\tag{5.15}
By definition, the difference s1s0s_{1}-s_{0} is precisely to the length of the curve between the points AA and BB, as we set out to show.
The arc length parameterization example can now serve as a starting point for the general demonstration. Having shown that the integral
γ0γ1R(γ)R(γ)dγ,(5.10)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma, \tag{5.10}
yields the length of the curve between the points AA and BB for one particular parameterization, we only need to prove that the above integral is independent of parameterization. In other words, we must show that if the curve were alternatively parameterized with the help of another variable ξ\xi, then
ξ0ξ1R(ξ)R(ξ)dξ=γ0γ1R(γ)R(γ)dγ,(5.16)\int_{\xi_{0}}^{\xi_{1}}\sqrt{\mathbf{R}^{\prime}\left( \xi\right) \cdot\mathbf{R}^{\prime}\left( \xi\right) }d\xi=\int_{\gamma_{0}} ^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot \mathbf{R}^{\prime}\left( \gamma\right) }d\gamma,\tag{5.16}
where the values ξ0\xi_{0} and ξ1\xi_{1} also correspond to the endpoints AA and BB, and ξ0<ξ1\xi_{0} \lt \xi_{1}. If the values of both variables γ\gamma and ξ\xi increase in the same direction along the curve, then ξ0\xi_{0} must correspond to AA and ξ1\xi_{1} to BB. Otherwise, ξ0\xi_{0} must correspond to BB and ξ1\xi_{1} to AA.
An important remark regarding our notation is in order. Even though the symbols R(γ)\mathbf{R}\left( \gamma\right) and R(ξ)\mathbf{R}\left( \xi\right) use the same letter R\mathbf{R}, they represent different, albeit related, functions. For example, the symbol R(1)\mathbf{R}\left( 1\right) may represents both R(γ)\mathbf{R}\left( \gamma\right) evaluated at γ=1\gamma=1 and R(ξ)\mathbf{R}\left( \xi\right) evaluated at ξ=1\xi=1. Naturally, those are generally completely different vectors. Nevertheless, the two functions R(γ)\mathbf{R}\left( \gamma\right) and R(ξ)\mathbf{R}\left( \xi\right) represent the same geometric quantity, i.e. the position vector R\mathbf{R}, and are therefore closely related enough to justify using the same letter. We can still easily distinguish the two functions by the letter representing the argument, and doing so is a common and necessary practice in Tensor Calculus that will be used extensively throughout our narrative.
Before we give the analytical argument that proves the identity
ξ0ξ1R(ξ)R(ξ)dξ=γ0γ1R(γ)R(γ)dγ,(5.16)\int_{\xi_{0}}^{\xi_{1}}\sqrt{\mathbf{R}^{\prime}\left( \xi\right) \cdot\mathbf{R}^{\prime}\left( \xi\right) }d\xi=\int_{\gamma_{0}} ^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot \mathbf{R}^{\prime}\left( \gamma\right) }d\gamma, \tag{5.16}
we would like to present a geometric argument that will give you an intuitive feel for it. The essence of the argument is that if the alternative parameterization R(ξ)\mathbf{R}\left( \xi\right) is such that the vector R(ξ)\mathbf{R}^{\prime}\left( \xi\right) is longer (or shorter) compared to R(γ)\mathbf{R}^{\prime}\left( \gamma\right) , then the interval of integration shrinks (or expands) correspondingly, leaving the value of the integral unchanged.
For an illustration, consider the two alternative parameterizations for the same curve in the following figure.
(5.17)
Notice that the relationship between γ\gamma and ξ\xi is
ξ=2γ.(5.18)\xi=2\gamma.\tag{5.18}
Whether we describe this change of variables as stretching or shrinking the parameterization, let us show that the old tangent vector R(γ)\mathbf{R}^{\prime}\left( \gamma\right) is twice the length of the new tangent vector R(ξ)\mathbf{R}^{\prime}\left( \xi\right) .
Consider the point on the curve that corresponds to γ=6\gamma=6 and ξ=12\xi=12. Next, identify the points that correspond to increasing each parameter by h=0.5h=0.5. The point that corresponds to γ+h=6.5\gamma+h=6.5 is located roughly half way between the points corresponding to γ=6\gamma=6 and γ=7\gamma=7. The point corresponding to ξ+h=12.5\xi+h=12.5 is located roughly only a quarter of the way between the same two points. Thus, for the same value of hh, the vector
R(γ+h)R(γ)h(5.19)\frac{\mathbf{R}\left( \gamma+h\right) -\mathbf{R}\left( \gamma\right) }{h}\tag{5.19}
is roughly twice as great as
R(ξ+h)R(ξ)h,(5.20)\frac{\mathbf{R}\left( \xi+h\right) -\mathbf{R}\left( \xi\right) }{h},\tag{5.20}
and, in the limit as hh approaches zero, we find
R(γ)=2R(ξ).(5.21)\mathbf{R}^{\prime}\left( \gamma\right) =2\mathbf{R}^{\prime}\left( \xi\right) .\tag{5.21}
Thus, the integrand in
γ0γ1R(γ)R(γ)dγ(5.22)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma\tag{5.22}
is twice the integrand in
ξ0ξ1R(ξ)R(ξ)dξ(5.23)\int_{\xi_{0}}^{\xi_{1}}\sqrt{\mathbf{R}^{\prime}\left( \xi\right) \cdot\mathbf{R}^{\prime}\left( \xi\right) }d\xi\tag{5.23}
for corresponding values of γ\gamma and ξ\xi. However, the interval of integration in the first integral is half that in the second integral for any given section of the curve. Thus, the two effects balance each other and, as a result, the two integrals yield the same value, as we set out to show.
Let us now turn to the formal proof of the fact that the integral
γ0γ1R(γ)R(γ)dγ,(5.10)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma, \tag{5.10}
is independent of the parameterization. Consider an alternative parameterization of the curve by the variable ξ\xi, where ξ0\xi_{0} corresponds to AA, ξ1\xi_{1} corresponds to BB. Furthermore, assume that ξ0<ξ1\xi_{0} \lt \xi_{1}, i.e. the two parameterizations have the same orientation, meaning that the values of γ\gamma and ξ\xi increase in the same direction along the curve. The proof for the case of opposite orientations is left as an exercise.
Suppose that γ\gamma and ξ\xi are related by the function ξ(γ)\xi\left( \gamma\right) , i.e.
ξ=ξ(γ),(5.24)\xi=\xi\left( \gamma\right) ,\tag{5.24}
where
ξ(γ0)=ξ0 and ξ(γ1)=ξ1.(5.25)\xi\left( \gamma_{0}\right) =\xi_{0}\text{ and }\xi\left( \gamma _{1}\right) =\xi_{1}.\tag{5.25}
For reasons that will become apparent shortly, it is essential to use the same letter ξ\xi to denote both the variable ξ\xi itself and the function that translates the values of γ\gamma to ξ\xi. Since the two parameterizations have the same orientation, ξ(γ)\xi\left( \gamma\right) is monotonically increasing and therefore has a positive derivative, i.e.
ξ(γ)>0.(5.26)\xi^{\prime}\left( \gamma\right) \gt 0.\tag{5.26}
The functions R(γ)\mathbf{R}\left( \gamma\right) and R(ξ)\mathbf{R}\left( \xi\right) are related by substituting the function ξ(γ)\xi\left( \gamma\right) into R(ξ)\mathbf{R}\left( \xi\right) , i.e.
R(γ)=R(ξ(γ)).(5.27)\mathbf{R}\left( \gamma\right) =\mathbf{R}\left( \xi\left( \gamma\right) \right) .\tag{5.27}
This identity makes it apparent why it was necessary to reuse the letter ξ\xi for the function relating the variables ξ\xi and γ\gamma. Had we denoted this function by a different symbol, say hh, then, in the resulting expression R(h(γ))\mathbf{R}\left( h\left( \gamma\right) \right) , it would be unclear whether R\mathbf{R} refers to R(γ)\mathbf{R}\left( \gamma\right) or R(ξ)\mathbf{R}\left( \xi\right) . In the expression R(ξ(γ))\mathbf{R}\left( \xi\left( \gamma\right) \right) , on the other hand, the appearance of the letter ξ\xi makes it clear that R\mathbf{R} stands for R(ξ)\mathbf{R}\left( \xi\right) .
The equation
R(γ)=R(ξ(γ))(5.27)\mathbf{R}\left( \gamma\right) =\mathbf{R}\left( \xi\left( \gamma\right) \right) \tag{5.27}
represents an identity in the independent variable γ\gamma and can therefore be differentiated with respect to γ\gamma. By the chain rule, we have
R(γ)=R(ξ(γ))ξ(γ).(5.28)\mathbf{R}^{\prime}\left( \gamma\right) =\mathbf{R}^{\prime}\left( \xi\left( \gamma\right) \right) \xi^{\prime}\left( \gamma\right) .\tag{5.28}
Substituting this result into the integral
γ0γ1R(γ)R(γ)dγ,(5.10)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma, \tag{5.10}
we find
γ0γ1R(γ)R(γ)dγ=γ0γ1R(ξ(γ))R(ξ(γ))ξ(γ)dγ,(5.29)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma=\int_{\gamma_{0} }^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \xi\left( \gamma\right) \right) \cdot\mathbf{R}^{\prime}\left( \xi\left( \gamma\right) \right) }\xi^{\prime}\left( \gamma\right) d\gamma,\tag{5.29}
where, in bringing ξ(γ)\xi^{\prime}\left( \gamma\right) from under the square root, we used the fact that ξ(γ)\xi^{\prime}\left( \gamma\right) is positive. The resulting integral is tailor-made for an application of the change-of-variables formula from ordinary Calculus, which yields
γ0γ1R(γ)R(γ)dγ=ξ0ξ1R(ξ)R(ξ)dξ,(5.30)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma=\int_{\xi_{0}} ^{\xi_{1}}\sqrt{\mathbf{R}^{\prime}\left( \xi\right) \cdot\mathbf{R} ^{\prime}\left( \xi\right) }d\xi,\tag{5.30}
as we set out to show.

5.2.1The rate of change of arc length

For a given parameterization γ\gamma, denote by s(γ)s\left( \gamma\right) the signed arc length with origin at γ=γ0\gamma=\gamma_{0}, i.e.
s(γ0)=0(5.31)s\left( \gamma_{0}\right) =0\tag{5.31}
and assume that ss and γ\gamma have the same orientation. Then, thanks to the result we have just established, s(γ)s\left( \gamma\right) can be expressed by the integral
s(γ)=γ0γR(γ)R(γ)dγ.(5.32)s\left( \gamma\right) =\int_{\gamma_{0}}^{\gamma}\sqrt{\mathbf{R}^{\prime }\left( \gamma^{\ast}\right) \cdot\mathbf{R}^{\prime}\left( \gamma^{\ast }\right) }d\gamma^{\ast}.\tag{5.32}
By the Fundamental Theorem of Calculus, the derivative s(γ)s^{\prime}\left( \gamma\right) is given by
s(γ)=R(γ)R(γ).(5.33)s^{\prime}\left( \gamma\right) =\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }.\tag{5.33}
Since s(γ)s\left( \gamma\right) is monotonically increasing, there also exists an inverse function γ(s)\gamma\left( s\right) that expresses the parameter γ\gamma in terms of the signed arc length ss. Its derivative is, of course, given by
γ(s)=1R(γ)R(γ).(5.34)\gamma^{\prime}\left( s\right) =\frac{1}{\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }}.\tag{5.34}
It is a common occurrence when working with inverse functions that the derivative of γ(s)\gamma\left( s\right) is expressed in terms of the values γ\gamma of the function γ(s)\gamma\left( s\right) rather than its independent variable ss. This may seem like an inconvenience, but it will soon prove to our advantage.
If the arc length ss has the opposite orientation relative to the parameter γ\gamma, then the function s(γ)s\left( \gamma\right) is given by the equation
s(γ)=γ0γR(γ)R(γ)dγ,(5.35)s\left( \gamma\right) =-\int_{\gamma_{0}}^{\gamma}\sqrt{\mathbf{R}^{\prime }\left( \gamma^{\ast}\right) \cdot\mathbf{R}^{\prime}\left( \gamma^{\ast }\right) }d\gamma^{\ast},\tag{5.35}
while its derivative is given by
s(γ)=R(γ)R(γ).(5.36)s^{\prime}\left( \gamma\right) =-\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }.\tag{5.36}
Then the derivative of the inverse function γ(s)\gamma^{\prime}\left( s\right) is given by
γ(s)=1R(γ)R(γ).(5.37)\gamma^{\prime}\left( s\right) =-\frac{1}{\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }}.\tag{5.37}
This completes our discussion of the arc length integral and we will now turn our attention to curvature.
Curvature is one of the central themes in Differential Geometry and we are about to take our initial steps towards its analytical description. Our analysis will be based exclusively on parameterizing the curve by its arc length ss.

5.3.1The curvature normal

Recall from earlier that the derivative of the position vector R(s)\mathbf{R} \left( s\right) is a tangent vector of unit length known as the unit tangent T\mathbf{T}, i.e.
T=R(s).(5.5)\mathbf{T}=\mathbf{R}^{\prime}\left( s\right) . \tag{5.5}
When we want to call attention to the fact that T\mathbf{T} is a function of ss, we will use the symbol T(s)\mathbf{T}\left( s\right) . Generally, the argument of a function may be included for several purposes. First, when the represented quantity is being differentiated, the argument indicates the independent variable with respect to which the differentiation is taking place. Second, the argument can be used to distinguish between two different functions denoted by the same later -- for example, T(s)\mathbf{T}\left( s\right) and T(γ)\mathbf{T}\left( \gamma\right) . Finally, it may be included simply to emphasize the fact that the symbol denotes a function rather than an isolated object.
Let us now consider the derivative T(s)\mathbf{T}^{\prime}\left( s\right) of the unit tangent vector T(s)\mathbf{T}\left( s\right) and denote it by the symbol B\mathbf{B}, i.e.
B=T(s).(5.38)\mathbf{B}=\mathbf{T}^{\prime}\left( s\right) .\tag{5.38}
In Chapter 4, we demonstrated that the derivative of a vector of constant length is orthogonal to it. Since T(s)\mathbf{T}\left( s\right) is a vector of unit (and therefore constant) length, its derivative B\mathbf{B}, is orthogonal to it, and since T\mathbf{T} represents the instantaneous direction of the curve, B\mathbf{B} is said to be orthogonal, or normal, to the curve.
Furthermore, since T(s)\mathbf{T}\left( s\right) is of constant length, its derivative measures solely the rate at which T(s)\mathbf{T}\left( s\right) changes direction. What is the underlying phenomenon responsible for T(s)\mathbf{T}\left( s\right) changing its direction? Of course, it is what we intuitively understand to be curvature. The greater the curvature, the greater the rate of change in the direction of T(s)\mathbf{T}\left( s\right) . The vector B\mathbf{B}, therefore, quantifies the concept of curvature. Thanks to its two signature properties -- being normal to the curve and being characteristic of curvature -- the vector B\mathbf{B} is known as the curvature normal.
(5.39)
Note that B\mathbf{B} can be expressed as the second derivative of the position vector R(s)\mathbf{R}\left( s\right) , i.e.
B=R(s).(5.40)\mathbf{B}=\mathbf{R}^{\prime\prime}\left( s\right) .\tag{5.40}
Thus, curvature is a second-derivative phenomenon.
Simply by imagining the way the unit tangent changes its direction as you travel along the curve, you should be able to convince yourself that the curvature normal B\mathbf{B} points in the "inward" direction, i.e. the direction towards which the curve is bending. The figure above shows B\mathbf{B} for a planar curve. The following figure shows the curvature normal B\mathbf{B} for a three-dimensional helical funnel.
(5.41)
Note that the displayed length of B\mathbf{B} is affected by our angle of view.
Another way to get a sense for the direction of B\mathbf{B} is to imagine yourself traveling by car with unit speed along the curve. The velocity of the car corresponds to the tangent vector T=R(s)\mathbf{T} =\mathbf{R}^{\prime}\left( s\right) , while the acceleration corresponds to the curvature normal B=T(s)\mathbf{B}=\mathbf{T}^{\prime}\left( s\right) . As you go around a bend, you will feel yourself being pulled in the outward direction by the apparent centrifugal force. This is so because your actual acceleration, i.e. B\mathbf{B}, points inward. Note that this effect is independent of the direction in which the car is travelling and we will now take a closer look at this phenomenon.

5.3.2A note on signs

Unlike the unit normal T\mathbf{T}, whose direction depends on the orientation of the parameterization, the curvature normal B\mathbf{B} is independent of it. This can be seen in a number of insightful ways. The first way is to consider the finite difference
ΔT=T(s+h)T(s)(5.42)\Delta\mathbf{T}=\mathbf{T}\left( s+h\right) -\mathbf{T}\left( s\right)\tag{5.42}
for two parameterizations with opposite orientations, as illustrated in the following figure.
(5.43)
Note that in both scenarios, the tangent vector turns inward as ss increases and therefore the vector ΔT\Delta\mathbf{T} also points in the inward direction. Thus, in the limit as h0h\rightarrow0, the before B=T(s)\mathbf{B=T} ^{\prime}\left( s\right) also points inward.
A related intuitive way to justify the independence of B\mathbf{B} from the orientation of the parameterization is to interpret the parameter ss as time and to imagine a hammer thrower in the act of spinning the "hammer", i.e. a metal ball attached by a wire to a handle. Prior to its release, the ball is kept in circular motion by the string's tension. The acceleration of the ball points in the direction of the tension force, i.e. inward, which is the case regardless of the direction in which the hammer is is spun.
(5.44)
Finally, the independence of B\mathbf{B} from the orientation of the parameterization can be demonstrated by a formal analytical calculation. Let s1s_{1} be an arc length parameterization that has the opposite orientation with respect to ss. Assuming ss and s1s_{1} share the same origin, we have
s=s1.(5.45)s=-s_{1}.\tag{5.45}
If we treat ss as a function s(s1)s\left( s_{1}\right) of s1s_{1}, then
ds(s1)ds1=1.(5.46)\frac{ds\left( s_{1}\right) }{ds_{1}}=-1.\tag{5.46}
Consider the two functions R(s)\mathbf{R}\left( s\right) and R(s1)\mathbf{R} \left( s_{1}\right) that represent the position vector with respect to ss and s1s_{1}. Then R(s1)\mathbf{R}\left( s_{1}\right) can be expressed as a composition of R(s)\mathbf{R}\left( s\right) and s(s1)s\left( s_{1}\right) , i.e.
R(s1)=R(s(s1)).(5.47)\mathbf{R}\left( s_{1}\right) =\mathbf{R}\left( s\left( s_{1}\right) \right) .\tag{5.47}
Differentiating the above identity with respect to s1s_{1}, we find
R(s1)=R(s)ds(s1)ds1=R(s).(5.48)\mathbf{R}^{\prime}\left( s_{1}\right) =\mathbf{R}^{\prime}\left( s\right) \frac{ds\left( s_{1}\right) }{ds_{1}}=-\mathbf{R}^{\prime}\left( s\right) .\tag{5.48}
In other words,
T(s1)=T(s)(5.49)\mathbf{T}\left( s_{1}\right) =-\mathbf{T}\left( s\right)\tag{5.49}
at corresponding values of ss and s1s_{1}, which confirms that the unit tangents associated with each parameterization point in the opposite directions.
Next, convert the above equation into an identity with respect to s1s_{1}, i.e.
T(s1)=T(s(s1)).(5.50)\mathbf{T}\left( s_{1}\right) =-\mathbf{T}\left( s\left( s_{1}\right) \right) .\tag{5.50}
Differentiating once again with respect to s1s_{1}, we find
T(s1)=T(s)ds(s11)ds1=T(s).(5.51)\mathbf{T}^{\prime}\left( s_{1}\right) =-\mathbf{T}^{\prime}\left( s\right) \frac{ds\left( s_{11}\right) }{ds_{1}}=\mathbf{T}^{\prime}\left( s\right) .\tag{5.51}
In summary,
T(s1)=T(s),(5.52)\mathbf{T}^{\prime}\left( s_{1}\right) =\mathbf{T}^{\prime}\left( s\right) ,\tag{5.52}
or
B(s1)=B(s),(5.53)\mathbf{B}\left( s_{1}\right) =\mathbf{B}\left( s\right) ,\tag{5.53}
as we set out to show. In general, odd-ordered derivatives of R(s1)\mathbf{R} \left( s_{1}\right) and R(s)\mathbf{R}^{\prime}\left( s\right) are opposite of each other while even-ordered derivatives are identical.

5.3.3The absolute curvature and the principal normal

The magnitude of B\mathbf{B} is known as the absolute curvature and is denoted by σ\sigma, i.e.
σ=lenB.(5.54)\sigma=\operatorname{len}\mathbf{B}.\tag{5.54}
The term absolute refers to the fact that σ\sigma is always nonnegative. However, absolute is often omitted and σ\sigma is referred to simply as curvature, although we have to exercise care since we will also work with the related concept of signed curvature κ\kappa for planar curves which may also be casually referred to as curvature, but may differ in sign from σ\sigma.
The unit vector P\mathbf{P} that points in the same direction as B\mathbf{B} is called the principal normal. Thus, the curvature normal B\mathbf{B} is the product of the absolute curvature σ\sigma and the principal normal P\mathbf{P}, i.e.
B=σP.(5.55)\mathbf{B}=\sigma\mathbf{P.}\tag{5.55}
The principal normal field for a planar curve is illustrated in the following figure.
(5.56)
Note, importantly, that the principal normal is undefined at points of inflection where it undergoes a nonremovable discontinuity. The following figure illustrates the principal normal P\mathbf{P} for a three-dimensional helical funnel.
(5.57)
By definition, the principal normal P\mathbf{P} is found in the normal plane, and thus singles out a particular direction in the two-dimensional normal space.
(5.58)
It can therefore act as an element in a natural basis for the normal plane. Later in this Chapter, we will supplement with a vector Q\mathbf{Q}, orthogonal to both P\mathbf{P} and T\mathbf{T}, that will complete the basis.

5.3.4Planar curves and signed curvature

Let us briefly turn our attention to planar curves. Whereas for curves in three dimensions we find a two-dimensional normal space, planar curves are characterized by a one-dimensional normal space. In other words, the normal direction is unique, and this is a common characteristic of hypersurfaces, i.e. geometric shapes whose dimension trails that of the surrounding space by 11, such as planar curves and two-dimensional surfaces in a three-dimensional space.
(5.59)
Thus, we know the direction (within the sign) of the principal normal P\mathbf{P} as soon as we have calculated the tangent T\mathbf{T}. In fact, for hypersurfaces, it is common to establish the unique normal direction prior to the analysis of curvature. This is usually accomplished by arbitrarily selecting one of two opposite normal directions. The unit vector N\mathbf{N} pointing in the selected direction is referred to as the unit normal. As in the case of the unit tangent, the article the indicates that the choice has already been made. The freedom to choose one of the two opposite unit normals is referred to as the choice of normal. In practice, it is common to choose a consistent normal direction so that the resulting normal field is globally continuous. The following figure shows two alternative choices of normal that result in continuous normal fields.
(5.60)
Recall that unlike the unit normal N\mathbf{N}, the curvature normal B\mathbf{B} is uniquely determined by the shape of the curve. Since B\mathbf{B} also points in the normal direction, it is colinear with the unit normal N\mathbf{N}. Thus, B\mathbf{B} is a scalar multiple κ\kappa of N\mathbf{N}, i.e.
B=κN.(5.61)\mathbf{B}=\kappa\mathbf{N.}\tag{5.61}
The number κ\kappa is known as the signed curvature. It is a special case of the beautiful concept of mean curvature that characterizes general hypersurfaces. Mean curvature will be described in a future book and, from the moment of its introduction, will play a crucial role in most of our subsequent explorations.
Crucially, the sign of κ\kappa depends on the choice of normal. It equals the magnitude of B\mathbf{B} when the latter points in the same direction as N\mathbf{N} and minus the magnitude of B\mathbf{B} otherwise. This fact can be expressed by the equation
κ=NB.(5.62)\kappa=\mathbf{N}\cdot\mathbf{B.}\tag{5.62}
The same equation in terms of the unit tangent T(s)\mathbf{T}\left( s\right) reads
κ=NT(s).(5.63)\kappa=\mathbf{N}\cdot\mathbf{T}^{\prime}\left( s\right) .\tag{5.63}
Much like the absolute curvature σ\sigma, the signed curvature κ\kappa is also often referred to simply as curvature, making the term curvature ambiguous for planar curves. To catalog the relationship between σ\sigma and κ\kappa, recall that
B=σP,(5.55)\mathbf{B}=\sigma\mathbf{P}, \tag{5.55}
where P\mathbf{P} is the principal normal. Note that both κN\kappa\mathbf{N} and σP\sigma\mathbf{P} equal the curvature normal B\mathbf{B}, i.e.
σP=B=κN.(5.64)\sigma\mathbf{P}=\mathbf{B}=\kappa\mathbf{N.}\tag{5.64}
Thus, σ=κ\sigma=\kappa when B\mathbf{B} (and P\mathbf{P}) point in the same direction as N\mathbf{N}, and σ=κ\sigma=-\kappa otherwise. The relationship among σ\sigma, κ\kappa, B\mathbf{B}, and N\mathbf{N} is illustrated in the following figure.
(5.65)
To reiterate, our entire discussion of signed curvature has been predicated on the choice of normal N\mathbf{N} having been arbitrarily made. Had we made the opposite choice of normal, κ\kappa would have had the opposite value. Thus, the phrase signed curvature with respect to the normal N\mathbf{N} is often used to describe κ\kappa. Meanwhile, the product κN\kappa\mathbf{N} is independent of the choice of normal since both κ\kappa and N\mathbf{N} change sign when the choice of normal is reversed.
Finally, note that the ambiguity in the choice of normal can be removed by coordinating the unit normal with the unit tangent. For example, we could agree to always choose N\mathbf{N} so that the set T,N\mathbf{T},\mathbf{N} is positively oriented. In other words, N\mathbf{N} is obtained from T\mathbf{T} by a 9090^{\circ} counterclockwise rotation. There is an advantage to this approach: choosing N\mathbf{N} in this way enables us to determine which way the curve is bending simply by examining the sign of κ\kappa. Namely, the curve is bending counterclockwise when κ>0\kappa\gt 0 and clockwise when κ<0\kappa \lt 0.
Curvature characterizes the rate at which a curve deviates from being straight. Torsion characterizes the rate which a curve deviates from being planar.

5.4.1The osculating plane

The plane spanned by the tangent and the principal normal directions T\mathbf{T} and P\mathbf{P} is known as the osculating plane. The verb to osculate comes from the Latin word osculum which means to kiss. While the tangent line captures the instantaneous direction of the curve, the osculating plane captures its instantaneous plane of "motion" as it accommodates both the "velocity" T\mathbf{T} and the "acceleration" B=σP\mathbf{B}=\sigma\mathbf{P}.

5.4.2The binormal

The orthonormal vectors T\mathbf{T} and P\mathbf{P} form a basis for the osculating plane. As a set, T\mathbf{T} and P\mathbf{P} are one vector short of spanning the entire three-dimensional space. This shortfall can be remedied by adding a unit vector Q\mathbf{Q} that is orthogonal to both T\mathbf{T} and P\mathbf{P}. Between the two vectors that satisfy this condition, choose Q\mathbf{Q} so that the set T,P,Q\mathbf{T},\mathbf{P},\mathbf{Q} is positively oriented. (The concept of the orientation of a set of vectors was discussed in Chapter 3.) We wish the vectors T\mathbf{T}, P\mathbf{P}, and Q\mathbf{Q} had arisen in alphabetical order, but alas! Fortunately, the set P,Q,T\mathbf{P},\mathbf{Q,T}, which features the same vectors in alphabetical order, has the same orientation as T,P,Q\mathbf{T},\mathbf{P},\mathbf{Q}.
(5.66)
The resulting vector Q\mathbf{Q} is known as the binormal vector or, simply, the binormal. The binormal can be expressed in terms of the unit tangent T\mathbf{T} and the principal normal P\mathbf{P} with the help of the cross product, i.e.
Q=T×P.(5.67)\mathbf{Q}=\mathbf{T}\times\mathbf{P.}\tag{5.67}
Collectively, the vectors T\mathbf{T}, P\mathbf{P}, and Q\mathbf{Q} are known as a Frenet or Frenet-Serret frame, but can also be referred to as a local frame. The principal normal P\mathbf{P} and the binormal Q\mathbf{Q} form a basis in the normal space.

5.4.3The definition of torsion

For the sake of brevity, we will now begin to drop the argument ss from most symbols and denote the derivatives by the subscript ss rather than a prime, i.e.
Ts=T(s)          (5.68)Ps=P(s)          (5.69)Qs=Q(s).          (5.70)\begin{aligned}\mathbf{T}_{s} & =\mathbf{T}^{\prime}\left( s\right)\ \ \ \ \ \ \ \ \ \ \left(5.68\right)\\\mathbf{P}_{s} & =\mathbf{P}^{\prime}\left( s\right)\ \ \ \ \ \ \ \ \ \ \left(5.69\right)\\\mathbf{Q}_{s} & =\mathbf{Q}^{\prime}\left( s\right) .\ \ \ \ \ \ \ \ \ \ \left(5.70\right)\end{aligned}
Recall that the derivative of the position vector R\mathbf{R} with respect to ss yields the unit tangent T\mathbf{T}. The derivative of T\mathbf{T}, in turn, leads to the concepts of absolute curvature σ\sigma and principal normal P\mathbf{P}. As we might expect, the derivative of P\mathbf{P} gives rise to yet another differential characteristic of the curve.
Let us represent Ps\mathbf{P}_{s} as a linear combination of T\mathbf{T}, P\mathbf{P}, and Q\mathbf{Q}, i.e.
Ps=αT+βP+γQ(5.71)\mathbf{P}_{s}=\alpha\mathbf{T}+\beta\mathbf{P}+\gamma\mathbf{Q}\tag{5.71}
and determine each component in turn.
The component α\alpha of Ps\mathbf{P}_{s} with respect to T\mathbf{T} is given by the dot product
α=PsT.(5.72)\alpha=\mathbf{P}_{s}\cdot\mathbf{T.}\tag{5.72}
By the product rule,
α=(PT)sPTs.(5.73)\alpha=\left( \mathbf{P}\cdot\mathbf{T}\right) _{s}-\mathbf{P} \cdot\mathbf{T}_{s}.\tag{5.73}
Since P\mathbf{P} is orthogonal to T\mathbf{T}, the first term vanishes and we have
α=PTs.(5.74)\alpha=-\mathbf{P}\cdot\mathbf{T}_{s}.\tag{5.74}
Recall that Ts\mathbf{T}_{s} is, by definition, the curvature normal B=σP\mathbf{B}=\sigma\mathbf{P}, i.e.
Ts=σP,(5.75)\mathbf{T}_{s}=\sigma\mathbf{P},\tag{5.75}
and, therefore,
α=σPP.(5.76)\alpha=-\sigma\mathbf{P}\cdot\mathbf{P}.\tag{5.76}
Finally, since P\mathbf{P} is unit length, its dot product with itself equals 11 and therefore
α=σ.(5.77)\alpha=-\sigma.\tag{5.77}
This identity makes a great deal of intuitive sense, especially for a planar curve. After all, the unit vectors P\mathbf{P} and T\mathbf{T} are locked in an orthogonal relationship. As a result, whatever the rate at which T\mathbf{T} is rotating -- which, of course, is proportional to σ\sigma -- P\mathbf{P} is rotating at the same rate, as illustrated in the following figure.
(5.78)
Therefore, the projection of Ts\mathbf{T}_{s} onto P\mathbf{P} equals the projection of Ps\mathbf{P}_{s} onto T-\mathbf{T}, which is what the above equation is telling us.
The coefficient β\beta of Ps\mathbf{P}_{s} with respect to P\mathbf{P}, given by the dot product
β=PsP,(5.79)\beta=\mathbf{P}_{s}\cdot\mathbf{P,}\tag{5.79}
is easily seen to be zero, i.e.
β=0,(5.80)\beta=0,\tag{5.80}
since P(s)\mathbf{P}\left( s\right) is of constant length and therefore Ps\mathbf{P}_{s} is orthogonal to P\mathbf{P}.
Finally, let us turn our attention to the final coefficient γ\gamma of Ps\mathbf{P}_{s} with respect to Q\mathbf{Q} given by
γ=PsQ.(5.81)\gamma=\mathbf{P}_{s}\cdot\mathbf{Q}.\tag{5.81}
Unlike the components of Ps\mathbf{P}_{s} with respect to T\mathbf{T} and P\mathbf{P}, its component with respect to the binormal Q\mathbf{Q} cannot be expressed in terms of the quantities that have already been introduced. After all, a curve in a three-dimensional space cannot be described by its curvature alone. The absolute curvature characterizes the shape of the curve within the osculating plane spanned by T\mathbf{T} and P\mathbf{P}. However, the derivative Ps\mathbf{P}_{s} is not necessarily contained in the osculating plane. Therefore, its component γ\gamma with respect to the binormal Q\mathbf{Q}, which is orthogonal to T\mathbf{T} and P\mathbf{P}, captures the rate at which the curve escapes the osculating plane and corresponds to a new geometric quantity.
The component of Ps\mathbf{P}_{s} with respect to the binormal Q\mathbf{Q} is known as the torsion τ\tau, i.e.
τ=PsQ.(5.82)\tau=\mathbf{P}_{s}\cdot\mathbf{Q}.\tag{5.82}
In particular, the full expansion of Ps\mathbf{P}_{s} with respect to T\mathbf{T}, P\mathbf{P}, and Q\mathbf{Q} reads
Ps=σT+τQ.(5.83)\mathbf{P}_{s}=-\sigma\mathbf{T}+\tau\mathbf{Q}.\tag{5.83}
Unlike the absolute curvature σ\sigma, the torsion τ\tau can be both positive and negative, as will be demonstrated in the upcoming example involving two helices of opposite orientations. Also note that the value of torsion does not depend on the orientation of the parameter ss. Indeed, recall when the parameterization is reversed, T\mathbf{T} changes its direction and P\mathbf{P} remains unchanged. Meanwhile Q\mathbf{Q}, being the cross product of T\mathbf{T} and P\mathbf{P}, also changes its direction. It is left as an exercise to show that the derivative Ps\mathbf{P}_{s} also changes its direction. Thus, since both Ps\mathbf{P}_{s} and Q\mathbf{Q} change their directions, their dot product, i.e. the torsion τ\tau, remains unchanged.

5.4.4An illustration of torsion with two oppositely-oriented helices

The shape of a spring is known as a helix. A helix is great for illustrating the interplay between curvature and torsion as it is characterized by constant values of both quantities. Helical shapes commonly encountered in everyday life include the aforementioned springs, threads on bolts, spiral staircases, and (with apologies to anyone reading these lines before 1940 or after 2025) slinkies.
A helix can have one of two orientations. Suppose that the axis of a helix is aligned with a ray pointing in the direction arbitrarily labeled as up. Then a right-handed helix twists in the counterclockwise direction as it goes up, while a left-handed helix twists in the clockwise direction. Threads on most bolts and screws are right-handed and are sometimes described by the mnemonic righty-tighty. Importantly, rigidly rotating a helix upside down does not change its orientation. In other words, one of the helices cannot be transformed into the other by turning it upside down. Instead, it will remain equivalent to itself. We know this from our everyday experience and will confirm it analytically by calculating the torsion of each helix.
  (5.84)
In the above figure, the helix on the left is right-handed while the one on the right is left-handed. As we discussed above, the orientation of the parameterization does not affect the value of torsion. Therefore, assume that the parameter ss increases in the upward direction for both helices. Then, for both shapes, the unit tangent T\mathbf{T} points slightly up, while the principal normal P\mathbf{P} points directly towards the axis in the strictly horizontal plane. (This property of P\mathbf{P} is intuitively clear from the fact that the upward "motion" is uniform and, thus, the "acceleration" has no vertical component.) By the right-hand rule, the binormal Q\mathbf{Q} points mostly up for the right-handed helix and mostly down for the left-handed helix.
Now, let us get a sense for the derivative Ps\mathbf{P}_{s} of the principal normal P\mathbf{P}. Since P\mathbf{P} remains in the horizontal plane for all values of ss, its derivative Ps\mathbf{P}_{s} is also found in that plane. Within that plane, P\mathbf{P} rotates in the counterclockwise direction for the right-handed helix and in the clockwise direction for the left-handed helix. Furthermore, we know that Ps\mathbf{P}_{s} is also contained in the plane spanned by T\mathbf{T} and Q\mathbf{Q}. Thus, let us examine how it is arranged in that plane relative to T\mathbf{T} and Q\mathbf{Q}. The following figure shows the plane spanned by T\mathbf{T} and Q\mathbf{Q} for both helices. Since P\mathbf{P} rotates in the counterclockwise direction for the right-handed helix, Ps\mathbf{P}_{s} points to the left. Correspondingly, Ps\mathbf{P}_{s} points to the right for the left-handed helix.
  (5.85)
Observe that for the right-handed helix, the component Ps\mathbf{P}_{s} with respect to Q\mathbf{Q}, i.e. the torsion τ\tau, is positive. Meanwhile, for the left-handed helix, it is negative.
This also proves that one type of helix cannot be transformed into the other by a rigid rotation. Indeed, since torsion depends only on the shape of the curve, it cannot be changed by a rigid rotation. Therefore, the torsion on the helix, say, on the left will remain positive if that helix were rotated upside down. Therefore, the upside-down version of the helix on the left is still fundamentally distinct from the helix on the right whose torsion is negative.
Finally, this is an opportune moment to reiterate an important point that we made at the beginning of the Chapter. Namely, despite the great geometric insight that we have been able to achieve, our geometric approach does not enable us to calculate either the absolute curvature σ\sigma or the torsion τ\tau for a specific curve -- at least, not easily. After all, the helix is the simplest conceivable three-dimensional curve with nonvanishing curvature and torsion and we do not yet have an effective tool for determining the numerical value of σ\sigma or τ\tau. This observation makes clear the need for a more robust analytical network. This will be achieved by the introduction of a coordinate system in the ambient space and the subsequent development of a tensor framework, which will enable us to work with the components of vectors rather than the vectors themselves since the components of vectors can be analyzed by the powerful methods of Calculus or -- when a reasonable analytical approach is not feasible -- numerical methods.

5.5.1The derivative of the binormal

It is only natural to wonder whether the derivative Qs\mathbf{Q}_{s} of the binormal Q\mathbf{Q} produces yet another differential characteristic of the curve that would join the ranks of curvature and torsion. To this end, let us decompose Qs\mathbf{Q}_{s} with respect to the local frame T,P,Q\mathbf{T,P,Q}, i.e.
Qs=αT+βP+γQ.(5.86)\mathbf{Q}_{s}=\alpha\mathbf{T}+\beta\mathbf{P}+\gamma\mathbf{Q}.\tag{5.86}
Since Q(s)\mathbf{Q}\left( s\right) is unit length, Qs\mathbf{Q}_{s} is orthogonal to Q\mathbf{Q} and therefore
γ=0.(5.87)\gamma=0.\tag{5.87}
The components α\alpha and β\beta of Qs\mathbf{Q}_{s} with respect to T\mathbf{T} and P\mathbf{P} are given by the dot products
α=QsT    and    β=QsP.(5.88)\alpha=\mathbf{Q}_{s}\cdot\mathbf{T}\text{ \ \ \ and\ \ \ \ }\beta =\mathbf{Q}_{s}\cdot\mathbf{P}.\tag{5.88}
In both cases, an application of the product rule will transfer the derivative from Qs\mathbf{Q}_{s} onto the other vector. For α\alpha, we have
α=QsT=(QT)sQTs(5.89)\alpha=\mathbf{Q}_{s}\cdot\mathbf{T}=\left( \mathbf{Q}\cdot\mathbf{T}\right) _{s}-\mathbf{Q}\cdot\mathbf{T}_{s}\tag{5.89}
Since Q\mathbf{Q} is orthogonal to T\mathbf{T}, the first term vanishes and thus
α=QTs.(5.90)\alpha=-\mathbf{Q}\cdot\mathbf{T}_{s}.\tag{5.90}
Recall that Ts=σP\mathbf{T}_{s}=\sigma\mathbf{P} and therefore
α=σQP.(5.91)\alpha=-\sigma\mathbf{Q}\cdot\mathbf{P}.\tag{5.91}
Finally, since Q\mathbf{Q} is orthogonal to P\mathbf{P}, we conclude that α\alpha vanishes, i.e.
α=0.(5.92)\alpha=0.\tag{5.92}
For the component β\beta of Qs\mathbf{Q}_{s} with respect to the principal normal P\mathbf{P}, the same approach yields the following chain of identities.
β=QsP          (5.93)(product rule)     =(QP)sQPs          (5.94)(recall that QP=0)    =QPs          (5.95)(recall that Ps=σT+τQ)    =Q(σT+τQ)          (5.96)(recall that QT=0 and lenQ=1)    =σQTτQQ                                   (5.97)=τ          (5.98)\begin{aligned}\beta & =\mathbf{Q}_{s}\cdot\mathbf{P}\ \ \ \ \ \ \ \ \ \ \left(5.93\right)\\\text{(product rule) \ \ \ \ } & =\left( \mathbf{Q}\cdot\mathbf{P}\right) _{s}-\mathbf{Q}\cdot\mathbf{P}_{s}\ \ \ \ \ \ \ \ \ \ \left(5.94\right)\\\text{(recall that }\mathbf{Q}\cdot\mathbf{P}=0\text{)\ \ \ \ } & =-\mathbf{Q}\cdot\mathbf{P}_{s}\ \ \ \ \ \ \ \ \ \ \left(5.95\right)\\\text{(recall that }\mathbf{P}_{s}=-\sigma\mathbf{T}+\tau\mathbf{Q} \text{)\ \ \ \ } & =-\mathbf{Q}\cdot\left( -\sigma\mathbf{T}+\tau \mathbf{Q}\right)\ \ \ \ \ \ \ \ \ \ \left(5.96\right)\\\text{(recall that }\mathbf{Q}\cdot\mathbf{T}=0\text{ and }\operatorname{len} \mathbf{Q}=1\text{)\ \ \ \ } & =\sigma\mathbf{Q}\cdot\mathbf{T} -\tau\mathbf{Q}\cdot \mathbf{Q\ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ }\ \ \ \ \ \ \ \ \ \ \left(5.97\right)\\& =-\tau\ \ \ \ \ \ \ \ \ \ \left(5.98\right)\end{aligned}
In summary, the complete decomposition of Qs\mathbf{Q}_{s} with respect to T\mathbf{T}, P\mathbf{P}, and Q\mathbf{Q} reads
Qs=τP.(5.99)\mathbf{Q}_{s}=-\tau\mathbf{P}.\tag{5.99}
Thus, the analysis of Qs\mathbf{Q}_{s} has not lead to a new differential characteristic of a curve. It would therefore appear that the curvature and the torsion capture all of the available information about the local behavior of a curve. In fact, as we demonstrate below, if the functions σ(s)\sigma\left( s\right) and τ(s)\tau\left( s\right) are stated a priori, they are sufficient to reconstruct the shape of the curve in a three-dimensional space.

5.5.2The statement of the Frenet equations

Let us combine the decompositions of Ts\mathbf{T}_{s}, Ps\mathbf{P}_{s}, and Qs\mathbf{Q}_{s} in terms of T\mathbf{T}, P\mathbf{P}, and Q\mathbf{Q} into a single set, i.e.
Ts=σP          (5.100)Ps=σT+τQ          (5.83)Qs=τP.          (5.99)\begin{aligned}\mathbf{T}_{s} & = \phantom{-} \sigma\mathbf{P}\ \ \ \ \ \ \ \ \ \ \left(5.100\right)\\\mathbf{P}_{s} & =-\sigma\mathbf{T}+\tau\mathbf{Q}\ \ \ \ \ \ \ \ \ \ \left(5.83\right)\\\mathbf{Q}_{s} & =-\tau\mathbf{P.} \ \ \ \ \ \ \ \ \ \ \left(5.99\right)\end{aligned}
Collectively, these equations are known as the Frenet formulas or the Frenet-Serret formulas. In matrix form, the equations read
[TsPsQs]=[σσττ][TPQ],(5.101)\left[ \begin{array} {c} \mathbf{T}_{s}\\ \mathbf{P}_{s}\\ \mathbf{Q}_{s} \end{array} \right] =\left[ \begin{array} {rrr} & \sigma & \\ -\sigma & & \tau\\ & -\tau & \end{array} \right] \left[ \begin{array} {c} \mathbf{T}\\ \mathbf{P}\\ \mathbf{Q} \end{array} \right] ,\tag{5.101}
and it becomes apparent how these formulas might generalize to higher-dimensional spaces. Indeed, this generalization will be accomplished in Chapter 20 on Riemannian spaces.

5.5.3The intrinsic equation of a curve

One of the insights provided by the Frenet equations is that the shape of a curve is fully specified by the curvature and torsion as functions of arc length. Suppose that the functions σ(s)\sigma\left( s\right) and τ(s)\tau\left( s\right) are given, along with the ambient location of one point on the curve, say, at s=s0s=s_{0}, and the curve's orientation at that point. In other words, let the values R0=R(s0)\mathbf{R}_{0}=\mathbf{R}\left( s_{0}\right) , T0=T(s0)\mathbf{T}_{0}=\mathbf{T}\left( s_{0}\right) , P0=P(s0)\mathbf{P}_{0} =\mathbf{P}\left( s_{0}\right) , and therefore Q0=Q(s0)=T0×P0\mathbf{Q}_{0} =\mathbf{Q}\left( s_{0}\right) =\mathbf{T}_{0}\times\mathbf{P}_{0}, be known. Then the entire curve can be reconstructed by solving the system of ordinary differential equations
[T(s)P(s)Q(s)]=[σ(s)σ(s)τ(s)τ(s)][T(s)P(s)Q(s)](5.101)\left[ \begin{array} {c} \mathbf{T}^{\prime}\left( s\right) \\ \mathbf{P}^{\prime}\left( s\right) \\ \mathbf{Q}^{\prime}\left( s\right) \end{array} \right] =\left[ \begin{array} {rrr} & \sigma\left( s\right) & \\ -\sigma\left( s\right) & & \tau\left( s\right) \\ & -\tau\left( s\right) & \end{array} \right] \left[ \begin{array} {c} \mathbf{T}\left( s\right) \\ \mathbf{P}\left( s\right) \\ \mathbf{Q}\left( s\right) \end{array} \right] \tag{5.101}
subject to the initial conditions
T(s0)=T0          (5.102)P(s0)=P0          (5.103)Q(s0)=Q0.          (5.104)\begin{aligned}\mathbf{T}\left( s_{0}\right) & =\mathbf{T}_{0}\ \ \ \ \ \ \ \ \ \ \left(5.102\right)\\\mathbf{P}\left( s_{0}\right) & =\mathbf{P}_{0}\ \ \ \ \ \ \ \ \ \ \left(5.103\right)\\\mathbf{Q}\left( s_{0}\right) & =\mathbf{Q}_{0}.\ \ \ \ \ \ \ \ \ \ \left(5.104\right)\end{aligned}
Once T\mathbf{T} is calculated, R\mathbf{R} can be reconstructed by solving the equation
R(s)=T(s)(5.105)\mathbf{R}^{\prime}\left( s\right) =\mathbf{T}\left( s\right)\tag{5.105}
subject to the initial condition
R(s0)=R0.(5.106)\mathbf{R}\left( s_{0}\right) =\mathbf{R}_{0}.\tag{5.106}
Of course, R\mathbf{R} could have been included among the unknowns, resulting in a system with 44 equations and 44 unknowns, but that would have diminished the elegance of the system and its closure under the functions T(s)\mathbf{T}\left( s\right) , P(s)\mathbf{P}\left( s\right) , and Q(s)\mathbf{Q}\left( s\right) .
Because σ(s)\sigma\left( s\right) and τ(s)\tau\left( s\right) , along with the initial conditions, are sufficient to describe the shape, the location, and the orientation of the curve, these functions are known as the intrinsic, or natural, equations of the curve. In the absence of initial conditions, the functions σ(s)\sigma\left( s\right) and τ(s)\tau\left( s\right) are still sufficient to determine the shape of the curve.

5.5.4The Frenet equations for a general parameterization

The foregoing analysis took fundamental advantage of the special parameterization of the curve with the help of the arc length ss. In practice, however, an arc-length parameterization is difficult to achieve for specific curves. Meanwhile, a different parameterization may be readily available. It is of interest, then, to adapt the above analysis to a parameterization with an arbitrary variable γ\gamma. Since the only parameter-dependent operation in our analysis is the derivative, all we need to do is express d/dsd/ds in terms of d/dγd/d\gamma.
Suppose that a quantity TT, scalar or vector, is defined on the curve. Consider it simultaneously as functions T(γ)T\left( \gamma\right) and T(s)T\left( s\right) of γ\gamma and ss. The functions T(γ)T\left( \gamma\right) and T(s)T\left( s\right) are related by the identity
T(s)=T(γ(s)),(5.107)T\left( s\right) =T\left( \gamma\left( s\right) \right) ,\tag{5.107}
where γ(s)\gamma\left( s\right) is dependence of the variable γ\gamma on the arc length ss. Then, by the chain rule, we find
ddsT(s)=ddγT(γ)γ(s).(5.108)\frac{d}{ds}T\left( s\right) =\frac{d}{d\gamma}T\left( \gamma\right) \gamma^{\prime}\left( s\right) .\tag{5.108}
Isolating the operators d/dsd/ds and d/dγd/d\gamma, we have
dds=γ(s)ddγ.(5.109)\frac{d}{ds}=\gamma^{\prime}\left( s\right) \frac{d}{d\gamma}.\tag{5.109}
Recall that the derivative γ(s)\gamma^{\prime}\left( s\right) is given by the equation
γ(s)=1R(γ)R(γ).(5.34)\gamma^{\prime}\left( s\right) =\frac{1}{\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }}. \tag{5.34}
This is where the fact that the derivative γ(s)\gamma^{\prime}\left( s\right) is naturally expressed as a function of γ\gamma rather than ss, is advantageous since it enables us to express d/dsd/ds explicitly in terms of γ\gamma and d/dγd/d\gamma, i.e.
dds=1R(γ)R(γ)ddγ.(5.110)\frac{d}{ds}=\frac{1}{\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }}\frac{d}{d\gamma}.\tag{5.110}
With the help of this equation, can now rewrite the Frenet equations for a curve subject to an arbitrary parameterization. Letting
V(γ)=R(γ)R(γ),(5.111)V\left( \gamma\right) =\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) },\tag{5.111}
we have
1VTγ=σP          (5.112)1VPγ=σT+τQ          (5.113)1VQγ=τP.          (5.114)\begin{aligned}\frac{1}{V}\mathbf{T}_{\gamma} & =\sigma\mathbf{P}\ \ \ \ \ \ \ \ \ \ \left(5.112\right)\\\frac{1}{V}\mathbf{P}_{\gamma} & =-\sigma\mathbf{T}+\tau\mathbf{Q}\ \ \ \ \ \ \ \ \ \ \left(5.113\right)\\\frac{1}{V}\mathbf{Q}_{\gamma} & =-\tau\mathbf{P}.\ \ \ \ \ \ \ \ \ \ \left(5.114\right)\end{aligned}
This form greatly increases the applicability of the Frenet formulas. Nevertheless, it remains fair to say that all equations presented so far have mostly theoretical applications due to the use of geometric vectors which have limited analytical capabilities. In order to facilitate practical calculations, it is necessary to refer the ambient space to a coordinate system in order to enable us to work with the components of vectors rather than the vectors themselves. This is the task that we will begin to undertake in the next Chapter.
Exercise 5.1What is the geometric interpretation of the integral
s0s1T(s)ds?(5.115)\int_{s_{0}}^{s_{1}}\mathbf{T}\left( s\right) ds?\tag{5.115}
Conclude that, for a closed curve, i.e. R(s0)=R(s1)\mathbf{R}\left( s_{0}\right) =\mathbf{R}\left( s_{1}\right) ,
s0s1T(s)ds=0.(5.116)\int_{s_{0}}^{s_{1}}\mathbf{T}\left( s\right) ds=\mathbf{0.}\tag{5.116}
Exercise 5.2Similarly, what is the geometric interpretation of the integral
s0s1σ(s)P(s)ds?(5.117)\int_{s_{0}}^{s_{1}}\sigma\left( s\right) \mathbf{P}\left( s\right) ds?\tag{5.117}
For this integral, too, conclude that for a closed curve,
s0s1σ(s)P(s)ds=0.(5.118)\int_{s_{0}}^{s_{1}}\sigma\left( s\right) \mathbf{P}\left( s\right) ds=\mathbf{0.}\tag{5.118}
Exercise 5.3Complete the proof of the fact that
γ0γ1R(γ)R(γ)dγ(5.10)\int_{\gamma_{0}}^{\gamma_{1}}\sqrt{\mathbf{R}^{\prime}\left( \gamma\right) \cdot\mathbf{R}^{\prime}\left( \gamma\right) }d\gamma\tag{5.10}
is independent of parameterization by analyzing an alternative parameterization ξ\xi whose orientation is opposite that of γ\gamma.
Exercise 5.4Show that the absolute curvature σ\sigma of a circle of radius RR is given by
σ=1R.(5.119)\sigma=\frac{1}{R}.\tag{5.119}
Exercise 5.5Show that the signed curvature κ\kappa of a circle of radius RR with respect to the outward normal is given by
κ=1R.(5.120)\kappa=-\frac{1}{R}.\tag{5.120}
Exercise 5.6Explain why the derivative Ps\mathbf{P}_{s} of the principal normal P\mathbf{P} changes direction when the orientation of the parameterization ss is reversed.
Exercise 5.7In Section 5.5.1, justify each step in the chain of identities leading to the coefficient β\beta.
Exercise 5.8Derive the third Frenet equation
Qs=τP(5.99)\mathbf{Q}_{s}=-\tau\mathbf{P} \tag{5.99}
by differentiating the identity
Q=T×P.(5.67)\mathbf{Q}=\mathbf{T}\times\mathbf{P.} \tag{5.67}
Exercise 5.9Show that the Frenet equations for a planar curve read
[TsPs]=[σσ][TP].(5.121)\left[ \begin{array} {c} \mathbf{T}_{s}\\ \mathbf{P}_{s} \end{array} \right] =\left[ \begin{array} {rr} & \sigma\\ -\sigma & \end{array} \right] \left[ \begin{array} {c} \mathbf{T}\\ \mathbf{P} \end{array} \right] .\tag{5.121}
Exercise 5.10Solve the planar Frenet equations for constant absolute curvature σ\sigma to show the resulting curve is a circle of radius 1/σ1/\sigma.
Send feedback to Pavel