An overview of this chapter’s contents and take-aways can be found here.
This introductory chapter tries to illustrate the value mathematics holds for the economist profession and deals with fundamental, overarching concepts that may be viewed as prerequisites for any mathematical application and study, regardless of its purpose or context. You might have come across some or even most of them during your undergraduate studies, but even if you have it may be worthwhile to take the time and review and re-organize them in your mind, and perhaps also refresh your memory a bit.
Table of Contents
Usually, it is easier to find motivation for studying abstract and complex matters if you know why you’re doing it. You may ask yourself, “Hey, I signed up for an Economics program! Why should I bother with a Math course, rather than reviews in Micro- or Macroeconomics, or perhaps Econometrics?” So, why is Mathematics of central importance to the economist profession?
As economists we work with mathematical models to describe economic problems and we use statistical and numerical estimation techniques to infer properties of theses models. Even if you do not considert doing research in economics later on in your professional life, you will most likely encounter problems that can only be described and solved using mathematics!
Mathematics can be thought of as a language. In this view, the vocabulary of mathematics consists of a large set of commonly agreed-upon mathematical notation (numbers, variables, functions, operators for e.g. addition () and multiplication (
), etc.) and symbols (e.g.
, etc.). Any combination of items in the vocabulary constitutes a statement, which can be judged whether it is meaningful and if so, also true. These statements can then be combined to form logical arguments that convey information about mathematical facts and relationships.
People “speaking” (or making use of) the mathematical language confine themselves to a rather narrow set of expressions. These expressions help you to more efficiently write down statements. The table below gives a brief overview of those symbols most important for economists. In addition, the sets of numbers and
, i.e. real, natural, complex, rational numbers and integers, are very important and should be familiar to you. Don’t worry if you can’t remember everything in the table just yet, you will see and use these often enough.
In the first column, the word quantifier, which refers to the first four symbols, may be new to you. They are important since we frequently make use of quantifying statements, that is, expressions of the form
which indicate whether a certain property holds for all (), some (
), exactly one (
) or none (
) of the elements considered. To see this in a simple example, consider the mathematical way of saying that all natural numbers are non-negative:
Now let us have a look at the table’s second column. The logical “or” has a meaning slightly different from standard English, and more precisely translates to “and/or”, i.e. it does not preclude that both statements are true, but requires at least one to be true. As an example, consider the true statement , which asserts that the number 1 is an element of the real numbers “or” an element of the natural numbers. The logical “not” reads as “it is not the case that” and inverts the meaning of a statement, asserting the exact opposite. We can frequently re-write these statements as more natural expressions, e.g.
as
, or
as
. This simplification abides exactly by the same logic as the one we also use in “normal” English where, for instance, “there are some people who own cats” is a more direct way of saying “not all people do not own cats”.
In our discussion of the table above, we have already considered statements without explicitly defining them. Formally, a statement is any combination of mathematical vocabulary. A statement may or may not be meaningful: reading it out verbally it gives you a grammatically correct English sentence. This makes assessing meaningfulness relatively intuitive and straightforward.
Furthermore, statements may or may not be true. Note that meaningful mathematical statements need not be true; consider e.g. or
. A statement that is not meaningfull can never be true: if there is no meaning, there is nothing to be contrasted against the universe of “true” circumstances. To (re-)familiarize yourself with these notions, consider Table 2, where
denotes “Alaska”,
“Berlin”, and
the set of European capitals.
Let’s get some more practice with notation and mathematical statements and have a look at the table below. Try to first cover columns 2 and 3 and see whether you can identify the verbal meaning of the mathematical statement and assess whether is true.
Note that the delimiting brackets are of crucial importance in the third and fourth statement, because only with them, it is clear that the implication () refers to the whole statement in parentheses rather than just to “
”. In the last example, on the other hand, they are just there to increase clarity and would probably be left out by many. Also note that “greater/smaller than” includes equality unless we use the prefix “strictly”. Similarly, we call
a weak (strict) inequality. As a last comment on the table above, while it may take some time to get used to the notation, you should be able to clearly see the notation’s value added by simply comparing the space needed for the mathematical and verbal statements.
If you feel like testing your understanding of the discussion thus far, you can take a short quiz found here.
Most of the time, it is rather easily assessed whether individual statements, as we have seen thus far, are true. The more essential part of mathematical analysis is how certain statements relate to each other. We call an argument an assertion of a relationship between two (sets of) statements which we call the premise(s) and conclusion(s), respectively. Typically, the assertion is either that the premises imply () the conclusion or that they are equivalent to it (
).
We can use basic logic to internally investigate whether the argument “makes sense”, i.e. whether the asserted relationship between premises and conclusion holds, while remaining agnostic about the plausibility of the premises. If it does, we call the argument valid. The establisished properties and definitions unambiguously determine whether an argument is valid or not. For instance, the argument that the premises “ants are taller than humans” and “humans are taller than elefants” imply the conclusion that “ants are taller than elefants” is indeed valid, since the conclusion logically follows from the premises.
If a valid argument additionally has true premises, it is called sound. It is worthwhile to stress that validity is always needed for soundness – an invalid argument can never be sound! Unlike with validity, the assessment of soundness, i.e. whether or not some premises are true, may be context-specific. For instance, the argument that if “ is a differentiable function” and if “any differentiable function is continuous”, then “
is a continuous function” is valid. Whether or not it is sound depends on the premises – the latter premise is, as we will see in Chapter 3, a general statement that is always true, whereas the former depends on how the concrete function
is defined in the context we are concerned with.
To familiarize ourselves with arguments, let us consider some examples (to evaluate soundness, assume that the geographical relationships we consider are those of the real world):
First, for Nr. 3, the premises do not preclude the conclusion. Therefore, provided that the premises are true, the conclusion may still be true as well. However, the premises do not give us enough information to assert that the conclusion must be true when the premises are true. The invalidity of Nr. 4 is for a similar reason, try to find out why exactly. Finally, Nr. 5 serves as an example for a valid statement of equivalence and a case where one statement (here: the conclusion) has multiple implications (here: the premises). Depending on who makes the argument (i.e. who “I” refers to precisely), this argument could also be sound, but given that you are reading this text in your browser with your eyes, it cannot be sound here.
Typically, mathematical theory is more concerned with argument validity rather than soundness. Theory provides us with theorems and propositions that tell us that “if this and that is true, then also some other property will be true”. You can find an abundance of examples in the remainder of the course, but to make the point very clear, let’s consider the so-called Weierstrass Extreme Value Theorem (its content is not important at this point, do not worry if this does not make sense yet), which states that “If (premise 1) is a continuous function and (premise 2)
has a compact domain then (conclusion)
must assume a global maximum and minimum.” Here,
is an unspecified, hypothetical function. For concrete functions, the premises may or may not be true, but this is not essential for the usefulness of this theorem and the validity of the statement.
On the other hand, if you are working on some exercise problems or writing an exam, you will frequently be given concrete contexts (in the example of the Weierstrass theorem: concrete functions) to work with. Then, you will likely refer to all the valid arguments that you know from your textbooks and try to make sound arguments with them. Say, for instance, you are given some utility function and are asked whether it has a global maximum. Then, if your argument is that by the Weierstrass Extreme Value Theorem, this function must have a global maximum, it depends on the precise function that you are given whether your argument is sound or not.
Above, we stated that the rules which determine whether an argument is valid or not come from “basic logic”. While the expression itself may give you some idea what we mean by this, thus far, it is not explicitly clear how to precisely think about this concept.
Basic logic can be thought of as the fundamental rules that determine whether a certain mathematical argument is valid.
To see this abstract elaboration in action, let us consider how basic logic helps us in the example of ants, humans and elephants, we can in fact proceed intuitively. We know logically that if one thing is larger than the other, and this other thing is again larger than a third thing, then the first thing must also be larger than the third – this is just common sense. Mathematically, all we do in this example is to compare positive real numbers with each other: the elements in the sets of ants’ (), humans’ (
) and elephants’ (
) numerical heights. We consider the argument
The basic mathematical reason that this relationship is true (i.e. the “mathematical common sense” that justifies the argument) is transitivity of the “strictly-greater-than” relation on the real numbers, namely that if for ,
and
, then also
.
In this rather simple example, transitivity of the ““-relation is the entire fundamental mathematical reason why this argument is valid. Still, it is already non-obvious how exactly this circumstance justifies validity of the argument. This is especially true for more complex arguments that depend on a multitude of fundamental mathematical facts. This is typically where mathematical proofs come in: they provide a step-wise decomposition of how fundamental mathematical circumstances make certain arguments valid or invalid.
Throughout their career, any economist will hear the words “necessary” and “sufficient” quite a lot. If you have been thinking thoroughly about the three logical arrows in the notation table, you will not have a hard time to understand what follows.
Suppose we are interested in some statement . A necessary condition for
must hold for
to be true. It need not guarantee truth of
. Thus,
is true only if the necessary condition is satisfied. A sufficient condition for
guarantees that
is true. However, it need not hold for
to be true. Thus,
is true if the sufficient condition is satisfied. Finally, an equivalent condition for
(i) must hold for
to be true and (ii) guarantees that
is true. Thus,
is true if and only if the equivalent condition is satisfied. By its definition, the equivalent condition is also both a necessary and a sufficient condition.
Let us consider an example, and let us define , where
is some function mapping from and to real numbers
, that we do not specify any further for now. Then, a necessary condition for
would be
.
can be true only if
is satisfied. A sufficient condition is
. If
holds, this guarantees that
does as well; however, there are many examlpes of functions
where
holds even though
is violated (for example, think about
). Finally, an equivalent condition is
.
If this is not fully clear to you now, think about whether is true for the following specific examples of
(you may consult the conditions
defined above):
In terms of our logical arrows, let be the condition. If
is necessary for
, then
is implied by
:
. If instead,
is sufficient for
, then
implies
:
. And if
is an equivalent condition for
, then
is equivalent to
:
.
If we want to establish , sufficient and equivalent conditions typically make us happy: their truth is enough to know that
is true. With a necessary condition
, on the other hand, we only know that
cannot be true unless
is also true (
). This may help disprove
: if
is not true, or respectively, the opposite of
,
, is true, then
is not true (and
is true):
. Thus, violation of the necessary condition implies violation of the statement of interest. Notice the relationship of negation and implication: we have just argued that
is equivalent to
. This means that, for any given implication, when considering the inverted/negated statements, you can always just “flip” the implication arrow. Go through the argument again and make sure that you logically understand why this works!
To give a you some practice, consider the United Kingdom (UK)’s definition of an economic recession, which states that an economy is in the state of recession whenever GDP growth has been negative for at least two quarters. Consider the following conditions – which ones are necessary, sufficient, equivalent or nothing at all for the German economy currently being in a recession?
In the mathematical context relevant to economists, you come across necessary and sufficient conditions mostly in optimization, where we frequently deal with them (mostly related to second derivatives) when investigating whether a solution constitutes a maximum, a minimum, or neither. Therefore, they are at the heart of, amongst others, utility or profit maximization, cost minimization, and also error/deviation minimization of statistical estimators.
When reading mathematical texts, you come across a range of “facts” with different names. If you are interested, you can find a brief overview of what sets apart Theorems, Propositions, Lemmas and Corollaries, which make up for almost all of these facts, below.
The most common “fact” is the proposition. It is a statement that is “interesting” by itself, and usually contains at least an important part or “setup”-result for the purpose of the text. Since by the nature of the word, some fact is proposed, propositions are always expected to come with a proof (but of course not references to propositions in other texts, as in “see Proposition 5 of Textbook XY”). Accordingly, all results labeled as “proposition” in the companion script of this course feature a proof allowing you to understand step by step why they are true.
A theorem is similar but distinct, as theorems are typically of greater importance than the proposition, either to the text itself or in the relevant mathematical context. For instance, a mathematical paper would probably call its two to three main results theorems and other related, more technical insights propositions. Moreover, any fact of central importance to a mathematical (sub-) field is likely to be called a theorem, take again the example of the Weierstrass Extreme Value Theorem.
Next, a lemma typically has no immediate value for the insights to be taken away from a text, but rather, it provides a “helper fact” that facilitates proving a proposition. As such, lemmas most frequently occur directly before propositions requiring rather complex, multi-step proofs, and their predominant value lies in organizing the structure of the line of reasoning presented as proof.
Finally, a corollary is something that follows rather immediately – without any or at most with one to two lines of proof – from one or more other facts. But just because corollaries are easy to establish given previous considerations does not mean they are not important, and some very important theorems are indeed corollaries!
What is true for all of the concepts mentioned here is that they give you a mathematical fact. These facts can be complex and/or unintuitive and it may be hard to immediately see why they are true. Naturally, you may therefore ask, when would we expect to see a proof? Well, for any of the concepts, when a text states them for the first time, they are expected to come with a proof immediately below it to allow the reader to judge upon their validity. Further, if the text’s main purpose is educational, proofs are also given for existing results so that they don’t fall from the sky for the reader. If the proof is not too essential for what the text wants to convey, you will frequently see reference to a resource giving the proof. Only if results are sufficiently well-established in the relevant mathematical context (e.g. the Weierstrass Extreme Value Theorem in the context of optimization), you will find that no proof is given at all.
As any good mathematical text should, let us begin our discussion of sets by defining what precisely we will be studying. Since this will be our first definition, the discussion below also outlines some general key insights into reading mathematical definitions.
Definition: Element, Set. A set is a collection of distinct objects, considered as a whole. An object in a set
is called an element or member of
, denoted
. For an object
that is no element of
, we write
.
It is important to make sure you know the meaning of every word in a definition. The emphasis here is on “the”, for mathematical expressions rarely have several meanings, as that could generate misunderstandings. The converse, however, is not true, as one can readily see from our definition where “element” and “member” are synonyms. The knowledge of these meanings is mostly gained by regular interaction with the words. In the above definition, for instance, the word “object” should be understood as “any entity that is of interest to the modeler.” Therefore, depending on the context, objects can be real numbers, but also functions, matrices, geometrical figures, or even sets themselves!
Moreover, in good mathematical definitions, no word is redundant, and the meaning does not go beyond what is written. In our example, the word “distinct” suggests that sets do not contain duplicates: thus, the collection may represent a set, while the one
may not. Moreover, “considered as a whole” suggests that the set itself should be seen as a distinct object. Conversely, the definition says nothing about the order of elements in a set, so that we may infer that the sets
and
are identical.
In terms of notation, you are likely familiar with the way the sets above are written: two curly braces, and within them the characterization of its elements. The word “characterization” is used deliberately rather than “list”, because typically, the sets we deal with are too big to list all its elements or even contain infinitely many of them, consider e.g. the set of natural numbers . More generally, we define sets by a mathematical statement as introduced in the previous section that characterizes the elements. How this works exactly can readily be seen from the definition of intervals given below. Note that it refers to the extended set of real numbers
that encompasses all real numbers,
, as well as
.
Definition: Real-valued Interval. A real-valued interval is a set that contains all in between two thresholds
,
. We denote
If ,
is called open, and if
, we call
closed. Else, we call
semi-open. If
, then the lower bound must be open. Conversely, if
, the upper bound must be open.
As can be seen, in terms of notation, a round bracket indicates that the threshold value is not included in the interval, whereas a square bracket indicates its inclusion.
In set theory, a key concept is the subset. For the sets and
, we say that
is a subset of
, denoted by
, whenever all elements of
are contained in
, formally (
).
is a proper subset of
,
, if all elements of
are contained in
but there is at least one element in
that is not an element of
, i.e.
.
The approach we adopt toward set theory is the so called “naive” approach. It is naive in the sense that it is not axiomatic. For an economist, there are no costs but many benefits to follow this simpler approach. In order to avoid paradoxes (the interested reader can have a look at Russell’s paradox), however, one needs to assume that every set we consider is itself a subset of a fixed, all encompassing, set called the universal (super)set, which we denote by . In addition, one defines an “encompassed by all” subset, called the empty set, and conventionally denoted
. For every set
, we thus have
. The empty set is always the same and contains no elements, while the universal set varies across applications, so that we may have
when considering sets of real numbers, and
for sets of real-valued vectors of length
.
Now it is time to consider some key concepts related to sets. To define them, let be arbitrary sets and
the universal superset.
As with real numbers (addition, subtraction, etc.), we can perform operations on sets:
These operations facilitate our lives greatly in many dimensions: e.g. the somewhat awkward definition of disjoint sets above, where we required that , can be simply re-written as
. The symbol “
” indicates a defining equality, and is used whenever we introduce a new object of interest. (Note that, in accordance with the introduction of “
” in the table on notation and symbols, you can read “Let
” as “let
(be) such that it is equal to the set
” In this sense, “
” is not a new symbol, but rather a combination of two familiar ones!) Alternatively, you will sometimes see “
“.
Many find it helpful to illustrate sets and operations on them using a “circle approach”. Here is an illustration of the set operations, where the circles denote the sets A and B, respectively:
Before moving to slightly more sophisticated issues related to sets, the following table gives an overview of the set notation discussed thus far:
So far, we have not yet explicitly addressed that elements of sets may be anything else than standard real numbers. To address this aspect, consider the power set: When (
) denotes a set, then the power set of
is
, i.e. the set of subsets of
. Note that
for any
, as the empty set is the “encompassed-by-all” set, as introduced above. To give an example,
. Note that this class of sets is subject to a different universal set than
. However, it is easily verified that
is a suitable universal set for the power sets of
,
, as
At times, it may be convenient to give the individual objects in the set an index, so that we may write ,
, or equivalently
(depending on your econometrics background, you may have seen that one writes samples of size
pairs of random variables in similar fashion, namely,
). Of course, we can use a more general index set, denote it by
, that need not be equal to
for an
. We distinguish finite, countable and uncountable index sets. The set is finite if (and only if) it contains only finitely many elements. The distinction between “countable” and “uncountable” is not too important here. When the elements of
are indeed sets, we can elegantly use the index set for short notations for multiple intersections or unions:
Finally, we say that the collection of sets
is pairwise disjoint whenever any two elements of
are disjoint, i.e.
.
As with the operations on real numbers, it is possible to establish a range of properties that set operations satisfy. Let us have a look at the ones most frequently used in economics:
Theorem: Properties of Set Operations. Let for a universal set
and
for an index set
, where
. The following properties hold:
(i) Commutativity: and
.
(ii) Associativity: and
.
(iii) Distributivity: and
.
(iv) Simple De Morgan Laws: and
.
(v) General De Morgan Laws: and
.
These rules are a good opportunity to re-familiarize yourself with the expressions Commutativity, Associativity and Distributivity, and they may also be helpful in developing a better intuition for sets using the circle-approach introduced above – take a piece of paper and see whether you can visually “prove” the simple De Morgan laws!
The last introductory section is concerned with functions and limits. It gives an introduction to functions using the concept of relations, partly for formal precision, but also to remind you that relations, as you may come across in your micro-oriented classes when studying (consumer) preference, are nothing fancy, but just a generalization of the concept of functions.
To understand the concept of relations, consider the Cartesian product of two sets
and
, defined as
Then, a binary relation from
to
is nothing but a subset of
:
, and if
, we say that
is an image of
under the relation
. The relation is binary because any
is either an element of
or not, and there is no (continuous) “degree of relatedness”. We write
or
, where
(1)
Note that the sets ,
, are a complete characterization of the relation
, this will be important in a second. Moreover, for any fixed
,
can be empty or contain multiple arguments. As an example, consider
, where the relation
is defined as the set
where we use the common notation . Then,
, so that
and
for any
. Another example that is frequently discussed in undergraduate economics courses (with varying degree of formality) are preference relations, where
contains vectors of goods quantities, and for a consumer
, the relation is given by
and if
, the consumer (weakly) prefers the consumption vector
to
.
Intuitively, it should be rather natural that we can view a “function” as introduced in high-school courses as a relation, since the values for are related to
through the function. Indeed, this is what we call a function also more formally: any relation that assigns exactly one value
to every argument
. So, if we call
a function, that means that for any
,
must be a single object (e.g. real number, but also vectors, matrices, etc., as we will see later), and not a set!
Let us go over the line of reasoning defining a function as a relation step by step. Once you have understood this, you will be familiar with the names and nature of all the fundamental concepts relevant for a function, including the domain, codomain, image and graph, which are very important for everything to follow!
A function that you are likely well-familiar with is the one of a rule which associates every element
in the domain
of
with a single element
in the codomain
of
. We write
This statement is a concise summary of all relevant aspects of : the domain
, the codomain
, and the rule
that maps
‘s into
‘s. Note that two functions are identical if and only if the mapping
and the domain
coincide; the codomain may well be different (consider e.g.
and
where
is the set of non-negative reals. Then,
and
are clearly identical). To see the connection to relations, consider the graph
of
,
Clearly, is a subset of
, since it contains only elements in
and adds the restriction
, which may exclude some elements. Like this, we can view the graph
as the relation from the domain
to the codomain
, since the set of
‘s related to any fixed
under
, denoted
above (cf. equation (1)), is simply
where the last equality is because is a set and sets do not contain duplicate elements. This highlights that the function assigns only one image
to any one
.
Mini-Excursion: We see here that when viewing relations as a generalization of functions, the set can be interpreted as a generalized image of
under
, in a fashion very similar to standard functions. So, if we wish to define a relation associating multiple images to arguments
, we can use the relation concept in a straightforward fashion to do so (an example are so-called correspondences, where the values are sets).
As take-away, one may summarize
Before moving on, a conceptual note. You may be used to calling “” a function, e.g. from high school. If so, you should stop doing this now. Indeed, people sometimes do this, especially at lower levels of mathematics, but this is arguably imprecise/wrong.
may refer to a specific element in the codomain of
, the value of
when evaluated at a concrete
, or, when considering
as a variable, as the mapping rule
of the function
(You may be familiar with this case from specific representations like “
“, which unambiguously summarizes the mapping rule). However, neither case provides sufficient information to fully characterize
(in the latter, it is still unclear what domain and codomain are), and you run into troubles related to notation when it comes to differentiation (see also the discussion in Chapter 3). To be formally precise, in everything to follow, we will call
the function,
an argument and the object
in the codomain of
the value of
at
, and
the mapping rule of
.
To conclude our investigations into functions here, let us consider some further important concepts that you will come across frequently in the function context. Again, you don’t need to memorize this by heart by now – just try to become familiar with the expressions.
For what follows, let be a function as defined above, and in addition, let
be another function. Then,
The word “range” is frequently used synonymously for the image of under
(also denoted as im
). Further, an alternative name for the preimage is “inverse image”, which may be somewhat misleading and easily confused with the image of the inverse function. Thus, let us not use this label, but be aware that some other texts and courses may do so.
The inverse function will be investigated more thoroughly later, but you can already note that (i) its existence depends crucially on the definition of the codomain as well as the mapping
, and (ii) that despite looking quite similar, the expressions
and
refer to fundamentally different concepts! One is a set in the domain of
that always exists, whereas the other is a value of the inverse function
in the domain
, which is only well-defined if
exists in the first place, i.e. if
is invertible (for the condition, see the list above)! To tell them apart more easily, one sometimes uses square brackets for (pre)images of sets and round ones for (inverse) images of single elements, as is done here. Make sure that you understand this difference!
As a last note on functions, the table below gives common rules for derivatives of functions where both domain and codomain are .
To conclude our investigations into the fundamental background concepts of mathematics that are relevant to the context of the economist, we consider the limit concept in relation to the real line, both for sequences of numbers and univariate, real-valued functions.
Let be a sequence of real numbers, i.e.
. Then, we call
the limit of this sequence if
Verbally, for any, and thus especially any arbitrarily small number , there exists a threshold
after which the sequence elements only deviate from
by less than
, such that eventually, as
, the sequence elements will lie arbitrarily close to
. If the limit
of the sequence
exists, we write
. Crucially, we also write that
if
i.e. if the sequence elements eventually exceed any arbitrarily large but fixed number . A similar characterization can be written down for
. Try to write it down on your own, and click the button below to compare your result.
Because function limits usually receive less attention in undergraduate economics programs than sequence limits, let us now study this issue, which is not quite the same, but as you will shortly see still highly similar.
When , we call
the limit of the function
at
, if
The concept is similar to the standard limit of a sequence: for any arbitrarily small , there must exist a neighborhood
,
such that
deviates from
by less than
on
. In other words, by choosing
sufficiently close to
, one may ensure that
deviates from
no more than
. We write
. Note that we need not have
, so that
can either be a boundary point (e.g.
when
is defined on
) or a point where
is not defined (e.g.
when
). Further, we adopt the convention that if for any sequence
, where
, so that
, it holds that
(
), then we write
(
).
To characterize the asymptotic behavior of a function with domain
or intervals unbounded to one side (e.g.
, etc.), one frequently considers the limits
and
. Here, it is important to know how these quantities are defined. We write
for a
if
Try to write down the analogous formal statement that defines as the left asymptote of
,
.
As with the limit at a point , we write
(
) if for any sequence
, where
, so that
, it holds that
(
), and analogously for
(
).
An important point is that need not necessarily hold. Consider, for instance,
and
, where
is not even defined (
). Next, consider the indicator function
on
that is equal to
if
and zero else. It is defined at
, i.e.
, but for any
and any
, there exists no
such that
for all
because
for
and
for
. Thus,
does not exist, and especially,
does not hold. Finally, even if the limit exists, the equation need not hold. Look at the function
with
that is equal to
at
and zero else. Then
.
Indeed, if , then
features a desirable property called continuity at
. We will have a more rigorous discussion of it later.
Definition: Continuity of Real Functions.
Consider a function ,
. Then,
(i) is called continuous at
if
.
(ii) is called continuous on the interval
if
:
.
An further concept that you may come across frequently is the one of left and right limits. The left (right) limit of at
is the value
takes “when moving towards
from the left (right)”. This is useful for two reasons: (i) we can characterize the behavior of functions like
at points
, here
, where the limit
is undefined, and (ii) the concept provides a rather straightforward method to disprove existence of the limit of
at
. Formally, we say that
is the right limit of
at
if
and is the left limit of
at
if
We write and
. Then, it is easily verified (for
, choose
, or respectively
) that the limit of
at
exists and is equal to
if and only if the right and left limits exist and
. Conversely, this implies that whenever
or either limit does not exist, then
does not exist as well. Try to use this method to show non-existence of the limit of
at
for the specific example of
and
. As a final remark, if they exist, proper limits (
) as well as left and right limits (
) are unique.
To conclude this introductory chapter, let us consider a few rules for limits. Some more simple facts are the following (the right column assumes that the respective limits exist):
Further, if is continuous, then
. Thus, if also
is continuous, then
. The next important fact is L’Hôpital’s rule for the limit of ratios:
Theorem: L’Hôpital’s Rule.
Let and
be two real valued differentiable functions on an open interval
and
, or
. Let
for all
,
. Suppose that
or
. Then, if
exists, it holds that
Thus, we can use derivatives and L’Hôpital’s rule if the product rule does not apply because at least one limit does not exist. Note that when the functions are sufficiently differentiable, you can apply this rule multiple times (i.e., higher order derivatives). An example is . With
, both the numerator and denominator approach
, and the limit product rule does not apply. However, by L’Hôpital’s rule, this limit corresponds to the limit of the derivative’s ratio,
.
A final, important rule with a quite memorable name is the following:
Theorem: Sandwich Theorem (Sequences).
Consider three real-valued sequences ,
and
such that
and
are convergent and
. Further, suppose that there exists
such that
. Then,
is convergent with
.
This theorem is frequently used to avoid involved mathematical considerations using the approach from the definition of the limit. Note that
or
need not necessarily depend on
, for instance, if we have
with
for all
, then we can also establish
from the sandwich theorem. Finally, the “
” part just tells us that it doesn’t matter for the limit if the inequality does not hold for some “early” elements of the sequences, in most applications, you might be lucky enough to choose
, i.e. the inequality holds for all
. As an example, consider the sequence
for
. We can bound
and since , the sandwich theorem allows us to conclude that
.
This theorem holds for limits of functions in an analogous way:
Theorem: Sandwich Theorem (Functions).
Consider three real-valued functions, ,
and
such that for a value
in their domain,
and
exist with
. Further, suppose that for any
in proximity to
, it holds that
. Then,
exists, and
.
Here, it may not be too clear what “in proximity to ” means precisely, at least not formally. To express this fact more formally, we need the distance concept we are to touch upon in the next chapter. As this has not been introduced this point, the vague statement given above shall suffice for now.
If you feel like testing your understanding of the concepts discussed in the second half of this chapter, you can take a short quiz found here.