Tải bản đầy đủ (.pdf) (9 trang)

TAL - A National Database of Questions - Classification is the Key pptx

Bạn đang xem bản rút gọn của tài liệu. Xem và tải ngay bản đầy đủ của tài liệu tại đây (55.5 KB, 9 trang )

1
Introduction 2
Motivation for Creating TAL 3
Why do we use questions? 3
An Automated Test Generator 5
An Approach to Classification 7
Practical classification problems 8
A General Classification Scheme 12
The Mathematics Subject Scheme 12
How to specify a question slot 14
Conclusions 15
Appendix A 15
Appendix B 16
References 17
Maths-CAA Series
May 2003
TAL - A National Database of Questions -
Classification is the Key
Contents
2
3
TAL - A National Database of Questions – Classification is the Key
FDTL awards the idea of shared assessment resources in this form was not
supported perhaps because the case was not made well enough.
At present TAL is used by the University of Bristol and Liverpool John Moore’s
University. As part of a European Leonardo project, ATHENA, the following
Universities: Università degli Studi di Perugia, CPE Lyon, University of Szeged,
Universidad Politécnica de Valencia all used TAL with their students. The
project set a mathematics test at university entrance level and translated the
test into the different languages. The students then all took the same test and
the results were compared. Any other university interested in taking part in this


project is welcome as are schools, which may find questions based on the SEFI
Core Zero syllabus of interest.
2. Motivation for Creating TAL
In the 1970s Bristol University staff used to mark students work every week.
The pressure of work eventually led to the abandonment of this practice and
there was a consequent fall in the mathematical competence of students. The
idea of occasional progress tests was formulated to address this fall in standards
but clearly if we could get a machine to mark students work every week then
all would be well. We conceived of a huge database of questions with students
set tests that were randomly generated from the database. Questions are ex-
pensive to generate and need continual updating. Many people however are
teaching roughly the same material and so the same questions could be widely
used. The concept of a National Database of questions was born. This concept
is much wider than mathematics and so the TAL project includes questions
from many different subject areas, although the largest contributors of ques-
tions are mathematics, chemistry and physiology. As TAL’s home is in Engi-
neering we are also generating a database of Engineering questions stemming
from a project, based in a new degree at Bristol called Engineering Design
(www.fen.bris.ac.uk/engdesign). The Engineering database aims to help stu-
dents understand engineering concepts; and is particularly aimed at questions
that do not require detailed calculation but require a feel for the subject.
2.1 Why do we use questions?
The main goal of education is that students should learn. To learn mathematics
they need to try their hand at problems. We want students to understand the
concepts of mathematics but such concepts are often quite complex. When
students can understand how to do a question involving a concept, the under-
standing may be quite one-dimensional, i.e. they can use the concept only in
that sort of question. In order to help students to have a richer understanding of
TAL - A National Database of Questions -
Classification is the Key

Jon Sims Williams and Mike Barry
Dept Engineering Maths, Bristol University
Email: ,
Abstract: This paper discusses a CAA system called TAL. TAL is unusual in that
it allows users to generate large numbers of equivalent tests from a specifica-
tion. The tests are generated from a database of questions and all questions
must be classified. Some of the difficulties involved in classifying questions
are discussed.
1. Introduction
The Test and Learn system, TAL is a database of questions with facilities for
staff to set tests and students to take tests. It was first conceived in 1994. We
built a system, which allowed students to run a programme on their PC linked
to a remote computer. The system collected a test from the remote computer,
ran the test locally on the PC and then sent the results back to the host compu-
ter. The system used the basic systems of computer-to-computer direct links
that were available at that time and worked very reliably. A bid to JISC was
successful (Higher Education Funding Committee JISC Technology Applica-
tions Project 2/352) in 1996. With this support we built a web-based version
that was first used in 1997. The task of populating the database with a large
enough set of questions was perceived to be too onerous a task for one univer-
sity or department and so the system was built to allow several universities to
use the same set of questions in the hope that participating staff would gradu-
ally add to the database and we would all benefit. This is not, of course, the
first time that the idea of sharing questions has been mooted and early in the
TAL project we visited Glyn James and Nigel Steele at Coventry University
and were given the paper results of a previous attempt to share questions.
Martin Greenhow of the University of Brunel has kindly offered his set of ques-
tions, Mathletics to be included in the database. The normal history of shared
sets of assessment questions is that several people are very keen and offer
questions at the outset; others are very keen to accept the sets of questions and

then the whole project dies. We hope that by having the database of questions
open for all to use remotely over the Internet, this project may yet succeed;
however it really needs full support and maybe central funding by the LTSN
subject centres. If the LTSN Generic Centre accepted the concept of a generic
database then that would be even better as we already support questions from
three subject areas, mathematics, chemistry and physiology. In the recent
5
TAL - A National Database of Questions – Classification is the Key
4
Jon Sims Williams and Mike Barry
that all the tests are similar.
3. An Automated Test Generator
We first conceived of the automatic test generator just picking questions ran-
domly and generating tests with about the same length and difficulty. So one
would choose a topic: “Differentiation” and then all the tests would be gener-
ated from questions with this classification. This was relatively simple to do
but pedagogically it did not work – teachers need to craft their tests so that the
questions are more precisely chosen. The problem of random choice showed
up when some tests had two or more questions that were similar, while others
had quite different questions.
Automatic test setting is wonderful for the teacher in that several tests are set
very quickly, but more control is needed in the formation of tests. The current
system allows the user to set a test by specifying the topic or topics to be used
for each question in the test. The concept is illustrated by the short test below:
Example test Design
Test Name: test1 Course Unit: EMAT1011
Setter: Jon Sims Williams Dates available: 3/5/02-15/5/02
Length of test: 11:34 +/- 2 minutes Facility of test: 60 % +/- 5%
Figure 1: A simple specification of a mathematics test.
Question No. Main Topic SubTopic1 SubTopic2

1
functions
understanding form of
graph
2
functions domain and/or range
simple algebraic
functions
3
differentiation practical application speed
4
differentiation
function of function
rule
trigonometric functions
concepts, we expose them to a wide variety of questions.
The goal is clear. We know that students find it helpful if they can measure
themselves up against a standard to see if they are doing well, and pedagogically
we want to expose students to a rich variety of problems to reinforce their
understanding. To reach this goal we need to motivate the students to do the
questions and we need to provide the questions.
There are two ways of providing questions:
1. Generate questions parametrically where the questions all have the same
structure but have different numbers.
2. Provide a large number of different sorts of questions on the same topic.
Method 1 has two advantages:
• Students gain confidence because they can practice until they succeed on
a familiar type of question
• It is much cheaper to generate questions parametrically
Method 2 gives students lots of practice but as each question is different they

do not learn how to answer questions by drill or formula but rather gain a more
multi-dimensional understanding of the topic. The disadvantages of this ap-
proach are the converse of using method 1. Most importantly, students may get
discouraged if the questions are all so different that they cannot build on the
understanding they get from successfully completing one question in the next
question they attempt. Ideally the questions lead on one from another increas-
ing the scope of understanding without being too challenging. A good teacher
can do this by hand selecting questions from the database. Automatic support
is more difficult; it could be done but at present we only collect and use data
on the relative difficulty of questions.
We use questions for motivation. However just providing questions is not enough.
Many students seem to need the motivation of getting some marks before they
will put the effort into trying questions. As soon as you give marks, students
complain if they think that other students are in some way cheating – yet to be
most effective in supporting learning we want to offer access to the tests from
students’ homes and halls of residence. Probably the most effective resolution
of this contradiction is to set a suite of tests all of which are on the same set of
topics and then tell students that one of these tests will be used for a supervised
test and the other are available for them to use for practice. One of TAL’s
special strengths is that it allows you to set tests according to a specification so
7
TAL - A National Database of Questions – Classification is the Key
6
Jon Sims Williams and Mike Barry
4. An Approach to Classification
The design of a classification scheme is primarily influenced by the needs of
users. There are various publicly available classifications of mathematics [2 -
6] but none of them has as its prime aim the need to classify questions for
teachers to use.
The users are setting a test from a large database of questions, so they must be

able to find questions that suit their needs. Teachers need to be able to find
questions on the subject that they are teaching and at the same time they need
to know if the questions will require any other expertise than just the specified
subject. So if we take some examples:
1. The basic skill of differentiating a polynomial.
2. The use of differentiation of a polynomial to find its maximum.
3. Given a simple model of the position of a ball expressed as a polynomial,
find out where the ball is travelling fastest.
4. Invent a simple model for a sledge sliding down a hill with a linear
resistance to motion. Find the point where the sledge travels fastest.
Now all these problems involve differentiation of a polynomial, so they could
all be classified under the same subject; however the questions require increas-
ing levels of modelling skills. In the classification scheme of the previous
section, these questions were classified as follows:
1. Mathematics/Calculus & Analysis/calculus/differentiation/polynomials
2. Mathematics/Calculus & Analysis/calculus/differentiation/max;min;
stationary points/polynomials
3. Mathematics/Calculus & Analysis/calculus/differentiation/max;min;
stationary points/Model Application/polynomials
In terms like “max;min; stationary points”, the semi-colon is used to mean
“and/or”.
If the user specified that he wanted to choose questions on …/differentiation
then all these questions would have been selected. The user could however
have excluded questions requiring the higher skills by excluding, for example,
questions classified as:
Mathematics/Calculus & Analysis/calculus/differentiation/max;min; sta-
tionary points.
This test has only four questions and realistically one should never set such
short tests as their reliability is low; however here we have chosen to set two
questions on functions and two on differentiation. Rather than just choose the

questions on differentiation randomly we have used two additional classifica-
tions to be more precise; so the first question is using differentiation to find a
speed in a practical application. The second question specifies that the differ-
entiation requires the understanding of how to differentiate functions of func-
tions and the functions concerned are trigonometric functions. So the setter of
the test has a clear idea about the sort of questions s/he is setting. Lastly two
additional requirements are that the test should take 11.34 ± 2minutes and the
facility of the test should be 60% ± 5%. The facility is the percentage of
students who normally can get a particular question right. This is averaged
over all the questions, and each lecturer can have a different set of facilities
and time to do for each class taught. The test generator searches the database
for suitable questions and generates as many tests as it can to satisfy the speci-
fication.
We like to set about 20 different tests satisfying the specification and in order
to do this it may be necessary to relax the specification a little and allow a
wider range of questions to be used for each question slot. The specification
can be relaxed by saying that the type of function used in the function of
function differentiation does not matter – just don’t specify it. Similarly we
could allow any sort of application of differentiation not just questions based
on speed. Other relaxations could be to specify another sort of question that
could be used for each question slot e.g. questions on differentiation using the
chain rule could be used as well as questions on the function of a function rule.
Figure 1 is illustrative of the idea of how tests are specified but MainTopic,
SubTopic1 and SubTopic2 have no particular meaning as will be shown in later
examples. The whole basis of this test generator is that all the questions are
classified and the classification of questions uses a tree structure. The ques-
tions slots in the example above as specified as:
1. Mathematics/Calculus & Analysis/functions/Understanding form of graph
2. Mathematics/Calculus & Analysis/functions/domain;range/simple
algebraic functions

3. Mathematics/Calculus & Analysis/calculus/differentiation/practical
application/ speed;acceleration
4. Mathematics/Calculus & Analysis/calculus/differentiation/ function of
function rule/ trigonometric functions
The ‘Mathematics’ is not redundant as the database contains questions on other
subjects. A semicolon is interpreted as ‘and/or’.
9
TAL - A National Database of Questions – Classification is the Key
8
Jon Sims Williams and Mike Barry
tions. If a teacher had taught the idea of ‘ill-conditioned’ as part of understand-
ing about solving equations and the errors that can occur he would want to find
the question classified under terms such as ‘solution of equations’. Now the
SEFI Core Zero syllabus, [2], operates at the school-university interface level
and has syllabus items that can be described by the classifications:
Algebra/Linear Laws/understand the terms ‘gradient’ and ‘intercept’ with
reference to straight lines
Algebra/Linear Laws/recognise when two lines are parallel
Algebra/linear Laws/obtain the solution of two simultaneous equations
in two unknowns using graphical and algebraic methods
These are all very much in the right area but incomplete as there is no refer-
ence in [2] to ill conditioning. If we introduced a classification under SEFI
Core Zero syllabus about ill-conditioning then one could no longer set tests
from it knowing that all questions were in the syllabus. The right response is
probably to say that one must know about ill-conditioned matrices to answer
this question and classify the question as:
Maths/Algebra/ Linear Algebra/Matrices/Matrix
properties/ill-conditioned (4.1.1)
It is possible that people would want to teach about ill conditioned equations
in a way that does not involve looking at matrices at all, so a classification:

Maths/Algebra/ Linear Algebra/linear equations/
ill-conditioned equations (4.1.2)
may also be viable.
Because of the need to be comprehensible to the target audience: teachers and
university lecturers, we have adopted a twin approach to classification of math-
ematics questions. Firstly, we have used the SEFI Core Zero curriculum for
engineering mathematics [2] edited by Leslie Mustoe and Duncan Lawson. As
part of a Leonardo de Vinci project called Athena 2000/1, a set of questions
based on the classification in [2] was generated and a TAL test on these ques-
tions was taken by students in different European countries. This syllabus is
useful for both school sixth forms and some first year university work. Sec-
ondly, we have adapted the Eric Weisstein’s World of Mathematics classifica-
tion [3]. This classification is aimed at helping users to find the meaning of
You will notice that there is a citation order used here where “max;min; sta-
tionary points” comes before “Model Application” or
“SimpleModelConstruction”. The normal rule in classification schemes is to
place “general” classifications before “special” ones. However
“SimpleModelConstruction” is a more general classification than “max;min;
stationary points” as it could be applied to all sorts of problems not just differ-
entiation problems, but we need the citation order to relate to the way in which
the teacher will think. Normally a teacher will choose to teach differentiation
and then when the subject is well established will start to ask students to use
the process of differentiation on “real” problems involving some sort of model.
Rather arbitrary decisions have been made, as a teacher might want to teach
modelling and then choose what sort of mathematical techniques could rea-
sonably be used by the students. At present we allow two classifications for
each question so it would be possible to implement both approaches.
The classification, as you can see, is mixing together a subject classification,
“Mathematics/Calculus & Analysis/calculus/differentiation/”, with a problem
type description such as “practical application” and the sort of application,

“speed;acceleration” and then sometimes there is information about the sort of
functions that are used in the question, e.g. trigonometric functions. The test
setter needs to be able to specify questions using all these sorts of data, but a
single tree structure containing all this is not ideal. A more general approach
to classifying questions is described in section 5.
4.1 Practical classification problems
In this section example questions are displayed and then an approach to as-
signing a classification is discussed.
Example 1
Two linear equations are said to be ‘ill-conditioned’ when:
Answer 1: They cannot be solved
Answer 2: They represent two parallel straight lines
Answer 3: They represent two nearly parallel straight lines
Answer 4: They represent lines of which one is parallel to either the x
or y axis
This question is posed in terms of equations and the answers are all in terms of
the geometric understanding of the equations. In the mathematical literature
however, the term ‘ill-conditioned’ applies to matrices rather than sets of equa-
11
TAL - A National Database of Questions – Classification is the Key
10
Jon Sims Williams and Mike Barry
Answer 2: The adjoint and the minor of the element a
ij
are identical
when i + j is odd
Answer 3: The cofactor and the minor of the element a
ij
are identical
when i + j is odd

Answer 4: The cofactor and the minor of the element a
ij
are identical
when i + j is even

This question needs to be classified under both minors and adjoints, but in the
classification hierarchy we have:
Algebra/Linear Algebra/Determinants/minor or cofactor
and
Algebra/Linear Algebra/Matrices/Matrix Operations/Adjoint
because the adjoint is the transpose of the matrix obtained by replacing each
element a
ij
by its cofactor A
ij
so it is a matrix operation.
So once again we need to decide on a major classification and then have sub-
classifications to show what additional knowledge is required. This question is
easy to sort out as one cannot fully understand it without understanding Adjoints
and one cannot understand Adjoints without understanding cofactors, so classi-
fying the question under Adjoints only is satisfactory.
This example points out a problem that questions often cross many boundaries
in a simple subject classification. We need to find a classification scheme that
allows a teacher to pick out questions that are on the topic being taught, but
then finds it easy to remove questions that require additional knowledge that
has not yet been taught.
It should be clear from these examples that the classification of questions is a
complex process. When it has been well done however, the setting of tests is
easy. Because the process is so complex we need a team working together
rather than rely on one or two people.

The classification, as you can see, is mixing together a subject classification,
“Mathematics/Calculus & Analysis/calculus/differentiation/”, with a problem
type description such as “practical application” and sort of application,
“speed;acceleration” and then sometimes there is information about the sort of
functions that are used in the question, e.g. trigonometric functions. The test
setter needs to be able to specify questions using all these sorts of data, but a
terms. So when we look up Linear Equation we find a definition in terms of y
= ax +b. Now, a part of what a teacher wants in terms of questions about
Linear Equations will be the definition; but s/he will also want to know how to
solve them and what goes wrong when you solve them etc. So we have built
upon the Weisstein’s classification in accordance with our experience of teachers
needs. Part of both the SEFI and adapted Weisstein’s syllabus are shown in the
Appendices.
The system allows multiple classifications so both of the interpretations (4.1.1)
and (4.1.2) are implemented. If you look at the Weisstein’s classification on
the Web you will find that (4.1.2) does not exactly fit.
Example 2
The equation of the straight line that passes through the point (1,2,3) and
is parallel to the line joining the points (2,3,1) and (5,4,2) is?
This question requires two ideas: the student needs to be able to form Displace-
ment Vectors, so that s/he knows the direction of the line and needs to be able
to write down the equation of a straight line using its direction and a point
through which is passes. So there are two possible classifications: Displace-
ment Vectors and Equations of lines. This question is really about equations of
lines and just incidentally requires one to know how to form displacement
vectors, so we would classify the question as:
Algebra/vector algebra/equations in vector form/equations of lines
But students also need to know about the displacement vectors so we put this
underneath equations of lines as other ‘required knowledge’.
Algebra/vector algebra/equations in vector form/equations of lines/dis-

placement vectors
A new scheme has been invented to avoid treating ‘required knowledge’ in this
way but it is not yet fully implemented.
Example 3
For an n x n matrix A, which of the following statements is true?
Answer 1: The adjoint and minus the minor of the element a
ij
are identi-
cal when i + j is even
13
TAL - A National Database of Questions – Classification is the Key
12
Jon Sims Williams and Mike Barry
Application Area - this tree allows the editor to describe the application
area e.g. physics/motion/speed;velocity.
In addition we use the following lists, where only one selection is made from
each list:
1. Modelling level: this classification only applies to certain questions. We
distinguish the following modelling levels:
• Simple Model Construction - a simple model must be constructed to
complete the question.
• Model Application - a model is provided as part of the question and
must be used
• Practical Application - this indicates that the student is expected to be
able to use the particular mathematics described in the Subject tree
pointer in a practical situation. Implicitly this will usually mean that
they have to use a standard model used in this sort of application.
2. Evaluation Type: This takes one of two values: Numeric or Algebraic, and
tells us if the question requires the answer in a numeric or algebraic form.
3. Bloom’s Educational Objectives: />mathematicsweb/show/Index.htm are a classification of the type of

question thus:
• Knowledge: states that the question simply checks if the student knows
something factual.
• Comprehension: usually refers to a question in which data is given and
the question checks if the student understands what it means.
• Application: The use of previously learned information in new and
concrete situations to solve problems that have single or best answers.
• Analysis: The breaking down of informational materials into their
component parts, examining (and trying to understand the
organizational structure of) such information to develop divergent
conclusions by identifying motives or causes, making inferences, and/or
finding evidence to support generalizations.
• Synthesis: Creatively or divergently applying prior knowledge and skills
to produce a new or original whole.
• Evaluation: Judging the value of material based on personal values/
opinions, resulting in an end product, with a given purpose, without real
right or wrong answers.

Lastly the numeric field is used to describe how theoretical the question is.
This is very important, as a question posed in an abstract manner on a simple
tree structure containing all this is not ideal. A more general approach to
classifying questions is described in the next section.
5. A General Classification Scheme
TAL is conceived as a question database for many different subjects, so we
have developed a database structure that allows each subject area to have its
own subject classification Essentially each question is classified under one or
more top level subjects e.g. mathematics and medicine. Then for each of
these top-level subjects the subject editor is able to specify three lists, three
trees and a numeric to describe questions. The questions can have multiple
descriptors taken from these lists and trees. Each list has a subject specific title

so that the users understand the interface in the language suited to the subject
area. So in mathematics we are interested in the set of function types used in
a question, but in chemistry they are not. Instead a list of question descriptors
has been proposed:
Question_descriptor = { Law or definition, Nomenclature, Qualitative test,
Reactions & reactivity, Structure, etc}
Rather than go into the details of the computer representation of this, the scheme
will be described from a mathematical viewpoint. The scheme described be-
low has been designed and built by a student but as yet it is not implemented
in the production system.
5.1 The Mathematics Subject Scheme
For mathematics we have selected three trees called:
Subject tree - for the subject tree we have adopted two schemes: one
based on Eric Weisstein’s World of Mathematics classification [3], this is
the same as has been described in most of the examples above. There are
considerable benefits from having a classification that is related to a syl-
labus and so in the ATHENA project we used the SEFI syllabus, [2]. If you
use the SEFI syllabus you know that the questions you find are in the
syllabus.
Function Type - this allows questions to be described in terms of the types
of functions used e.g. {simple algebraic functions, trigonometric func-
tions, exponential/log functions, inverse/hyperbolic functions}. Since this
structure is a tree it allows a fine description of function types although
only types given in the list are used in classification at present.
15
TAL - A National Database of Questions – Classification is the Key
14
Jon Sims Williams and Mike Barry
place questions that the user does not find appropriate. However it would be
desirable to be able to control the ‘facility’ and perhaps the ‘time_to_do’ of

some of the question slots in the test; so we propose to allow the user to specify
the minimum facility and maximum length of questions in a question slot.
Finally the test setter can select the average facility and length of the whole
test and how accurately these two conditions must be satisfied. So the test can
be specified as running for an hour plus or minus 5 minutes and similarly the
average facility can be required to lie in the range (55,65) for example. The
test compiler will then find all the questions that match the specifications for
each question slot and try to generate the number of tests requested.
6. Conclusions
The TAL system has been running since 1997. Initially a very small number of
questions were available to set mathematics tests but now we have nearly
2000 questions mainly suited to first year university mathematics. There are
another 2000 questions on other subjects. Automatic test setting is wonderful
in that it saves considerable staff time and the availability of multiple tests
allows students to get practice and be motivated. It is however easy to be
sloppy about setting tests automatically and we have proposed some addi-
tional controls in section 5 to make it easier to build appropriate tests. For the
TAL database to be a really good resource it needs editors for every sub- area of
mathematics to maintain the standards and encourage development of new
questions. It is doubtful if any university can really maintain such a resource
and so a collaborative effort is needed with several universities collaborating
in its development.
Appendix A - A part of the SEFI classification
Analysis and Calculus
Rates of change and differentiation
Average & instantaneous rates of change
Definition of derivative at point
Derivative as instantaneous change rate
Derivative as gradient at a point
Difference between derivative & derived function

Use notations: dy/dx, f(x), y etc.
Use table of simple derived functions
Recall derived functions for simple fns
Use multiple & sum differentiation rules
subject may be quite inappropriate. The scheme used for this Theoretical Nu-
meric is:
0 or 1 a pure numbers question
2 the question is about basic rules or nomenclature of the subject
3 or 4 there is a simple parameter in the question
5 or 6 the question requires a simple theoretical understanding of the
subject
7,8 or 9 the question requires a proof of some theoretical property
We have had very little experience in using these although quite a number of
questions have been classified on this scale.
5.2 How to specify a question slot
Section 3 describes the way tests are set with the existing system. In this
section we describe the use of the new classification scheme. Tests are set by
specifying a number of question slots sufficient to form a test of the required
length. If the questions used have an average time_to_do of 5 minutes then 12
question slots will be needed to form a test that is an hour long. Each question
slot is defined by specifying:
• The subject or a disjunction of subjects from the subject tree
• The set of permitted function types
• If the question is to be an application problem then choose the application
area
• If the problem is a modelling type of question then choose the level of
modelling
• The Bloom category for the question and the enumeration type {numerical
or algebraic} can be set
• Finally a constraint on the level of the theoretical numeric can be set.

When the question slot has been specified the database is searched to find all
the questions that satisfy the specification and the additional subject knowl-
edge that is implicit in the set of questions are displayed. The test setter may
then select any of these subject areas to exclude them the test.
From our experience of using automatic test setting we find that just control-
ling the content of question slots is not sufficient for good test design. At
present the system allows the user to look at each question in each of the tests
set and, provided there are spare questions, to swap spare questions in to re-
17
TAL - A National Database of Questions – Classification is the Key
16
Jon Sims Williams and Mike Barry
Theory
Total differential
Vector Derivative
Increasing & Decreasing
Monotone Decreasing
Monotone Increasing
Monotonic Function
Etc.
Mean-Value Theorems
etc.
9. References
[1] TAL – the Bristol University “Test and Learn System” at
www.tal.bris.ac.uk.
[2] The SEFI core curriculum for engineering mathematics http://
learn.lboro.ac.uk/mwg/core.html
[3] Eric Weisstein’s World of Mathematics classification http://
mathworld.wolfram.com/
[4] MathematicsWeb - />show/Index.htt this is basically about Journals but has a 3 level

classification e.g. maths/applied maths/numerical analysis.
[5] Zentralblatt MATH - />[6] 2000 Mathematics Subject Classification />[7] Bloom’s Educational Objectives />mathematicsweb/show/Index.htt
Use the product rules of differentiation
Use the quotient rules of differentiation
Chain rule for differentiation
Relation of gradient and derivative
Equation of tangent & normal to graph
Stationary points, maximum and minimum values
Find if function is.increasing using differential
Define a stationary point of a function
Distinguish turning and stationary point
Locate a turning point using derivative
Classify Turning Points by 1st derivative
Find second derivative of simple functions
Classify stationary Points by 2nd derivative
Appendix B - A part of the General Classification Scheme
Calculus and Analysis
Calculus
Differentiation
Chain;product;ratio Rules
Differentiability
Directional derivative
Equations of Lines
Function of function rule
Gradients;slopes in R2
higher derivatives
Implicit;log differentiation
Introductory
Maclaurin series
Max;min or stationary points

Global Maximum
Inflexion point
Local Minimum
Saddle point
Numerical differentiation
Parametric equations
Partial Derivative
Second derivative test
Tangents;normals
Taylors series

×