1. Introduction and Basic Concepts
Titrimetry
(also known as titration or volumetric analysis) is a fundamental quantitative
analytical technique used to determine the amount or concentration of a
substance in a sample by reacting it with a measured volume of a standard
reagent. In a typical titration experiment, a solution of known concentration
(the titrant) is gradually added from a burette to a solution containing
the analyte until the reaction between them is complete[^1]. The point at which
the analyte has completely reacted with the titrant is the equivalence point,
defined by the condition that chemically equivalent amounts of reactant have
been mixed. In practice, an observable signal is used to indicate that the
equivalence point has been reached – this observable change is called the end
point of the titration[^2]. The end point may be signaled by a visible
indicator color change or by an instrumental reading (such as a sudden change
in voltage, current, or other physical property). The difference between the
end point and the true equivalence point represents a small titration error,
which analysts seek to minimize by choosing an appropriate detection
method[^2]. Titration techniques are valued for their accuracy and simplicity,
and they continue to be widely used in laboratories as reliable quantitative
methods – especially when coupled with modern instrumental end-point
detection[^3]. Notably, titrimetric analysis has a long history: it has been
practiced since at least the 18th century and was formally described in
textbooks by the mid-19th century[^4]. Despite the advent of many other
analytical technologies, titration remains a definitive method in analytical
chemistry due to its precision, cost-effectiveness, and robustness.
Chemists classify titrations according to the type of chemical reaction
involved and the method of end-point detection. The most common classes of
titrimetric methods are acid–base titrations, oxidation–reduction (redox)
titrations, complexometric titrations, and precipitation titrations, each of
which relies on a different reaction chemistry[^1]. In an acid–base titration,
an acidic analyte is titrated with a basic titrant (or vice versa) until
neutralization occurs; a pH-sensitive indicator or pH electrode is used to
signal the completion of the reaction. Redox titrations involve electron
transfer reactions and often use oxidation-state indicators or self-indicating
reagents (e.g. permanganate ion) to mark the end point. Precipitation
titrations rely on the formation of an insoluble precipitate during the reaction
– for example, chloride ion can be titrated with silver nitrate until silver
chloride precipitates, and the first excess of silver is detected by a colored
precipitate with an indicator[^5]. Complexometric titrations involve
formation of a stable complex between the analyte (usually a metal ion) and the
titrant (often a chelating agent like EDTA); these use specialized indicators
that form colored complexes with the metal ion, such that the color changes
when the metal is fully bound by the titrant[^6]. Each of these titration types
has specific indicators and conditions to ensure a sharp end point. In cases
where a suitable visual indicator exists, classical titrations can be performed
with simple laboratory glassware. Where visual indicators are not available or
the color change is too subtle, instrumental methods are employed to
determine the end point by monitoring a physical property of the solution (such
as electrical potential, conductivity, absorbance, etc.), as discussed later.
It should be noted that the term titrimetric analysis is
slightly broader than volumetric analysis. Classic titration techniques
measure the amount of titrant by volume (hence volumetric analysis), but
titrimetric methods can also measure the amount of titrant by other means – for
example, by mass in gravimetric titrations or by electrical charge in
coulometric titrations. In other words, titration does not strictly require
volume measurement; what matters is that an equivalent amount of reagent is
added to fully react with the analyte[^3]. In practice, however, most
titrations in the laboratory are volumetric, using calibrated glassware to
deliver precise volumes of titrant. The versatility of titrimetry lies in its
ability to yield accurate and precise results with relatively simple apparatus
and straightforward calculations based on reaction stoichiometry. This has made
titration a cornerstone of quantitative chemistry education and a workhorse
technique in industrial quality control and academic research.
2. Practical Foundations
of Titrimetry
Performing
a titration requires careful technique and properly calibrated equipment. The
core apparatus in a classical titration includes a burette (a graduated
glass tube with a stopcock) to deliver the titrant, a flask (typically an
Erlenmeyer flask or beaker) containing the sample solution, and often a pipette
or volumetric flask to measure and transfer the sample aliquot. Supporting
equipment like a burette stand and a white tile (placed under the flask to
better observe color changes) are also commonly used. The reagents include the
titrant solution of known concentration (also called a standard solution)
and an indicator if a visual end point is used. Figure 1 illustrates a
typical titration setup and procedure, from filling the burette with titrant to
the point where an indicator changes color signaling the end point.
In this simple acid–base titration example, the burette delivers a standard
NaOH solution into an acidic sample with phenolphthalein indicator, resulting
in a pink color at the end point. Titration experiments do not require highly
sophisticated instruments; the essential requirements are accurate volume
measurements and a means of detecting the end point[^7]. Nonetheless, careful
handling is vital: the burette must be clean and free of air bubbles, the
volumes must be read at eye level to avoid parallax error, and the titrant
should be added slowly (especially near the expected end point) with continuous
swirling of the sample flask to ensure thorough mixing. These practical
considerations help improve the accuracy and reproducibility of titrimetric
analyses.
A critical aspect of titrimetry is the use of standard solutions
– reagents of accurately known concentration. Such solutions are typically
prepared using a primary standard, which is a compound pure enough and
stable enough that it can be weighed out directly to prepare a solution of
known concentration[^8]. Primary standards should have a known formula, high
purity, stability in air (non-hygroscopic, etc.), and reasonably high molar
mass (to minimize weighing errors). Examples of primary standard substances
include anhydrous sodium carbonate (for acid titrations), potassium hydrogen
phthalate (KHP, for base titrations), silver nitrate (for halide titrations), and
potassium dichromate (for redox titrations). Not all reagents can serve as
primary standards; for instance, sodium hydroxide pellets absorb moisture and
carbon dioxide from air, and solutions of NaOH slowly react with CO₂, so NaOH
is not used as a primary standard. Instead, NaOH solutions are secondary
standard solutions: their concentrations must be determined by
standardizing against a primary standard (e.g. KHP) before use[^8]. Similarly,
reagents like potassium permanganate (which can decompose over time) act as
secondary standards standardized by titration with a primary standard such as
oxalic acid or sodium oxalate[^9]. Accurate standardization is an essential
step in titrimetry because the reliability of the analytical result directly
depends on knowing the titrant concentration precisely.
Once a standardized titrant is available, a typical titration procedure
involves the following steps: (1) Sample preparation: A measured volume
of the sample solution (or a dissolved solid sample) is placed into the
titration flask. Sometimes the sample is pre-treated (for example, by adding a
buffer or a reagent to adjust pH, or an indicator is added at this stage). (2) Titrant
addition: The titrant is slowly added from the burette to the sample
solution. The solution is continuously mixed by swirling. As the titration
progresses, the analyst watches for the end-point signal (indicator color
change or instrument reading). (3) Detection of end point: When the end
point is reached (e.g., the indicator just changes color persistently), the
titration is stopped. The volume of titrant delivered is recorded by noting the
burette reading before and after titration. (4) Calculation: Using the
titrant volume and concentration, along with the reaction stoichiometry, the
amount of analyte in the sample is calculated. The fundamental calculation is
based on the reaction’s mole ratio: at the equivalence point, moles of titrant
= moles of analyte (if the reaction is 1:1, otherwise the stoichiometric ratio
is used). For example, if an acid HA is titrated with NaOH, the point of
neutralization satisfies moles HA = moles NaOH added; from the volume of NaOH
and its molarity, one computes the acid concentration. These computations yield
results like the concentration of the analyte or the purity of a substance,
often reported with respect to the sample volume or mass.
In some cases, a back-titration (or residual titration) is
performed instead of a direct titration. A back-titration is useful when the
reaction between the analyte and titrant is slow or does not have a clear end
point, or when the analyte is in a non-soluble form. In a back-titration, a
known excess amount of a standard reagent is added to the sample to fully react
with the analyte; then the excess of that reagent is titrated with a second
standard titrant. The difference between the amount added and the amount
back-titrated corresponds to the analyte. This indirect approach can often
improve accuracy for certain systems. Whether using direct or back titration,
the reliability of titrimetry comes from combining stoichiometric reactions
with precise volumetric (or other) measurements, making it one of the classical
yet powerful methods of analysis.
3. Titrimetry with
Chemical Endpoint Determination
In
classical titrations, the completion of the reaction is signaled by a chemical
indicator or some inherent property of the reacting system, without the
need for electronic instruments. Such methods rely on human observation of a
change in the solution – typically a color change, the appearance/disappearance
of turbidity, or some other visible event. The choice of indicator or detection
method is tailored to the type of titration reaction. We discuss the main types
of titrations with chemically determined end points below.
3.1 Acid–Base
Titrations (Neutralization Titrations)
Acid–base
titrations involve the reaction of hydronium ions (H^+) and hydroxide ions
(OH^−) to form water. The analyte is an acid or base, and the titrant is a
standard base or acid of known concentration. The most common indicators for
these titrations are pH indicators – weak organic acids or bases that
exhibit different colors in their protonated and deprotonated forms. The
indicator is chosen such that its color transition range overlaps the pH change
at the equivalence point of the titration. For example, phenolphthalein (a
common indicator) is colorless in acidic solution and turns pink in basic
solution; it changes color around pH 8–10, making it suitable for titrating
strong acids with strong bases (where the equivalence pH is ~7-9). Another
example is methyl orange, which is red in acidic solution and yellow in
alkaline solution, with a transition range around pH 3.1–4.4, useful for strong
acid–weak base titrations. A wide variety of pH indicators are available, each
with a distinct transition range and color change[^10]. The titration curve of
an acid–base titration (plot of pH vs. titrant volume) typically shows a sharp
change in pH near the equivalence point, which justifies the use of an
indicator that changes color in that steep region. At the end point, the sudden
color change (for instance, the first permanent appearance of a faint pink in
phenolphthalein for an acid titration) signifies that the amount of titrant
added is chemically equivalent to the amount of acid/base in the sample.
Acid–base titrimetry is widely used for determining the concentrations of acids
(e.g. acidity of vinegar) or bases (alkalinity of water, ammonia content,
etc.), and for assays of industrial products. It is simple, rapid, and accurate
when proper indicators are used and is a staple method in analytical chemistry
education.
3.2 Precipitation Titrations
In a precipitation
titration, the reaction between titrant and analyte produces an insoluble
precipitate. A classic example is the titration of chloride ions with a
standard silver nitrate solution (known as argentometric titration). The
reaction Ag^+ + Cl^− → AgCl(s) removes chloride from solution as solid silver
chloride. The challenge in precipitation titrations is detecting the exact
point at which the analyte is fully precipitated and the next drop of titrant
produces a slight excess of titrant in solution. This is often achieved by
using an indicator that responds to the first excess of titrant. In the
chloride titration example, one common indicator is chromate ion (CrO_4^2−) in
the form of potassium chromate added to the analyte solution. During the
titration, as long as chloride is present, Ag^+ preferentially precipitates it
as white AgCl. Once chloride is exhausted, additional Ag^+ reacts with chromate
indicator to form a red-brown precipitate of silver chromate (Ag_2CrO_4),
signaling the end point by a distinct color change in the precipitate
appearance[^5]. This method is known as Mohr’s method for chloride. Other precipitation
titration indicators include adsorption indicators such as fluorescein
derivatives (used in Fajans method): these are organic dyes that change color
when they adsorb onto the surface of the precipitate, which happens when a
slight excess of titrant appears (changing the charge of the precipitate
surface). An example is the titration of chloride with AgNO_3 using
dichlorofluorescein; near the end point, AgCl precipitate particles adsorb the
dye anion and a discernible color shift (usually to pink) indicates the end
point. Precipitation titrations are used for halides (Cl^−, Br^−, I^−), certain
metal ions (e.g. using sulfate precipitation as BaSO_4), and other ions that
form insoluble salts. They require the formation of a precipitate with well-behaved
solubility characteristics and a clear indication of slight excess titrant.
Modern methods sometimes monitor changes in solution turbidity or use
photometric measurements for more precise end-point detection, but classical
visual indicators remain effective in many cases.
3.3 Complexometric Titrations
Complexometric
titrations are based on the formation of a soluble but well-defined complex
between the analyte (typically a metal ion) and the titrant (usually a
multidentate ligand). The most important complexometric titrations involve EDTA
(ethylenediaminetetraacetic acid) or its disodium salt as the titrant,
which can form stable 1:1 complexes with many divalent and trivalent metal
ions. These titrations are widely used to determine water hardness (calcium and
magnesium content), metal ion concentrations in solution, and composition of
metal alloys, among other applications. The end-point detection in EDTA
titrations usually relies on metal ion indicators – dyes that form colored
complexes with the metal ion. A classic example is Eriochrome Black T
(EBT) for calcium/magnesium: EBT forms a wine-red complex with Mg(II) or Ca(II)
in solution. When EDTA is added, it preferentially binds the metal ions
(stronger complex), freeing the indicator. At the equivalence point, all metal
ions are sequestered by EDTA, and the indicator reverts to its free form, which
is a different color (blue in the case of Eriochrome Black T). Thus, the color
change from wine-red to blue indicates that the metal has been completely
chelated by EDTA[^6]. Different metal–indicator combinations are used depending
on the metal of interest (e.g., Calmagite for calcium/magnesium, Murexide for
calcium, Xylenol orange for various metals, etc.). The choice of pH and buffer
is critical in complexometric titrations because metal-ligand binding and
indicator color transitions are often pH-dependent. Complexometric titrimetry
provides a convenient and accurate way to measure metal ion concentrations in
solutions. It is more selective than simple precipitation, and by controlling
pH and using masking agents, one can often titrate specific metals in the
presence of others.
3.4 Redox
Titrations (Oxidation–Reduction Titrations)
Redox
titrations are based on oxidation–reduction reactions between the titrant and
analyte. These titrations find extensive use in analyzing oxidizing or reducing
agents – for example, determining the iron(II) content with a standard
permanganate solution, or the hydrogen peroxide content by titration with
permanganate or dichromate. End-point detection in redox titrations can
sometimes be achieved without an external indicator if one of the reactants or
products is colored. A prime example is the permanganate titration
(permanganometry): KMnO_4 is a strong oxidizing agent and has a deep purple
color. When used as a titrant to oxidize, say, Fe^2+ to Fe^3+, the MnO_4^− is
reduced to nearly colorless Mn^2+. As long as Fe^2+ (analyte) remains, each
addition of permanganate is decolorized. Once all Fe^2+ is consumed, the first
slight excess of MnO_4^− imparts a persistent pale pink or purple tint to the solution,
signaling the end point. Thus, permanganate is self-indicating in many
titrations and requires no separate indicator[^11]. In cases where the titrant
or analyte are not strongly colored, redox indicators can be used. Redox
indicators are compounds that have different colors in their oxidized and
reduced forms. For example, ferroin (a complex of phenanthroline with
iron) is often used in cerium(IV) titrations: its color changes from red (Fe^2+
form) to pale blue (Fe^3+ form) at the end point when the indicator itself gets
oxidized by excess Ce^4+. Another ubiquitous indicator is starch for
iodine-based titrations: in iodometry (where iodine is produced or consumed in
the reaction), a few drops of starch solution are added; starch forms an
intense blue complex with elemental iodine. During a titration of, e.g., iodine
with thiosulfate, the disappearance of the blue starch-iodine color indicates
that iodine has been consumed. Then, when a slight excess of thiosulfate is
added past equivalence, the blue color does not reappear upon mixing,
confirming the end point. Starch is extremely sensitive (able to detect trace
iodine), so it is usually added near the end of an iodometric titration to
avoid a prematurely intense color. As in acid–base systems, the immediate
vicinity of the redox titration end point is where the indicator undergoes its
color change, which should coincide with the completion of reaction[^11]. Redox
titrations encompass a broad range of analyses: common examples include the
dichromate titration of iron (using barium diphenylamine sulfonate as
indicator), iodometric titrations for copper or chlorine (using starch
indicator), and bromate or cerium(IV) titrations for various organics and
inorganics. They are indispensable in industrial analysis (e.g., determining
oxidizing agent strength, food preservative content like sulfites, etc.) and
often have well-established standard methods.
In
all these titrations with chemical end point detection, success depends on
selecting an indicator or signaling reaction that changes sharply at the true
equivalence point. The development of theories by Wilhelm Ostwald and others in
the late 19th century greatly advanced the understanding of indicators
(especially acid–base indicators), allowing chemists to tailor indicator choice
to the titration curve of a given reaction. The proper use of indicators,
combined with good technique, permits visual titrations to achieve excellent
accuracy (often within 0.1% relative error for concentration determinations).
However, visual methods do have limitations, such as the subjectivity of color
perception and the requirement that the solution and indicator not obscure the
end point (e.g., highly colored or opaque sample solutions can be problematic).
These limitations motivate the use of instrumental end-point detection methods,
which are discussed next.
4. Titrimetry with
Physical Endpoint Determination
Rather
than relying on the human eye and a chemical indicator, many titrations use instrumental
measurements to detect the end point. In these methods, a physical property
of the solution that changes significantly during the titration is monitored
with a suitable sensor or device. Instrumental end-point detection offers
greater objectivity and often higher precision, as it is not subject to human
color perception or the need for a sharply visible change. The titrations are
often carried out in the same way (adding titrant until reaction completion),
but the end point is determined by a sudden change in an electrical or optical
signal recorded by the instrument. The common types of physical end-point
detection in titrimetry include electrochemical methods, optical methods, and
thermometric methods. Key examples are outlined below:
- Potentiometric Titrations: These use a
voltage (electrical potential) measurement to find the end point.
Typically, a pair of electrodes is placed in the titration solution –
often a sensing (indicator) electrode that is responsive to the analyte or
a related ion, and a reference electrode. For example, in an acid–base
titration, a glass pH electrode (indicator electrode) and a reference
electrode can be used to monitor the solution’s pH continuously as titrant
is added. The potential difference between the electrodes corresponds to
the solution pH (via the Nernst equation). As the titration progresses,
the measured electrode potential (or pH) changes gradually and then
rapidly near the equivalence point, producing a titration curve. The
equivalence point can be determined by finding the inflection point of the
pH vs. volume curve or the volume at which the slope (or first derivative)
is maximized. Potentiometric titrations are not limited to acid–base
reactions; they are also used for redox titrations (with an appropriate
redox electrode measuring potential), precipitation titrations (using
specific ion electrodes, e.g. a silver electrode for halides), and
complexometric titrations. The end point in potentiometry is often
identified by a sharp change in potential[^12]. Because the measuring
instrument (pH meter or potentiometer) detects the end point, no visual
indicator is needed – an advantage for colored or turbid solutions.
Potentiometric titration is one of the most versatile and widely used
instrumental titration methods.
- Conductometric Titrations: These rely on
measuring the electrical conductance (or its inverse, resistance)
of the solution during the titration. The conductance depends on the ionic
composition of the solution. As the titration reaction proceeds, ions are
consumed and/or produced, changing the solution’s conductivity. A
conductivity cell (usually two metal electrodes with an AC current)
measures the conductance. A classical example is the titration of a strong
acid with a strong base: initially, the solution has high conductance due
to H^+ and other ions; as NaOH is added, H^+ is neutralized to water
(which is weakly ionized), so conductance drops. After the equivalence
point, excess OH^− from the titrant increases the conductance again.
Plotting conductance vs. titrant volume yields two linearly varying
regions intersecting at the equivalence point. Conductometric titrations
are particularly useful when no suitable indicator exists or the solution
is colored. They are applied in acid–base titrations (especially of weak
acids or bases in absence of good indicators), precipitation titrations
(where the disappearance/appearance of ionic species affects
conductivity), etc. Unlike potentiometry, conductometry does not require a
specific ion-selective electrode, only a general conductivity probe. The
end point is determined by the change in slope of the conductance curve.
One consideration is that conductance measurements can be influenced by
temperature and mobility of ions, so temperature control is important for
accuracy[^12].
- Amperometric and Biamperometric Titrations: These techniques involve measuring an electric current
flowing through the solution under an applied voltage. In amperometric
titration, a constant potential is applied between two electrodes and
the current is measured as titrant is added. The current is related to the
oxidation or reduction of the titrant or analyte at the electrode surface.
A notable example is the titration of chloride with AgNO_3 using a pair of
silver electrodes: before the equivalence point, Cl^− is present and can
carry current by being oxidized at the anode (Ag -> Ag^+ and electron)
and reduced at cathode (Ag^+ + e^- -> Ag) – essentially the silver
electrodes dissolve/plate in presence of chloride. When Cl^− is depleted
at equivalence, the current drops sharply because the solution no longer
supports that electrochemical reaction (this setup is called biamperometric
or dead-stop end point detection with polarized electrodes). Thus, the
titration end point is indicated by a sudden change (often a minimum) in
current. Amperometric titrations can also be conducted with one
indicator electrode at a fixed potential (where an analyte or titrant is
oxidized/reduced) and a reference electrode, measuring current flow that
changes once one reactant is consumed. These methods are especially
helpful for redox systems and for detecting end points in precipitation
titrations (e.g., Karl Fischer titration in one mode uses bipotentiometric
end detection of excess iodine). They offer high sensitivity and are used
when visual indicators are inadequate. For instance, the Karl Fischer
titration for water content employs a biamperometric end-point detection:
two platinum electrodes detect the point at which excess iodine (generated
in the reagent) appears, causing a sharp rise in current – that signals
that all water has been consumed and iodine is free[^13].
- Coulometric Titrations: A coulometric
titration is somewhat different in that no standard titrant solution is
added; instead, a titrant is generated in situ by an electrical current,
and the amount of titrant is determined by the total electrical charge
(coulombs) passed. This method is governed by Faraday’s law, which relates
charge to the amount of substance reacted. Coulometric titrations often
use constant-current electrolysis to produce a titrant at a known rate
(for example, generating I_2 from iodide, or OH^− from water electrolysis)
and the time or total charge to reach the end point is measured. The end
point may be detected by an indicator electrode or by the same kinds of
signals as above (e.g., a sudden change in voltage or current when
titration is complete). One famous application is the coulometric Karl
Fischer titration for water: iodine is generated coulometrically and
reacts with water in the presence of sulfur dioxide and a base (Karl
Fischer reagent); when water is depleted, excess iodine is detected and
the total charge used to produce iodine corresponds to the water content.
Coulometric titrations are extremely useful for very small quantities of
analyte (trace analysis) because one can deliver extremely small amounts
of titrant by controlling the current and time rather than trying to
manipulate tiny volumes. The accuracy of coulometric titration is high
since it is based on electric charge measurement, often eliminating the
need for a standard titrant solution altogether[^12]. The results are
calculated directly from the charge passed at the equivalence point.
- Photometric (Spectrophotometric) Titrations: These titrations use optical measurements (such as absorbance
of light at a specific wavelength) to monitor the progress of the reaction
and detect the end point. Rather than observing an indicator by eye, a
photometric titration quantitatively measures absorbance changes
associated with the consumption or formation of a colored species. For
instance, in a complexometric titration of metal ions, one could use a
UV-Vis spectrophotometer to track the decrease of the metal-indicator
complex’s color intensity as EDTA is added (the absorbance drops until the
indicator is displaced from the metal at equivalence). Alternatively, if
either the titrant or analyte or product has a distinct absorption, that
can be monitored – for example, following the absorbance of permanganate’s
purple color in a redox titration, which diminishes until the equivalence
point and then increases when permanganate is in excess. The end point is
determined as the volume at which the absorbance vs. volume curve shows a
breakpoint or inflection. Photometric titration can be done manually by taking
aliquots and measuring in a spectrometer, or automatically with flow cells
and fiber optic probes dipped in the solution. A specialized variant is
the colorimetry using an optrode (optical sensor) which Metrohm and
others have developed, replacing visual detection with an electronic eye
for color change[^14]. The advantage of photometric methods is that they
can detect end points even if the color change is slight or invisible to
the human eye, and they can be automated for continuous monitoring.
- Thermometric Titrations: These are less
common but rely on measuring the temperature change of the solution
during the titration. Many reactions either release heat (exothermic) or
absorb heat (endothermic). In a thermometric titration, a sensitive
thermometer or thermistor probe tracks the solution temperature. At the
equivalence point, the rate of temperature change often shifts because the
dominant reaction is complete and further addition of titrant may produce
a different reaction or simply dilute/cool the solution. A well-known
example is the titration of strong acids and bases, which is exothermic –
the solution warms as neutralization occurs, and after the equivalence
point, adding more titrant (base) may cause cooling (since excess base
dilution is usually endothermic). Plotting temperature vs. titrant volume
yields a curve where the equivalence point is identified by a change in
the slope. Thermometric titration has the benefit of not requiring any
indicator or special electrode; it only needs a thermometer and proper
insulation to detect small temp changes. It has been used for certain fast
reactions and in cases where other methods are not feasible, though its
applications are more niche compared to the above methods.
In
summary, physical end-point detection in titrimetry provides alternatives that
can increase accuracy and enable titrations in situations where visual methods
fail. Table 1 summarizes some instrumental end-point methods and the
property measured. Each method requires appropriate instrumentation (pH meter,
conductivity meter, amperometric setup, spectrophotometer, etc.) but many
modern titration systems integrate one or more of these detection modes. By
automating the detection of the end point, instrumental titrations reduce the
subjectivity associated with indicators and often allow the titration data to
be recorded and analyzed (e.g., plotting a full titration curve). This
capability leads us to the topic of instrumental titrimetry, where
entire titration procedures are managed by instruments.
[^1]: Encyclopædia Britannica, “Titration” – definition of
titration as a quantitative analytical process using a standard solution added
from a burette[1][2].
[^2]: Encyclopædia Britannica, “Titration” – explanation of
equivalence point vs. end point and titration error[3].
[^3]: IUPAC Compendium of Analytical Nomenclature (Orange Book) –
Titrimetric analysis remains widely used in quantitative analysis, especially
with instrumental endpoints; note that titrimetric and volumetric
are not strict synonyms, since titrant amount can be measured by volume or mass
or charge[4][5].
[^4]: C. K. Zacharis, American Pharmaceutical Review 2024 –
Titration is an established technique in use since the 1800s, with the first
titrimetric methods textbook published in 1855 (by Friedrich Mohr)[6][7].
[^5]: Encyclopædia Britannica, “Titration” – example of a
precipitation titration (chloride with silver nitrate) where the end point is
indicated by the appearance of a colored precipitate (silver chromate) when
using chromate indicator (Mohr’s method)[8].
[^6]: Encyclopædia Britannica, “Titration” – discussion of
complexometric EDTA titrations and use of dyes forming colored complexes with
metal ions that change color at the end point[9][10].
[^7]: J. Clifton, ReAgent Science Blog (2024) – Basic titration
apparatus includes a burette, stand, flask, and an indicator; titration is a
straightforward experiment requiring simple equipment and careful technique[11].
[^8]: IUPAC Gold Book, “standard solution” – definition of
primary standard (high-purity substance used to prepare standard solution) and
secondary standard (solution standardized by a primary standard)[12].
[^9]: NCERT Chemistry Laboratory Manual, Vol. XI – Examples of
primary and secondary standards: e.g. sodium carbonate, potassium dichromate,
KHP as primary standards; NaOH and KMnO₄ as secondary standards that must be
standardized before use[13].
[^10]: MilliporeSigma (Regina, Analytical Techniques, 2016) –
Common acid–base indicators and their pH transition ranges (e.g. litmus: red at
pH <5, blue at pH >8; phenolphthalein: colorless in acid, pink in base
around pH 8.3–10)[14].
[^11]: Encyclopædia Britannica, “Titration” – redox titration
indicators act analogously to acid–base indicators, changing color upon
oxidation or reduction at the end point (e.g. distinct colors for oxidized vs.
reduced forms)[15].
[^12]: Encyclopædia Britannica, “Titration” – overview of
instrumental titration methods: potentiometric (measuring voltage),
conductometric (conductance), amperometric (current), and coulometric
titrations (measuring total charge) for end-point detection[16][17].
[^13]: M. Messuti, TestOil Blog (2012) – Karl Fischer moisture
titration was invented in 1935 by Karl Fischer; it uses an electrochemical
end-point (biamperometric detection of excess iodine) and can be done in
volumetric or coulometric modes for trace water analysis[18][19].
[^14]: Metrohm Application Notes – Photometric titration with an
optical sensor (Optrode) replaces subjective visual end-point detection with an
objective measurement of absorbance or transmission change, improving end-point
accuracy (e.g. in determinations of water hardness or acidity)[[20][21]].
5. Instrumental Titrimetry
The incorporation of
instrumentation into titrimetric analysis has greatly enhanced the precision,
convenience, and capabilities of titration methods. Instrumental titrimetry
refers to titration techniques that employ electronic instruments to control
the titration and/or detect the end point. Over the past century, titration has
evolved from a purely manual operation with glass burettes and color indicators
to sophisticated automated systems that can deliver titrant, sense the end
point electronically, and compute results with minimal human intervention. This
evolution was driven by the need to eliminate human error (both in detecting
end points and in reading burettes) and to handle large numbers of analyses
more efficiently.
Apparatus Development: Early titrations in the
18th and 19th centuries relied on simple devices – for example, François
Descroizilles in 1791 devised one of the first burettes (a simple graduated
cylinder with a stopcock) for acid–base titrations[^5]. In 1824, Joseph Louis
Gay-Lussac improved the burette design by adding a side tube and introduced the
terms burette and pipette into analytical vocabulary[^5]. By
1845, Étienne Ossian Henry had developed a more modern form of burette
resembling those used today[^5]. Karl Friedrich Mohr, a German chemist, further
refined titration hardware (introducing the Mohr burette with a clamp and tip)
and wrote an influential textbook in 1855 that standardized titration methods[^4][^8].
These advances in glass apparatus made manual titration a mainstream
quantitative technique by the late 19th century. However, even with good
apparatus, manual titrations suffered from certain limitations: the analyst had
to judge a color change by eye and read volumes by sight, steps prone to
subjective interpretation and small errors.
Recording Titrators: The first half of the
20th century saw the introduction of electronic devices like the pH electrode
(invented by Fritz Haber and Z. Klemensiewicz in 1906, improved by Arnold
Beckman into the first pH meter in 1934) and other ion-selective electrodes.
These allowed continuous monitoring of solution conditions during a titration.
By mid-20th century, laboratories began to use recording titrators – essentially
a combination of a burette, a mechanical or electronic volume delivery system,
and a chart recorder attached to an electrode. For example, a pH titrator could
automatically plot the pH curve on paper as titrant was added. This provided a
permanent record of the titration curve and a more objective determination of
the equivalence point (by later analysis of the curve). During the 1920s–30s,
there were even attempts at automated titration: as early as 1929, researchers
had built devices to automate acid–base titrations to an electrical end point.
However, these were not widespread until later. In the 1950s and 1960s, as
electronics and control systems advanced, commercial titration systems emerged
that could automatically detect the end point – for instance, by using a preset
mV jump in potentiometric titration to stop the burette.
End-Point Detection Titrators: Instrumental
titrators in the later 20th century were designed to perform titrations to a
predefined end-point criterion without requiring the analyst to watch the
reaction. One approach was the endpoint titrator, where a specific
sensor (pH, conductivity, photometric, etc.) would trigger a stop when a
certain value was reached. For example, an autotitrator might dispense titrant
until the pH meter reads 7.00 (for a neutralization) or until a certain
millivolt potential is observed in a redox titration. These instruments often
featured an electronic burette (sometimes a motor-driven syringe or pump) and
an input from an electrode or photodiode. Once the end point condition was met,
the device would stop titrant addition. This eliminated the guesswork around
color indicators and reduced variability between different operators[^19].
Additionally, these titrators could calculate the result immediately based on
the volume delivered at the endpoint, streamlining the analysis.
Digital and Automated Titration Systems: Since
the late 20th century, titrimetry has fully embraced automation and digital
control. Modern automatic titrators are microprocessor-controlled
instruments that handle most aspects of the titration: they can fill and
dispense titrant precisely (with automatic burettes often accurate to 0.001 mL
or better), stir the solution, record the sensor response, determine the
endpoint (either by fixed threshold or more sophisticated curve analysis), and
compute the analyte concentration using stored formulas. These systems often
have touch-screen interfaces and can store multiple titration methods (programs
for different analyses) which can be easily recalled[^19][^20]. They also
provide data logging and can output results to computers or LIMS (Laboratory
Information Management Systems). A significant benefit of automation is
improved precision and repeatability – the titrant dispensing
systems in modern autotitrators can be much more precise than a human operator,
and the endpoint detection is consistent across runs. Automated titration also
enhances safety and throughput: since the instrument can run unattended once
started, an analyst can set up multiple titrations (on multi-sample titrators
or by sequential operation) and walk away, freeing time for other tasks. Many
instruments also include features like automatic cleaning and rinsing of
burettes, and some have multiple burettes for handling different titrants in
sequence.
One specific branch of instrumental titrimetry is the development of Karl
Fischer titrators for water determination, which exemplifies a specialized
automated titration. Karl Fischer titration, invented in 1935, initially was a
manual titration with visual detection of the endpoint (using iodine and
starch). Modern Karl Fischer titrators are fully automated devices – they
perform either volumetric titration (dispensing an iodine-containing reagent
until a bipotentiometric sensor detects excess iodine, indicating all water is
consumed) or coulometric titration (electrogenerating iodine until endpoint)
with a high degree of automation and precision[^13]. They often come with
automated syringes, integrated magnetic stirrers, and microprocessor control to
calculate moisture content directly. This development highlights how
instrumentation has extended the applicability of titrations to new areas (such
as trace water analysis at ppm levels, which would be difficult by purely
manual means).
Another innovation in instrumental titrimetry is the coupling of
titration with flow analysis systems. Techniques like Flow Injection
Analysis (FIA) and Sequential Injection Analysis (SIA) were developed in the
1970s–1980s to automate wet-chemical analysis. In FIA, a sample is injected
into a carrier stream and can be made to react with a titrant in a controlled
way, with a detector (often photometric or electrochemical) measuring the
result. While FIA is not a titration in the classical sense (since it often
relies on reaching a steady-state signal rather than a true equivalence point),
certain configurations called flow injection titrations use a burette to
add titrant to a flowing sample until a detector threshold is reached. These
systems can greatly increase sample throughput for routine analyses. Similarly,
automated titrators can be equipped with sample changers (autosamplers) to
titrate many samples in sequence, which is invaluable in industrial quality
control labs.
Instrumental titrimetry has thus transformed titration from an artisan
skill to a highly reproducible analytical procedure. By addressing the primary
shortcomings of manual titration – namely, subjective endpoint detection and
manual data handling[^19] – modern instruments ensure that titration results
are consistent between different operators and laboratories. For instance,
automatic potentiometric titrators eliminate color change subjectivity and
record the exact volume and potential at endpoint, improving both accuracy and
traceability of results. They also reduce transcription errors by automatically
calculating and storing the results[^19][^20]. With proper maintenance
(particularly of electrodes and burette calibration), automated titrators can
deliver very high precision, often better than 0.1% relative standard
deviation.
In summary, instrumental titrimetry encompasses the use of pH meters,
ion-selective electrodes, photometers, and automated buretting systems to
perform titrations. It represents the marriage of classical chemical reactions
with modern sensors and control systems. The result is a suite of analytical
methods that maintain the core advantages of titration – exacting
stoichiometric accuracy and simplicity of chemistry – while mitigating many of
the practical limitations. Automated titration systems are now standard
equipment in many laboratories, reflecting the enduring importance of
titrimetric analysis in the modern analytical toolkit.
6. Overview of the
History of Titrimetry
The
development of titrimetry is deeply intertwined with the growth of analytical
chemistry and the need for accurate quantitative methods. The origins of
titration date back over two centuries. Here we outline some key milestones and
figures in the history of titrimetric analysis:
- Early Foundations (18th Century): The
concept of determining an unknown by reacting it with a measured amount of
reagent emerged in the 18th century. One early description of a
titration-like procedure is attributed to Étienne François Geoffroy
in 1729, who is often credited with the first account of a true
titration[^9]. By the mid-1700s, chemists were exploring neutralization
for quantitative analysis: in 1756, Scottish physician Francis Home
used a colored indicator (infusion of cochineal) to determine the strength
of limewater (an alkali) by adding acid until the color changed – arguably
one of the first recorded uses of an indicator in titration. Another
pioneer, English chemist William Lewis, conducted experiments in
the 1760s titrating potash (impure K_2CO_3 from wood ashes) with acid to
determine its alkali content, improving the consistency of alkali supply
for industries[^9]. These early efforts were limited by the lack of
precise tools, but they laid the groundwork for volumetric analysis as a
quantitative technique.
- Volumetric Analysis Invented (Late 18th – Early 19th Century): The birth of titrimetry as a recognized method is usually
credited to François Antoine Henri Descroizilles, a French chemist.
In the 1790s (circa 1791 or 1795 in different accounts), Descroizilles
developed an apparatus he called the berrette (an early burette)
and conducted titrations to determine the “degree of saturation” of
solutions – notably, he titrated alkaline solutions against sulfuric acid
using a colored indicator to judge completion[^5]. He applied this method,
for example, to quantify the amount of chlorine in bleaching liquor, an
important process at that time. Descroizilles’ work essentially introduced
volumetric analysis as a practical tool. Following him, another Frenchman,
Joseph-Louis Gay-Lussac, made significant contributions. In 1824,
Gay-Lussac introduced an improved burette design with a side arm
(sometimes called an alkalimeter) for easier use, and he coined the
terms burette and pipette in print[^5]. Gay-Lussac also
formulated titrimetric methods for analytes like silver (Gay-Lussac’s
method for silver assay by titration with salt solution) and published
procedures for standardizing solutions (he used the word “titrer”, meaning
to determine concentration, from which titration is derived[^5]).
By 1828, the term “titration” was in use in the context of determining
concentrations[^5]. These developments in France firmly established the
utility of titration in analytical chemistry.
- Mid-19th Century Advances: The mid-1800s
saw titrimetry flourish and spread through Europe. German chemist Karl
Friedrich Mohr is a central figure of this era. Mohr improved
volumetric techniques and apparatus – he devised the Mohr pinchcock
burette and introduced visual indicators for various titrations. In 1855,
Mohr published “Lehrbuch der chemisch-analytischen Titrirmethode”
(“Textbook of Analytical Chemistry Titration Methods”), which was the
first comprehensive treatise on titrimetric analysis[^4][^8]. This book
systematized titration methods (acid-base, argentometric, etc.) and
greatly popularized volumetric analysis in laboratories worldwide. Many
classic titration methods bear the names of 19th-century chemists: Karl
Mohr himself (Mohr’s method for chloride with chromate indicator[^5]),
Jacob Volhard (who in 1874 developed Volhard’s method, a
back-titration for halides using thiocyanate in presence of iron
indicator), and Johann Heinrich Wilhelm Ferdinand Wacker (Wacker’s
titration for manganese). Another notable contribution was by Justus
Liebig, who applied titration in agricultural chemistry (developing a
titration for cyanide, among others). The 19th century also introduced acid-base
indicators systematically: Robert Wilhelm Bunsen and Henry
Roscoe studied indicators; later, around 1884–1888, Wilhelm Ostwald
(a founder of physical chemistry) explained indicator action with his
theory of ionization, allowing rational choice of indicators for
titrations. The Kjeldahl method (1883) for nitrogen analysis in organic
compounds is an example of a back-titration (ammonium produced is measured
by titration) that became a standard method, underscoring titration’s
importance in quantitative analysis of that era.
- Emergence of Redox and Complexometric Titrations: Oxidation-reduction titrations were developed in the late 19th
and early 20th centuries as more was understood about redox chemistry. Permanganate
titration (permanganometry) was introduced by Friedrich Mohr and
others for determining iron, calcium, etc., and became widely adopted; Dichromate
titration for iron was introduced by Jean-Baptiste Dumas and later
optimized. Iodometry (titrations involving iodine) was developed in
the nineteenth century (notably by Karl Friedrich Mohr and others) and
proved very versatile for analyzing oxidizing agents like copper(II),
chlorine, and more. Complexometric titration using EDTA is a comparatively
later development – EDTA was first synthesized in 1935, but its analytical
use blossomed after 1945 when chemists such as Gerold Schwarzenbach
in Zurich explored its ability to titrate metal ions with indicators.
Schwarzenbach’s work in the 1940s and 1950s established the principles of
EDTA titration and introduced many metallochromic indicators, greatly
expanding the scope of titrimetry to virtually all metal ions.
- pH Concept and Buffering (20th Century):
The understanding of acids and bases was revolutionized by Søren P. L.
Sørensen, who introduced the pH scale in 1909. This concept, along
with mass-action theory, allowed precise calculation of titration curves
and improved indicator selection. The development of the glass electrode
for pH by Haber and Müller (1906) and its commercial production by Beckman
(1930s) provided a crucial tool that directly fed into titrimetric
practice by enabling potentiometric titrations. With these tools,
titration could be monitored electronically, which paved the way for
automated endpoint detection.
- Instrumentation and Automation (Mid-20th Century): A significant historical milestone was the automation of
titration. As early as the 1930s, there were automated titrators described
in the literature (e.g., an automated acid–base titrator that used
conductance to end the titration). The true rise of automated titration
systems occurred in the 1950s-1960s. In the mid-1960s, companies like
Metrohm (Switzerland) and Radiometer (Denmark) introduced commercial
automatic titrators that combined burettes, stirrers, and electronic
endpoints[^20]. By the 1970s, automatic titrators were capable of
inflection-point detection (using the first or second derivative of
titration curve) and could handle a variety of titration types. This
period also saw the introduction of coulometric titration (notably
the Karl Fischer coulometric titrator in the 1970s) which extended
titration to trace analysis of water and other species.
- Modern Developments (Late 20th – 21st Century): Titration has continued to advance with technology. Modern
autotitrators feature computer interfaces, high precision dispensing (with
digital stepper motors or pistons meeting ISO volumetric standards), and
often multiple detection modes (combined pH, redox, photometric detection
in one unit). Software improvements allow gran plot or partial derivative
calculations to determine endpoints in complex titrations automatically. Flow
Injection Analysis (1975), introduced by Ruzicka and Hansen, although
not a direct titration, influenced how solutions could be handled in
automated systems and led to continuously monitored titrations and
high-throughput assay systems. Today, titrators are commonly connected to
computers or networked for data management, and features like autosamplers
and robust data logging meet the needs of regulated industries
(pharmaceutical, environmental monitoring, etc.). Even with these
high-tech enhancements, the fundamental chemical basis of titrimetry
remains unchanged from the days of Gay-Lussac and Mohr – a testament to
the enduring soundness of the titration principle.
In
conclusion, titrimetry’s history spans from rudimentary experiments with
color-changing vegetable extracts in the 1700s to fully automated,
computer-controlled systems in modern laboratories. Each era of development –
introduction of burettes, standard solutions, theoretical understanding of
equilibria, electrochemical sensors, and automation – has built upon the
previous, preserving the core idea: that a quantitative reaction with a known
reagent can reveal how much of a substance is present. Titration’s longevity
and continual adaptation underscore its importance. It remains a key method
taught in chemistry curricula and employed daily in labs worldwide for its
reliability, accuracy, and the direct insight it provides into chemical
quantities through simple reactions.
Bibliography
- Encyclopædia Britannica – “Titration.” Encyclopædia Britannica Online. Last
updated Dec 27, 2025. (Definition, types of titrations, equivalence vs.
end point)[1][22]
- IUPAC Orange Book – Compendium of Analytical Nomenclature, Section on
Titrimetric Analysis. IUPAC Analytical Chemistry Division. (General
principles of titrimetry and terminology)[4][5]
- IUPAC Gold Book – Definition of Standard Solution, Primary and Secondary
Standard. IUPAC Compendium of Chemical Terminology, 2014. (Definitions
of primary/secondary standard in titration)[12]
- Zacharis, C. K. (2024). “Instrument-Based Testing: A More Modern and Robust
Approach to Titration.” American Pharmaceutical Review 24(3), June
1, 2024. (Discussion of manual vs. automated titration, historical notes
on first titration textbook in 1855)[6][7]
- Clifton, Jessica (2024). “Who Invented Titration in Chemistry?” ReAgent
Science Blog, Jan 3, 2024. (Historical overview of titration:
Descroizilles’s first burette in 1791, Gay-Lussac’s contributions in 1824,
origin of the term “titrer” in 1543, Mohr’s textbook in 1855)[23][24]
- Johansson, Axel (1988). “The development of the titration methods: Some
historical annotations.” Analytica Chimica Acta 206, 97–109.
(Historical account crediting Geoffroy in 1729 for first titration, and
outlining contributions of Lewis, Descroizilles, Gay-Lussac, Ostwald)[25]
- NCERT Chemistry
Laboratory Manual (Class XI) – Experiment on
Titrimetric Analysis. National Council of Educational Research and
Training, India. (Practical guidelines on titration, examples of primary
and secondary standards like Na₂CO₃, KHP vs. NaOH, KMnO₄)[13]
- Regina (MilliporeSigma) (2016). “Instrumental Techniques – Titration.”
(Illustrated overview of titration principles, common indicators and their
pH ranges)[14]
- Encyclopædia Britannica – “Titration”, extended entry (visual indicators in redox,
complexometric, precipitation titrations). (Examples of indicator color
changes: litmus, phenolphthalein; Mohr’s method with chromate; redox
indicators)[26][15]
- Metrohm Application
Bulletin – Photometric Titrations with Optrode.
Metrohm AG (2010s). (Describes replacement of visual end point with
photometric sensor for automated titration, improving precision)[21]
- TestOil Knowledge
Center – Messuti, M. (2012). “Karl Fischer Water
Test: Quantifies the Amount of Water.” (Explains Karl Fischer titration,
invented 1935, and its electrochemical end-point detection for water)[18][19]
- SelectScience Interview – Haslam, C. (2024). “Embracing automated titration in the lab.”
(Expert interview noting automated titration first developed in mid-1960s,
and advantages of modern autotitrators such as Thermo Orion series)[27][28]
[1] [3]
[8]
[9]
[10]
[15]
[16]
[17]
[22]
[26]
Titration | Definition, Types, & Facts | Britannica
https://www.britannica.com/science/titration
[2] [11]
[23]
[24]
Who Invented Titration? | The Science Blog
https://www.reagent.co.uk/blog/who-invented-titration/
https://media.iupac.org/publications/analytical_compendium/Cha06sec1.pdf
[6] [7] Instrument Based Testing: A More Modern and
Robust Approach to Titration | American Pharmaceutical Review - The Review of
American Pharmaceutical Business & Technology
[12] IUPAC Gold Book - standard solution
https://goldbook.iupac.org/terms/view/S05924/pdf
[13] ncert.nic.in
https://ncert.nic.in/pdf/publication/sciencelaboratorymanuals/classXI/chemistry/kelm206.pdf
[14] Regina_2016 - Instrumental Techniques - Titration
[18] [19]
Karl Fischer Water Test: Quantifies the Amount of Water - TestOil
https://testoil.com/routine-testing/quantifying-the-amount-of-water-karl-fischer-water-test/
[20] Recognizing the Endpoints of Automated Titrations
https://www.azom.com/article.aspx?ArticleID=20337
[21] Recognition of endpoints (EP) - Metrohm
https://www.metrohm.com/en/discover/blog/20-21/recognition-of-endpoints--ep-.html
[25] The development of the titration methods : Some historical annotations
- ScienceDirect
https://www.sciencedirect.com/science/article/abs/pii/S000326700080834X
[27] [28]
Embracing automated titration in the lab

