Guidelines for Canadian Drinking Water Quality: Guideline Technical Document – Chloramines

Download the alternative format
(PDF format, 910 KB, 67 pages)

Organization: Health Canada

Type: Guideline

Published: 2020-02-07

Cat.: H144-13/15-2019E-PDF

ISBN: 978-0-660-32161-5

Pub.: 190278

Table of Contents

Part I. Overview and Application

1.0 Guideline

It is not considered necessary to establish a maximum acceptable concentration for chloramines in drinking water, based on the low toxicity of monochloramine at concentrations found in drinking water. Any measures taken to limit the concentration of chloramines or their by-products in drinking water supplies must not compromise the effectiveness of disinfection.

2.0 Executive summary

The term "chloramines" refers to both inorganic and organic chloramines. This document focuses on inorganic chloramines, which consist of monochloramine, dichloramine and trichloramine. Unless specified otherwise, the term "chloramines" will refer to inorganic chloramines throughout the document.

Chloramines are found in drinking water mainly as a result of treatment, either intentionally as a disinfectant in the distribution system, or unintentionally as a by-product of the chlorination of drinking water in the presence of natural ammonia. As monochloramine is more stable and provides longer-lasting disinfection than chlorine, it is commonly used in the distribution system as a secondary disinfectant. Chlorine is more effective as a primary disinfectant and is used in the treatment plant. Chloramines have also been used in the distribution system to help reduce formation of common disinfection by-products such as trihalomethanes and haloacetic acids. However, chloramines also react with natural organic matter to form other disinfection by-products.

All public drinking water supplies should be disinfected, unless specifically exempted by the responsible authority. Disinfection is an essential component of public drinking water treatment; the health risks associated with disinfection by-products are much less than the risks from consuming water that has not been adequately disinfected. Where chloramines are used in a Canadian drinking water supply, the residual is typically below 4 mg/L in the distribution system.

This guideline technical document focuses on the health effects related to exposure to chloramines in drinking water supplies, also taking into consideration taste and odour concerns. It does not review the benefits or the processes of chloramination; nor does it assess the health risks related to exposure to by-products formed as a result of the chloramination process. Health Canada, in collaboration with the Federal-Provincial-Territorial Committee on Drinking Water, does not consider it necessary to establish a guideline for chloramines in drinking water based on health considerations. It also does not consider it necessary to establish an aesthetic objective, since levels commonly found in drinking water are within an acceptable range for taste and odour, and since protection of consumers from microbial health risks is paramount.

2.1 Health effects

The International Agency for Research on Cancer and the United States Environmental Protection Agency (U.S. EPA) have classified monochloramine as "not classifiable as to its carcinogenicity to humans" based on inadequate evidence in animals and in humans. The information on dichloramine and trichloramine is insufficient to establish any link with unwanted health effects in animals or in humans. These forms are also less frequently detected in drinking water. Studies have found minimal effects in humans and animals following ingestion of monochloramine in drinking water, with the most significant effect being decreased body weight gain in animals. However, this effect is due to reduced water consumption caused by taste aversion.

2.2 Exposure

Human exposure to chloramines primarily results from their presence in treated drinking water; monochloramine is usually the predominant chloramine. Intake of monochloramine and dichloramine from drinking water is not expected through either skin contact or inhalation. Intake of trichloramine from drinking water might be expected from inhalation; however, it is relatively unstable in water and is only formed under specific conditions (at very high chlorine to ammonia ratios or under low pH), which are unlikely to occur in treated drinking water. Consequently, exposure to chloramines via inhalation and skin contact during showering or bathing is expected to be negligible.

2.3 Analysis and treatment

Although there are no standardized methods to directly measure chloramines, there are several such methods to measure total and free chlorine. The results from these methods can be used to calculate the levels of combined chlorine (or chloramines). There are also online and portable analysers that can directly measure monochloramine and total and free ammonia with limited interferences.

For municipal plants, a change in distribution system disinfectant (such as changing the disinfectant residual to chloramine) can impact water quality. When considering conversion to chloramines, utilities should assess the impacts on their water quality and system materials, including the potential for corrosion, nitrification and formation of disinfection by-products.

In treatment plants using free chlorine or chloramines for disinfection, chloramines may be found in drinking water at the treatment plant, in the distribution system and in premise plumbing. For consumers that find the taste of chloramines objectionable, there are residential drinking water treatment devices that can decrease concentrations of chloramines in drinking water. However, removal of the disinfectant residual is not recommended.

2.4 International considerations

Drinking water quality guidelines, standards and/or guidance from other national and international organizations may vary due to the science available at the time of assessment, as well as differing policies and approaches, including the choice of key study, and the use of different consumption rates, body weights and allocation factors.

Several organizations have set guidelines or regulations for chloramines in drinking water, all based on the same study which found no health effects at the highest dose administered.

The U.S. EPA has established a maximum residual disinfectant level of 4 mg/L for chloramines, recognizing the benefits of adding a disinfectant to water on a continuous basis and of maintaining a residual to control for pathogens in the distribution system. The World Health Organization and Australia National Health and Medical Research Council both established a drinking water guideline of 3 mg/L for monochloramine. The European Union has not established a guideline value for either monochloramine or chloramines.

3.0 Application of the guideline

Note: Specific guidance related to the implementation of drinking water guidelines should be obtained from the appropriate drinking water authority in the affected jurisdiction.

Chloramines are formed when chlorine and ammonia are combined in water and comprise three chemical species: monochloramine (NH2Cl), dichloramine (NHCl2) and trichloramine (NCl3). The relative amounts formed are dependent on numerous factors, including pH, chlorine:ammonia ratio (Cl2:NH3-N), temperature, contact time and mixing. When chloramines are used as a disinfectant in drinking water systems, the desired species is monochloramine. When treatment processes are optimized for monochloramine stability (Cl2:NH3-N weight ratio of 4.5:1–5:1, pH >8.0), almost all of the chloramines are present as monochloramine. Since chloramines can also be formed when ammonia is present in source water, utilities should characterize their source water to assess the presence of and variability of ammonia levels. When utilities are considering conversion from chlorine to chloramines, they should assess the impacts on their water quality and system materials, including the potential for corrosion, nitrification and the formation of disinfection by-products.

Maintenance of adequate disinfectant residual, through the establishment of monitoring and mitigation plans, will minimize bacterial regrowth in the distribution system. Through these plans, a drop in monochloramine level, suggestive of unexpected changes in water quality, can be more quickly detected. Specific requirements for chloramine residual concentrations are set by the regulatory authority and may vary among jurisdictions. Monochloramine, used as a secondary disinfectant, should be applied so as to maintain a stable residual concentration throughout the distribution system. The appropriate amount of disinfectant needed to maintain water quality in the distribution system will depend on (among other factors), the characteristics of the distribution system, the species of bacteria, the presence of biofilms, the temperature, the pH, the amount of biodegradable material and the nutrients (C:N:P) present in the treated water. Water utilities should be aware that a minimum target chloramine residual of "detectable" will not be sufficient to effectively limit bacterial growth in the distribution system. Regular monitoring of distribution system water quality (e.g., disinfectant residual, microbial indicators, turbidity, pH) and having operations and maintenance programs in place (e.g., water mains cleaning, cross-connection control, replacements and repairs, minimizing water age and managing storage) are important to ensure that drinking water is transported to the consumer with minimum loss of quality.

Depending on the water system, chloramine residual concentrations of >1.0 mg/L may be required to maintain lower general microbiological populations, to reduce coliform occurrences and to control biofilm development. Some utilities may require monochloramine concentrations that are much higher than this to address their specific distribution system water quality. Nitrification in the distribution system is also a potential problem for municipal systems that chloraminate. The concerns for utilities from nitrification are the depletion of the disinfectant residual, increased bacterial growth and biofilm development in the distribution system, and could lead to decreased pH, especially in low alkalinity water, which can result in corrosion issues. When used as part of a program for nitrification prevention and control, suggested best operational practices for a chloramine residual are 2 mg/L leaving the treatment plant and preferably greater than 1.5 mg/L at all monitoring points in the distribution system. Information on strategies for controlling nitrification can also be found in the guideline technical document for ammonia.

Most Canadian drinking water supplies maintain a chloramine residual range below 4 mg/L in the distribution system. At these concentrations, tastes and odours related to chloramines are generally acceptable to most consumers. Individual sensitivities in the population are widely variable, but generally, taste and odour complaints occur at levels of 3–3.7 mg/L of monochloramine. Although taste and odour concerns do not make the water unsafe to consume, they should be taken into account during the selection of operational and management strategies for the water treatment and distribution systems. The primary concern when using monochloramine for secondary disinfection is maintain the microbiological safety of the drinking water supply during distribution in order to protect public health.

Taste and odour issues can be indicators that operational changes may be required to address causal issues such as water age, loss of monochloramine stability, formation of dichloramine. Utilities should establish operational targets for a disinfectant residual concentration appropriate for their system. The concentration should allow them to meet their water quality objectives such as microbial protection, minimal formation of disinfection by-products, nitrification prevention, biological stability and corrosion control.

Dialysis treatment providers at all levels—large facilities/hospitals, small community facilities, mobile units, providers for independent/home dialysis—as well as aquarium owners should be notified that water is chloraminated.

3.1 Monitoring

Utilities using chloramines for secondary disinfection should, at minimum, monitor total and free chlorine residual daily in water leaving the treatment plant and throughout the distribution system. Disinfectant residual sampling should be conducted at the point of entry (baseline) and throughout the distribution system. This ensures that the target chloramine level is being applied at all times and provides a comparison for residual levels observed throughout the distribution system. Sample locations should be chosen to represent all areas of the distribution system. Key points for sampling also include the entry point to the distribution system (baseline), storage facilities, upstream and downstream of rechloramination booster stations, in areas of low flow or high water age, in areas of various system pressures, in mixed zones (blend of chlorinated and chloraminated water) and in areas with various sizes and types of pipe material. Utilities need to carefully consider impacts on the distribution system prior to blending chloraminated and chlorinated waters to meet their specific situation so as to ensure that water will have acceptable disinfection. Some utilities should also consider increasing the frequency of sampling during warmer months (August to October are typical nitrification months). A useful strategy is to target an increased number of remote locations throughout the drinking water distribution system, including dead ends and areas of low flow, to provide a more representative assessment of residual achieved and to identify any problem areas. Dedicated sampling taps are an ideal approach for residual sampling. Hydrants can also be used for residual sampling, following best management practices and utilizing proper equipment to obtain an appropriate sample. Additional samples can be added for investigative purposes. Having operators well-trained in the use of field testing methods for free and total chlorine is important for ensuring the accuracy of measurements.

For water utilities that are chloraminating, it is important to monitor for nitrification events. It is recommended that system-specific monitoring plans be developed that include actions to be taken under various water quality conditions relating to nitrification. This plan should include monitoring of finished water and of locations throughout the distribution system including storage facilities and areas of low flow and high water age. Parameters to monitor should include monochloramine, total chloramine, free and total chlorine, free and total ammonia, nitrite, nitrate, adenosine triphosphate (ATP), temperature, pH and alkalinity. Results should be graphed to evaluate trends and establish alert and action levels that are system-specific. The percent of monochloramine to total chloramine should also be calculated and graphed to determine if organic chloramines are present. Organic chloramines are undesirable because they are an ineffective biocide (see Sections 4.4.3 and 6.0). By evaluating trends, water utilities can quickly observe any increase or decrease in concentrations and take appropriate action, if required (e.g., enhanced monitoring, reformation of chloramines, boost residual). The monitoring frequency should be established in conjunction with the responsible drinking water authority. An example monitoring plan may include testing twice per month between December and May and increasing to once per week leading up to and during the nitrification season (June through October). Changes in the trends of nitrification parameters should also trigger more frequent monitoring. Utilities that undertake comprehensive preventive measures and have baseline data indicating that nitrification does not occur in the system may conduct less frequent monitoring. Monitoring of general microbiological populations in drinking water distribution systems is also useful for assessing system water quality. The frequency of monitoring should be determined based on a system-specific assessment and ideally should be sufficient to characterize short-term and long-term temporal (e.g., hours, days, weeks, months) changes.

More information on monitoring for nitrite, nitrate, ammonia and microbiological populations can be found in the appropriate Health Canada guideline technical and guidance documents for each of these parameters available on Health Canada's website.

Part II. Science and Technical Considerations

4.0 Identity, use and sources in the environment

Chloramines are oxidizing compounds containing one or more chlorine atoms attached to a nitrogen atom. In the literature, the term "chloramine" refers to both inorganic and organic chloramines. Health effects of organic chloramines (formed when dissolved organic nitrogen reacts with free chlorine or inorganic chloramine (Lee and Westerhoff, 2009)) are beyond the scope of the present document and will not be discussed. Throughout the document, the term "chloramines" will refer only to inorganic chloramines, unless otherwise specified.

Inorganic chloramines consist of three chemically related compounds: monochloramine, dichloramine and trichloramine. Only mono- and dichloramine are very soluble in water. The volatility varies depending on the compound, with trichloramine being the most volatile. Their physical properties are provided in the table below:

Table 1. Physical and chemical properties of inorganic chloramine compounds
Parameter Inorganic chloramine compounds
Monochloramine Dichloramine Trichloramine
Synonym ChloramideTable 1 footnote d ChlorimideTable 1 footnote d Nitrogen trichlorideTable 1 footnote b
CAS No. 10599-90-3 3400-09-7 10025-85-1
Molecular formula NH2Cl NHCl2 NCl3
Molecular weightTable 1 footnote a 51.48 85.92 120.37
Water solubilityTable 1 footnote a Soluble Soluble Limited to hydrophobic
Boiling point 486°C (predicted)Table 1 footnote e 494°C (predicted)Table 1 footnote e NA
pKTable 1 footnote a 14 ± 2 7 ± 3 NA
Henry's Law constant—Kaw (estimated, at 25°C)Table 1 footnote c 0.00271 0.00703 1
Vapour pressure (at 25°C) 1.55 × 10–7 PaTable 1 footnote c 8.84 × 10–8 PaTable 1 footnote c 19.99 kPaTable 1 footnote b

NA = not available

Chloramines have been used for almost 90 years as disinfectants to treat drinking water. Although chloramines are less efficient than free chlorine in killing or inactivating pathogens, they generate no trihalomethanes (THMs) or haloacetic acids (HAAs). They are also more stable than free chlorine, thus providing longer disinfection contact time within the drinking water distribution system. Because of these properties, chloramines are mainly used as secondary disinfectants to maintain a disinfectant residual in the distribution system and are generally not used as primary disinfectants (Shorney-Darby and Harms, 2010).

Of the three chloramines, monochloramine is the preferred species for use in disinfecting drinking water because of its biocidal properties and relative stability, and because it rarely causes taste and odour problems when compared with dichloramine and trichloramine (Kirmeyer et al., 2004).

Chloramines do not occur naturally (IARC, 2004). They may be intentionally produced or generated as by-products of drinking water chlorination, including in groundwater systems that undergo chlorination in the presence of natural ammonia, as well as in chlorinated wastewater effluents (WHO, 2004; Hach, 2017).

For disinfection purposes, chloramines are formed through a process called chloramination (U.S. EPA, 1999). Chloramination involves the addition of ammonia (NH3) to free aqueous chlorine (hypochlorous acid, HOCl). This mixture can lead to the formation of inorganic compounds, such as monochloramine, dichloramine, and trichloramine (NHMRC, 2011).

NH3 + HOCl        ->      NH2Cl (monochloramine) + H2O
NH2Cl + HOCl      ->      NHCl2 (dichloramine) + H2O
NHCl2 + HOCl      ->      NCl3 (trichloramine) + H2O

Chloramine speciation mainly depends on the chlorine to ammonia ratio (Cl2:NH3-N) and the pH, but also depends on the temperature, contact time and mixing (Randtke, 2010). Cl2:NH3-N ratios of ≤5:1 by weight (equivalent to ratios of ≤1:1 by mole) are optimum for monochloramine formation. The Cl2:NH3-N ratio by weight is defined as the amount of chlorine added in proportion to the amount of ammonia added (in milligrams); all the Cl2:NH3-N ratios presented in the following document are reported by weight. In idealized conditions, ratios between 5:1 and 7.6:1 favour dichloramine production, whereas trichloramine is produced at higher ratios. Monochloramine formation occurs under neutral to alkaline conditions (pH 6.5–9.0) (monochloramine formation occurs most rapidly at a pH of 8.3), whereas acidic conditions are optimum for the formation of dichloramine (pH 4.0–6.0) and trichloramine (pH <4.4) (Kirmeyer et al., 2004).

Under typical drinking water treatment conditions (pH 6.5–8.5) and with a Cl2:NH3-N ratio of <5:1 (a ratio between 4.5:1 and 5:1 is typically accepted as optimal for chloramination), both monochloramine and dichloramine are formed with a much higher proportion of monochloramine (AWWA, 2006). For example, when water is chlorinated with a Cl2:NH3-N ratio of 5:1 at 25°C and pH 7.0, the proportions of monochloramine and dichloramine are 88% and 12%, respectively (U.S. EPA, 1994a). For its part, trichloramine can be formed in drinking water at pH 7.0 and 8.0, but only if the Cl2:NH3-N ratio is increased to 15:1 (Kirmeyer et al., 2004). Thus, under usual water treatment conditions, monochloramine is the principal chloramine encountered in drinking water. Elevated levels of dichloramine and trichloramine in drinking water may occur, but are due to variations in the quality of the raw water (e.g., pH changes) or to accidental changes in the Cl2:NH3-N ratio (Nakai et al., 2000; Valentine, 2007).

Chloramines (mono-, di- and trichloramines) can be found in media other than drinking water. In swimming pools, for example, they are disinfection by-products (DBPs) incidentally formed from the decomposition, via chlorination, of organic-nitrogen precursors, such as urea, creatinine and amino acids, originating from human excretions (e.g., sweat, feces, skin squama, urine) (Li and Blatchley, 2007; Blatchley and Cheng, 2010; Lian et al., 2014).

Chloramines are also formed when wastewater effluents or cooling waters with free ammonia present are treated with chlorine (U.S. EPA, 1994a). In the food industry, they may result from the reaction between hypochlorite and nitrogen compounds coming from the proteins released by vegetables or animals (Massin et al., 2007). In a home, chloramine fumes (a combination of monochloramine and dichloramine forming a noxious gas) can be produced when bleach and ammonia are accidentally mixed for cleaning purposes (Gapany-Gapanavicius et al., 1982).

4.1 Environmental fate

This section presents the environmental fate of chloramines in the water phase. The decomposition of monochloramine in water is affected by many factors with the rate increasing due to: increased temperature and inorganic carbon, increase of Cl2:NH3-N ratio causing formation of dichloramine and trichloramine from monochloramine, initial chloramine concentration, and decreased pH (Wilczak et al., 2003b). Using decay experiments, Vikesland et al. (2001) reported that at pH 7.5, monochloramine has a half-life of over 300 h at 4oC, whereas it decreases to 75 h at 35oC. Autodecomposition of aqueous monochloramine to dichloramine will occur by one of two pathways: hydrolysis and acid-catalyzed disproportionation, both of which are described in Wilczak et al. (2003b).

When the pH is neutral, trichloramine in water will slowly decompose by autocatalysis to form ammonia and HOCl (U.S. EPA, 1994a). Trichloramine has limited solubility. Since it is extremely volatile, it will volatilize into air (U.S. EPA, 1994b; Environment Canada and Health Canada, 2001). By contrast, according to the physico-chemical properties listed in Table 1, mono- and dichloramine are very soluble in water and not very volatile.

4.1.1 Impact of chloramines on aquatic life

Chloramines enter the Canadian aquatic environment primarily through municipal wastewater release (73%) and release of drinking water (14%) (e.g., water main flushing and breaks, and storage discharges); other minor sources are possible (Pasternak et al., 2003). Release of chloraminated water (total chlorine of 2.53 mg/L and 2.75 mg/L) as a result of a drinking water main break reportedly caused two large fish kills in the Lower Fraser River watershed (Nikl and Nikl, 1992). To mitigate the impact of chlorine or chloramines, aquarium owners must ensure the use of proper aeration or chlorine/chloramine quenching (Roberts and Palmeiro, 2008).

4.2 Terminology

This section provides definitions for some relevant terms used in this document, as adapted from the American Water Works Association (AWWA, 1999; Symons et al., 2000):

4.3 Chemistry in aqueous media

The objectives of chloramination are to maximize monochloramine formation, to minimize free ammonia, and to prevent excess dichloramine formation and breakpoint chlorination. Chloramine formation is governed by the reactions of ammonia (oxidized) and chlorine (reduced); its speciation is principally determined by the pH and the Cl2:NH3-N weight ratio. The reaction rate of monochloramine formation depends on the pH, the temperature, and the Cl2:NH3-N weight ratio. Ideally, a weight ratio of 4.5:1–5:1 will help minimize free ammonia and reduce the risk of nitrification (AWWA, 2006).

The breakpoint chlorination curve can be used to illustrate the ideal weight ratio where monochloramine production can be maximized. For a utility wishing to produce monochloramine, the breakpoint ratio should be determined experimentally for each water subply (AWWA, 2006). Figure 1 shows an idealized breakpoint curve that occurs between pH 6.5 and 8.5 (Spon, 2008). Initially, monochloramine is formed, and once the Cl2:NH3-N weight ratio is greater than 5:1, monochloramine formation decreases because no free ammonia is available to react with the free chlorine being added. The reaction of free chlorine with monochloramine leads to the formation of dichloramine. When high enough Cl2:NH3-N weight ratios are achieved, breakpoint chlorination will occur. The breakpoint curve is characterized by the "hump and dip" shape (Figure 1). Dichloramine undergoes a series of decomposition and oxidation reactions to form nitrogen-containing products, including nitrogen, nitrate, nitrous oxide gas and nitric oxide (AWWA, 2006). Trichloramine, or nitrogen trichloride, is an intermediate during the complete decomposition of chloramines. Its formation depends on the pH and the Cl2:NH3-N weight ratio and may appear after the breakpoint (Kirmeyer et al., 2004; Randtke, 2010; AWWA, 2013).

After breakpoint, free chlorine is the predominant chlorine residual, not monochloramine. However, the reaction rate of breakpoint chlorination is determined by the formation of monochloramine and the formation and decay rates of dichloramine and trichloramine, reactions that are highly dependent on pH. The theoretical Cl2:NH3-N weight ratio for breakpoint chlorination is 7.6:1; however, the actual Cl2:NH3-N ratio varies from 8:1 to 10:1, depending on the pH, the temperature, and the presence of reducing agents, such as iron, manganese, sulphide and organic compounds. Reducers create a chlorine demand; i.e., they compete with the free chlorine added, potentially limiting the chlorine available to react with ammonia (Kirmeyer et al., 2004; AWWA, 2006; Muylwyk, 2009). It is therefore important that each utility generates a site-specific breakpoint curve experimentally. The breakpoint curve is determined by analysing water for various forms of chlorine at different chlorine dosages (AWWA, 2006).

Automated systems can be used to monitor and maintain the weight ratio. For example, the Regional Municipality of York (2019) in Ontario uses an automated system which measures free chlorine and then performs calculations to determine required ammonia dose based on many variables. The goal is to maximize the proportion of monochloramine to total chloramine.

Figure 1: Idealized breakpoint chlorination curve (modelled after Griffin and Chamberlin, 1941; Spon, 2008; and Randtke, 2010). At first, monochloramine is formed until the Cl2:NH3-N weight ratio exceeds 5:1 after which formation decreases. Then dichloramine is formed. Once the Cl2:NH3-N weight ratio is high enough, breakpoint chlorination occurs and the breakpoint curve is characterized by the "hump and dip" shape.

Figure 1. Idealized breakpoint chlorination curve (modelled after Griffin and Chamberlin, 1941; Spon, 2008; and Randtke, 2010). At first, monochloramine is formed until the Cl<sub>2</sub>:NH<sub>3</sub>-N weight ratio exceeds 5:1 after which formation decreases. Then dichloramine is formed. Once the Cl<sub>2</sub>:NH<sub>3</sub>-N weight ratio is high enough, breakpoint chlorination occurs and the breakpoint curve is characterized by the "hump and dip" shape. Text description follows.
Figure 1 - Text Description

A graph that shows the idealized breakpoint chlorination curve for achieving maximum monochloramine formation while also minimizing free ammonia. The verticle axis shows concentration with no units and the horizontal axis shows a scale of chlorine to ammonia weight ratios (Cl2:NH3-N) ranging from 1 to 9. The graph shows that at a weight ratio ranging from 4.5:1 to 5:1 (the ideal zone), monochloramine formation and total chlorine are at a maximum while free ammonia reaches a concentration of zero. As the weight ratio increases, dichloramine formation begins increasing after a ratio of about 5.2:1, reaches a maximum at about 5.8:1 and drops to zero at about 7.2:1, where monochoramine formation also reaches zero. At a ratio of about 7.7:1, break point chlorination is depicted with the total chlorine curve reaching its minimum concentration. Beyond break point chlorination and beyond a ratio of 9:1, both total chlorine and trichloramine formation steadily increase.

4.4 Application to drinking water treatment

Chloramines are used in water treatment, generally as a secondary disinfectant for the distribution system; they can also form unintentionally as by-products of the chlorination of drinking water in the presence of natural ammonia. Chloramines can be used to help reduce the formation of common disinfection by-products such as trihalomethanes and haloacetic acids in the distribution system; however, they can also react with natural organic matter to form other disinfection by-products. Chloramines may impart tastes and odours to treated water, but to a lesser extent compared to free chlorine.

4.4.1 Chloramines in water treatment

The mechanisms by which monochloramine inactivates microbiological organisms are not fully understood (Jacangelo et al., 1991; Coburn et al., 2016). It has been suggested that free chlorine and chloramines react with different functional groups in the cell membrane (LeChevallier and Au, 2004). For monochloramine, the proposed mode of action is the inhibition of such protein-mediated processes as bacterial transport of substrates, respiration, and substrate dehydrogenation (Jacangelo et al., 1991; Coburn et al., 2016). Experiments with bacteria indicated monochloramine was most reactive with sulphur-containing amino acids (LeChevallier and Au, 2004; Rose et al., 2007). Monochloramine did not severely damage the cell membrane or react strongly with nucleic acids. It is a more selective reactant than free chlorine and seems to act in more subtle ways at concentrations found in drinking water (Jacangelo et al., 1991). Inactivation with monochloramine appears to require reactions at multiple sensitive sites (Jacangelo et al., 1991).

4.4.2 Primary disinfection

Primary disinfection is the application of a disinfectant in drinking water treatment with a primary objective of achieving the necessary microbial inactivation. It is important to allow for sufficient contact time to ensure that breakpoint chlorination has occurred and primary disinfection is complete.

Monochloramine is much less reactive than free chlorine, has lower disinfecting power and is generally not used as a primary disinfectant because it requires extremely high CT valuesFootnote 1 to achieve the same level of inactivation as free chlorine (Jacangelo et al., 1991, 2002; Taylor et al., 2000; Gagnon et al., 2004; LeChevallier and Au, 2004; Rose et al., 2007; Cromeans et al., 2010).

4.4.3 Secondary disinfection

Secondary disinfection may be applied to the treated water as it leaves the treatment facility or at rechloramination points throughout the distribution system, to introduce and maintain a disinfectant residual in the drinking water distribution system.

The main function of the residual is to protect against microbial regrowth (LeChevallier and Au, 2004). The disinfectant residual can also serve as a sentinel for water quality changes. A drop in residual concentration can provide an indication of treatment process malfunction, inadequate treatment, a break in the integrity of the distribution system, biofilm growth or water age (LeChevallier, 1998; Haas, 1999; AWWA, 2017).

Monochloramine is slower to react than free chlorine and, in treated drinking water, can provide a more stable and longer-lasting disinfectant residual in the distribution system (Jacangelo et al., 1991; U.S. EPA, 1999; LeChevallier and Au, 2004; Cromeans et al., 2010). However, combined chlorine residuals do not work as well as sentinels of potential post-treatment contamination events than free chlorine residuals. Declines in combined chlorine measurements may not always be large enough or rapid enough to alert utilities that a contamination problem has occurred within the distribution system (Snead et al., 1980; Wahman and Pressman, 2015). Also, a drop in residual may be due to nitrification rather than post-treatment contamination (Wahman and Pressman, 2015). Disinfectant residual monitoring should be conducted alongside other parameters as part of broader programs for microbiological quality and nitrification.

Organic chloramines provide little to no disinfection (Feng, 1966; Donnermair and Blatchley, 2003). In bench-scale experiments, Lee and Westerhoff (2009) investigated the difference between chlorination and chloramination on organic chloramine development in water containing natural organic matter. Organic chloramines are formed more rapidly through the process of chlorination (significant amounts after a 10 minute contact time) than through chloramination (minimal amount after 2 hours). It is common practice to provide sufficient contact time with chlorine to achieve breakpoint chlorination (which is important to ensure proper primary disinfection) followed by ammonia addition. The risk of generating organic chloramines is an important consideration, although less important than achieving primary disinfection targets.

4.4.4 Formation of chloraminated disinfection by-products

Chloramines are often used as an alternative secondary treatment to free chlorine in order to meet DBP compliance based on HAAs and THMs; however, chloramines also react with natural organic matter (NOM) to form other DBPs such as iodinated disinfection by-products (I-DBPs) and nitrosamines (Richardson and Ternes, 2005; Charrois and Hrudey, 2007; Hua and Reckhow, 2007; Richardson et al., 2008; Nawrocki and Andrzejewski, 2011). Hydrazine can also form as a result of abiotic reactions of ammonia and monochloramine (Najm et al., 2006).

I-DBPs are more readily formed in chloraminated systems. Monochloramine oxidizes iodide to hypoiodous acid quickly, but the reaction with NOM is slow, allowing sufficient time for the formation of I-DBPs (Singer and Reckhow, 2011). Chlorine and ozone can also oxidize iodide to hypoiodous acid; however, the iodide is further oxidized to iodate, forming only trace to minimal amounts of I-DBPs (Hua and Reckhow, 2007). In bench-scale formation experiments using simulated raw water, Pan et al. (2016) found that chloraminated water formed more polar I-DBPs than water treated with either chlorine dioxide or chlorine. The authors also noted that as pH increased (from 6 to 9), the formation of polar I-DBPs decreased. Water quality factors such as pH and ratios of dissolved organic carbon, iodide and bromide have been demonstrated to play an important role in determining the species and abundance of iodated trihalomethanes formed under drinking water conditions (Jones et al., 2012).

N-Nitrosodimethylamine (NDMA) is a nitrogen-containing DBP that may be formed during the treatment of drinking water, particularly during chloramination and, to a lesser extent, chlorination (Richardson and Ternes, 2005; Charrois and Hrudey, 2007; Nawrocki and Andrzejewski, 2011). The key to controlling the formation of NDMA lies in limiting its precursors, including dichloramine (Health Canada, 2011). Additionally, Krasner et al. (2015) demonstrated that several pre-oxidation technologies were effective in destroying the watershed-derived NDMA precursors (ozone > chlorine > medium pressure UV > low pressure UV > permanganate). Uzen et al. (2016) observed that site-specific factors such as upstream reservoirs, wastewater discharge, and mixing conditions can affect NDMA formation potential and should be characterized for each individual site. More detailed descriptions of precursors and treatment options can be found in Krasner et al. (2015), Woods et al. (2015), and Uzen et al. (2016). Cationic polymers containing diallyldimethylammonium chloride, used in water treatment, can also act as a source of NDMA precursors (Wilczak et al., 2003a).

Under certain conditions, hydrazine can form through a reaction between ammonia and monochloramine. Najm et al. (2006) found that at low concentrations of free ammonia-nitrogen (<0.5 mg/L) and at pH <9, less than 5 ng/L of hydrazine was formed, but that increasing either ammonia concentrations or pH also increased hydrazine formation. Davis and Li (2008) obtained 13 samples from six chloraminated drinking water utilities and found hydrazine above the detection limit of 0.5 ng/L in 7 of the samples (0.53–2.5 ng/L). Hydrazine production was found to be greatest where treatment processes had high pH (e.g., lime softening). Several treatment practices were identified to minimize hydrazine production, including delaying chloramination until the pH was adjusted (i.e., recarbonation step) and managing the Cl2:NH3-N ratio to minimize the free ammonia concentration (Najm et al., 2011).

4.4.5 Taste and odour considerations

Consumer concerns regarding chloramines in drinking water are often related to taste and odour issues, although the taste and odour of chloramines are generally less noticeable and less offensive to consumers than those of free chlorine. The principal chloramine species, monochloramine, normally does not contribute significantly to the objectionable taste and odour of drinking water when present at concentrations of less than 5 mg/L (Kirmeyer et al., 2004). Di- and tri- chloramines are more likely to cause complaints, especially if they comprise more than 20% of the chloramine concentration in the drinking water (Mallevialle and Suffett, 1987).

Several studies conducted with panels or volunteers to determine the taste and odour thresholds of chloramines in water showed that the perception of taste and odour was highly variable. Krasner and Barrett (1984) used linear regression of data compiled from a trained panel of moderate- to highly-sensitive individuals to derive a taste threshold of 0.48 mg/L and an odour threshold of 0.65 mg/L for monochloramine. Only the most sensitive panelists could detect monochloramine in the range of 0.5–1.5 mg/L (Krasner and Barrett, 1984). By contrast, a taste threshold of 3.7 mg/L was determined using untrained volunteers from the public (Mackey et al., 2004). Similarly, Shorney-Darby and Harms (2010) found that concentrations of 5 mg/L of monochloramine in drinking water were unlikely to cause taste and odour complaints. Lubbers and Bianchine (1984) found a wide variability in individual perception of chloramine taste. Although a dose of 24 mg/L was slightly (6/10) to very (2/10) unpleasant to most volunteer test subjects (n = 10), one subject could not detect a taste and another did not find it objectionable.

By contrast, the presence of dichloramine and trichloramine was detected at much lower concentrations. Krasner and Barrett (1984) determined the taste and odour thresholds for sensory detection of dichloramine to be 0.13 mg/L and 0.15 mg/L, respectively. Objectionable tastes and odours were noted between 0.9 and 1.3 mg/L and to a lesser extent at 0.7 mg/L (Krasner and Barrett, 1984). Shorney-Darby and Harms (2010) stated that taste and odour complaints were not expected when dichloramine was below 0.8 mg/L. However, the Metropolitan Water District of Southern California felt that 0.5 mg/L was a better cut-off (Krasner and Barrett, 1984). A similar odour threshold concentration was seen for trichloramine at 0.02 mg/L (Shorney-Darby and Harms, 2010). Utilities can address taste and odour issues through a variety of operational strategies to address water age, disinfection demand, hydraulic issues (such as dead ends and low-flow areas), bacteria growth and dosing issues (Kirmeyer et al., 2004).

Optimizing treatment for monochloramine production reduces the potential to form dichloramine and trichloramine, resulting in water with the least flavour. Reactions of chloramines with organic compounds in water can form by-products that also cause taste and odour issues.

Operational strategies to reduce taste and odour issues include treating the water to remove taste and odour precursors, flushing the distribution system and reducing the water age in the distribution system.  Available studies (see Section 9.0) and surveys have not indicated evidence of adverse health effects associated with exposure to monochloramine at concentrations used in drinking water disinfection. Although levels commonly found in drinking water are within an acceptable range for taste and odour, individual sensitivities regarding the acceptability of water supplies can vary. In addition, where elevated chloramine concentrations are required in order to maintain an effective disinfectant residual throughout the distribution system, the median taste thresholds may be exceeded. Therefore it is important that utilities contemplating a conversion to monochloramine remain aware of the potential for taste and odour concerns during the selection of operational and management strategies.

Evaluating consumer acceptability and tracking consumer complaints are important for verifying water quality under a risk-based approach to drinking water management, such as a Water Safety Plan approach (WHO, 2005; Bartram et al., 2009). Communicating with consumers on issues such as taste and odour is a key part of assessing and promoting the acceptability of drinking water supplies to the public. Guidance materials to help utilities  that use chloramination develop programs for communication and consumer feedback is available elsewhere (Whelton et al., 2007).

5.0 Exposure

No environmental data were found for inorganic chloramines in sediments, soils and ambient air (Environment Canada and Health Canada, 2001). As a result, drinking water is considered the primary source of exposure for this assessment.

5.1 Water

Chloramines can be measured as combined chlorine residuals, corresponding to the difference between total chlorine residual and free chlorine residual. This method has limitations because the combined chlorine value does not determine the individual concentrations of monochloramine, dichloramine, trichloramine and organochloramines present in drinking water, and because the free chlorine measurement is not always accurate in the presence of high levels of chloramines. Individual chloramines can be differentiated using multi-stage procedures, but interferences such as organic chloramines may in some cases result in misleading measurements (e.g., overestimation of monochloramine concentrations) (Lee et al., 2007; Ward, 2013) (see Section 6.1 for additional information). Monochloramine can also be measured directly using portable analysers and online continuous analysers (see Section 6.2) with less interference.

Limited provincial data are reported in Table 2 for both those systems where chloramines are intentionally produced through chloramination and those systems where chloramines are produced by chlorinating in the presence of background ammonia. Generally, levels of chloramines or combined chlorine are below 3 mg/L; only a few values exceeding 3 mg/L were reported.

Table 2. Residual disinfectant levels in some distribution systems across Canada
Province
(Year)
# of sites Range: min–max
(No. detects/total no. of samples)
Total Chlorine Free Chlorine Chloramines Mono- Di- Tri-
BCTable 2 footnote a
(2015)
37 0.01–1.86 ppm
(1913/1923)
0–0.97
ppm (1908/1918)
--- 0.6–1.51 mg/L
(51/51)
0–0.7 µg/L
(47/51)
---
BCTable 2 footnote a
(2016)
37 0.03–1.89 ppm
(1931/1931)
0–0.93 ppm
(1918/1929)
--- 0–1.3 mg/L
(50/51)
0–0.61 µg/L
(48/50)
---
ONTable 2 footnote b,Table 2 footnote c
(2012–2017)
108 --- --- 0–2.64 mg/L
(1498/1500)
--- --- ---
ONTable 2 footnote d
(2016)
2 --- --- 1.26–2.20 mg/L (1460/1460) 1.21–2.03 mg/L (1460/1460) no data no data
QCTable 2 footnote e
(2013–2015)
3Table 2 footnote f --- --- 0.07–1.80
mg/L (7/7)
--- --- ---
QCTable 2 footnote e
(2013–2015)
5Table 2 footnote g 0.01–4.6 mg/L (3432/3432) 0–4.2 mg/L
(2888/3432)
0–3.24 mg/L
(3423/3432)
--- --- ---
QCTable 2 footnote e
(2013–2015)
11Table 2 footnote i 0–7.05 mg/L
(2916/2924)
0–5.5 mg/L
(2685/2924)
0–2.47 mg/L
(2916/2924)
--- --- ---
SKTable 2 footnote h (2006–2015) 18 0.62–3.24 mg/L
(25/25)
0.01–1.18 mg/L
(24/24)
0.01–3.9 mg/L
(28/28)
1.28–3.11 mg/L
(24/24)
--- ---

--- = no data

5.2 Air

Chloramines may be encountered in the ambient air of food industry facilities that typically use large quantities of disinfecting products. For example, total concentrations of chloramines (mainly trichloramine) have been reported in the ambient air of green salad processing plants (e.g., 0.4–16 mg/m3; Hery et al., 1998) and of turkey processing plants (e.g., 0.6–1 mg/m3 average concentrations; Kiefer et al., 2000).

5.3 Swimming pools and hot tubs

Chloramines are found in indoor, and to a lesser extent in outdoor swimming pool water and their surrounding atmosphere as DBPs of chlorination (Simard et al., 2013). In general, swimmers ingest 14 ml (geometric mean) of pool water per swim with geometric means of 38.2 ml, 22.1 ml and 10.4 ml being ingested by children, teens and adults, respectively (Dufour, et al., 2017). Simard et al. (2013) showed that indoor pool water contained 739 µg/L (range: 311 µg/L to 1723 µg/L) of total chloramines with 55% as mono-, 12% as di- and 33% as trichloramines (calculated as 404, 88 and 243 µg/L, respectively based on average concentration) while the average concentration of total chloramines in outdoor swimming pools was 142 µg/L (range: 8 µg/L to 854 µg/L) and consisted of 51% mono-, 16% di- and 33% trichloramines (calculated as 72, 23 and 47 µg/L, respectively based on average concentration). Using the above values, monochloramine intake while swimming by children, teens and adults can be calculated as 3 to 15 µg, 2 to 9 µg and 1 to 4 µg, respectively.

Due to its high volatility and low solubility, trichloramine is the predominant species present in the swimming pool atmosphere. Numerous papers report trichloramine concentrations in the atmosphere of indoor swimming pools (including water parks), with mean concentrations ranging from approximately 114 µg/m3 to 670 µg/m3 (Carbonnelle et al., 2002; Thickett et al., 2002; Jacobs et al., 2007; Dang et al., 2010; Parrat et al., 2012). The levels of airborne trichloramine are influenced by such factors as number of swimmers, organic compounds (mainly urine and sweat) introduced into the water by swimmers, air ventilation, as well as water temperature, circulation and movement (splashing, waves, etc.) (Carbonnelle et al., 2002; Jacobs et al., 2007; Parrat et al., 2012).

5.4 Multiroute exposure through drinking water

The physico-chemical characteristics of mono- and di- chloramine (e.g. they are water soluble but not volatile; see Table 1) indicate that dermal and inhalation are unlikely routes of exposure. Conversely, trichloramine is very volatile and not soluble in water. In addition, trichloramine is relatively unstable in water and is only formed beyond breakpoint (see Figure 1) or under low pH, conditions that are unlikely to occur in treated drinking water. Therefore under normal usage conditions, the ratio of trichloramine to total chloramine is very low and would not contribute significantly to the dermal or inhalation routes of exposure. Consequently, exposure to chloramines via inhalation and dermal routes during showering or bathing is expected to be negligible.

6.0 Analytical methods

Monochloramine can be measured via laboratory methods or through online and portable analysers. Laboratory methods calculate chloramines as the difference between total and free chlorine. The chloramines calculated in this manner include monochloramine, dichloramine, trichloramine and organic chloramines.

Organic chloramines are formed when dissolved organic nitrogen reacts with either free chlorine or inorganic chloramines (Lee and Westerhoff, 2009). Organic chloramines are known interfering agents for both the laboratory amperometric and the N,N-diethyl-p-phenylenediamine (DPD) methods (APHA et al., 2012). Wahman and Pressman (2015) highlighted that organic chloramines can result in the overestimation of monochloramine. Lee and Westerhoff (2009) estimated that utilities are likely to overestimate chloramine residuals by approximately 10% as a result of interference from organic chloramines. Similarly, in Gagnon et al. (2008), a pipe loop distribution system study identified that organic chloramines comprised approximately 10–20% of the total chlorine residual in a chloraminated system. As the chloramines calculated through laboratory methods may contain organic chloramines, the monochloramine residual and therefore the disinfection ability may be overestimated.

Monochloramine can be directly measured through handheld and online analysers based on the indophenol method. The advantage of this method is that interference from organic chloramines and possible overestimation of monochloramine is avoided.

6.1 Laboratory methods

There are no standardized methods to directly measure chloramines. There are, however, several standardized methods to measure total and free chlorine (Table 3). Free chlorine is the sum of chlorinated species that do not contain ammonia or organic nitrogen (i.e., Cl2, HOCl, OCl-, and Cl3-), whereas combined chlorine is the sum of chlorine species that are combined with ammonia (NH2Cl, NHCl2, and NCl3) (Randtke, 2010). Since total chlorine is often used as an assumed proxy for combined chlorine (chloramines), it is important that free chlorine be measured to validate the assumption that none is present. Equation 1 below shows how chloramines (inorganic + organic) can be determined by subtracting free chlorine from total chlorine:

Combined chlorine (chloramines) = Total chlorine – Free chlorine (1)

Method detection limits (MDLs) are dependent on the sample matrix, instrumentation, and selected operating conditions and will vary between individual laboratories. Analyses should be carried out as directed by the responsible drinking water authority in the affected jurisdiction.

When using the DPD colorimetric test, it is important to ensure that field staff is well-trained to do both free and total chlorine measurements. This ensures that false positive results are not inadvertently reported (Spon, 2008). Users should consult with the manufacturer regarding method interferences, interfering substances and any associated corrective steps that may be necessary. Several methods (SM 4500 D, 4500 F, and 4500 G; refer to Table 3) have additional steps (beyond total, free and combined chlorine) that can be used to distinguish between the various chloramine species. Both dichloramine and trichloramine are relatively unstable, and their formation reactions do not proceed to completion under typical drinking water conditions (Randtke, 2010). Specific instructions for mitigating effects of interfering agents (including interference from other chlorine species), optimal analytical performance (including the use of reagent blanks), and reaction times for accurate sample readings are available in the method documents.

Table 3. Standardized methods for total, free and combined chlorine measurement in drinking water
Method
(Reference)
Methodology Residual determined
(MDL)
Comments
ASTM D1253
(ASTM2014)
Amperometric titration Total, free, combined (NA) Reaction is slower at pH >8 and requires buffering to pH 7.
A maximum concentration of 10 mg/L is recommended.
Interferences include cupric, cuprous and silver ions; trichloramine; some N-chloro compounds; chlorine dioxide; dichloramine; ozone; peroxide; iodine; bromine; ferrate; Caro's acid.
SM 4500-Cl D
(APHA et al., 2017)
Amperometric titration Total, free, combined (NA) Dilutions are recommended for concentrations above 2 mg/L. Interferences are trichloramine, chlorine dioxide, free halogens, iodide, organic chloramines, copper, and silver.
Monochloramine can interfere with free chlorine and dichloramine can interfere with monochloramine.
Method can also be used to characterize species (monochloramine and dichloramine).

SM 4500-Cl G (colorimetric)
(APHA et al., 2017)

SM 4500-Cl F
(ferrous)
(APHA et al., 2017)

N,N-diethyl-p-phenylenediamine (DPD)

Total, free, combined
(10 µg/L)

Total, free, combined
(18 µg/L)

Interferences include oxidized manganese, copper, chromate, iodide, organic chloramines.

Method can also be used to characterize species (monochloramine, dichloramine, and trichloramine) in a laboratory setting.

Hach 10260 Rev 1.0
(HACH, 2013)
DPD Chemkey Total (0.04 mg/L), free (0.04 mg/L), combined (NA) Interferences include acidity >150 mg/L CaCO3, alkalinity >250 mg/L as CaCO3 highly buffered or extreme pH samples, bromine, chlorine dioxide, iodine, ozone, organic chloramines, peroxides, oxidized manganese, oxidized chromium.
NA: not available

6.2 Online and portable analysers

Commercial online and portable analysers are available to directly quantify monochloramine; analysis is based on indophenol methods (e.g., SM 4500-NH3 G) (APHA et al., 2017). When measuring the chloramines concentration by the standardized methods discussed in the previous section, the interference of organic chloramines can provide an overestimation of the monochloramine concentration. Indophenol methods for monochloramine residual measurements have been shown to be the most reliable (Lee et al., 2007; Ward, 2013). The portable analysers utilize the indophenol method and have the advantages of directly measuring monochloramine concentration and of eliminating most interferences. Online continuous analysers are also available to measure monochloramine and free and total ammonia.

To make accurate measurements using these units, water utilities should develop a quality assurance and quality control (QA/QC) program such as those outlined in SM 3020 (APHA et al., 2017). In addition, periodic verification of results using an accredited laboratory is recommended. Water utilities should check with the responsible drinking water authority in the appropriate jurisdiction to determine if results from these units are acceptable.

6.3 Sampling considerations

Key points for sampling include the entry point to the distribution system (baseline), storage facilities, upstream and downstream locations of rechloramination booster stations, areas of low flow, areas of high water age, areas of various system pressures, mixed zones (blended water of chlorinated and chloraminated water), and areas with various sizes and types of pipe materials. Utilities need to carefully consider impacts on the distribution system prior to blending chloraminated and chlorinated waters to meet their specific situation so as to ensure that water will have acceptable disinfection (AWWA, 2006). Dedicated sampling taps are an ideal approach for residual sampling (AWWA, 2013). Hydrants can be used for residual sampling following best management practices and utilizing appropriate equipment to obtain a proper sample (U.S. EPA, 2016a). Alexander (2017) recommended targeting remote locations of a distribution system, suggesting that it is preferable to target more locations with fewer samples at those locations vs.  more samples at fewer locations, and recommended taking additional investigative samples. An investigation of disinfectant residuals in Flint, Michigan, revealed that the previous number and location of sampling sites (10 sites) were not sufficient to detect problem areas. These few locations demonstrated adequate disinfectant (chlorine) residual; however, expanding the number and location of sample sites to more representative locations (an additional 24 sites) revealed that the disinfectant residual was problematic (Pressman, 2017). Although chloramine residuals are consumed less readily in the distribution system, it is still important to have an adequate and representative set of sampling sites. Resources outlining approaches for determining the number and location of sampling points for monochloramine disinfectant residual monitoring in the distribution system are available elsewhere (Louisiana Department of Health and Hospitals, 2016).

Alexander (2017) highlighted the importance of proper field testing techniques. Sample vials can become scratched (during transportation in a truck) or dirty, leading to inaccurate readings. Additionally, plastic vials are prone to the formation of fine bubbles, which can be resolved with slow inversion of the sample. It is important that operators be aware of the challenges of disinfectant residual sampling.

Sampling programs should be reviewed annually to examine historical data, water use patterns/changes, as well as any changes in water treatment or distribution system operation (AWWA, 2013).

7.0 Treatment technology and distribution system considerations

As chloramines are added to drinking water to maintain a residual concentration in the distribution system, or are formed as a by-product of the chlorination of drinking water in the presence of natural ammonia, they are expected to be found in drinking water at the treatment plant, as well as in the distribution and plumbing systems. When considering chloramination, operators may wish to also consult Khiari (2018) for more information. This reference provides a good overview of chloramination considerations as well as links to various Water Research Foundation reports on varied topics associated with chloramination.

7.1 Municipal scale

Monochloramine CT requirements are one to many orders of magnitude greater than those required by free chlorine to achieve similar levels of inactivation of heterotrophic bacteria, E. coli, nitrifying bacteria, enteric viruses and Giardia cysts (LeChevallier and Au, 2004; Wojcicka et al., 2007; Cromeans et al., 2010; Health Canada, 2012b, 2019a, 2019b). Reported CT values also demonstrate that similar to free chlorine, monochloramine is not effective for inactivation of Cryptosporidium oocysts (LeChevallier and Au, 2004; Health Canada, 2019a).

Given the operational benefits of secondary disinfection, operators should strive to maintain a stable disinfectant residual throughout the system. Regularly monitoring distribution system water quality (e.g., disinfectant residual, microbial indicators, turbidity, pH) and having operational and maintenance programs in place (water mains cleaning, cross-connection control, replacements and repairs) are important for ensuring that drinking water is transported to the consumer with minimum loss of quality (Kirmeyer et al., 2001, 2014).

7.1.1 Disinfectant residual and microbial control

To control bacterial regrowth and biofilm formation, a strong, stable disinfectant residual is needed that is capable of reaching the ends of the distribution system. The amount of disinfectant needed to maintain water quality depends on the characteristics of the distribution system, the species of bacteria, the type of disinfectant used, the temperature, the pH, the amount of biodegradable organic material present and available nutrients (C:N:P) (Kirmeyer et al., 2004). When applying monochloramine as a disinfectant, utilities should be aware of their target disinfectant residual value, and adjustments should be made to address monochloramine demand and decay. Manuals have been developed to guide water utilities in establishing goals and disinfection objectives. These manuals recommend that utilities set system-specific disinfectant residual targets based on their water quality objectives and system characteristics, and that they ensure chloramine concentrations leaving the treatment plant are sufficient to achieve their established target residual (Kirmeyer et al., 2004; AWWA, 2013).

Full-scale and laboratory-scale studies have examined the effectiveness of monochloramine residual concentrations in controlling biofilm growth in drinking water distribution systems. Camper et al. (2003) observed that a monochloramine residual of 0.2 mg/L did not completely control biofilm growth on coupons of epoxy, ductile iron, polyvinyl chloride (PVC) or cement. Similarly, Pintar and Slawson (2003) noted that maintaining a low monochloramine residual of 0.2–0.6 mg/L was not sufficient to inhibit the establishment of biofilms containing nitrifying bacteria and heterotrophic bacteria on PVC coupons. Wahman and Pressman (2015) noted that when using inorganic chloramines (monochloramine) a "detectable" total chlorine residual is not sufficient to effectively limit bacterial growth in the distribution system. In full-scale investigations conducted by Norton and LeChevallier (1997) and Volk and LeChevallier (2000), water samples from drinking water systems maintaining monochloramine residual concentrations of >1.0 mg/L maintained lower heterotrophic bacterial counts and had substantially fewer coliform occurrences than systems with lower monochloramine levels. In laboratory experiments, a monochloramine residual of >1.0 mg/L reduced viable bacteria counts by greater than 1.5 log for biofilms grown on PVC, galvanized iron, copper and polycarbonate (LeChevallier et al., 1990; Gagnon et al., 2004; Murphy et al., 2008). When biofilms were grown on iron pipe, a monochloramine residual of >2.0 mg/L was needed to reduce bacterial numbers by greater than 2 log (LeChevallier et al., 1990; Ollos et al., 1998; Gagnon et al., 2004). Monochloramine residual concentrations of >2 mg/L may similarly be required to control the development of nitrifying biofilms (see Section 7.2.4).

Monochloramine is believed to penetrate biofilms better because of its low reactivity with polysaccharides—the primary component of the biofilm matrix (Vu et al., 2009; Xue et al., 2014). In disinfection transport experiments using chlorine/monochloramine-sensitive microelectrodes, monochloramine penetrated biofilm faster and further than free chlorine but did not result in complete inactivation (Lee et al., 2011, Pressman et al., 2012). Work by Xue et al. (2013, 2014) suggests that monochloramine disinfection efficacy and persistence may be affected by the combination of polysaccharide extracellular polymeric substances (EPSs) obstructing monochloramine reactive sites on bacterial cells and protein EPS components reacting with monochloramine. Free chlorine, on the other hand, has decreased biofilm penetration but acts more effectively where it does penetrate (Lee et al., 2011, 2018). These findings highlight the challenges for drinking water utilities that chloraminate but employ free chlorine application periods for nitrification control and also help explain why nitrification is so difficult to stop once it has started (Lee et al., 2011, 2018).

An understanding of system-specific factors that may make it difficult to achieve target residuals is important for maximizing the effectiveness of strategies for secondary disinfection.

7.1.2 Considerations when changing disinfection practices to chloramination

A change in disinfectant (oxidant), including changing secondary disinfection to chloramines, can impact water quality. It is strongly recommended that site-specific water quality/corrosion studies be conducted to capture the complex interactions of water quality, distribution system materials and treatment chemicals used in each individual water system. Some considerations when applying chloramine treatment include nitrification, biofilm, residual loss, the formation of DBPs, the degradation of elastomer materials (Section 7.2.7) and the impacts the distribution system/premise plumbing materials have on corrosion. Zones, within the distribution system, where mixing with chlorinated water may occur should also be considered. The complexities and potential consequences (e.g., corrosion) of changing disinfection practices are further discussed in Section 7.2.5 and additional guidance can be found in AWWA (2006b) and U.S. EPA (1999).

7.1.3 Presence of ammonia in source water

Ammonia may be present in source water and strategies to treat it will depend on many factors, including the characteristics of the raw water supply, the source and concentration of ammonia (including variation), the operational conditions of the specific treatment method and the utility's treatment goal. In some cases, utilities will form chloramines as a strategy to remove naturally occurring ammonia in the raw water supply, while others may use breakpoint chlorination. In both cases, frequent monitoring of relevant parameters (ammonia; combined, total and free chlorine) is needed to ensure that objectives are achieved at all times. More details on ammonia removal technologies and strategies, including biological treatment, can be found in Health Canada (2013a).

For source water containing organic nitrogen, the organic nitrogen will react with chlorine to form organic chloramines. In almost all cases, organic chloramines have no disinfection properties. Organic chloramines will appear as dichloramine when measured through forward titration using either amperometric or DPD methods (Randtke, 2010). As organic chloramines interfere with chloramination analytical methods, the chloramine residual may be overestimated if using amperometric or DPD methods.

7.2 Distribution system considerations

The use of chloramines as a secondary disinfectant can have impacts on the distribution system water quality, including corrosion, nitrification, microbial regrowth and the growth of opportunistic pathogens.

7.2.1 Residual loss

Chloramines auto-decompose through a series of reactions (Jafvert and Valentine, 1992). However, decomposition of chloramines can also be impacted by interaction with pipe materials and NOM, resulting in loss of residual.

Vikesland and Valentine (2000) demonstrated a direct reaction between Fe(II) and the monochloramine molecule, ultimately forming ammonia. Iron oxides play an autocatalytic role in the oxidation of Fe(II) by monochloramine (Vikesland and Valentine, 2002). In bench-scale experiments, Westbrook and DiGiano (2009) compared the rates of chloramine loss at two pipe surfaces: cement-lined ductile iron and tuberculated, unlined cast iron pipe. Chloramine decay occurred faster—from 3.8 mg/L to <1 mg/L in 2 h—for the unlined cast iron pipe compared with the cement-lined ductile iron pipe—from 3.7 mg/L to 1.5 mg/L in 2 h.

Duirk et al. (2005) described two reaction pathways between chloramines and NOM: a fast reaction and a slow reaction where chloramines were hypothesized to hydrolyze to HOCl. The slow reaction was reported to account for the majority of the chloramine loss in this mechanism. Zhou et al. (2013) reported that the mixing ratio of chloramines impacted their decay rate in the presence of NOM; a higher decay rate was observed for chloramines mixed in a ratio of 3:1 than for those mixed at 4:1. Wilczak et al. (2003b) suggested that the nature and characteristics of NOM are important factors in determining how reactive chloramines are with NOM. Additionally; the authors observed that bacterial cell fragments shed from full-scale biologically active granular activated carbon (GAC) filters exerted a significant chloramine demand.

Carbonate has been observed to accelerate monochloramine decay by acting as a general acid catalyst (Vikesland et al., 2001). The same study also showed that nitrite exerted a significant long-term monochloramine demand and demonstrated that although bromide affects chloramine loss, at low concentrations (<0.1 mg/L), bromide did not have a significant role (Vikesland et al., 2001). Wahman and Speitel (2012) explored the important role of HOCl in nitrite oxidation and found that it peaked in the pH range of 7.5–8.5. A 5:1 Cl2:NH3-N ratio was also associated with increased nitrite oxidation (compared with a 3:1 Cl2:NH3-N ratio). Increased oxidation of nitrite by HOCl was also associated with lower nitrite concentration (0.5 mg N/L compared with 2 mg N/L) and decreased monochloramine concentrations (1 mg Cl2/L compared with 4 mg Cl2/L).

Factors such as mixing ratios, pH, and temperature have also been shown to play an important role in chloramine decay rates. In experiments by Zhou et al. (2013), the chloramine decay rate was shown to increase over time (72 h). At pH >7, the decay rate was higher for chloramines mixed at a ratio of 3:1 than for those mixed at 4:1. However, at pH <7, the ratio did not play an obvious role. Chloramine loss was greater at low pH. Generally, chloramine loss increased with increasing temperature and the decay rate was higher for chloramines mixed at a ratio of 3:1 than at 4:1. Research conducted by the Philadelphia Water Department similarly showed monochloramine decay over a 25-day period was slow at lower water temperatures (38% decay at 7°C) and increased at higher temperatures (84% decay at 30°C) (Kirmeyer et al. 2004). Web-based chloramine formation and decay models have been developed by Wahman (2015), based on the unified models described by Jafvert and Valentine (1992) and Vikesland et al. (2001), as well as NOM reactions described by Duirk et al. (2005). A user-input model is also publicly available (U.S. EPA, 2016b).

7.2.2 Rechloramination and temporary breakpoint chlorination

It is important to have a comprehensive distribution management plan to control nitrification and to maintain sufficient disinfectant residual. Recombining the (released) free ammonia in the distribution system by booster chloramination can be used to maintain the ratio near 5:1 throughout the system (Wilczak, 2006). Rechloramination stations can be designed to perform specific actions based on the need to either reform and/or boost chloramines in the water. The free ammonia measured in the water and the desired chloramine residual are used to calculate the amounts of chlorine and ammonia for dosing. The rechloramination station can then either: do nothing; reform chloramines; boost chloramines; or reform and boost chloramines (Meteer and Horsley, 2015). Booster chloramination requires proper control to ensure excess ammonia is not added as this may lead to nitrification (AWWA, 2013).

Previously, periodic free chlorination was conducted in an attempt to stop nitrification. However, it has been shown that this practice is ineffective for long-term nitrification control as the microbial communities are initially sensitive to the change but then return to their stable state once chloramination resumes and nitrification re-commences (Vikesland et al., 2006 ; AWWA, 2013; Gomez-Alvarez et al., 2016). In cases where temporary switches to chlorine were used, the practice may have ranged from a short-term targeted breakpoint of an isolated area of the distribution system to a longer, system-wide chlorine burnout, depending on whether nitrification is localized or widespread (Skadsen and Cohen, 2006; AWWA, 2013). In a full-scale study, the process of switching from chloramines to chlorine and back to chloramines created a breakpoint chlorination reaction, which resulted in a low chlorine residual "transitional front" in the distribution system. Although the HPC temporarily decreased, THMs increased upon switching back to chloramines. While conducting breakpoint chlorination experiments in distribution systems, Rosenfeldt et al. (2009) observed that incorporating systematic flushing was important to achieving chlorine residual.

7.2.3 Opportunistic pathogens

The microbiological ecology of distribution systems and piped plumbing supplies can be influenced by factors such as the disinfection strategy, the operational and water quality parameters, the system materials and the age of the system (Baron et al., 2014; Ji et al., 2015). There are differences in the relative abundance of certain bacterial groups in bulk water and biofilms in the distribution system when providing a disinfectant residual using monochloramine compared with using free chlorine (Norton et al., 2004; Williams et al., 2005; Hwang et al., 2012; Chiao et al., 2014; Mi et al., 2015).

Distribution system and premise plumbing biofilms can serve as reservoirs for opportunistic premise plumbing pathogens (OPPPs) such as Legionella pneumophila, non-tuberculous mycobacteria (e.g., M. avium, M. intracellulare), Pseudomonas aeruginosa and Acinetobacter baumanii (Fricker, 2003; Falkinham, 2015). In developed countries, Legionella and non-tuberculous mycobacteria are increasingly recognized as important causes of waterborne disease for vulnerable sections of the population (Wang et al., 2012, Beer et al., 2015; Falkinham, 2015).

The use of chloramines in the distribution system may contribute to environmental conditions that are less favourable for Legionella and more favourable for non-tuberculous mycobacteria (Norton et al., 2004; Williams et al., 2005; Flannery et al., 2006; Moore et al., 2006; Gomez-Alverez et al., 2012, 2016; Revetta et al., 2013; Baron et al., 2014; Chiao et al., 2014; Gomez-Smith et al., 2015; Mancini et al., 2015). The free-living amoeba Naegleria fowleri is a drinking water pathogen of emerging concern in the southern United States as a result of infections linked to municipal drinking water supplies in Arizona and Louisiana (Bartrand et al., 2014). N. fowleri causes primary amoebic meningoencephalitis (PAM), an extremely rare, but often fatal disease that has been predominantly associated with water inhalation during recreational water activities at warm freshwater locations (Yoder et al., 2012; Bartrand et al., 2014). Two cases in Louisiana represent the first time disinfected tap water had been implicated in N. fowleri infection (Yoder et al., 2012). Strategies for the control of N. fowleri at impacted water utilities that use chloramines include recommendations to: maintain a minimum monochloramine residual of 0.5 mg/L throughout the distribution system, and monitor for nitrification (Robinson and Christy, 1984; NHMRC, 2011; Louisiana Department of Health and Hospitals, 2016).

A well-maintained distribution system is an important component in minimizing microbiological growth in distributed water in premise plumbing systems. Utilities also need to be aware that changes in the microbiological diversity of drinking water systems can occur with changes to disinfection strategies. An understanding of the effects different disinfectants and source waters may have on drinking water system ecology will help maximize the effectiveness of disinfection strategies and minimize such unintended consequences as the potential enrichment of specific microbial groups (Williams et al., 2005; Baron et al., 2014).

7.2.4 Nitrification

Nitrification, the microbiological process whereby ammonia is sequentially oxidized to nitrite and nitrate by ammonia-oxidizing bacteria and nitrite-oxidizing bacteria, respectively, is a significant concern for utilities using chloramines for secondary disinfection. Free ammonia can be present as a result of treatment (e.g., improper dosing or incomplete reactions) or released during chloramine demand and decay (Cunliffe, 1991; U.S. EPA, 2002).

Water quality problems caused by nitrification include the formation of nitrite and nitrate, loss of disinfectant residual, bacterial regrowth and biofilm formation, DBP formation, and decreases in pH (especially in low alkalinity waters) and alkalinity that can lead to corrosion issues, including the release of lead and copper (U.S. EPA, 2002; Zhang et al., 2009, 2010; AWWA, 2013). Growth of nitrifying bacteria leads to a loss in disinfectant residual and increased biofilm production, which further escalates chlorine demand, ammonia release and bacterial regrowth (Kirmeyer et al., 1995; Pintar and Slawson, 2003; Scott et al., 2015). Cometabolism of monochloramine by ammonia-oxidizing bacteria and reactions between chloramines and nitrite are also significant mechanisms for the loss of chloramine residual (Health Canada, 2013b; Wang et al., 2014, Wahman et al., 2016).

Factors favouring nitrification include warm water temperatures, long detention times, and the presence of high levels of organic matter (exerting a chloramine demand that promotes free ammonia release) in the distribution system (Kirmeyer et al., 2004; AWWA, 2013). Nitrification episodes have typically been reported during the summer months when water temperatures range from 20°C to 25°C (Pintar et al., 2000; AWWA, 2013). However, nitrification can also occur in distribution systems at low temperatures (e.g., <10°C) in areas with long (days to weeks) detention times (Pintar et al., 2000; AWWA, 2013).

In general, to prevent nitrification it is recommended that utilities maintain a minimum monochloramine concentration of 2 mg/L leaving the treatment plant and preferably greater than 1.5 mg/L at all monitoring points in the distribution system (U.S. EPA, 1999; Norton and LeChevallier, 1997; AWWA, 2013). Once it has started, nitrification cannot be easily stopped, even by applying high chloramine doses. Nitrifying bacteria have been detected in drinking water distribution systems in the presence of chloramine residuals exceeding 5 mg/L (Cunliffe, 1991; AWWA, 2013). Nitrifying biofilms may harbour viable bacteria even after complete penetration with monochloramine at elevated concentrations (Wolfe et al., 1990; Pressman et al., 2012). Utilities should be aware of the chemistry of their water and system materials before considering conversion from chloramines to free chlorine and vice versa (AWWA, 2013). Development of nitrification monitoring programs is critical for utilities that use chloramines (AWWA, 2013, 2017). When establishing monitoring programs for nitrification in the distribution system, the most useful parameters to monitor over time to understand baseline trends and to monitor for any changes are: chlorine (free and total), monochloramine, total and free ammonia, nitrite, nitrate, temperature, pH, percent of monochloramine, total chloramine and alkalinity (AWWA, 2013; Ballantyne and Meteer, 2018; Regional Municipality of York, 2019). ATP measurements or HPC, as indicators of biostability or microbiological populations, are also considered valuable (AWWA, 2017).

When using flushing for nitrification prevention, it is important that the most appropriate flushing velocity be used to avoid disturbing and releasing legacy deposits into the bulk water. Improper flushing techniques can stir up and potentially spread contaminants around the flushed area or deeper into the distribution system, thus increasing public health risk. Hill et al. (2018) reported that the following conditions can disturb legacy deposits: excessive flushing rate or velocity; insufficient flushing rate or velocity; lack of directional control; and inadequate flush duration. It is important that water utilities identify and implement the most appropriate flushing technique. Automatic flushing stations are recommended if the goal is to turnover bulk water in an area due to water age or poor circulation (Hill et al., 2018).

Detailed information on the causes of nitrification and its monitoring, prevention and mitigation are available elsewhere (AWWA, 2013; Ballantyne and Meteer, 2018). It is important to note that even the stringent control of excess free ammonia and the maintenance of a proper Cl2:NH3-N ratio may not always be effective in preventing nitrification. This is due to the fact that chloramines in the distribution system will start to decay based on water quality conditions and water age, releasing free ammonia into the water (AWWA, 2013).

7.2.5 Lead and copper release

Several studies have compared the impact of chlorine and chloramines on lead release. In a study where three utilities had converted from chlorine to chloramines, Dyksen et al. (2007) found that one utility had decreased lead levels, that a second utility experienced an increase in lead release that was below the U.S. EPA Lead and Copper Rule (LCR) action level, while the third utility experienced increases that exceeded the LCR action level. As part of an all-encompassing strategy to reduce lead exposure and provide effective corrosion control, the San Francisco Public Utilities Commission increased the pH of its low alkalinity, soft water supply (pH 8.6–9.4). As a result, San Francisco did not experience lead and copper release; nor did it impact the utility's ability to meet the LCR action levels following conversion to chloramines from chlorine (Wilczak et al., 2010).

Conversely, conversion from chlorine to chloramines as a secondary disinfectant was implicated as a cause of increased lead release in the Washington, DC water crisis in 2004, in which measured lead levels were greater than 300 μg/L  (Renner, 2004; Dyksen et al., 2007). High oxidation–reduction potential (ORP) conditions favour the formation and accumulation of lead dioxide (PbO2) scales (Schock and Gardels, 1983; Schock et al., 1996). Since chlorine is a powerful oxidant, switching from chlorine to chloramines reduced the oxidizing potential of the distributed water and destabilized the PbO2 scale, resulting in increased lead release (Schock and Giani, 2004; Lytle and Schock, 2005). A number of studies confirmed that stable PbO2 deposits could be readily formed and subsequently destabilized and reduced to (soluble) Pb(II) in a short period (weeks to months) under realistic conditions of distribution system pH, ORP and alkalinity (Edwards and Dudi, 2004; Lytle and Schock, 2005; Switzer et al., 2006; Valentine and Lin, 2009). Valentine and Lin (2009) suggested that for chloraminating systems, this type of lead release could be managed through having higher pH (>8), having carbonate levels >50mg/L, minimizing the monochloramine dose and using a lower chlorine:ammonia (Cl:N) molar ratio of 0.7:1. However, the authors noted that this lower ratio could enhance nitrification if adequate residual is not maintained. They also noted that NOM was a more significant agent for reducing lead oxide than monochloramine. In pipe loop experiments (water with <5 mg/L of CaCO3 and a pH of 7.3, zinc orthophosphate of 0.8 mg/L as PO4), lead release was significantly higher (approximately one order of magnitude) after a 30-min and 24-h stagnation period for the chloraminated (1 mg/L and 5 mg/L residual) pipe loops compared with chlorinated (1 mg/L) pipe loops (Woszczynski et al., 2013). Similarly, in simulated partial lead service line replacement experiments, sodium silicate-treated water at pH 8 with a monochloramine residual of 3 mg/L was found to increase lead release compared with the chlorinated water (1 mg/L) (Zhou et al., 2015).

Edwards and Dudi (2004) found that high initial concentrations of lead were leached from in-line devices such as meters and shut-off valves, and they decreased exponentially over the 58-day test period. The authors found that more lead was leached from brass treated with chloramines than that treated with chlorine, although this was only significant at the 85% confidence level, which the authors attributed to the small sample size. In Lin et al. (1997), in the absence of corrosion inhibitors, chloramines were found to suppress copper release and increase lead release from brass (3% lead content) compared with chlorine. The authors also examined lead and copper release from brass in the presence of zinc orthophosphate, sodium phosphate blend, long-chain phosphate blend, and a commercial sodium phosphate; they found that lead and copper release were elevated in the chloraminated systems compared with the chlorinated systems.

A study assessed the leaching potential for several components tested under NSF International (NSF)/American National Standards Institute (ANSI) Standard 61 protocols with both chlorinated and chloraminated water. This study also included results from a benchmarking study undertaken to assess 28 waters (chlorinated and chloraminated), collected from utilities across the United States, against the NSF/ANSI Standard 61 exposures waters (pH 5 and pH 10 water [Section 8] and pH 8 water [Section 9]) with respect to the leaching of lead, copper, and zinc from brass samples (rods). Results indicated that neither chlorine nor chloramines were a dominant factor in the metal release of the brass rods (Sandvig et al., 2012). Laboratory experiments were conducted to quantify the levels of lead leached from seven commercially available low-lead brass alloys containing lead at 0.25% or less. The tests were conducted with two different waters, including an NSF /ANSI Standard 61 section 9 test water and a low-pH, low-alkalinity, chloraminated water expected to be aggressive with respect to lead leaching. The concentrations of lead leached from all low-lead alloys were below 1 µg/L under both leaching conditions for the 4-week duration of the experiment (Triantafyllidou and Edwards, 2010).

Douglas et al. (2004) determined that nitrification within the distribution system was the cause of a reduction in pH that resulted in high lead levels in an area of the city with lead service lines. Concentrations of lead at the tap (10–15 µg/L for flowing samples) were seen in the distribution system where a reduction in pH (from 8.5 to a range of 7.8–8.2) was observed. Nitrification can occur within premise plumbing, causing a reduction in pH resulting in the possible release of copper and lead from brass fittings (Zhang et al., 2008, 2009, 2010). A study by Zhang et al. (2009) showed that the extent of pH reduction depended on the alkalinity of water, with higher pH reduction occurring at lower alkalinity.

Limited information has been reported in the literature about the effect of disinfectants on copper. In a 2007 utility survey, 2 of 11 participating utilities reported copper data before and after conversion from chlorine to chloramines (Dyksen et al., 2007). The two utilities both reported decreases in copper release post-chloramination. A bench-scale study by Rahman et al. (2007) examined the effects of three disinfectants (chlorine, chlorine dioxide, and chloramines) under two pH (7.2 and 8.5) and alkalinity (10 or 100 mg CaCO3/L) conditions. The authors found that there was no significant difference in dissolved copper release for the disinfectant-free control compared with any of the disinfectants used in the study. Generally, the study found that the application of disinfectant lowered the copper release, with the lowest total copper release being found under the pH 8.5, 10 mg CaCO3/L and 1.0 mg Cl2/L conditions.

As part of a study assessing the leaching potential for brass components tested under NSF/ANSI Standard 61 protocols using chlorinated and chloraminated water, utilities throughout the United States were surveyed on copper and lead release issues. The survey responses indicated few problems with copper and lead releases when chloramines were used for secondary disinfection. However, the authors cautioned that their observations did not specifically examine switching from chlorine to chloramines (Sandvig et al., 2012).

Boyd et al. (2008) studied the effects of changing disinfectants from free chlorine to chloramines and vice versa on leaching rates of metals and concentrations of metals from lead, brass, and copper components in the distribution system. Local tap water was used and water quality parameters (disinfectant residual, pH, alkalinity, and orthophosphate concentration) were monitored and maintained. The authors conducted pipe loop testing using new copper tubing and bronze piping (as a surrogate for standard brass) with copper–bronze, lead–bronze, and lead–copper galvanic couplings. Copper concentrations in unpassivated copper and bronze pipe loops were sensitive to free chlorine and chloramines, but the effects were transient and not related to a specific disinfectant.

Some treatment processes can increase lead levels in drinking water by changing water quality parameters which can impact lead release. As such, any changes made to the treatment process, particularly those that affect water quality parameters such as pH, alkalinity and ORP (e.g., changing residual disinfectant from chlorine to chloramines), should be accompanied by close monitoring for lead release to assess the need for corrosion control. For a more detailed discussion on corrosion control, see Health Canada (2009).

7.2.6 Iron

There is limited information on the impacts of chloramines on iron. However, in a utility survey by Dyksen et al. (2007), three utilities that had converted from chlorine to chloramines responded with data on "red water." Generally, discoloured water complaints decreased following the switch to chloramines; however, two utilities observed an initial increase in complaints for several months before they decreased to below complaint levels observed during chlorine treatment. The implications of discoloured water events should not be ignored as trace inorganics such as lead as well as arsenic, chromium, uranium, manganese, and vanadium have been associated with iron scale release (Friedman et al., 2010; Camara et al. 2013; Masters and Edwards, 2015; Trueman and Gagnon, 2016). For a more detailed discussion on corrosion control, see Health Canada (2009).

7.2.7 Elastomeric materials

The effect of chloramines on gasket material is a common concern. In a study by Seidel et al. (2005), 16% of utilities reported that they experienced gasket deterioration with the use of chloramines. Rockaway et al. (2007a) conducted a series of accelerated degradation experiments at 23°C, 45°C, and 75°C and at 1 mg/L, 30 mg/L, and 60 mg/L of chlorine and chloramines for six different elastomer materials: natural rubber, neoprene rubber, ethylene propylene diene monomer-peroxide-cured (EPDM-P), ethylene propylene diene monomer-sulphur-cured (EPDM-S), styrene butadiene rubber (SBR), and nitrile rubber. Generally, elastomers were more susceptible to degradation in the presence of chloramines compared with chlorine. Elastomers exposed to chloramines exhibited greater loss of hardness and tensile strength and increased welling. The authors evaluated the overall performance of the materials and their sensitivity to degradation and rated the EPDM-P and EPDM-S as the least sensitive, the natural rubber, SBR, and nitrile rubber as sensitive, and the neoprene rubber as acutely sensitive. Nagisetty et al. (2014) determined that degrading elastomers can release organic materials that are the result of leaching of additives from the elastomer itself or by-products of the reaction of chloramines and the elastomer. Using accelerated degradation processes at 40 mg/L and 30 mg/L chloramines, they found that a total of 18 different organic compounds were released from natural rubber, SBR, and EPDM-S after 3 and 30 days. Bonds (2004) compared the degradation of fluorocarbons (FKM), SBR, EPDM, nitriles, and neoprenes as mechanical joints and push-on joints exposed to a chloramine solution (110 mg/L) over 364 days. The author reported that with the exception of FKM, all the materials showed significant signs of deterioration when exposed in the sheet (large surface area) configuration, but that the same extent of degradation was not observed when the surface area was small (as in the fitted gasket formation). Rockaway et al. (2007a, 2007b) stressed that the selection of elastomer material is an engineering decision, as it required assessment of the suitability of each material for the intended environment as well as other considerations for the consequences of critical failure. The possibility of release of other compounds and degradation by-products should also be considered.

7.3 Residential scale

Generally, it is not necessary to use drinking water treatment devices with municipally treated water. The use of residential-scale treatment devices on municipally treated water is based primarily on individual choice. Although private residential drinking water treatment devices may be an option for reducing concentrations of chloramines in drinking water if the consumer finds the taste objectionable, removal of the disinfectant is not recommended. Treatment devices forremoval of chloramines for aesthetic concerns are covered under NSF/ANSI Standard 42, which establishes minimum requirements for materials, design and construction, and performance of drinking water treatment systems that are designed to reduce specific aesthetic-related (taste and odour) contaminants. Systems certified under NSF/ANSI Standard 42 must be able to reduce an influent concentration of 3.0 ± 0.3 mg/L monochloramine to ≤0.5 mg/L monochloramine (NSF/ANSI, 2018).

Health Canada does not recommend specific brands of drinking water treatment devices, but it strongly recommends that consumers use devices that have been certified by an accredited certification body as meeting the appropriate NSF/ANSI drinking water treatment unit standards. These standards have been designed to safeguard drinking water by helping to ensure the material safety and performance of products that come into contact with drinking water. Certification organizations provide assurance that a product conforms to applicable standards and must be accredited by the Standards Council of Canada (SCC). In Canada, the following organizations have been accredited by the SCC to certify drinking water devices and materials as meeting NSF/ANSI standards (SCC, 2019):

8.0 Kinetics and metabolism

No information on kinetics or metabolism relating to dermal or inhalation exposure to either mono-, di- or trichloramine was identified in the literature. The only available kinetics studies were oral and in vitro studies that used monochloramine. As an oxidant, monochloramine is unlikely to undergo absorption, distribution and excretion intact (National Research Council, 1987). Rather, it will react with organic and inorganic substrates in saliva and stomach fluid to form organic chloramines and iodinated compounds (National Research Council, 1987; Scully and White, 1991). Metabolism studies demonstrated that the chloride ion and not monochloramine is absorbed into the bloodstream, distributed into the tissues and excreted in the urine (Abdel-Rahman et al., 1983).

8.1 Absorption

Based on in vitro time persistence studies and estimated chemical reactions in saliva and gastric juice, monochloramine is unlikely to be absorbed intact following ingestion (Scully and White, 1991; Kotiaho et al., 1992). By adding water containing monochloramine at levels found in drinking water (typically 1–3 ppm) to samples of human saliva, Kotiaho et al. (1992) showed that the reaction of monochloramine with saliva was concentration dependent. At a low concentration of 0.7 ppm, monochloramine remained in human saliva for up to 5 min, while at a concentration of 1.8 ppm it persisted for up to 50 min. At higher concentrations (3.6–15 ppm), monochloramine did not completely react with saliva even after 2 h (Kotiaho et al., 1992). These measured decay times of monochloramine in saliva and the short residence time in the mouth suggest that the majority of monochloramine in treated drinking water likely reaches the stomach unchanged.

However, depletion of monochloramine in stomach fluid was so rapid (within 30 sec for concentrations of 0.7–15 ppm) that intact monochloramine is unlikely to enter systemic circulation (Kotiaho et al., 1992). A metabolism study in male Sprague–Dawley rats (n = 4) given 36Cl radiolabelled monochloramine (1.1 mg/rat) orally showed that 36Cl- appeared in the bloodstream and that chlorite and chlorate were not seen as metabolites. Chloride reached a peak plasma level of 10.3 µg/L 8 h post-dosing (Abdel-Rahman et al., 1983). The absorption rate constant was 0.278 h-1 and the absorption half-life was 2.5 h for the labelled chloride.

8.2 Distribution

Abdel-Rahman et al. (1983) examined the distribution of the 36Cl portion of radiolabelled monochloramine given orally to male rats (1.1 mg/rat) and found plasma contained the highest amount (3.15 µg36Cl/g of tissue), followed by whole blood (2.66 µg/g), skin (2.13 µg/g), testes (2.09 µg/g) packed cells (1.90 µg/g), bone marrow (1.82 µg/g), kidney (1.62 µg/g), lung (1.58 µg/g), stomach (1.53 µg/g), thyroid (1.36 µg/g), thymus (1.36 µg/g), duodenum (1.20 µg/g), spleen (1.11 µg/g), carcass (0.77 µg/g), liver (0.74 µg/g), ileum (0.59 µg/g), and fat (0.18 µg/g).

8.3 Metabolism

Ingested monochloramine is quickly broken down in the stomach (Kotiaho et al., 1992). The reaction products formed are influenced by the concentration of monochloramine, the pH and composition of stomach fluid, and diet (Scully and White, 1991). The pH of the stomach ranges from 1 to 8 (Scully and White, 1991). At a pH near 1, the protonated form of monochloramine is more prevalent and more likely to transfer its chlorine to amino acids present in the gastric juice, thereby producing N-chlorinated amino acids (Scully and White, 1991). Monochloramine may also generate iodinated organic compounds in gastric juice and saliva in the presence of nutrients such as tyrosine, 4-aminobenzoic acid, arachidonic acid, and folic acid (Bercz and Bawa, 1986). Formation of dichloramine, trichloramine or molecular chlorine did not occur in gastric juice in vitro (Kotiaho et al., 1992). The presence of food would likely increase reaction rates of organics with monochloramine, resulting in a shorter half-life for monochloramine in saliva and stomach fluid (Scully and White, 1991). A metabolism study by Abdel-Rahman et al. (1983) that focused on chloride, chlorite, and chlorate as metabolites found only chloride present in the bloodstream of rats following oral dosing with 36Cl radiolabelled monochloramine (1.1 mg/rat).

The results obtained by Kotiaho et al. (1992), Abdel-Rahman et al. (1983), and Scully and White (1991) suggest that any observed toxicity associated with the ingestion of monochloramine is likely due to the formation of reaction products in the stomach rather than absorption of intact inorganic chloramines.

8.4 Excretion

Abdel-Rahman et al. (1983) showed that monochloramine is partly excreted in the urine as chloride. Excretion data for male Sprague–Dawley rats (n = 4) given 36Cl radiolabelled monochloramine (1.1 mg/rat) showed only 0.40% and 0.08% of the administered dose was excreted in the urine and feces, respectively, in the first 24 h post-treatment. After 120 h, the values were 25.15% in urine and 1.98% in feces. No 36Cl was detected in expired air at any time point (Abdel-Rahman et al., 1983).

9.0 Health effects

The following discussion on health effects will focus primarily on studies conducted via the oral route, since it is the most relevant route of exposure for chloramines. Most available studies are for monochloramine, the predominant species in drinking water. Dichloramine and trichloramine are infrequently detected in drinking water; only one oral study for each found in the literature. Minimal to no effects were seen in both humans and animals following ingestion of monochloramine, with the most significant effect (decreased body weight gain) attributed to a reduction in water intake caused by decreased palatability from high levels of monochloramine.

9.1 Effects in humans

Despite more than 90 years' history of using monochloramine as a drinking water disinfectant, only a few clinical and case–control studies and surveys are available from the literature (U.S. EPA, 2009; Vermont Department of Health, 2012). Results from these studies did not show any adverse health effects, although the limitations of the studies (e.g., weak exposure assessment, chloramination used as control, sample bias) do not allow for definitive conclusions to be drawn.

By contrast, numerous studies have been published regarding the toxicity of chloramines via inhalation. These studies are related to accidental intoxication from chloramine gas (monochloramine and dichloramine) due to improper handling of cleaning agents or occupational/recreational exposure to trichloramine from its use in food industry facilities or its formation in situ in swimming pools, scenarios that are not relevant to drinking water exposure (Reisz and Gammon, 1986; Tanen et al., 1999; Cohle et al., 2001). Additionally, several papers have reported negative health effects (decreased haptoglobin, decreased hemoglobin, formation of Heinz bodies, hemolytic anemia, methemoglobinemia) in hemodialysis patients when tap water containing chloramines was used for dialysis (Eaton et al., 1973; Kjellstrand et al., 1974; Klein, 1986; Fluck et al., 1999; Weinstein et al., 2000; Junglee et al., 2010). These hematological changes were not observed in healthy volunteers deliberately ingesting chloraminated drinking water indicating that ingested monochloramine is unlikely to enter systemic circulation intact (Lubbers et al., 1982, 1984; Kotiaho et al., 1992). As such, dialysis was not considered a relevant route of exposure.

9.1.1 Acute toxicity

Only two short-term studies were located regarding human health effects following ingestion of drinking water containing monochloramine. These did not show any adverse effects at doses of up to 24 mg/L.

In phase I of a double-blind acute rising-dose tolerance study, no effects in liver, thyroid function, kidney function, hematology or overall health were observed in male volunteers consuming distilled water containing up to 24 mg/L of monochloramine (Lubbers et al., 1982). Males (10/group, aged 21–35) with normal methemoglobin levels, thyroid function and glutathione levels drank two 500 ml portions of either organic-free demineralized deionized water alone or water containing monochloramine, with a 4-hour interval between portions. Treatment days occurred every 3 days for a total of 5 treatment days over a period of 13 days. Doses of monochloramine were increased over the course of the treatments, beginning with 0.01 mg/L on Treatment Day 1 and increasing to 24 mg/L by the end of treatment. A battery of tests, including blood and urine biochemistry (e.g., cholesterol, triglycerides, triiodothyronine [T3]), hematology (blood cell counts, hemoglobin, and haptoglobin and methemoglobin levels), cell morphology, and a physical examination found no difference between pre- and post-treatment values (Lubbers et al., 1982).

In a randomized, controlled parallel trial study by Wones et al. (1993), thyroid and lipid effects were generally absent, except for a slight increase in apolipoprotein B levels at 15 ppm/day of monochloramine. Three groups of healthy men (16/group) were given distilled water and were fed a diet high in fat and cholesterol for the first 4 weeks of the study to establish a baseline. For the 4-week treatment period, all subjects continued on the study diet and were given 1.5 L/day of water containing monochloramine at 0 ppm (control), 2 ppm or 15 ppm. Blood levels were taken at the end of each 4-week period to assess lipid (total cholesterol, high-density lipoprotein [HDL], low-density lipoprotein [LDL], triglycerides, apolipoprotein A1, A2 and B) and thyroid (thyroid-stimulating hormone [TSH], T3, thyroxine [T4]) metabolism. No changes were observed in thyroid function in either treatment group. No changes in lipid function were seen except for a slight, statistically significant increase in apolipoprotein B (a component of LDL cholesterol) observed at the 15 ppm (15 mg/L) dosage. The significance of increased apolipoprotein B independent of changes in LDL cholesterol is unknown. Study limitations included relatively brief baseline and treatment periods, and previous consumption of chlorinated drinking water by most subjects.

9.1.2 Subchronic and chronic toxicity and carcinogenicity

Except for a slight increase (although within normal range) in T3 seen at the 5 mg/L/day dosage (Lubbers et al., 1984), long-term studies did not find any association between chloramination and health effects, including dermal symptoms, upper respiratory effects and bladder cancer (Zierler et al., 1986, 1988; McGeehin et al., 1993; Weintraub et al., 2006; CDC, 2008; Vermont Department of Health, 2012). Most long-term chloramine studies compared the effects of different disinfection treatment methods (e.g., chlorination vs. chloramination) on specific health outcomes (e.g., bladder cancer) or used chloramination exposure as the control group. Other shortcomings of these studies included lack of exposure characterization, previous or concurrent exposure to chlorine, and recall bias, making the results unsuitable for risk assessment (Zierler et al., 1986, 1988; McGeehin et al., 1993; Weintraub et al., 2006; CDC, 2008; Vermont Department of Health, 2012).

In a continuation of an earlier study (described in Section 9.1.1), two groups of healthy male volunteers (n = 10/group; aged 21–35) were given unlimited organic-free demineralized deionized water to drink and asked to refrain from drinking tap water. One group also consumed 500 ml of water containing 5 ppm of monochloramine daily for 12 weeks followed by an 8-week recovery period (Lubbers et al., 1984). The same battery of tests was used as used in the previous study (Lubbers et al., 1982). The only statistically significant effect was an increase in T3 uptake which changed over time when compared with the control group; however, this change remained within the normal T3 range and was clinically irrelevant.

No direct link between reported health effects and chloramination could be established in two surveys conducted in the United States following the switch from chlorine to chloramination as a secondary disinfectant by local public utilities (Weintraub et al., 2006; CDC, 2008).

After the local water utility switched from chlorine to chloramine as its secondary disinfectant, the San Francisco Health Department received a small number of dermatitis complaints (Weintraub et al., 2006). These complaints prompted a questionnaire-based study. Despite widespread publicity, only 17 respondents completed the questionnaire from a total population of 2.4 million in the service area. Of these, 11 had histories of allergies while 8 had histories of skin issues (e.g., hives, shingles, eczema). Of the 14 that sought medical treatment, no diagnoses were recalled by respondents. Since the complaints were heterogeneous (no consistency between individual complaints) and respondents had underlying or pre-existing conditions that plausibly explained the reported symptoms, the effects were likely unrelated to chloramination, according to the authors. Study shortcomings included use of a convenience sample, small sample size, no dermatological exam, and no validation of the questionnaire. Media coverage linking health complaints to chloramination may have influenced symptom reporting (Weintraub et al., 2006).

Similarly, Vermont's Champlain Water District and the Public Health Service of the Centers for Disease Control and Prevention (CDC) conducted a joint investigation into health complaints (n = 74) such as upper respiratory tract symptoms, watery eyes/nose, scratchy throat, gastrointestinal ailments, and skin rashes/itching following the water district's change to chloramines as a secondary disinfectant (CDC, 2008). The switch to monochloramine was widely publicized and a mass media campaign against the use of chloramination was undertaken by a citizens' group called People Concerned about Chloramine (PCAC) prior to the switch. Both home visits and telephone interviews (46/50 subjects met the inclusion criteria) were conducted with persons expressing health concerns. Twenty-five (54%) respondents had at least one pre-existing health condition (allergies/asthma, skin issues, diabetes, autoimmune disorders, cardiac history, respiratory, arthritis, cancer). Of the 32 who sought medical help, 8 received a diagnosis (five cases of psoriasis/eczema and one each of fluid in lungs, writer's rash and surface eye irritation). Twenty-three of the 34 participants who changed their water use habits to treat their symptoms showed improvement. Chloramine concentrations of 1.8–2.3 mg/L were reported at the distribution entry point. Consumer exposure was not measured but was expected to be lower than in the distribution system. Symptoms may have been biased by self-reporting, the media campaign and the presence of PCAC members at home visits. No clear link could be drawn between treatment and symptoms (CDC, 2008).

A follow-up survey of 173 health care providers, including family practice and primary care physicians and specialists, was conducted by the Department of Health in Vermont to assess the prevalence of health problems related to monochloramine use. Of the 81 surveys returned, no patients were reported to have an illness directly related to monochloramine use, while only two patients were considered to have an underlying disease that was exacerbated by chloraminated water use. Eleven providers reported they were unsure if patient complaints were related to the water, while 59 providers reported chloraminated water did not cause patient complaints. It was concluded that some people may be sensitive to monochloramine in the water (Vermont Department of Health, 2012).

No association was seen between chloraminated drinking water and bladder cancer when compared with chlorinated drinking water in three studies (Zierler et al., 1986, 1988; McGeehin et al., 1993).

McGeehin et al. (1993) investigated the possible association between bladder cancer and drinking water disinfection methods using a population-based, case–control study in Colorado. Cases (327 subjects with bladder cancer) and controls (261 subjects with other cancers, except colorectal and lung) were identified through the Colorado Central Cancer Registry and were limited to living, white subjects, aged 21 to 84. Residential and water source histories and information on potential confounders were obtained through telephone interviews. Residences were linked to water utilities information (e.g., water source, disinfection type) and used to create exposure profiles. Logistic regression was used to adjust for cigarette smoking, consumption of tap water and coffee, and medical history. The mean person-year level of exposure to chloraminated surface water was lower (p = 0.02) for cases (mean of 14.8 years) than for controls (mean of 18.8 years). The risk of bladder cancer decreased with increasing duration of exposure to chloraminated surface water (p < 0.01) with a risk estimate of 0.7 (95% confidence interval [CI] 0.4–1.1) for subjects who consumed chloraminated water for 21–40 years, while those who consumed chloraminated water for more than 40 years had an odds ratio (OR) of 0.6 (95% CI 0.4–1.0). The decreased risk seen likely corresponded to the decrease in years of exposure to chlorinated water and not to any protective effect of chloramination, since there is no biological evidence that suggests chloramination inhibits neoplastic transformation of bladder epithelium. Some potential shortcomings of the study included changes to disinfection methods made by some water utilities over time, the assumption that water parameters and individual consumption patterns were constant over time, gaps in historical data on water treatment and in individual exposure profiles, and the potential for recall bias.

Using Massachusetts Department of Public Health death records, Zierler et al. (1986) examined mortality patterns of state residents (≥45 years of age) consuming either chlorine-treated or chloramine-treated drinking water. Subjects were assigned to either group based on residence at time of death. No correlation was seen between cancer mortality and consumption of chloraminated water, but the chloramine group did have a slight excess of deaths from pneumonia and influenza (standardized mortality ratio [SMR] of 118; CI 116–120 for chloraminated water vs. SMR of 98; CI 95–100 for chlorinated water). Bladder cancer deaths were moderately associated with chlorine-treated (SMR of 105; CI 99–111) but not chloramine-treated (SMR of 91; CI 88–98) drinking water. The results should be considered qualitative, since the study had a number of shortcomings (e.g., exposure history based on residence at time of death, error in assigning disease status, failure to assess previous residence histories, length of exposure and water quality) and may have been influenced by unidentified or uncontrolled confounding factors. The authors also suggested that deaths from pneumonia may have been a secondary cause of death, with cancer or some other underlying illness likely being the primary cause.

In a follow-up case–control study that considered exposure duration, Zierler et al. (1988) again found no correlation between bladder cancer mortality and consumption of chloraminated surface water. The study looked at data on 614 individuals who died of primary bladder cancer and who had been exposed to surface water treated by chlorination or chloramination for either a lifetime or 50% of residence. The control group (1,074 individuals) consisted of five disease groups: cardiovascular disease, cerebrovascular disease, lung cancer, chronic obstructive pulmonary disease, and lymphatic cancer. Potential confounders (e.g., age, sex, pack-years of smoking, occupation, socioeconomic status) were provided through proxy interviews and were controlled by a multiple logistic regression model. The frequency of bladder cancer mortality increased among lifetime residents of communities receiving chlorinated drinking water compared with residents of communities receiving chloraminated drinking water. However, the U.S. EPA (1994b) pointed out that it is impossible to find control subjects who have not been exposed to chlorine or chloramines, and that the relationship between chloraminated water consumption and bladder cancer incidence cannot be defined based on the results of these studies.

9.1.3 Developmental and reproductive toxicity

In the only developmental study available, chloramination was less likely to cause adverse pregnancy outcomes than chlorination. Aschengrau et al. (1993) conducted a case–control study comparing adverse pregnancy outcomes (congenital anomalies, stillbirths, neonatal deaths) with a variety of community drinking water quality parameters (source, treatment, routine chemistry). Residential address at the time of pregnancy outcome or during the first trimester was matched to routine water quality analyses conducted by the Massachusetts Department of Environmental Protection and used to characterize exposure. After adjusting for potential confounding factors (e.g., maternal race, age, previous birth with congenital anomaly, medical insurance coverage type, alcohol consumption, water source, residence location), chlorinated surface water for drinking was associated with an increase in both stillbirths (adjusted OR = 2.6; 95% CI 0.9–7.5) and major malformations (adjusted OR = 1.5; 95% CI 0.7–2.1) when compared with chloraminated drinking water. A number of study shortcomings (such as the lack of a non-treated control group, no consideration of other maternal health issues, the rough estimates of individual exposure, no water consumption data, no measurement of chloramine levels) make the results difficult to interpret (Aschengrau et al., 1993).

9.2 Effects on experimental animals

The main effects seen in animal studies using monochloramine were decreased body weight gain, decreased organ weights and alterations of blood parameters. However, paired water studies indicate that these effects are likely caused by decreased water consumption and not by the toxicity of monochloramine (Daniel et al., 1990, 1991; Poon et al., 1997).

No dermal studies were identified for any of the chloramines. The majority of available oral studies in animals were for monochloramine; only one subchronic oral study was available for both di- and trichloramine. No inhalation studies were found for either mono- or dichloramine. Two acute inhalation studies were available for trichloramine (Barbee et al., 1983; Gagnaire et al., 1994).

9.2.1 Acute toxicity

Acute oral exposure to monochloramine caused stomach irritation at high doses while its effects on blood parameters (e.g., glutathione [GSH]) were transient (Abdel-Rahman et al., 1984; Kato et al., 1997; Nishiwaki et al., 1997; Umeda et al., 1999). Inhalation of high doses of trichloramine resulted in severe respiratory irritation (Barbee et al., 1983; Gagnaire et al., 1994).

9.2.1.1 Oral

No significant effects on blood parameters were observed in a limited number of acute oral studies using monochloramine, although stomach effects were seen in high-dose studies.

Monochloramine given orally to male rats (5/dose) at high concentrations (20 mM, 60 mM, and/or 120 mM, corresponding to approximately 1,000 mg/L, 3,100 mg/L, and 6,200 mg/L) was shown to irritate the gastric mucosa and impair the ulcerogenic/repair process in the stomach, with doses of ≥60 mM (corresponding to 11.25 mg/kg bw, calculated) inducing severe mucosal lesions (Kato et al., 1997; Nishiwaki et al., 1997; Umeda et al., 1999).

In another study examining several blood parameters, male Sprague–Dawley rats (4/dose) were gavaged with a single dose of monochloramine in deionized water (0, 10 mg/L, 20 mg/L or 40 mg/L) (Abdel-Rahman et al., 1984). Osmotic fragility was unaffected, whereas increases in blood GSH, an antioxidant, were noted at 20 mg/L and 40 mg/L 15 min after dosing; they continued for up to 60 min but returned to normal after 2 h.

9.2.1.2 Inhalation

Barbee et al. (1983) reported an inhalation LC50 (1 h) of 112 ppm (550 mg/m3) in rats exposed to trichloramine. Pulmonary edema and clear fluid in the lungs and trachea were seen in the animals that died following exposure. No mortality was seen at the lowest concentration of 58 ppm. At the highest concentration of 157 ppm, rats displayed effects such as eye irritation, excessive lacrimation, salivation, gasping and inactivity, as well as convulsions and nasal discharge.

In another inhalation study that evaluated upper airway irritation from airborne chemicals, male OF1 mice (n ~ 8/concentration) were exposed oronasally to 0.19–5.0 ppm of trichloramine for 60 min (Gagnaire et al., 1994). The RD50, an exposure concentration producing a 50% respiratory rate decrease and an indicator of respiratory irritation, was calculated as 2.5 ppm. Once exposure was discontinued, recovery was fairly rapid except for the highest dose where mice had a slower recovery.

9.2.2 Subchronic exposure
9.2.2.1 Monochloramine: decreased body weights and blood effects

Daniel et al. (1990) exposed Crl:CD BR Sprague–Dawley rats (10/sex/dose) to 0, 25, 50, 100 and 200 mg/L of monochloramine in drinking water (equivalent to 0, 1.8, 3.4, 5.8, and 9.0 mg/kg bw per day for males and 0, 2.6, 4.3, 7.7, and 12.1 mg/kg bw per day for females) for 90 days. A significant dose-related reduction in daily water consumption was seen in both sexes at all doses. In addition, a significant reduction in body weight gain (males at ≥50 mg/L, and females at 200 mg/L) and in final body weight (both sexes at the highest dose) was noted. Changes in some hematological parameters and in absolute and relative organ weights were also noted in the absence of histopathological effects. Clinical chemistry results showed a dose-dependent decrease in the level of calcium at all doses in treated males, whereas an increase was seen in females at the highest dose only. The hematological and clinical chemistry changes were either not dose-dependent, not biologically significant or within the normal range for rats of that age and strain. A no-observed-adverse-effect level (NOAEL) of 100 mg/L (equivalent to 7.7 mg/kg bw per day for females and 5.8 mg/kg bw per day for males) was proposed in the study based on final body weight. The authors suggested conducting a matched watering and feeding study (where controls are given reduced water to match the treated group) to help distinguish between systemic toxic effects and effects on weight gain due to taste aversion.

Decreases in body weight gain (≥50 mg/L in males and ≥100 mg/L in females), final body weight (at ≥100 mg/L in both sexes), water consumption (≥12.5 mg/L for females; ≥100 mg/L for males) and food consumption (females only at 100 ppm and above) were also observed in B6C3F1 mice exposed to monochloramine at 0, 12.5, 25, 50, 100 and 200 mg/L (equivalent to 0, 2.5, 5, 8.6, 11.1, and 15.6 mg/kg bw per day for males and 0, 2.8, 5.3, 9.2, 12.9, and 15.8 mg/kg bw per day for females) in their drinking water for 90 days (Daniel et al., 1991). Changes in relative and/or absolute organ weights at ≥100 ppm were seen in the absence of compound-related gross or microscopic lesions. Dose-dependent increases in white blood cells were seen in female mice at 25 mg/L and above; similarly, increases were seen in males but were not dose-dependent. In addition, significant decreases in the percentage of neutrophils were seen in both male and female mice at the top two doses, and an increase in the number of lymphocytes was seen in males only at the top two doses. Other hematological and clinical chemistry changes were also noted at 12.5 mg/L and above but were inconsistent and not dose-dependent. The authors report a NOAEL of 50 mg/L (8.6 mg/kg bw per day for males and 9.2 mg/kg bw per day for females) based on decreased organ weights, decreased weight gain and decreased consumption of food and water. They indicated that the observed effects could be a consequence of the decreased water consumption associated with taste aversion (Daniel et al., 1991).

The impact of taste aversion and reduced water intake on body and organ weights was assessed by Poon et al. (1997) using male Sprague–Dawley rats (10/dose) exposed to 0 or 200 ppm of monochloramine (equivalent to 0 or 21.6 mg/kg bw per day, as calculated by the study authors) for 13 weeks. A paired-water control (with restricted water intake) and a control group (given bicarbonate-buffered water ad libitum) were used. Over the course of the study, the monochloramine group consistently drank less water (p < 0.05) than the control group given water ad libitum. Both the monochloramine group and the paired-water control group showed similar reduced body weight gain, final body weights and food consumption. Changes in organ weights and minor changes in biochemistry, hematology, immunological and histopathology were also seen in the treated group as well as in the paired-water controls. The authors concluded that these changes were related to the reduced water and food consumption seen in these two groups and not a result of monochloramine toxicity.

Body weight loss (50 ppm and above) and decreased water consumption (25 ppm and above) were observed in male A/J mice (12/dose) administered monochloramine in drinking water (0, 2.5, 25, 50, 100 or 200 ppm) for 30 days (Moore et al., 1980). No evidence of hemolysis or other blood effects was reported.

By contrast, adult African green monkeys (five males and seven females) exposed to monochloramine in drinking water up to 100 mg/L per day (10 mg/kg bw per day) using a rising dose protocolFootnote 2 did not show adverse effects on body weight or on thyroid function (serum thyroxine [T4] concentrations), hematology, hematological oxidative functions or serum chemistry (Bercz et al., 1982). Three other chemicals (chlorine dioxide, sodium chlorite and sodium chlorate) were also tested on the same monkeys with a 9-week rest period before a new chemical was introduced. No separate control group was used; however, the monkeys served as their own control. Similarly, no signs of overt toxicity, decreased body weight gain or hematological changes were observed in rats exposed to monochloramine in drinking water (up to 100 mg/L) for 45 days in a study sponsored by the U.S. EPA's Health Effects Research Laboratory (Bull, 1980). However, a decrease in the amount of methemoglobin in blood was noted at the highest dose, contrary to the authors' expectations of an increase, based on observations in patients dialyzed with chloraminated tap water (Eaton et al., 1973; Kjellstrand et al., 1974; Klein, 1986; Fluck et al., 1999; Junglee et al., 2010).

Body weights were significantly decreased starting at 3 months at the highest dose in male Sprague–Dawley rats (4/dose) exposed to monochloramine in drinking water at 0, 1, 10 or 100 mg/L for up to 12 months (Abdel-Rahman et al., 1984). Blood parameters were also investigated; however, the observed changes (increased blood osmotic fragility and decreased blood GSH) were not consistent among the treatment groups and were not dose- or time-dependent. Other parameters (decreased red blood cell count and hematocrit) were seen at mid-time points but not at 10 months. Hemoglobin and mean corpuscular hemoglobin concentrations were reduced at 100 mg/L after 10 months; this, according to the authors, was an indication of erythrocyte damage. No data on water consumption was provided in this study; however, the authors noted significant decreases in body weights at 100 mg/L, which may have had effects on blood parameters, as seen in the previous studies cited above.

9.2.2.2 Monochloramine: immunotoxic effects

Immunotoxic effects of monochloramine were investigated by Exon et al. (1987). Male Sprague–Dawley rats (12/dose) were exposed to monochloramine concentrations in drinking water (0, 9, 19 or 38 ppm) from weaning (3 weeks old) for 9 weeks. A reduction in relative and absolute spleen weight at the highest dose and an increased production of prostaglandin E2 by adherent resident peritoneal cells were observed at the two highest doses. A decrease (inverse dose-dependent) in serum antibody synthesis at the two lowest doses (9 and 19 ppm) was also observed. A review by the U.S. EPA (1994a) indicated that the biological importance of these immunotoxic effects from monochloramine was not clear, as there was no correlation among the various immunological endpoints studied. They also noted deficiencies in some of the methodologies used (e.g., ELISA test to analyse antibodies). No other remarkable findings were noted with respect to body weights or in the battery of other immunoassays used (Exon et al., 1987).

No immunotoxic effectsFootnote 3 related to monochloramine exposure were observed when female B6C3F1 mice (8/dose) were exposed to monochloramine in drinking water at 0, 2, 10, 20, 100 and 200 ppm (0, 0.4, 2, 4, 20 and 40 mg/kg bw per day) for 28 days (NTP, 2000; Guo et al., 2011).

In a 13-week drinking water study, a small, statistically significant increase in Concanavalin A-mediated lymphocyte transformation activity (a measure of cell agglutination) was observed in male Sprague–Dawley rats (10/group) given 200 ppm (21.6 mg/kg bw per day) of monochloramine in drinking water compared with paired-water controls (Poon et al., 1997). A second control group was given bicarbonate-buffered water ad libitum throughout the study. Both the monochloramine group and the paired-water control group showed similarly reduced body weight gain which the authors concluded was related to the reduced water and food consumption seen in these two groups. The authors also concluded that monochloramine was not likely to be a significant immunotoxicant, given that there were no effects observed on relative thymus weight and that no treatment-related histopathological changes in the thymus and spleen were observed.

9.2.2.3 Dichloramine and trichloramine

Only one subchronic study assessed the oral toxicity of di-and trichloramine in rats. Biochemical and mild to minimal histological effects were seen at doses of 0.2 ppm and above in females and males in the absence of clinical signs or weight gain changes.

Sprague–Dawley rats (10/sex/dose) were exposed to dichloramine (0, 0.2, 2, 20 and 200 ppm; calculated as 0.019, 0.19, 1.9 and 18 mg/kg bw per day for males and 0.025, 0.26, 2.5 and 24 mg/kg bw per day for females) or trichloramine (0, 0.2, 2, 10 and 90 ppm; calculated as 0.020, 0.23, 1.1, and 9.6 mg/kg bw per day for males and 0.028, 0.29, 1.3 and 13 mg/kg bw per day for females) in drinking water for 13 weeks (Nakai et al., 2000). Dichloramine produced minimal histological effects (epithelial hyperplasia) in the gastric cardia at ≥2 ppm (≥0.19 mg/kg bw per day) in males, and at ≥20 ppm (≥2.5 mg/kg bw per day) in females. Minimal to mild effects were seen in the kidney and thyroid but without any clear dose–response link, changes in relative organ weight or clinical manifestations. Other relatively mild effects, considered adaptive, were also seen in the liver. Trichloramine exposure produced minimal to mild histological effects in the thyroid at 2 ppm and above in both males (≥0.23 mg/kg bw per day) and females (≥0.29 mg/kg bw per day). Minimal to mild changes (considered adaptive by the authors) in the liver of males were noted at ≥0.2 ppm (≥0.02 mg/kg bw per day), whereas at 90 ppm (13 mg/kg bw per day), females showed increases in hepatic enzyme activity. At the highest dose, minimal histopathological changes in the kidney were also seen in males and increased relative kidney weights were observed in both males and females. Although histological changes in the kidney were noted in females at all doses, no dose–response relationship in incidence or severity was observed.

It must be noted that water consumption was decreased in both studies and that the minimal effects reported did not have clinical manifestations over the dose range examined.

9.2.3 Long-term exposure and carcinogenicity

The National Toxicological Program, exposed F344/N rats and B6C3F1 mice (n = 70/sex/species/group) to monochloramine in drinking waterFootnote 4 (0, 50, 100 and 200 ppm) for 2 years (NTP, 1992; Dunnick and Melnick, 1993). The calculated weighted mean equivalent dosesFootnote 5 were 0, 2.6, 4.8 and 8.7 mg/kg bw per day in male rats; 0, 2.7, 5.2, and 9.5 mg/kg bw per day in female rats; 0, 7.4, 14.0, and 23.8 mg/kg bw per day in male mice; and 0, 7.6, 14.2 and 24.2 mg/kg bw per day in female mice. Interim sacrifices (10/sex/species/dose) were conducted at weeks 14 and 66. Both rats and mice showed a dose-related decrease in water consumption throughout most of the study that was attributed to palatability.

In treated F344/N rats, feed consumption was similar to that of the controls. However, mean body weights in the high-dose rat group (both sexes) were decreased (5–10%) compared with the low- and mid-dosed rats. Changes in absolute and relative liver, brain and/or kidney weight were also observed but were considered by the authors to be related to the lower body weights and not biologically significant. In female F344/N rats, there was a marginal, statistically significant increase in the incidence of mononuclear cell leukemia (MNCL) at the top two doses, but this increase was not dose-dependent, since the responses at these two doses were very similar: 8/50 (16%), 11/50 (22%), 15/50 (30%), and 16/50 (32%) for control, low-, mid- and high-dose groups, respectively. In addition, this marginal increase was within the range (although in the upper range) reported for historical controls (16–33%) and no reduced latency of MNCL in treated female rats was noted. According to NTP (1992), the evidence was weak in support of an association between MNCL in female rats and the consumption of chloraminated water. No increase in MNCL was observed in male rats. No other clinical findings, hematological effects, effects on survival, or gross microscopic lesions were attributable to the consumption of chloraminated water. A NOAEL was set at 8.7 mg/kg bw per day (200 ppm) based on the absence of biologically significant adverse effects in male rats exposed at the highest concentration tested. Although lower body weights were observed at this level of exposure, they were likely caused by the unpalatability of the drinking water.

In mice (both sexes), there were also dose-related decreases in mean body weights during most of the study. Feed consumption was similar in treated male mice compared with controls and was only slightly lower in high-dose females compared with controls. Changes in absolute and relative organ weights (kidney, liver and brain) were also observed in the high-dose mice and were also considered to be related to the lower body weights. No other clinical findings, effects on survival rates or gross microscopic lesions were attributable to the consumption of chloraminated water.

At the time of the study, the NTP (1992) concluded that under conditions of the bioassay, there was equivocal evidence of carcinogenicity in female F344/N rats based on MNCL but no evidence of carcinogenicity in either male F344/N rats or in B6C3F1 mice of either sex. Beginning in 2006, NTP discontinued the use of F344 rats in carcinogenicity studies because of the high and variable background incidence of MNCL in F344 rats, the species- and strain-specificity of MNCL and its questionable significance to humans, especially when rates fall within normal historical control values (King-Herbert and Thayer, 2006; Maronpot et al., 2016).

In a rat liver foci assay, chloramine failed to induce an increase in γ-glutamyl-transpeptidase foci (GGT foci; an indicator of carcinogenicity) in male Sprague–Dawley rats given 14.75 mg/kg of chloramine in drinking water (Herren-Freund and Pereira, 1986). Based on this study, monochloramine did not show any potential for tumour initiation. Negative results were also seen using concentrated drinking water samples disinfected with monochloramine in another liver foci assay in rats (Miller et al., 1986).

9.2.4 Genotoxicity

The literature on the genotoxicity of chloramines is limited to four in vitro studies and two in vivo studies; these studies investigated the genetic toxicology of monochloramine only. Monochloramine was not genotoxic in in vivo studies or in a modified Ames test but was weakly mutagenic in Bacillus substilisassay.

9.2.4.1 In vitro findings

In in vitro studies, monochloramine was shown to be weakly mutagenic in Bacillus subtilis but not mutagenic in Salmonella typhimurium TA100 in a modified Ames test without metabolic activation (Shih and Lederberg, 1976a; Thomas et al., 1987).

Monochloramine induced single-strand breaks in Bacillus subtilis (Shih and Lederberg, 1976b), double-strand plasmid DNA breakage, DNA fragmentation and DNA double-strand breaks, as well as chromatin condensation in rabbit gastric mucosal cells and human gastric cancer cells (Suzuki et al., 1997, 1998; Shibata et al., 1999).

9.2.4.2 In vivo findings

Monochloramine did not induce chromosomal damage in vivo in either bone marrow or sperm of orally dosed mice.

No increases in micronucleated polychromatic erythrocytes or in chromosomal aberrations were seen in the bone marrow of Swiss CD-1 mice (5/sex/dose) gavaged with 1 ml of 0, 50 mg/L, 100 mg/L, or 200 mg/L of monochloramine per day for 5 days and then sacrificed (Meier et al., 1985). In the same study, single, gavage doses of monochloramine (0, 50 mg/L, 100 mg/L, or 200 mg/L) given to Swiss CD-1 mice (4/sex/dose) also did not cause any chromosomal aberrations.

Abdel-Rahman et al. (1984) investigated DNA synthesis in various organs (via incorporation of [3H] thymidine into the nuclei) by administering methyl thymidine intraperitoneally to rats treated with 0, 1 mg/L, 10 mg/L, or 100 mg/L of monochloramine for 3 months. Increased DNA synthesis (as indicated by increased [3H] thymidine uptake into  testes cell nuclei) was noted at 100 mg/L after the 3-month treatment, whereas increased uptake into kidney and spleen cell nuclei were noted at the two lower doses (1 mg/L and 10 mg/L) only. Monochloramine did not increase the number of sperm-head abnormalities in male B6C3F1 mice (10/group) gavaged with 1 ml of 0, 50 mg/L, 100 mg/L, or 200 mg/L of monochloramine per day for 5 days and then sacrificed (Meier et al., 1985).

9.2.5 Reproductive and developmental toxicity

No development or reproductive effects were seen in two limited oral rat studies using monochloramine in drinking water. No studies were located for di- and trichloramine.

Adult male (12/dose) and female Long–Evans rats (24/dose) were gavaged with monochloramine in deionized water at doses of 0, 2.5, 5, and 10 mg/kg bw per day for 66 days (males) and 76 days (females) before and during mating. Dosing continued throughout gestation and up to day 21 of lactation for females (Carlton et al., 1986). No changes in body weight gain, hematology, fertility, reproductive organ histopathology or sperm parameters (including motility, morphology and count) were seen in adults. Lactation, litter viability, litter size, mean pup weight and day of eye opening were not affected up to postnatal day (PND) 21. Several female pups were retained until PNDs 28–40 for observation and showed no effect on average days of vaginal patency. Statistical analysis of the data is not provided in the study.

No significant external or soft tissue anomalies or resorption were seen in fetuses of female Sprague–Dawley rats (6/dose) exposed to monochloramine in drinking water (0, 1 mg/L, 10 mg/L, or 100 mg/L) for 2½ months prior to mating and up to gestation day 20 (Abdel-Rahman et al., 1982). Information on statistical analyses, maternal weight, drinking water intake, and monochloramine intake were not provided in the study. The study also used a small number of animals per dose and reported fetal effects based on individual pups rather than by litter or maternal unit.

10.0 Classification and assessment

The International Agency for Research on Cancer (IARC, 2004) has classified monochloramine as "not classifiable as to its carcinogenicity to humans" (Group 3) based on inadequate evidence in humans and in experimental animals. Similarly, the U.S. EPA (1994a) classified monochloramine as not classifiable as to human carcinogenicity (Group D), based on inadequate evidence of cancer in both humans and animals. The NTP (1992) study concluded that there was equivocal (marginal) evidence of carcinogenic activity in female rats based on an increase in the incidence of MNCL and that there was no evidence of carcinogenic activity of chloraminated water in male rats, male mice, or female mice. Although an increased incidence of MNCL in female rats was observed, there was no evidence of a dose–response relationship and no evidence of a temporal relationship between increasing dose and incidence of tumours. In addition, it was noted that MNCL has a high spontaneous rate of occurrence in female F344 rats and that the levels reported in the NTP (1992) study were within the historical control range (U.S. EPA, 1994a). The use of F344 rats by NTP has since been discontinued in carcinogenicity studies because of the high and variable background incidence of MNCL in F344 rats, the species- and strain-specificity of MNCL and its questionable significance to humans, especially when rates fall within normal historical control values (King-Herbert and Thayer, 2006; Maronpot et al., 2016).

The databases for di- and trichloramine were weak, with only two subchronic oral or drinking water studies available. Although additional studies are available for trichloramine via inhalation, these pertain to occupational and swimming pool exposures, which occurred under different conditions than those found in drinking water systems. It is unlikely that trichloramine would be produced during regular residential use of chloraminated drinking water (e.g., during showering or bathing).

The database for monochloramine is better characterized and consists of studies conducted orally, with several studies dosing in drinking water. In drinking water, monochloramine is the predominant chloramine present while dichloramine and trichloramine are infrequently detected.

A decrease in animal body weight gain was observed in a number of subchronic and chronic studies conducted using monochloramine with rats and mice. These effects occurred at doses of 50–200 ppm (2.6–9.5 mg/kg bw per day for rats; 7.4–24.2 mg/kg bw per day for mice). Several authors have suggested that this response is probably linked to decreased water consumption by rodents, most likely due to aversion to the taste of high levels of chloramine in drinking water, and thus not considered biologically significant (Daniel et al., 1990, 1991; NTP, 1992). This opinion is also shared by the U.S. EPA (1994b) and the WHO (2011). Poon et al. (1997) demonstrated this effect by adding a paired-water control to their study which found changes in body weights, organ weights, and minor changes in biochemistry, hematology, immunological and histopathology in the treated group as well as in the paired-water controls. The authors concluded that these changes were related to the reduced water and food consumption seen and not as a result of monochloramine toxicity.

Some possible immunological effects have been reported, although the biological significance of these effects is unclear. No treatment-related developmental or reproductive effects have been observed in rats exposed to monochloramine in drinking water in limited studies. It should also be noted that monochloramine rapidly breaks down in the stomach (see Section 9.2.3), forming other potentially toxic compounds depending on the stomach contents and pH.

In humans, results from available studies and surveys did not indicate any adverse health effects associated with exposure to monochloramine in drinking water. Because there were limitations with these studies (e.g., weak exposure assessment, chloramination used as control, sample bias), no definitive conclusions could be drawn. There have been no published reports of dermal or ocular irritation in humans following bathing or showering with chloraminated tap water. In addition, no information is available on any potential systemic toxicity that can be caused by exposure to chloramine via the dermal route.

A toxic endpoint from exposure to monochloramine (or di- or trichloramine) has not been identified due to the lack of toxicity observed in rodent and human studies. Hence, it is not deemed necessary to establish a Health Based Value for chloramines in drinking water.

10.1 Aesthetic considerations

Chloramines are in general less noticeable and less offensive than free chlorine.Monochloramine normally does not contribute significantly to objectionable taste and odour issues when present at concentrations typically found in drinking water; however, the presence of dichloramine and trichloramine is more likely to cause customer taste and odour complaints. These complaints can be an important source of information for utilities, and operational strategies to reduce tastes and odours include treating water to remove taste and odour precursors, flushing the distribution system and reducing detention time and water age in the distribution system. The main objective of monochloramine is to provide secondary disinfection, protecting public health by maintaining the microbiological safety of the drinking water supply during distribution. Therefore, where elevated chloramine concentrations are required in order to maintain an effective disinfectant residual throughout the distribution system, the taste thresholds may be exceeded.

10.2 International considerations

Several organizations have set guidelines or regulations pertaining to the concentration of chloramines in drinking water, all based on the absence of toxicity in the same chronic 2-year study (NTP, 1992), which indicated a NOAEL of 200 ppm, the highest dose of the study. None of the agencies applied an additional uncertainty factor for possible carcinogenicity, as the carcinogenic effects reported in the study were equivocal and the tumour incidences were within the range observed in historical controls.

The U.S. EPA (1998) has established a maximum residual disinfectant level (MRDL) and a maximum residual disinfectant level goal (MRDLG) for chloramines, both set at 4.0 mg/L (4,000 μg/L) and measured as combined total chlorine (as Cl2). The U.S. EPA (1998) considers MRDLs to be enforceable standards, analogous to maximum contaminant levels (MCLs), which recognize the benefits of adding a disinfectant to water on a continuous basis and of maintaining a residual to control for pathogens in the distribution system. The MRDL and MRDLG are the same because there are no limitations imposed by the analytical methods or treatment technology.

The World Health Organization (2004, 2011) established a drinking water guideline of 3 mg/L (or 3,000 μg/L) for monochloramine based on the same NTP (1992) study. WHO (2011) noted that there was insufficient data to establish a guideline for the other two forms of inorganic chloramines, di-and trichloramine. The Australian drinking water guideline for monochloramine is also set at 3.0 mg/L (NHMRC, 2011), based on the same NTP (1992) study. The European Union has not established a guideline value for chloramines and only a handful of European countries use monochloramine as a disinfectant to treat drinking water.

11.0 Rationale

The primary concern when using monochloramine for secondary disinfection is to maintain the microbiological safety of the drinking water supply during distribution in order to protect public health. Health risks from chloramines or from any of its DBPs are much lower than the risks from consuming water that has not been adequately disinfected. Minimal to no effects were seen in both humans and animals following ingestion of monochloramine, with the most significant effect being decreased body weight gain in animal studies. However, this effect is attributed to a reduction in water consumption caused by taste aversion to drinking water with high levels of monochloramine. The information on dichloramine and trichloramine is insufficient to establish any link with unwanted health effects in animals or in humans. These forms are also not frequently detected in drinking water.

Based on the lack of toxicity observed in rodent and human studies, Health Canada in collaboration with the Federal-Provincial-Territorial Committee on Drinking Water has deemed that there is no need to establish a guideline for chloramines in drinking water. It has also determined that an aesthetic objective is not necessary, since levels commonly found in drinking water are within an acceptable range for taste and odour, and since protection of consumers from microbial health risks is paramount.

Where chloramines are used as a drinking water disinfectant, it is recommended that their concentration be determined on a system-specific basis to ensure effectiveness of disinfection and maintenance of an appropriate residual while minimizing by-product formation and aesthetic concerns.

12.0 References

Appendix A: List of acronyms

ANSI
American National Standards Institute
APHA
American Public Health Association
bw
body weight
CDC
Centers for Disease Control and Prevention (United States)
CI
confidence interval
Cl2:NH3-N
chlorine:ammonia weight ratio
CT
concentration × time
DBPs
disinfection by-products
DPD
N,N-diethyl-p-phenylenediamine
EPA
Environmental Protection Agency (United States)
GSH
glutathione
HAAs
haloacetic acids
HOCl
hypochlorous acid
HPC
heterotrophic plate count
IARC
International Agency for Research on Cancer
I-DBPs
iodinated disinfection by-products
MDL
Method detection limit
MNCL
mononuclear cell leukemia
MRDL
maximum residual disinfectant level (United States)
MRDLG
maximum residual disinfectant level goal (United States)
NA
not available
NDMA
N-nitrosodimethylamine
NOAEL
no-observed-adverse-effect-level
NOM
natural organic matter
NSF
NSF International
NTP
National Toxicology Program (United States)
OPPPs
opportunistic premise plumbing pathogens
OR
odds ratio
T3
triiodothyronine
T4
thyroxine
THMs
trihalomethanes
WHO
World Health Organization

Footnotes

Footnote 1

A measure of a chemical's efficacy as a disinfectant and is the residual concentration of a disinfectant, measured in mg/L (C) multiplied by the disinfectant contact time, measured in minutes (T).

Return to footnote 1 referrer

Footnote 2

The study also looked at other drinking water chemicals.

Return to footnote 2 referrer

Footnote 3

Immunotoxic endpoints included Natural Killer Cell Activity assay; IgM antibody titers; hemolytic plaque assay; red blood cell counts; white blood cell counts and differentials; and organ weights (including thymus and spleen).

Return to footnote 3 referrer

Footnote 4

Drinking water was buffered, charcoal-filtered and deionized.

Return to footnote 4 referrer

Footnote 5

Sample calculation taken from NTP (1992) appendix L (tables L3, L4, L7 and L8) dose per day: [(13weeks/total number of weeks) X mean value for weeks 1–13)] + [(39 weeks/total number of weeks) * mean value for weeks 14– 52] + [(49 weeks/total number of weeks) X mean value for weeks 53–101)] = calculated weighted mean equivalent doses

Return to footnote 5 referrer

Page details

Date modified: