How to Measure Flash Point of a Substance

Flash Point

Flash Point is the lowest temperature at which vapor of a volatile combustible substance can ignite in the air in the presence of an ignitable source. At flash point, more vapor may not be produced rapidly enough to sustain combustion. Hence, the fire is mostly not sustained. At some temperature above the flash point, the liquid will produce enough vapor to support combustion. (This temperature is known as the fire point).

It is used to measure how volatile a liquid is and the level of risk it could pose. They are measured by heating a liquid to specific temperatures under controlled conditions and then applying a flame.

This test is done using an:

  • Open cup test apparatus
  • A close cup test apparatus or
  • Both

How It Is Measured

There are two basic types of flash point measurement: Open cup and Closed cup.

Open cup measurement:

In open cup devices, the sample is contained in an open cup which is heated and, at intervals, a flame brought over the surface. The measured flash point will vary with the height of the flame above the liquid surface and, at sufficient height, the measured flash point temperature will coincide with the fire point. The best-known example is the Cleveland Open Cup (COC).

Closed cup measurement:

There are two types of closed cup testers: Non-equilibrial, such as Pensky-Martens, where the vapors above the liquid are not in temperature equilibrium with the liquid, and equilibrial, such as Small Scale (commonly known as Setaflash), where the vapors are deemed to be in temperature equilibrium with the liquid. In both these types, the cups are sealed with a lid through which the ignition source can be introduced. Closed cup testers normally give lower values for the flash point than an open cup (typically 5–10 °C or 9–18 °F lower) and are a better approximation to the temperature at which the vapor pressure reaches the lower flammable limit.

READ ALSO  Fire Suppression Systems, Types & How It Works

Generally, a flash point is an empirical measurement rather than a fundamental physical parameter. The measured value will vary with equipment and test protocol variations, including temperature ramp rate (in automated testers), the time allowed for the sample to equilibrate, sample volume, and whether the sample is stirred.

Read AlsoHow to use a fire blanket

History

The use of the flashpoint as a measure of a liquid’s hazardousness dates from the 19th century. Before gasoline became important, kerosene was the main petroleum product (used mainly as fuel for lamps and stoves), and there was a tendency on the part of petroleum distillers to leave as much as possible of the commercially worthless gasoline in the kerosene to sell more product. This adulteration of kerosene with highly volatile gasoline caused numerous fires and explosions in storage tanks and oil lamps. Legal measures were instituted to curb the danger, test methods were prescribed and minimum flash points were set.

Flash Point Of Some Substances

  • Automotive gasoline, −43 °C (−45 °F)
  • Ethyl alcohol, 13 °C (55 °F)
  • Automotive diesel fuel, 38 °C (100 °F)
  • Kerosene, 42–72 °C (108–162 °F)
  • Home heating oil, 52–96 °C (126–205 °F)
  • SAE 10W-30 motor oil, 216 °C (421 °F)

Related Posts

Key difference between FLASH POINT and FIRE POINT

What is the difference between ignition temperature and fire point

Flammable and Combustible Liquid – Hazards

What Is Fire Point

11 Common Fire Hazards In The Workplace

Fire Tetrahedron; Unique Difference from the Fire Triangle

4 Methods Of Fire Extinguishing You Should Know