Friday, September 23, 2011

Optical Mapping

In Chapter 7 of the 4th edition of Intermediate Physics for Medicine and Biology, Russ Hobbie and I mention an optical technique that is used to measure the transmembrane potential in the heart.
Experimental measurements of the transmembrane potential often rely on the use of a voltage sensitive dye whose fluorescence changes with the transmembrane potential [Knisley et al. (1994); Neunlist and Tung (1995); Rosenbaum and Jalife (2001)].
This method, often called optical mapping, has revolutionized cardiac electrophysiology, because it allows you to use optical methods to make electrical measurements. If you want to learn more, take a look at the book Optical Mapping of Cardiac Excitation and Arrhythmias, by David Rosenbaum and Jose Jalife (2001). The chapters in this book were written by the stars of this field.
  1. Optical Mapping: Background and Historical Perspective. Guy Salama.
  2. Mechanisms and Principles of Voltage-Sensitive Fluorescence. Leslie M. Loew.
  3. Optical Properties of Cardiac Tissue. William T. Baxter.
  4. Optics and Detectors Used in Optical Mapping. Kenneth R. Laurita and Imad Libbus.
  5. Optimization of Temporal Filtering for Optical Transmembrane Potential Signals. Francis X. Witkowski, Patricia A. Penkoske, and L. Joshua Leon.
  6. Optical Mapping of Impulse Propagation within Cardiomyocytes. Herbert Windisch.
  7. Optical Mapping of Impulse Propagation between Cardiomyocytes. Stephan Rohr and Jan P. Kucera.
  8. Role of Cell-to-Cell Coupling, Structural Discontinuities, and Tissue Anisotropy in Propagation of the Electrical Impulse. André G. Kléber, Stephan Rohr, and Vladimir G. Fast.
  9. Optical Mapping of Impulse Propagation in the Atrioventricular Node: 1. Todor N. Mazgalev and Igor R. Efimov.
  10. Optical Mapping of Impulse Propagation in the Atrioventricular Node: 2. Guy Salama and Bum-Rak Choi.
  11. Optical Mapping of Microscopic Propagation: Clinical Insights and Applications. Albert L. Waldo.
  12. Mapping Arrhythmia Substrates Related to Repolarization: 1. Dispersion of Repolarization. Kenneth R. Laurita, Joseph M. Pastore, and David S. Rosenbaum.
  13. Mapping Arrhythmia Substrates Related to Repolarization: 2. Cardiac Wavelength. Steven Girouard and David S. Rosenbaum.
  14. Video Imaging of Cardiac Fibrillation. Richard A. Gray and José Jalife.
  15. Video Mapping of Spiral Waves in the Heart. William T. Baxter and Jorge M. Davidenko.
  16. Video Imaging of Wave Propagation in a Transgenic Mouse Model of Cardiomyopathy. Faramarz Samie, Gregory E. Morley, Dhjananjay Vaidya, Karen L. Vikstrom, and José Jalife.
  17. Optical Mapping of Cardiac Arrhythmias: Clinical Insights and Applications. Douglas L. Packer.
  18. Response of Cardiac Myocytes to Electrical Fields. Leslie Tung.
  19. New Perspectives in Electrophysiology from The Cardiac Bidomain. Shien-Fong Lin and John P. Wikswo, Jr..
  20. Mechanisms of Defibrillation: 1. Influence of Fiber Structure on Tissue Response to Electrical Stimulation. Stephen B. Knisley.
  21. Mechanisms of Defibrillation: 2. Application of Laser Scanning Technology. Stephen M. Dillon.
  22. Mechanisms of Defibrillation: 3. Virtual Electrode-Induced Wave Fronts and Phase Singularities; Mechanisms of Success and Failure of Internal Defibrillation. Igor R. Efimov and Yuanna Cheng.
  23. Optical Mapping of Cardiac Defibrillation: Clinical Insights and Applications. Douglas P. Zipes.
For those who are tired of reading, two videos have recently been published in the Journal of Visualized Experiments that explain the technique step-by-step. One video is about studying the rabbit heart and the other is about the mouse heart. These excellent video clips were filmed in the laboratory of Igor Efimov, of Washington University in Saint Louis.

My former graduate student, Debbie Janks, is now a post doc in Efimov’s lab. Regular readers of this blog may recognize Janks’ name, as she provides many insightful comments following these blog entries. Janks studied optical mapping from a theoretical perspective when she was here at Oakland University. She published a nice paper that examined the question of averaging over depth during optical mapping. The optical method does not measure the transmembrane potential at the tissue surface. Rather, light penetrates some distance into the tissue, and the optical signal is a weighted average of the transmembrane potential over depth. Janks looked at the effect of this averaging during an electrical shock. Rather than explaining the whole story, I will present it as a new homework problem. That way, you can figure it out for yourself. Enjoy.
Section 7.10

Problem 47 1/2 The signal measured during optical mapping, V, is a weighted average of the transmembrane potential, Vm(z), as a function of depth, V=∫0Vm(z)w(z)dz, where w(z) is a normalized weighting function. Suppose the light decays with depth exponentially, with an optical length constant δ. Then w(z) = exp(−z/δ)/δ. Often a shock will cause Vm(z) to fall off exponentially with depth, Vm(z)=Vo exp(z/λ), where Vo is the transmembrane potential at the tissue surface and λ is the electrical length constant (see Sec. 6.12).
(a) Perform the required integration to find an analytical expression for the optical signal, V, as a function of Vo, δ and λ.
(b) What is V in the case δ much less than λ? Explain this result physically.
(c) What is V in the case δ much greater than λ? Explain this result physically.
(d) For which limit do you obtain an accurate measurement of the transmembrane potential at the surface, V=Vo?
In cardiac tissue, δ is usually on the order of a millimeter, whereas λ is more like a quarter of a millimeter, so averaging over depth significantly distorts the measured signal. For a more detailed analysis of this problem, see Janks and Roth (2002).

Friday, September 16, 2011

Does cell biology need physicists?

The American Physical Society has an online journal, Physics, with the goal of making recent research accessible to a wide audience. The journal website states:
Physics highlights exceptional papers from the Physical Review journals. To accomplish this, Physics features expert commentaries written by active researchers who are asked to explain the results to physicists in other subfields. These commissioned articles are edited for clarity and readability across fields and are accompanied by explanatory illustrations.
One recent paper that caught my eye was an essay written by Charles Wolgemuth, titled “Does Cell Biology Need Physicists?” Wolgemuth asks key questions in the introduction to his essay.
The past has shown that cell biologists are extremely capable of making great progress without much need for physicists (other than needing physicists and engineers to develop many of the technologies that they use). Do the new data and new technological capabilities require a physicist’s viewpoint to analyze the mechanisms of the cell? Is physics of use to cell biology?
Later in the essay, Wolgemuth asks his central question in a more specific way:
It is possible that the physics that cells must deal with is slave to the reactions; i.e., the protein levels and kinetics of the biochemical reactions determine the behavior of the system, and any physical processes that a cell must accomplish are purely consequences of the biochemistry. Or, could it be that cellular biology cannot be fully understood without physics?
Readers of the 4th edition of Intermediate Physics for Medicine and Biology are likely to scream “Yes!” to these questions. I too enthusiastically answer yes, but I agree with Wolgemuth that it is proper to ask such basic questions occasionally.

I should add that Russ Hobbie and I tend to look primarily at macroscopic phenomena in Intermediate Physics for Medicine and Biology, such as the biomechanics of walking with a cane, the interpretation of an electrocardiogram, or the algorithm required to form an image of the brain using a CAT scan. We occasionally look at events on the atomic scale, but for the most part we ignore molecular biophysics. Yet, the cellular scale is an interesting intermediate level that is becoming a fertile field for the applications of physics to biology. Indeed, I examined this issue when discussing the textbook Physical Biology of the Cell last year in this blog. The discussion that Russ and I give to fluid dynamics, diffusion, and bioelectricity in Intermediate Physics for Medicine and Biology is relevant to cellular topics.

To answer his question, Wolgemuth provides five examples in which physics provides key insights into cellular biology: 1) Molecular motors, 2) Cellular movement, 3) How cells swim, 4) Cell growth and division, and 5) How cells interact with the environment. One of my favorite parts of the essay is the consideration of potential pitfalls for physicists in biology.
Fifteen years ago, around the time that I began working in biophysics, there were very few collaborations between physicists and cell biologists, especially if the physicists were theorists. Theory was, and still is to a good degree, a word that should be avoided in the presence of biologists. Those of us who use math and computers to try to understand how cells work tend to call ourselves modelers instead of theorists. My suspicion is that many of the first physicists and mathematicians who tried to develop models for how biology works attempted to be too abstract or too general. As physicists we like to try to find universal laws, and though there are undoubtedly general principles at play in cell biology, it is likely that there are no real universal rules. Evolution need not find only one way to do something but more often probably finds many. Rather than search out generalities, we will serve biology better if we deal with specifics. As Aharon Katchalsky, who is largely credited with bringing nonequilibrium thermodynamics to biology, purportedly said, “It is easier to make a theory of everything than a theory of something.”

In recent years, physicists have done a much better job at addressing specific problems in biology. However, there still remains a divide between the two communities. Indeed, good physical biology that comes out of the physics community often goes unnoticed or is under appreciated. The burden falls on us to properly convey our work so as to be accessible to biologists. We need to make conscious efforts at communication and dissemination of our results. Two useful approaches toward this end are to publish in broader audience journals that reach both communities, and for papers that contain theoretical analyses to provide a qualitative description of the modeling in the main text, while leaving the more mathematical details for the appendices or supplemental material (for further discussion of this topic, see Ref. [55]). It is also of prime importance to maintain and to forge new connections between physicists and biologists.
Wolgemuth comes closest to answering his own questions near the end of the essay, where he predicts
To be truly successful, we must provide an understanding of biology that spans the gorge from biochemistry and genetics to cellular function, and do it in such a way that our models and experiments are not only informative about physics, but directly impact biology.

Cell biology is awaiting these descriptions. And it may be that physicists are the most able to draw these connections between the protein level description of cellular biology that currently dominates and a more intuitive, yet still quantitative, description of the behavior of cells and their responses to their environments.

Friday, September 9, 2011

Radon Transform

In Chapter 12 of the 4th Edition of Intermediate Physics for Medicine and Biology, Russ Hobbie and I introduce the Radon transformation. It consists of finding the projections F(θ, x') at different angles θ from a function f(x,y). But why is it called the “Radon” transformation, and does it have anything to do with the radioactive gas radon discussed in Chapter 16?

Well, it has nothing to do with the element radon. Instead, and predictably, the term honors Johann Radon, the Austrian mathematician who investigated this transformation. In “A Tribute to Johann Radon” in the IEEE Transactions on Medical Imaging (Volume 5, Page 169, 1986, reproduced long after his death to honor his memory) Hans Hornich wrote
With the death in Vienna on 25 May 1956 of Dr. Johann Radon, Professor of the University of Vienna, not only the mathematical world and Austrian science but also the German Mathematical Union has suffered a severe loss, as have also many other scientific bodies of which the deceased was a prominent member, and who spent most of his teaching life in German universities.

Radon was born in the small town of Tetschen in Bohemia near the border of Saxony on December 16, 1887. He studied at Vienna University where, alongside Mertens and Wirtinger, Escherisch above all was the great influence on Radon’s development: Escherisch had, as one of the first in Austria, imparted to his students the world of ideas of Weierstrass and his rigorous foundations of analysis. Through Escherich, Radon was led next to variational calculus….

A few years later appeared his “Habilitationsschrift” “Theory and application of absolute additive weighting functions” (S. Ber. math. naturw., Kl. K. Akad. Wiss. Wien II Abt., vol. 122, pp. 1295–1438, 1913), which played a leading role in the development of analysis; the Radon integral and the Radon theorem laid the foundations of functional analysis. As an application Radon somewhat later treated the first and second boundary value problem of the logarithmic potential in a very general way.
The Radon transformation has important applications in medical imaging, and plays a crucial role in computed tomography, positron emission tomography, and single photon emission tomography. I found a nice layman’s description of the Radon transform in an essay at the website http://www.ams.org/samplings/feature-column/fcarc-tomography, written by Bill Casselman.
The original example of this sort of technology [involving a collaboration between medicine and mathematics], and the ancestor of many of these technologies, is what is now called computed tomography, for which Allan Cormack, a physicist whose research became more and more mathematical as time went on, laid down the theoretical foundations around 1960. He shared the 1979 Nobel prize in medicine for his work in this field.

In fact the basic idea of tomography had been discovered for purely theoretical reasons in 1917 by the Austrian mathematician Johann Radon, and it had been rediscovered several times since by others, but Cormack was not to know this until much later than his own independent discovery. The problem he solved is this: Suppose we know all the line integrals through a body of varying density. Can we reconstruct the body itself? The answer, perhaps surprisingly, is that we can, and furthermore we can do so constructively. In practical terms, we know that a single X-ray picture can give only limited information because certain things are obscured by other, heavier things. We might take more X-ray pictures in the hope that we can somehow see behind the obscuring objects, but it is not at all obvious that by taking a lot—really, a lot—of X-ray pictures we can in effect even see into objects, which is what Radon tells us, at least in principle. Making Radon’s theorem into a practical tool was not a trivial matter.”
You can listen to a lecture on tomography and inverting the Radon transform here.

Friday, September 2, 2011

Fraunhofer Diffraction

Last week’s blog entry discussed Fresnel diffraction, which Russ Hobbie and I analyzed in the 4th edition of Intermediate Physics for Medicine and Biology when we examined the ultrasonic pressure distribution produced near a circular piezoelectric transducer. This week, I will analyze diffraction far from the wave source, known as Fraunhofer diffraction, named for the German scientist Joseph von Fraunhofer.

The mathematics of Fraunhofer diffraction is a bit too complicated to derive here, but the gist of it can be found by inspecting the first equation in Section 13.7 (Medical Uses of Ultrasound), found at the bottom of the left column on page 351. The pressure is found by integrating 1/r times a cosine function over the transducer face. When you are far from the transducer, r is approximately a constant and can be taken out of the integral. In that case, you just integrate cosine over the transducer area. This becomes similar to the two-dimensional Fourier transform defined in Chapter 12 (Images). The far field pressure distribution produced by a circular transducer given in Eq. 13.40 is the same Bessel function result as derived in Problem 10 of Chapter 12.

The intensity distribution in Eq. 13.40 is known as the Airy pattern, after the English scientist and mathematician George Biddell Airy. As shown in Fig. 13.15, the pattern consists of a central peak, surrounded by weaker secondary maxima. The Airy pattern occurs during imaging using a circular aperture, such as when viewing stars through a telescope. Two adjacent stars appear as two Airy patterns. Distinguishing the two stars is difficult unless the separation between the images is greater than the separation between the peak of the Airy pattern and its first zero. This is called the Rayleigh criterion, after Lord Rayleigh. Rayleigh (1842–1919, born John William Strutt)—one of those 19th century English Victorian physicists I like so much—did fundamental work in acoustics, and published the classic textbook Theory of Sound.

Friday, August 26, 2011

Fresnel Diffraction

In Section 13.7 of the 4th edition of Intermediate Physics for Medicine and Biology, Russ Hobbie and I discuss the medical uses of ultrasound. One important problem we analyze is the pressure distribution produced by a piezoelectric transducer.
There are some important features of the radiation pattern from a transducer which we review next. Consider a circular transducer, the surface of which is oscillating back and forth in a fluid… Each small element of the vibrating fluid creates a wave that travels radially outward, the points of constant phase being expanding hemispheres. The amplitude of each spherical wave decreases as 1/r, the intensity falling as 1/r2. We want the pressure at a point z on the axis of the transducer. It is obtained by summing up the effect of all the spherical waves emanating from the face of the transducer….

The [average intensity] is plotted in Fig. 13.13 for a fairly typical but small transducer (a = 0.5 cm, f = 2 MHz)... Close to the transducer there are large oscillations in intensity along the axis: there are corresponding oscillations perpendicular to the axis, as shown in Fig. 13.14. The maxima and minima form circular rings. This is called the near field or Fresnel zone… The depth of the Fresnel zone is approximately a2/λ [where a is the radius of the transducer, λ is the wavelength, and f is the frequency].
Figure 13.13 from the 4th edition of Intermediate Physics for Medicine and Biology, showing Fresnel diffraction.
Fig, 13.13.
The calculated intensity along the axis, as shown in our Fig. 13.13, is interesting. In the Fresnel zone, the intensity has many points where it is zero. In Intermediate Physics for Medicine and Biology we calculate why this happens mathematically, but it is illuminating to describe what is happening physically. Basically, this is a result of wave interference. Our statement that “each small element of the vibrating fluid creates a wave that travels radially outward” is often called Huygens principle. Each point on the face of the transducer produces such a wavelet. To understand the pressure distribution, we must examine the phase relationship among these various wavelets. Very near the face of the transducer, the waves that contribute significantly to the pressure are in phase; they all interfere constructively and you get a maximum (evaluate Eq. 13.39 at z = 0 and you get a nonzero constant). However, as you move away, more distant points on the transducer face contribute to the pressure on the axis, and these points may be out of phase with the pressure produced by the point at the center. For some value of z the in-phase and out-of-phase wavelets interfere destructively, resulting in zero intensity. Increase z a little more, and not only do the in-phase points at the center and the out-of-phase points just away from the center contribute to the pressure, but so do some in-phase points even farther from the center. When you add it all up, you get a net constructive interference and a non-zero intensity. And so it goes, as you move out farther and farther along the z axis.

Figure 13.14 from the 4th edition of Intermediate Physics for Medicine and Biology, showing Fresnel diffraction.
Fig. 13.14.
The radial distribution of the intensity is surprisingly rich and complex, given the rather simple integral that underlies the behavior. If you want to explore the radial distribution in more detail, go to the excellent website http://wyant.optics.arizona.edu/fresnelZones/fresnelZones.htm, where you can perform these calculations yourself. You can adjust the parameters as you wish and create plots such as those in Fig. 13.14, and also produce grayscale images of the full intensity distribution that provide much insight. The website was produced with optics in mind, so you have to put in strange looking parameters to model ultrasound. To reproduce the middle panel of Fig 13.14, input 770,000 for the wavelength in nm, 10,000 for the aperture diameter in microns, and 15.75 for the observation distance in mm. To my eye, the agreement between the website’s calculation and Fig. 13.14 is impressive. At small values of z the plots get very complex and beautiful. For the same wavelength and aperture, I like the richness of z = 5 mm, and for z = 4 mm you get a fairly uniform brightness except for a dramatic dark spot right at the center. It reminds me of Poisson’s spot, which I discussed in the September 17, 2010 entry in this blog, about Augustin-Jean Fresnel. Indeed, the physics behind the calculations in Fig. 13.14 and Poisson’s spot in optics are nearly identical. The circular aperture is a classic problem in Fresnel diffraction. You can find a more detailed discussion of this topic in the textbook Optics (4th edition), by Eugene Hecht. (My bookshelf contains the first edition, by Hecht and Zajac, that I used in my undergraduate optics class at the University of Kansas).

If you want to be clever, you could make the ultrasound transducer vibrate only at those radii that result in constructive interference along the axis, and have it remain stationary at radii that cause destructive interference. (Of course, this would mean you would have to design your transducer face cleverly so concentric rings vibrate, separated by rings that do not, which might make constructing the transducer more difficult.) Using such a trick eliminates the dark spots along the z axis, increasing the intensity there. This method is commonly used to focus light waves, and is called a zone plate. It has been used occasionally with ultrasound.

Friday, August 19, 2011

The Nonlinear Poisson-Boltzmann Equation

Last week’s blog entry was about the Gouy-Chapman model for a charged double layer at an electrode surface. The model is based on the Poisson-Boltzmann equation (Eq. 9.10 in the 4th edition of Intermediate Physics for Medicine and Biology). One interesting feature of the Poisson-Boltzmann equation is that it is nonlinear. In applications when the thermal energy of ions in solution is much greater than the energy of the ions in an electrical potential, the equation can be linearized (Eq. 9.13). That is not always the case.

Homework problem 9 in Chapter 9 of Intermediate Physics for Medicine and Biology was added in the 4th edition. It begins
Problem 9 Analytical solutions to the nonlinear Poisson-Boltzmann equation are rare but not unknown. Consider the case when the potential varies in one dimension (x), the potential goes to zero at large x, and there exists equal concentrations of monovalent cations and anions. Chandler et al. (1965) showed that the solution to the 1-d Poisson-Boltzmann equation, d2ζ/dx2=sinh(ζ), is…
You will need to get a copy of the book to see this lovely solution. It is a bit too complicated to write in this blog, but it involves the exponential function, the hyperbolic tangent function, and the inverse hyperbolic tangent function. I like this homework problem, because you can solve both the nonlinear and linear equations exactly, with the same boundary conditions, and compare them to get a good intuitive feel for the impact of the nonlinearity. I admit, the problem is a bit advanced for an intermediate-level book, but upper-level undergraduates or graduate students studying from our text should be up to the challenge.

The full citation to the paper by Knox Chandler, Alan Hodgkin, and Hans Meves mentioned in the problem is
Chandler, W. K., A. L. Hodgkin, and H. Meves (1965) “The Effect of Changing the Internal Solution on Sodium Inactivation and Related Phenomena in Giant Axons,” Journal of Physiology, Volume 180, Pages 821–836.
I always thought it odd that one finds a really elegant analytical solution to the nonlinear Poisson-Boltzmann equation in a paper about sodium channel inactivation in a squid nerve axon (with Nobel Prize-winning physiologist Alan Hodgkin as a coauthor). The solution is buried in the discussion (in a section set of in a smaller font than the rest of the paper). The reason for its appearance is that Chandler et al. found changes in membrane behavior with intracellular ion concentration, and postulated that the measured voltage drop between the inside and outside of the axon consisted of a voltage drop across the membrane itself (which affects the ion channel behavior) and a voltage drop within a double layer adjacent to the membrane. It is the double layer voltage that they model using the Poisson-Boltzmann equation.

Nowadays, the nonlinear Poisson-Boltzmann equation is typically solved using numerical methods. See, for example, the paper that Russ Hobbie and I cite in Intermediate Physics for Medicine and Biology, written by Barry Honig and Anthony Nicholls: “Classical Electrostatics in Biology and Chemistry,” Science, Volume 268, Pages 1144–1149, 1995 (it now has over 1500 citations in the literature). Their abstract states
A major revival in the use of classical electrostatics as an approach to the study of charged and polar molecules in aqueous solution has been made possible through the development of fast numerical and computational methods to solve the Poisson-Boltzmann equation for solute molecules that have complex shapes and charge distributions. Graphical visualization of the calculated electrostatic potentials generated by proteins and nucleic acids has revealed insights into the role of electrostatic interactions in a wide range of biological phenomena. Classical electrostatics has also proved to be a successful quantitative tool yielding accurate descriptions of electrical potentials, diffusion limited processes, pH-dependent properties of proteins, ionic strength-dependent phenomena, and the solvation free energies of organic molecules.
Such calculations continue to be an active area of research. See, for example, “The Role of DNA Shape in Protein-DNA Recognition” by Remo Rohs, Sean West, Alona Sosinsky, Peng Liu, Richard Mann and Barry Honig (Nature, Volume 461, Pages 1248–1253, 2009).

Friday, August 12, 2011

Gouy and Chapman

Sometimes when I am studying physics, I run across a model or equation with names attached to it, and I wonder “just who are these people?” For example, in Chapter 9 of the 4th edition of Intermediate Physics for Medicine and Biology, Russ Hobbie and I discuss the Gouy-Chapman model.
In this section we study one model for how ions are distributed at the interface in Donnan equilibrium. The model was used independently by Gouy and by Chapman to study the interface between a metal electrode and an ionic solution. They investigated the potential changes along the x axis perpendicular to a large plane electrode. The same model is used to study the charge distribution in a semiconductor.
So who are Gouy and Chapman? I can tell they lived a long time ago, because in the next section we write “In an ionic solution, ions of opposite charge attract one another. A model of this neutralization was developed by Debye and Huckel a few years after Gouy and Chapman developed [their] model.” I know Peter Debye worked in the early days of quantum mechanics, so Gouy and Chapman had to be a bit earlier.

As if often the case with scientific developments in the 19th century, a Frenchman and an Englishman share credit for the discovery. Louis Georges Gouy (1854–1926) was a French experimental physicist from the University of Lyon. He is best known as the inventor of the Gouy balance, a device for measuring magnetic susceptibility. Gouy had an interest in Brownian motion at a time when the atomic nature of matter was still an open question. Russ and I discuss Brownian motion in Section 4.3 of our book.
[The] movement of microscopic-sized particles, resulting from bombardment by much smaller invisible atoms, was first observed by the English botanist Robert Brown in 1827 and is called Brownian motion.
Albert Einstein wrote a fundamental paper on Brownian motion in 1905, his miraculous year. In a subsequent paper on the same topic, Einstein began (Annalen der Physik, Volume 19, Pages 371–381, 1906)
Soon after the appearance of my paper on the movements of particles suspended in liquids demanded by the molecular theory of heat, Siedentopf (of Jena) informed me that he and other physicists—in the first instance, Prof. Gouy (of Lyons)—had been convinced by direct observation that the so-called Brownian motion is caused by the irregular thermal movements of the molecules of the liquid.
Gouy’s model sought to explain the electrical potential around small Brownian particles, or colloids. He described the double layer of charge that develops at the surface, with one charge layer bound to the surface of the particle, and a layer of counterions in the surrounding fluid.

David Leonard Chapman (1869–1958) was an English physical chemist at Oxford. He was interested in the theory of detonation in gasses, and developed the Chapman-Jouget condition describing their behavior. About three years after Gouy, Chapman derived a model describing the double layer at a charged surface. The Gouy-Chapman model is an application of what is now known as the Poisson-Boltzmann equation (Eq. 9.13 in our book).

In 1924, physicist Otto Stern extended the Gouy-Chapman model by noting that ions can’t be represented as point charges when they are within a few ion radii of the surface. This leads to the Stern layer of immobile counter-ions right next to the surface, and a diffuse layer of counter-ions whose concentration decays exponentially.

Friday, August 5, 2011

Fisher-Kolmogorov equation

The two volumes of Mathematical Biology, by James Murray, superimposed on Intermediate Physics for Medicine and Biology.
Mathematical Biology,
by James Murray.
In the 4th edition of Intermediate Physics for Medicine and Biology, Russ Hobbie and I discuss many of the important partial differential equations of physics, such as Laplace’s equation, the diffusion equation, and the wave equation. One lesser-known PDE that we don’t discuss is the Fisher-Kolmogorov equation. However, our book supplies most of what you need to understand this equation.

In Section 2.10, we examine the logistic equation, an ordinary differential equation governing population growth,

du/dt = b u (1 − u) .

For u much less than one, the population grows exponentially with rate b. As u approaches one, the population levels off near a steady state value of u = 1. Our Eq. 2.28 gives an analytical solution to this nonlinear equation.

In Section 4.8, we drive the diffusion equation, which for one dimension is

du/dt = D d2u/dx2 .

This linear partial differential equation is one of the most famous in physics. It describes diffusion of particles, and also the flow of heat by conduction. D is the diffusion constant.

To get the Fisher-Kolmogorov equation, just put the logistic equation and the diffusion equation together:

du/dt = D d2u/dx2 + b u (1 u) .

The Fisher-Kolmogorov equation is an example of a “reaction-diffusion equation.” Russ and I discuss a similar reaction-diffusion equation in Homework Problem 24 of Chapter 4, when modeling intracellular calcium waves. The only difference is that we use a slightly more complicated reaction term rather than the logistic equation.

Mathematical Biology, by James Murray, with Intermediate Physics for Medicine and Biology.
Mathematical Biology,
by James Murray.
In his book Mathematical Biology, James Murray discusses the Fisher-Kolmogorov equation in detail. He states
The classic simplest case of a nonlinear reaction diffusion equation … is [The Fisher-Kolmogorov equation]… It was suggested by Fisher (1937) as a deterministic version of a stochastic model for the spatial spread of a favoured gene in a population. It is also the natural extension of the logistic growth population model discussed in Chapter 11 when the population disperses via linear diffusion. This equation and its travelling wave solutions have been widely studied, as has been the more general form with an appropriate class of functions f(u) replacing ku(1u). The seminal and now classical paper is that by Kolmogoroff et al. (1937)…. We discuss this model equation in the following section in some detail, not because in itself it has such wide applicability but because it is the prototype equation which admits travelling wavefront solutions. It is also a convenient equation from which to develop many of the standard techniques for analyzing single-species models with diffusive dispersal.”
The Fisher-Kolmogorov equation was derived independently by Ronald Fisher (1890–1962), an English biologist, and Andrey Kolmogorov (1903–1987), a Russian mathematician. The key original papers are
Fisher, R. A. (1937) “The Wave of Advance of Advantageous Genes.” Annals of Eugenics, Volume 7, Pages 353–369.

Kolmogoroff, A., I. Petrovsky, and N. Piscounoff (1937) “Etude de l’equation de la diffusion avec croissance de la quantite de matiere et son application a un probleme biologique.” Moscow University Mathematics Bulletin, Volume 1, Pages 1–25.

Friday, July 29, 2011

The terahertz

In the opening section of the 4th edition of Intermediate Physics for Medicine and Biology, Russ Hobbie and I provide a table of common prefixes used in the metric system:
gigaG 109
megaM106
kilo k103
milli m10−3
microμ10−6
nanon10−9
pico p10−12
femto f10−15
attoa10−18
We went all the way down to “atto” on the small side, but stopped at “giga” on the large side. I now wish we had skipped “atto” and instead included “tera,” or “T”, corresponding to 1012. Why? Because the prefix tera can be very useful in optics when discussing the frequency of light.

To better appreciate why, take a look at the letter to the editor “Let’s Talk TeraHertz!” in the April 2011 issue of my favorite journal, the American Journal of Physics. There Roger Lewis—from the melodious University of Wollongong—argues that the terahertz is superior to the nanometer when discussing light. Lewis writes
…the terahertz shares the desirable properties of the nanometer as a unit in teaching optics… Like the nanometer, the terahertz conveniently represents visible light to three digits in numbers that fall in the midhundreds… The terahertz has other desirable properties that the nanometer lacks. First, the frequency is a more fundamental property of light than the wavelength because the frequency does not change as light traverses different media, whereas the wavelength may. Second, the energy of a photon is directly proportional to its frequency… The visible spectrum is often taken to span 400–700 nm, corresponding to 749–428 THz, falling in the octave 400–800 THz. …
I suspect that the reason I have always preferred wavelength over frequency when discussing light is that the nanometer provides such a nice, easy-to-remember unit to work with. Had I realized from the start that terahertz offered an equally useful unit for discussing frequency, I might naturally think in terms of frequency rather than wavelength. Incidentally, Planck’s constant is 0.00414 (or about 1/240) in the units of eV/THz.

After reading Lewis’s letter, I checked Intermediate Physics for Medicine and Biology to see how Russ and I characterized the properties of visible light. On page 360 I found our Table 14.2, which lists the different colors of the electromagnetic spectrum in terms of wavelength (nm), energy (eV) and frequency. We didn’t explicitly mention the unit THz, but we did list the frequency in units of 1012 Hz, so terahertz was there in every way but name. As a rule I don’t like to write in my books, but nevertheless I suggest that owners of Intermediate Physics for Medicine and Biology take a pencil and replace “(1012 Hz)” in Table 14.2 with “(THz)”.

Russ and I discuss the terahertz explicitly in Chapter 14 about Atoms and Light.
14.6.4 Far Infrared or Terahertz Radiation

For many years, there were no good sources or sensitive detectors for radiation between microwaves and the near infrared (0.1–100 THz; 1 THz = 1012 Hz). Developments in optoelectronics have solved both problems, and many investigators are exploring possible medical uses of THz radiation (“T rays”). Classical electromagnetic wave theory is needed to describe the interactions, and polarization (the orientation of the E vector of the propagating wave) is often important. The high attenuation of water in this frequency range means that studies are restricted to the skin or surface of organs such as the esophagus that can be examined endoscopically. Reviews are provided by Smye et al. (2001), Fitzgerald et al. (2002), and Zhang (2002).
The citations are to
Fitzgerald, A. J., E. Berry, N. N. Zinonev, G. C. Walker, M. A. Smith and J. M. Chamberlain (2002) “An Introduction to Medical Imaging with Coherent Terahertz Frequency Radiation,” Physics in Medicine and Biology, Volume 47, Pages R67–R84.

Smye, S. W., J. M. Chamberlain, A. J. Fitzgerald and E. Berry (2001) “The Interaction Between Terahertz Radiation and Biological Tissue,” Physics in Medicine and Biology, Volume 46, Pages R101–R112.

Zhang, X-C. (2002) “Terahertz Wave Imaging: Horizons and Hurdles,” Physics in Medicine and Biology, Volume 47, Pages 3667–3677.
So not only is the terahertz useful when talking about visible light, but also it is useful if working in the far infrared, when the frequency is about 1 THz. Such “T rays” (I hate that term) are being used nowadays for imaging during airport security and as a tool to study cell biology and cancer.

Friday, July 22, 2011

Euler: The Master of Us All

Euler: The Master of Us All,  by William Dunham, superimposed on Intermediate Physics for Medicine and Biology.
Euler: The Master of Us All,
by William Dunham.
Swiss mathematician Leonhard Euler (1707–1783) is a fascinating man. I discussed him once before in this blog, during an entry about the book e: The Story of a Number. Euler’s name never appears in the 4th edition of Intermediate Physics for Medicine and Biology, but his influence is there.
William Dunham describes Euler’s life and work in his book Euler: The Master of Us All. In the Preface, Dunham writes
This book is about one of the undisputed geniuses of mathematics, Leonhard Euler. His insight was breathtaking, his vision profound, his influence as significant as that of anyone in history. Euler contributed to long-established branches of mathematics like number theory, analysis, algebra, and geometry. He also ventured into the largely unexplored territory of analytic number theory, graph theory, and differential geometry. In addition, he was his century’s foremost applied mathematician, as his work in mechanics, optics, and acoustics amply demonstrates. There was hardly an aspect of the subject that escaped Euler’s penetrating gaze. As the twentieth-century mathematician Andre Weil put it, “All his life…he seems to have carried in his head the whole of the mathematics of his day, both pure and applied.”
In Chapter 11 of Intermediate Physics for Medicine and Biology, Russ Hobbie and I discuss one of Euler’s best known contributions, his relationship between the exponential function, trigonometric functions, and complex numbers.
The numbers that we have been using are called real numbers. The number i = √−1 is called an imaginary number. A combination of a real and imaginary number is called a complex number. The remarkable property of imaginary numbers that make them useful in this context is that e = cosθ + i sinθ.
Dunham wrote about this identity:
“From these equations,” Euler noted with evident satisfaction, “we understand how complex exponentials can be expressed by real sines and cosines.” His enthusiasm has been echoed by mathematicians ever since. Few would argue that Euler’s identity is among the most beautiful formulas of all.
Euler didn’t invent complex numbers, but he did contribute significantly to their development, including a derivation of this gem (“which seems extraordinary to me,” wrote Euler)

ii = e−π/2.

Dunham’s book gives examples of Euler’s contributions to number theory, logarithms, infinite series, analytic number theory, complex variables, algebra, geometry, and combinatorics. For instance, Dunham describes an discovery Euler made when in his 20s.
One of his earliest triumphs was a solution of the so-called “Basel Problem” that perplexed mathematicians for the better part of the previous century. The issue was to determine the exact value of the infinite series

1 + 1/4 + 1/9 + 1/16 + 1/25 + … + 1/k2 + … .

… The answer was not only a mathematical tour de force but a genuine surprise, for the series sums to π2/6. This highly non-intuitive result made the solution all the more spectacular and its solver all the more famous.
As he grew older, Euler slowly became blind. His accomplishments despite his handicap remind me of Beethoven composing his majestic 9th symphony after going deaf. Dunham writes about Euler
Although unable to see, he not only maintained but even increased his scientific output. In the year 1775, for instance, he wrote an average of one mathematical paper per week. Such productivity came in spite of the fact that he now had to have others read him the contents of scientific papers, and he in turn had to dictate his work to diligent scribes. During his descent into blindness, he wrote an influential textbook on algebra, a 775-page treatise on the motion of the moon, and a massive, three-volume development of integral calculus, the Institutiones calculi integralis. Never was his remarkable memory more useful than when he could see mathematics only in his mind’s eye.

That this blind and aging man forged ahead with such gusto is a remarkable lesson, a tale for the ages. Euler’s courage, determination, and utter unwillingness to be beaten serves, in the truest sense of the word, as an inspiration for mathematician and non-mathematician alike. The long history of mathematics provides no finer example of the triumph of the human spirit.
Dunham concludes
Euler left behind a legacy of epic proportions. So prolific was he that the journal of the St. Petersburg Academy was still publishing the backlog of his papers a full 48 years after his death. There is hardly a branch of mathematics—or for that matter of physics—in which he did not play a significant role.
 Listen to William Dunham talk about Leonhard Euler.
https://www.youtube.com/embed/fEWj93XjON0