Friday, May 14, 2021

A Bifurcation Diagram for the Heart

In Figure 10.26 of Intermediate Physics for Medicine and Biology, Russ Hobbie and I plot a bifurcation diagram for the logistic map: xj+1 = a xj (1 − xj).

A bifurcation diagram for the logistic map. Figure 10.26 in Intermediate Physics for Medicine and Biology.
A bifurcation diagram for the logistic map, showing 300 values of xj for values of a between 1 and 4. Figure 10.26 in Intermediate Physics for Medicine and Biology.

The bifurcation diagram summarizes the behavior of the map as a function of the parameter a. Some values of a correspond to a steady state, others represent period doubling, and still others lead to chaos.

When I teach Biological Physics, I don’t introduce chaos using the logistic map. Instead, I solve IPMB’s Homework Problem 41, about cardiac restitution and the onset of fibrillation.

Homework Problem 41 from Chapter 10 of Intermediate Physics for Medicine and Biology.

While Problem 41 provides insight into chaos and its relation to cardiac arrhythmias, Russ and I don’t draw a bifurcation diagram that summarizes how the action potential duration, APD, depends on the cycle length, CL (the time between stimuli, it’s the parameter analogous to a in the logistic map). In this post I present such a diagram.

A bifurcation diagram associated with Homework Problem 41 in Intermediate Physics for Medicine and Biology.
A bifurcation diagram associated with Homework Problem 41 in Intermediate Physics for Medicine and Biology. The plot shows 20 values of APDj for values of CL between 100 and 400 ms.

I don’t have the software to create a beautiful diagram like in Fig. 10.26, so I made one using MATLAB. It doesn’t have as much detail as does the diagram for the logistic map, but it’s still helpful.

The region marked 1:1 (for CL = 310 to 400 ms) implies steady-state behavior: Each stimulus excites an action potential with a fixed duration. Transients existed before the system settled down to a steady state, so I discarded the first 10,000 iterations before I plotted 20 values of APDj (j = 10,001 to 10,020). 

Between CL = 283 and 309 ms the system predicts alternans: the response to the stimulus alternates between two APDs (long, short, long, short, etc.). Sometimes this is called a 2:2 response. Alternans are occasionally seen in the heart, and are usually a sign of trouble.

From CL = 154 to 282 ms the response is 2:1, meaning that after a first stimulus excites an action potential the second stimulus occurs during the refractory period and therefore has no effect. The third stimulus excites another action potential with the same duration as the first (once the transients die away). This is a type of period doubling; the stimulus has period CL but the response has period 2CL. In cardiac electrophysiology, this behavior resembles second-degree heart block.

For a CL of 153 ms or shorter, the system is chaotic. I didn’t explore the diagram in enough detail to tell if self-similar regions of steady-state behavior exist within the chaotic region, as occurs for the logistic map (see Fig. 10.27 in IPMB).

A bifurcation diagram is a useful way to summarize the behavior of a nonlinear system, and provides insight into deadly heart arrhythmias such as ventricular fibrillation.

Friday, May 7, 2021

Servoanalysis of Carotid Sinus Reflex Effects on Peripheral Resistance

In Chapter 10 of Intermediate Physics for Medicine and Biology, Russ Hobbie and I discuss feedback and control. Homework Problem 12 analyzes the feedback circuit that controls blood pressure.

Problem 12 from Chapter 10 of Intermediate Physics for Medicine and Biology.
Problem 12 from Chapter 10 of Intermediate Physics for Medicine and Biology.

The reference to the article by Allen Scher and Allan Young is

Scher AM, Young AC (1963) “Servoanalysis of Carotid Sinus Reflex Effects on Peripheral Resistance,” Circulation Research, Volume 12, Pages 152–165.
I downloaded this paper to learn more about their experiment. Below are excerpts from their introduction.
The baroceptors of the carotid sinus (and artery) and the aortic arch are the major sense organs which reflexly control the systemic blood pressure. Since the demonstration of the reflex function of these receptors… there has been much work on the responses of the blood pressure, heart, and peripheral vessels to changes in pressure in the carotid arteries and the aorta…. In our study, we subjected the isolated perfused carotid sinus to maintained pressures at different levels… and measured the resultant systemic pressures and pressure changes.
Two variables were measured: the systemic pressure (Russ and I call this the arterial pressure, part, in the homework problem) and the pressure in the carotid sinus (psinus). Let’s consider them one at a time.

Below I have drawn a schematic diagram of the circulatory system, consisting of the pulmonary circulation (blood flow through the lungs, pumped by the right side of the heart) and the systemic circulation (blood flow to the various organs such as the liver, kidneys, and brain, pumped by the left side of the heart). Scher and Young measured the arterial pressure in the systemic circulation. Most of the pressure drop occurs in the arterioles, capillaries, and venules, so you can measure the arterial pressure in any large artery (such a the femoral artery in the leg) and it is nearly equal to the pressure produced by the left side of the heart. Arterial pressure is pulsatile, but Scher and Young used blood reservoirs to even out the variation in pressure throughout the cardiac cycle, providing a mean pressure. 

A schematic diagram of the circulatory system.
The circulatory system.

My second drawing shows the carotid sinus, a region near the base of the carotid artery (the artery that feeds the brain) that contains baroceptors (nowadays commonly called baroreceptors; pressure sensors that send information about the arterial pressure to the brain so it can maintain the proper blood pressure). Scher and Young isolated the carotid sinus. They didn’t remove it completely from the animal—after all, they still needed to supply blood to the brain to keep it alive—but it was effectively removed from the circulatory system. In the drawing below I show it as being separate from the body. However, the nerves connecting the baroreceptors to the brain remain intact, so changes in the carotid sinus pressure still signal the brain to do whatever’s necessary to adjust the systemic pressure.

The carotid sinus was attached to a feedback circuit, similar to the voltage clamp used by Hodgkin and Huxley to study the electrical behavior of a nerve axon (see Sec. 6.13 of IPMB). I drew the feedback circuit as an operational amplifier (the green triangle), but this is metaphor for the real instrument. An operational amplifier will produce whatever output is required to keep the two inputs equal. In an electrical circuit, the output would be current and the inputs would be voltage. In Scher and Young’s experiment, the output was flow and the inputs were pressure. Specifically, one of the inputs was the pressure measured in the carotid sinus, and the other was a user-specified constant pressure (po in the drawing). The feedback circuit set psinus = po, allowing the sinus pressure to be specified by the experimenter.
A schematic diagram to represent the feedback circuit that controlled the sinus pressure.
A schematic diagram to represent the feedback circuit that controlled the sinus pressure.
 
Once this elaborate instrumentation was perfected, the experiment itself was simple: Adjust psinus to whatever value you want by varying po, wait several seconds for the system to come to a new equilibrium (so psinus and part have adjusted to a new constant value), and then measure part. Scher and Young obtained a plot of part versus psinus, similar to that given in our homework problem.

As always, details affect the results.
  • Any contribution from pressure sensors in the aortic arch was eliminated by cutting the vagus nerve. Only baroreceptors in the carotid sinus contributed to controlling blood pressure.
  • Scher and Young performed experiments on both dogs and cats. The data in Homework Problem 12 is from a cat.
  • The blood reservoirs acted like capacitors in an electrical circuit, smoothing changes with time.
  • In some experiments, a dog was given a large enough dose of anesthetic that the nerves sending information from the sinus baroreceptors to the brain were blocked. In other experiments, the nerve from the baroreceptors to the brain was cut. In both cases, the change in part with psinus disappeared.
  • Many of Scher and Young’s experiments examined how the feedback circuit varied with time in response to either a step change or a sinusoidal variation in psinus. All of these experiments were ignored in the homework problem, which considers steady state. 
  • Often my students are confused by Problem 12. They think there is only one equation relating part and psinus, but to solve a feedback problem they need two. To resolve this conundrum, realize that when the carotid sinus is not isolated but instead is just one of many large arteries in the body, its pressure is simply the arterial pressure and the second equation is psinus = part.
  • The study provided hints about how the brain adjusted arterial pressure—by changing heart rate, stroke volume, or systemic resistance—but didn’t resolve this issue. 
  • The experiments were performed at the University of Washington School of Medicine in Seattle
  • Allen Scher was a World War II veteran, serving as a Marine in the Pacific. He contributed to our understanding of the electrocardiogram, and was a coauthor on the Textbook of Physiology, cited often as “Patton et al. 1989” in IPMB.
  • Scher was born on April 17, 1921 and died May 12, 2011 at the age of ninety. Recently we celebrated the the hundred-year anniversary of his birth. Happy birthday, Dr. Scher!

Friday, April 30, 2021

A Dozen Electrocardiograms Everyone Should Know

In Chapter 7 of Intermediate Physics for Medicine and Biology, Russ Hobbie and I discuss the electrocardiogram. Today, I present twelve ECGs that everyone should know. I’ve drawn them in a stylized and schematic way, ignoring differences between individuals, changes from beat to beat, and noise.

When I taught medical physics at Oakland University, my lecture on ECGs was preceded by a lesson on cardiac anatomy. If some anatomical terms in this post are unfamiliar, I suggest reviewing the Texas Heart Institute website.

1. Normal Heartbeat

The electrocardiogram (ECG).

Electrocardiograms are plotted on graph paper that consists of large squares, each divided into a five-by-five grid of small squares. The horizontal axis is time and the vertical axis is voltage. Each large square (or box) corresponds to a fifth of a second and a half of a millivolt.

The normal ECG contains three deflections: a small P wave associated with depolarization of the atria, a large narrow QRS complex associated with depolarization of the ventricles, and a T wave associated with repolarization of the ventricles.

A normal heart rate ranges from 60 to 100 beats per minute. The ECG below repeats every four boxes, so the time between beats is 0.8 seconds or 0.0133 minutes, which is equivalent to a heart rate of 75 beats per minute. 

An ECG of a normal heartbeat.
Normal heartbeat.

2. Sinus Bradycardia

A sinus bradycardia is a slow heart beat (slower than 60 beats per minute, or five boxes). The term sinus means the sinoatrial node is causing the slow rate. This node in the right atrium is the heart’s natural pacemaker. Other than its slow rate, the ECG looks normal. A sinus bradycardia may need to be treated by drugs or an implantable pacemaker, or it may represent the healthy heartbeat of a trained athlete who pumps a large amount of blood with each beat. 

An ECG of a sinus bradycardia.
Sinus bradycardia.

3. Sinus Tachycardia

A sinus tachycardia is the opposite of a sinus bradycardia. It’s a fast heart beat (faster than 100 beats per minute, or three boxes). The rapid rate arises because the sinoatrial node paces the heart too quickly. A sinus tachycardia may trigger other more severe arrhythmias we discuss later.

An ECG of a sinus tachycardia.
Sinus tachycardia.

4. Atrial Flutter

During atrial flutter a reentrant circuit exists in the atria. That is, a wave front propagates in a loop, constantly chasing its tail. The nearly sinusoidal, small-magnitude signal in the ECG originates in the atria. Every few rotations around the circuit, the wave front passes through the atrioventricular node—the only connection between the atria and the ventricles—giving rise to a normal-looking QRS complex (the T wave is buried in the atrial signal).

An ECG of atrial flutter.
Atrial flutter.

5. Atrial Fibrillation

Atrial fibrillation is similar to atrial flutter, except the wave fronts in the atrium are not as well organized, propagating in complicated, chaotic patterns resembling turbulence. The part of the ECG arising from the atria looks like noise. Occasionally the wave front passes through the atrioventricular node and produces a normal QRS complex. During atrial fibrillation the ventricles do not fill with blood effectively because the atria and ventricles are not properly synchronized.

An ECG of atrial fibrillation.
Atrial fibrillation.

6. First-Degree Atrioventricular Block

In first-degree atrioventricular block, the time between the end of the P wave and the start of the QRS complex is longer than it should be. Otherwise, the ECG appears normal. This rarely results in a problem for the patient, but does imply that the atrioventricular node is not healthy and trouble with it may develop in the future.

An ECG of first-degree atrioventricular block.
First-degree atrioventricular block.

7. Second-Degree Atrioventricular Block

In second-degree atrioventricular block, the P waves appear like clockwork. The signal often passes through the atrioventricular node to the ventricles, but occasionally it does not. Different types of second-degree block exist. Sometimes the delay between the P wave and the QRS complex gets progressively longer with each beat until propagation through the atrioventricular node fails. Other times, the node will periodically drop a beat; for example if every second beat fails you have 2:1 AV block. Still other times, the node fails sporadically. 

An ECG of second-degree atrioventricular block.
Second-degree atrioventricular block.

8. Third-Degree Atrioventricular Block

If your atrioventricular node stops working entirely, you have third-degree atrioventricular block (also known as complete heart block). A ventricular beat exists because some other part of the conduction system (say, the Bundle of His or the Purkinje fibers) serves as the pacemaker. In this case, the P wave and QRS complex are unsynchronized, like two metronomes set to different rates. In complete heart block, the ventricles typically fire at a slow rate. Sometimes a patient will intermittently go in and out of third-degree block, causing fainting spells.

An ECG of third-degree atrioventricular block.
Third-degree atrioventricular block.

9. Premature Ventricular Contraction

Sometimes the heart will beat with a normal ECG, and than sporadically have an extra ventricular beat: a premature ventricular contraction. Usually these beats originate within the ventricle, so they are not distributed through the cardiac conduction system and therefore produce a wide, bizzare QRS complex. As long as these premature contractions are rare they are not too dangerous, but they risk triggering more severe ventricular arrhythmias.

An ECG of premature ventricular contraction.
Premature ventricular contraction.

10. Ventricular Tachycardia

A ventricular tachycardia is a rapid heartbeat arising from a reentrant circuit in the ventricles. The ECG looks like a series of premature ventricular contractions following one after another. The VT signal typically has a large amplitude, and the atrial signal is often too small to be seen. This is a serious arrhythmia, because at such a fast rate the heart doesn’t pump blood effectively. It’s not lethal itself, but can become deadly if it decays into ventricular fibrillation.

An ECG of ventricular tachycardia.
Ventricular tachycardia.

11. Ventricular Fibrillation

In ventricular fibrillation, different parts of the ventricles contract out of sync, resulting in the heart quivering rather than beating. A heart in VF does not pump blood, and the patient will die in ten to fifteen minutes unless defibrillated by a strong electric shock. Ventricular fibrillation is the most common cause of sudden cardiac death.

An ECG of ventricular fibrillation.
Ventricular fibrillation.

12. Asystole

In asystole, the heart has no electrical activity, so the ECG is a flat line. Asystole is the end stage of ventricular fibrillation, when the chaotic electrical activity dies away and nothing remains.

An ECG of asystole.
Asystole.


Once you master these twelve ECGs, you’ll be on your way to understanding the electrical behavior of the heart. If you want to learn more, I suggest trying the SkillStat six-second ECG game, which includes 27 ECGs. It’s fun.

Friday, April 23, 2021

Electric and Magnetic Fields From Two-Dimensional Anisotropic Bisyncytia


Page 223 of Intermediate Physics for Medicine and Biology.

Figure 8.18 on page 223 of Intermediate Physics for Medicine and Biology contains a plot of the magnetic field produced by action currents in a slice of cardiac tissue. The measured magnetic field contours have approximately a four-fold symmetry. The experiment by Staton et al. that produced this data was a tour de force, demonstrating the power of high-spatial-resolution biomagnetic techniques. 
 
Sepulveda NG, Wikswo JP (1987) “Electric and Magnetic Fields From Two-Dimensional Anisotropic Bisyncytia,” Biophysical Journal, Volume 51, Pages 557-568, superimposed on Intermediate Physics for Medicine and Biology.
Sepulveda and Wikswo (1987).
In this post, I discuss the theoretical prediction by Nestor Sepulveda and John Wikswo in the mid 1980s that preceded and motivated the experiment.
Sepulveda NG, Wikswo JP (1987) “Electric and Magnetic Fields From Two-Dimensional Anisotropic Bisyncytia,” Biophysical Journal, Volume 51, Pages 557-568.
Their abstract is presented below.
Cardiac tissue can be considered macroscopically as a bidomain, anisotropic conductor in which simple depolarization wavefronts produce complex current distributions. Since such distributions may be difficult to measure using electrical techniques, we have developed a mathematical model to determine the feasibility of magnetic localization of these currents. By applying the finite element method to an idealized two-dimensional bisyncytium [a synonym for bidomain] with anisotropic conductivities, we have calculated the intracellular and extracellular potentials, the current distributions, and the magnetic fields for a circular depolarization wavefront. The calculated magnetic field 1 mm from the tissue is well within the sensitivity of a SQUID magnetometer. Our results show that complex bisyncytial current patterns can be studied magnetically, and these studies should provide valuable insight regarding the electrical anisotropy of cardiac tissue.
Sepulveda and Wikswo assumed the tissue was excited by a brief stimulus through an electrode (purple dot in the illustration below), resulting in a circular wave front propagating outward. The transmembrane potential coutours at one instant are shown in red. The assumption of a circular wave front is odd, because cardiac tissue is anisotropic. A better assumption would have been an elliptical wave front with its long axis parallel to the fibers. Nevertheless, the circular wave front captures the essential features of the problem.

If the tissue were isotropic, the intracellular current density would point radially outward and the extracellular current density would point radially inward. The intracellular and extracellular currents would exactly cancel, so the net current (their sum) would be zero. Moreover, the net current would vanish if the tissue were anisotropic but had equal anisotropy ratios. That is, if the ratios of the electrical conducivities parallel and perpendicular to the fibers were the same in the intracellular and extracellular spaces. The only way to produce a net current (shown as the blue loops in the illustration below) is if the tissue has unequal anisotropy ratios. In that case, the loops are four-fold symmetric, rotating clockwise in two quadrants and counterclockwise in the other two.

Current loops produce magnetic fields. The right-hand-rule implies that the magnetic field points up out of the plane in the top-right and the bottom-left quadrants, and down into the plane in the other two. The contours of magnetic field are green in the illustration below, and the peak magnitude for a 1 mm thick sheet of is about one fourth of a nanotesla.

Jut for fun, I superimposed the transmembrane potential, net current density, and magnetic field plots in the picture below.

Notes:
  1. The measurement of the magnetic field is a null detector of unequal anisotropy ratios. In other words, in tissue with equal anisotropy ratios the magnetic field vanishes, so the mere existence of a magnetic field implies the anisotropy ratios are unequal. The condition of unequal anisotropy ratios has many implications for cardiac tissue. One is discussed in Homework Problem 50 in Chapter 7 of IPMB
  2. If the sheet of cardiac tissue is superfused by a saline bath, the magnetic field distribution changes.
  3. Wikswo was a pioneer in the field of biomagnetism. In particular, he developed small scanning magnetometers that had sub-millimeter spatial resolution. He was in a unique position of being able to measure the magnetic fields that he and Sepulveda predicted, which led to the figure included in IPMB
  4. I was a graduate student in Wikswo’s laboratory when Sepulveda and Wikswo wrote this article. Sepulveda, a delightful Columbian biomedical engineer and a good friend of mine, worked as a research scientist in Wikswo’s lab. He was an expert on the finite element method—the numerical technique used in his paper with Wikswo—and had written his own finite element code that no one else in the lab understood. He died a decade ago, and I miss him. 
  5. Sepulveda and Wikswo were building on a calculation published in 1984 by Robert Plonsey and Roger Barr (“Current Flow Patterns in Two Dimensional Anisotropic Bisyncytia With Normal and Extreme Conductivities,” Biophys. J., 45:557-571). Wikswo heard either Plonsey or Barr give a talk about their results at a scientific meeting. He realized immediately that their predicted current loops implied a biomagnetic field. When Wikswo returned to the lab, he described Plonsey and Barr’s current loops at a group meeting. As I listened, I remember thinking “Wikswo’s gone mad,” but he was right.
  6. Two years after their magnetic field article, Sepulveda and Wikswo (now with me included as a coauthor) calculated the transmembrane potential produced when cardiac tissue is stimulated by a point electrode. But that’s another story.
I’ll give Sepulveda and Wikswo the last word. Below is the concluding paragraph of their article, which looks forward to the experimental measurement of the magnetic field pattern that was shown in IPMB.
The bidomain model of cardiac tissue provides a tool that can be explored and used to study and explain features of cardiac conduction. However, it should be remembered that “a model is valid when it measures what it is intended to measure” (31). Thus, experimental data must be used to evaluate the validity of the bidomain model. This evaluation must involve comparison of the model's predictions not only with measured intracellular and extracellular potentials but also with the measured magnetic fields. When the applicability of the bidomain model to a particular cardiac preparation and the validity and reliability of our calculations have been determined experimentally, this mathematical approach should then provide a new technique for analyzing normal and pathological cardiac activation.
Members of John Wikswo's laboratory at Vanderbilt University in the mid 1980s.
Members of Wikswo's lab at Vanderbilt University in the mid 1980s: John Wikswo is on the phone, Nestor Sepulveda is in the white shirt and gray pants, and I am on the far right. The other people are Frans Gielen (the tall guy with arms crossed on the left), Ranjith Wijesinghe (between Gielen and Sepulveda), Peng Zhang (between Wikswo and Sepulveda), and Pat Henry (knelling).

Friday, April 16, 2021

The Code Breaker: Jennifer Doudna, Gene Editing, and the Future of the Human Race

The Code Breaker, by Walter Isaacson, superimposed on Intermediate Physics for Medicine and Biology.
The Code Breaker,
by Walter Isaacson

My favorite authors are Simon Winchester, David Quammen, and Walter Isaacson. This week I read Isaacson’s latest book: The Code Breaker: Jennifer Doudna, Gene Editing, and the Future of the Human Race. I would place it alongside The Eighth Day of Creation and The Making of the Atomic Bomb as one of the best books about the history of science

In his introduction, Isaacson writes

The invention of CRISPR and the plague of COVID will hasten our transition to the third great revolution of modern times. These revolutions arose from the discovery, beginning just over a century ago, of the three fundamental kernels of our existence: the atom, the bit, and the gene.

The first half of the twentieth century, beginning with Albert Einstein’s 1905 papers on relativity and quantum theory, featured a revolution driven by physics. In the five decades following his miracle year, his theories led to atom bombs and nuclear power, transistors and spaceships, lasers and radar.

The second half of the twentieth century was an information-technology era, based on the idea that all information could be encoded by binary digits—known as bits—and all logical processes could be performed by circuits with on-off switches. In the 1950s, this led to the development of the microchip, the computer, and the internet. When these three innovations were combined, the digital revolution was born.

Now we have entered a third and even more momentous era, a life-science revolution. Children who study digital coding will be joined by those who study genetic code.
Early in the book, Isaacson describes Francisco Mojica’s discovery that bacteria have “CRISPR spacer sequences”: strands of DNA that serve as an immune system protecting them from viruses.
As we humans struggle to fight off novel strains of viruses, it’s useful to note that bacteria have been doing this for about three billion years, give or take a few million centuries. Almost from the beginning of life on this planet, there’s been an intense arms race between bacteria, which developed elaborate methods of defending against viruses, and the ever-evolving viruses, which sought ways to thwart those defenses.

Mojica found that bacteria with CRISPR space sequences seems to be immune from infection by a virus that had the same sequence. But bacteria without the spacer did get infected. It was a pretty ingenious defense system, but there was something even cooler: it appeared to adapt to new threats. When new viruses came along, the bacteria that survived were able to incorporate some of that virus’s DNA and thus create, in its progeny, an acquired immunity to that new virus. Mojica recalls being so overcome by emotion at this realization that he got tears in his eyes. The beauty of nature can sometimes do that to you.

The Code Breaker focuses on the life and work of Jennifer Doudna, who won the 2020 Nobel Prize in Chemistry. However, the star of the book is not Doudna, nor Emmanuelle Charpentier (who shared the prize with Doudna), nor Mojica, nor any of the other scientific heroes. The star is RNA, the molecule that carries genetic information from DNA in the nucleus to the cytoplasm where proteins are produced.

By 2008, scientists had discovered a handful of enzymes produced by genes that are adjacent to the CRISPR sequences in a bacteria’s DNA. These CRISPR-associated (Cas) enzymes enable the system to cut and paste new memories of viruses that attack the bacteria. They also create short segments of RNA, known as CRISPR RNA (crRNA), that can guide a scissors-like enzyme to a dangerous virus and cut up its genetic material. Presto! That’s how the wily bacteria create an adaptive immune system!
Doudna and Charpentier’s Nobel Prize resulted from their developing the CRISPR-Cas9 system into a powerful technique for gene editing.
The study of CRISPR would become a vivid example of the call-and-response duet between basic science and translational medicine. At the beginning it was driven by the pure curiosity of microbe-hunters who wanted to explain an oddity they had stumbled upon when sequencing the DNA of offbeat bacteria. Then it was studied in an effort to protect the bacteria in yogurt cultures from attacking viruses. That led to a basic discovery about the fundamental workings of biology. Now a biochemical analysis was pointing the way to the invention of a tool with potential practical uses. “Once we figured out the components of the CRISPR-Cas9 assembly, we realized that we could program it on our own,” Doudna says. “In other words, we could add a different crRNA and get it to cut any different DNA sequence we chose.”

Several other themes appear throughout The Code Breaker

  • The role of competition and collaboration in science, 
  • How industry partnerships and intellectual property affect scientific discovery, 
  • The ethics of gene editing, and
  • The epic scientific response to the COVID-19 pandemic

I’m amazed that Isaacson’s book is so up-to-date. I received my second dose of the Pfizer-BioNTech vaccine last Saturday and then read The Code Breaker in a three-day marathon. My arm was still sore while reading the chapter near the end of the book about RNA Covid vaccines like Pfizer’s.

There’s a lot of biology and medicine in The Code Breaker, but not much physics. Yet, some of the topics discussed in Intermediate Physics for Medicine and Biology appear briefly. Doudna uses x-ray diffraction to decipher the structure of RNA. Electroporation helps get vaccines and drugs into cells. Electrophoresis, microfluidics, and electron microscopy are mentioned. I wonder if injecting more physics and math into this field would supercharge its progress. 

CRISPR isn’t the first gene-editing tool, but it increases the precision of the technique. As Winchester noted in The Perfectionists, precision is a hallmark of technology in the modern world. Quammen’s book Spillover suggests that humanity may be doomed by an endless flood of viral pandemics, but The Code Breaker offers hope that science will provide the tools needed to prevail over the viruses.

I will close with my favorite passage from The Code Breaker: Isaacson’s paean to curiosity-driven scientific research.

The invention of easily reprogrammable RNA vaccines was a lightning-fast triumph of human ingenuity, but it was based on decades of curiosity-driven research into one of the most fundamental aspects of life on planet earth: how genes encoded by DNA are transcribed into snippets of RNA that tell cells what proteins to assemble. Likewise, CRISPR gene-editing technology came from understanding the way that bacteria use snippets of RNA to guide enzymes to chop up dangerous viruses. Great inventions come from understanding basic science. Nature is beautiful that way.

 

.
“How CRISPR lets us edit our DNA,” a TED talk by Jennifer Doudna. 

Nobel Lecture, Jennifer Doudna, 2020 Nobel Prize in Chemistry.

Friday, April 9, 2021

The Vitamin D Questions: How Much Do You Need and How Should You Get It?

A bottle of vitamin D capsules, on Intermediate Physics for Medicine and Biology.
Vitamin D
Two years ago my doctor recommended I start taking vitamin D. I’m annoyed at having to take a supplement every day, but I do what the doc says.

Your body needs exposure to ultraviolet light to produce its own vitamin D, but too much UV light causes skin cancer. Russ Hobbie and I address this trade-off in Section 14.10 of Intermediate Physics for Medicine and Biology.

There has been an alarming increase in the use of tanning parlors by teenagers and young adults. These emit primarily UVA, which can cause melanoma. Exposure rates are two to three times greater than solar radiation at the equator at noon (Schmidt 2012). Many states now prohibit minors from using tanning parlors. Proponents of tanning parlors point out that UVB promotes the synthesis of vitamin D; however, the exposure to UVB in a tanning parlor is much higher than needed by the body for vitamin D production. Tanning as a source of Vitamin D is no longer recommended at any age level (Barysch et al. 2010).
To learn more about how much vitamin D people need, I read an article by Deon Wolpowitz and Barbara Gilchrest.
Wolpowitz D, Gilchrest BA (2006) “The Vitamin D Questions: How Much Do You Need and How Should You Get It?Journal of the American Academy of Dermatology, Volume 54, Pages 301–317.

Below are excerpts from their conclusion.

Given the scarcity of naturally occurring vit [vitamin] D in many otherwise adequate diets, human beings may once have depended on unprotected exposure to natural sunlight as the primary environmental source of vit D, at least during those periods of the year when sunlight can produce [previtamin] D3 in the skin... However, chronic unprotected exposure to carcinogenic UV radiation in sunlight not only results in photoaging, but also greatly increases the risk of skin cancer. This risk is further exacerbated by the extended lifespan of human beings in the 21st century. Fortunately, there is a noncarcinogenic alternative—intestinal absorption of vit D-fortified foods and/or dietary supplements…

All available evidence indicates that younger, lighter-skinned individuals easily maintain desirable serum 25-OH vit D levels year-round by incidental protected sun exposure and customary diet. Daily intake of two 8-oz glasses of fortified milk or orange juice or one standard vit or incidental protected exposure of the face and backs of hands to 0.25% minimum erythema [reddening of the skin] dose of UVB radiation 3 times weekly each generates adequate serum 25-OH levels by classic criteria. Dietary supplementation of vit D is efficient, efficacious, and safe. Thus, it is prudent for those at high statistical risk for vit D deficiency, such as patients who are highly protected against the sun [or old folks like me], to take daily supplemental vit D (200-1000 IU) with concurrent dietary calcium to meet current and future RDA [recommended daily allowance] levels.
So, maybe my physician was right when she put me on vitamin D. But I decided to double my dose in the winter. Is that a good idea? A recent paper out of South Korea has some interesting data.
Park SS, Lee YG, Kim M, Kim J, Koo J-H, Kim CK, Um J, Yoon J (2019) “Simulation of Threshold UV Exposure Time for Vitamin D Synthesis in South Korea,” Advances in Meteorology, Volume 2019, Article 4328151.
Sang Seo Park and his colleagues analyze UV exposure, and estimate the threshold exposure time at noon to sunlight in Seoul for vitamin D synthesis (blue) and erythema (red). 

Threshold exposure time to sunlight in Seoul for vitamin D synthesis (blue) and erythema (red). from Park et al. (2019) Simulation of Threshold UV Exposure Time for Vitamin D Synthesis in South Korea. Adv. Meteorol., 2019:4328151.
Threshold exposure time to sunlight in Seoul for vitamin D synthesis (blue) and erythema (red). From Park et al. (2019).

This is a semilog plot, so the difference between July and January is about a factor of ten. I think increasing my vitamin D dose in the winter is, if anything, conservative. I could probably do without the supplement in the summer. Rochester (42° latitude) is farther north than Seoul (38°), so the seasonal effect might be even greater where I live.

In conclusion, I think Russ and I are correct to stress the harmful effects of UV light in IPMB. If you’re worried about not getting enough vitamin D, take a supplement in the winter.

Friday, April 2, 2021

The Boltzmann Distribution Applied to a Harmonic Oscillator

In Chapter 3 of Intermediate Physics for Medicine and Biology, Russ Hobbie and I discuss the Boltzmann distribution. If you have a system with energy levels of energy En populated by particles in thermal equilibrium, the Boltzmann distribution gives the probability of finding a particle in the nth level.

A classic example of the Boltzmann distribution is for the energy levels of a harmonic oscillator. These levels are equally spaced starting from a ground state and increasing without bound. To see the power of the Boltzmann distribution, solve this new homework problem.
Section 3.7

Problem 29½. Suppose the energy levels, En, of a system are given by

En = n ε,     for  n = 0, 1, 2, 3, …

where ε is the energy difference between adjacent levels. Assume the probability of a particle occupying the nth energy level, Pn, obeys the Boltzmann distribution
where A is a constant, T is the absolute temperature, and k is the Boltzmann constant.

(a) Determine A in terms of ε, k, and T. (Hint: the sum of the probabilities over all levels is one.)
(b) Find the average energy E of the particles. (Hint: E = ∑PnEn.)
(c) Calculate the heat capacity C of a system of N such particles. (Hint: U = N E and C = dU/dT.)
(d) What is the limiting value of C for high temperatures (kT >> ε)? (Hint: use the Taylor series of the exponential.)
(e) What is the limiting value of C for low temperatures (kT << ε)?
(f) Sketch a plot of C versus T.
You may need these infinite series

1 + x + x2 + x3 + ⋯ = 1/(1x) 
x + 2x2 + 3x3 + ⋯ = x/(1x)2 

This is a somewhat advanced problem in statistical mechanics, so I gave several hints to guide the reader. The calculation contains much interesting physics. For instance, the answer to part (e) is known as the third law of thermodynamics. Albert Einstein was the first to calculate the heat capacity of a collection of harmonic oscillators (a good model for a crystalline solid). Theres more physics than biology in this problem, because most of the interesting behavior occurs at cold temperatures but biology operates at hot temperatures.

If you’re having difficulty solving this problem, here
s one more hint:

enx = (ex)n

Enjoy!

Friday, March 26, 2021

Cooling by Radiation

In Chapter 14 of Intermediate Physics for Medicine and Biology, Russ Hobbie and I discuss thermal radiation. If you’re a black body, the net power you radiate, wtot, is given by Eq. 14.41

wtot = S σSB (T4Ts4) ,                (14.41)

where S is the surface area, σSB is the Stefan-Boltzmann constant (5.67 × 10−8 W m−2 K−4), T is the absolute temperature of your body (about 310 K), and Ts is the temperature of your surroundings. The T4 term is the radiation you emit, and the Ts4 term is the radiation you absorb.

The fourth power that appears in this expression is annoying. It means we must use absolute temperature in kelvins (K); you get the wrong answer if you use temperature in degrees Celsius (°C). It also means the expression is nonlinear; wtot is not proportional to the temperature difference TTs.

On the absolute temperature scale, the difference between the temperature of your body (310 K) and the temperature of your surroundings (say, 293 K at 20 °C) is only about 5%. In this case, we simplify the expression for wtot by linearizing it. To see what I mean, try Homework Problem 14.32 in IPMB.
Section 14.9 
Problem 32. Show that an approximation to Eq. 14.41 for small temperature differences is wtot = S Krad (TTs). Deduce the value of Krad at body temperature. Hint: Factor T4Ts4 =  (TTs)(…). You should get Krad = 6.76 W m−2 K−1.
The constant Krad has the same units as a convection coefficient (see Homework Problem 51 in Chapter 3 of IPMB). Think of it as an effective convection coefficient for radiative heat loss. Once you determine Krad, you can use either the kelvin or Celsius temperature scales for TTs, so you can write its units as W m−2 °C−1.
 
Air and Water, by Mark Denny, superimposed on Intermediate Physics for Medicine and Biology.
Air and Water,
by Mark Denny.
In Air and Water, Mark Denny analyzes the convection coefficient. In a stagnant fluid, the convection coefficient depends only on the fluid’s thermal conductivity and the body’s size. For a sphere, it is inversely proportional to the diameter, meaning that small bodies are more effective at convective cooling per unit surface area than large bodies. If the body undergoes free convection or forced convection (for both cases the surrounding fluid is moving), the expression for the convection coefficient is more complicated, and depends on factors such as the Reynolds number and Prandtl number of the fluid flow. Denny gives values for the convection coefficient as a function of body size for both air and water. Usually, these values are greater than the 6.76 W m−2 °C−1 for radiation. However, for large bodies in air, radiation can compete with convection as the dominate mechanism. For people, radiation is an important mechanism for cooling. For a dolphin or mouse, it isn’t. Elephants probably make good use of radiative cooling.
 
Finally, our analysis implies that when the difference between the temperatures of the body and the surroundings is small, a body whose primary mechanism for getting rid of heat is radiation will cool exponentially following Newton’s law of cooling.

Friday, March 19, 2021

The Carr-Purcell-Meiboom-Gill Pulse Sequence

The most exciting phrase to hear in science, the one that heralds new discoveries, is not “Eureka!” but “That’s funny...” 

Isaac Asimov

In Section 18.8 of Intermediate Physics for Medicine and Biology, Russ Hobbie and I describe the Carr-Purcell pulse sequence, used in magnetic resonance imaging.

When a sequence of Ï€ [180° radio-frequency] pulses that nutate M [the magnetization vector] about the x' axis are applied at TE/2, 3TE/2, 5TE/2, etc., a sequence of echoes are formed [in the Mx signal], the amplitudes of which decay with relaxation time T2. This is shown in Fig. 18.19.
The Carr-Purcell pulse sequence, as shown in Fig. 18.19 of Intermediate Physics for Medicine and Biology.
Fig. 18.19  The Carr-Purcell pulse sequence.
All π pulses nutate about the x' axis.
The envelope of echoes decays as et/T2.

Russ and I then discuss the Carr-Purcell-Meiboom-Gill pulse sequence.
One disadvantage of the CP [Carr-Purcell] sequence is that the Ï€ pulse must be very accurate or a cumulative error builds up in the successive pulses. The Carr-Purcell-Meiboom-Gill sequence overcomes this problem. The initial Ï€/2 [90° radio-frequency] pulse nutates M about the x' axis as before, but the subsequent [Ï€] pulses are shifted a quarter cycle in time, which causes them to rotate about the y' axis.
 
The Carr-Purcell-Meiboom-Gill pulse sequence, as shown in Fig. 18.21 of Intermediate Physics for Medicine and Biology.
Fig. 18.21  The Carr-Purcell-Meiboom-Gill pulse sequence.

The first page of Meiboom, S. and Gill, D. (1958) “Modified Spin-Echo Method for Measuring Nuclear Relaxation Times.” Rev. Sci. Instr. 29:688–691, superimposed on Intermediate Physics for Medicine and Biology.
Meiboom, S. and Gill, D. (1958)
“Modified Spin-Echo Method for
Measuring Nuclear Relaxation Times.”
Rev. Sci. Instr.
29:688–691.
Students might enjoy reading the abstract of Saul Meiboom and David Gill’s 1958 article published in the Review of Scientific Instruments (Volume 29, Pages 688-691).
A spin echo method adapted to the measurement of long nuclear relaxation times (T2) in liquids is described. The pulse sequence is identical to the one proposed by Carr and Purcell, but the rf [radio-frequency] of the successive pulses is coherent, and a phase shift of 90° is introduced in the first pulse. Very long T2 values can be measured without appreciable effect of diffusion.
This short paper is so highly cited that it was featured in a 1980 Citation Classic commentary, in which Meiboom reflected on the significance of the research.
The work leading to this paper was done nearly 25 years ago at the Weizmann Institute of Science, Rehovot, Israel. David Gill, who was then a graduate student… , set out to measure NMR T2-relaxation times in liquids, using the well-known Carr-Purcell pulse train scheme. He soon found that at high pulse repetition rates adjustments became very critical, and echo decays, which ideally should be exponential, often exhibited beats and other irregularities. But he also saw that on rare and unpredictable occasions a beautiful exponential decay was observed... Somehow the recognition emerged that the chance occurrence of a 90° phase shift of the nuclear polarization [magnetization] must underlie the observations. It became clear that in the presence of such a shift a stable, self-correcting state of the nuclear polarization is produced, while the original scheme results in an unstable state, for which deviations are cumulative. From here it was an easy step to the introduction of an intentional phase shift in the applied pulse train, and the consistent production of good decays.
The key point is that the delay between the initial Ï€/2 pulse (to flip the spins into the xy plane) and the string of Ï€ pulses (to create the echoes) must be timed carefully (the pulses must be coherent). Even adding a delay corresponding to a quarter of a single oscillation changes everything. In a two-tesla MRI scanner, the Larmor frequency is 83 MHz, so one period is 12 nanoseconds. Therefore, if the timing is off by just a few nanoseconds, the method won’t work.

Initially Gill didn’t worry about timing the pulses precisely, so usually he was using the error-prone Carr-Purcell sequence. Occasionally he got lucky and the timing was just right; he was using what’s now called the Carr-Purcell-Meiboom-Gill sequence. Meiboom and Gill “somehow” were able to deduce what was happening and fix the problem. Meiboom believes their paper is cited so often because it was the first to recognize the importance of maintaining phase relations between the different pulses in an MRI pulse sequence.

In his commentary, Meiboom notes that
Although in hindsight the 90° phase shift seems the logical and almost obvious thing to do, its introduction was triggered by a chance observation, rather than by clever a priori reasoning. I suspect (though I have no proof) that this applies to many scientific developments, even if the actual birth process of a new idea is seldom described in the introduction to the relevant paper.
If you’re a grad student working on a difficult experiment that’s behaving oddly, don’t be discouraged if you hear yourself saying “that’s funny...” A discovery might be sitting right in front of you!