Medical students turn to peer-support groups for assistance: A Q&A

Photo by rawpixel

School can be overwhelming, especially medical school. But Stanford Medicine offers many different forms of mental health support, including a peer-to-peer support program for medical students called Ears 4 Peers. To learn more, I spoke with Dina Wang-Kraus, MD, a Stanford psychiatry and behavioral sciences resident and co-founder of the program.

What inspired you to start the Ears 4 Peers program?

“In 2012, I was a first-year medical student and I was noticing that a significant number of my classmates were experiencing compassion fatigue and burnout. We were encouraged to reach out to the counseling and psychology services but there was some hesitancy, either from busy schedules or anxieties surrounding stigma. So, Norma Villalon, MD, and I decided to found a peer-to-peer support program. I started a similar program in college at Johns Hopkins, called A Place to Talk.

The hope was to have near-peers — those who were just walking in your shoes — provide support. Our goal was to bridge the distance students often feel when in a competitive, challenging situation. We may have been adults in our mid-twenties to forties, but we were only in the infancy of our training.

Rebecca Smith-Coggins, MD, is our faculty adviser and leader. From day one, she’s believed in our cause.”

What are some issues the program addresses?

“We receive calls regarding issues like academic stress, interpersonal relationship conflicts, imposter syndrome, intimate partner violence, Stanford Duck syndrome and suicidal thoughts. We also receive calls from students feeling lonely, disconnected and homesick, especially around finals, holidays and medical board exams. And some students call hoping to be referred for additional support.”

How are Ears 4 Peers mentors selected and trained?

“Ears 4 Peers mentors are nominated by their peers or self-nominated. They complete an application to tell us more about themselves, what draws them to this type of work and what they hope to gain from the experience.

We’re very lucky to have the support of Alejandro Martinez, PhD, the Associate Dean of Students for the Stanford undergraduate campus. He and his team designed a curriculum specifically for Stanford School of Medicine.”

What role do you play in the program now?

“As a resident, I’ve transitioned out of being an official Ears 4 Peers mentor but I continue to remain actively involved in near-peer mentoring for medical students. Two years ago as an intern in psychiatry, I worked with Jessi Gold, MD, to inaugurate Stanford’s  Medical Student Reflection Groups. Each group is made up of four to 10 medical students who commit to joining for six to 12 months. We meet every other week, and groups are facilitated by psychiatry residents trained in group therapy and psychotherapy. As resident physicians, we remain near-peers; however, we’re able to facilitate a different kind of support and personal growth given our psychiatry training.

Stanford students are welcome to reach out to me at sdwangkraus@stanford.edu to learn more.”

What advise can you give medical students and residents?

“I recall medical school to be an exhilarating time, but it also felt like I was drinking from Niagara Falls, one cup at a time. There were times when I felt overwhelmed and even burnt out.

We see a lot of beauty and humility in medicine, but there are also times when we see a lot of tragedy and suffering. Having peer-support, knowing that I was not alone, was empowering and liberating — and it continues to be.”

This is a reposting of my Scope blog story, courtesy of Stanford School of Medicine.

Shedding New Light on Luminous Blue Variable Stars

3D Simulations Disperse Some of the Mystery Surrounding Massive Stars

ResizedImage449270-LVB
A snapshot from a simulation of the churning gas that blankets a star 80 times the sun’s mass. Intense light from the star’s core pushes against helium-rich pockets in the star’s exterior, launching material outward in spectacular geyser-like eruptions. The solid colors denote radiation intensity, with bluer colors representing regions of larger intensity. The translucent purplish colors represent the gas density, with lighter colors denoting denser regions. Image: Joseph Insley, Argonne Leadership Computing Facility.

Three-dimensional (3D) simulations run at two of the U.S. Department of Energy’s national laboratory supercomputing facilities and the National Aeronautics and Space Administration (NASA) have provided new insights into the behavior of a unique class of celestial bodies known as luminous blue variables (LBVs) — rare, massive stars that can shine up to a million times brighter than the Sun.

Astrophysicists are intrigued by LBVs because their luminosity and size dramatically fluctuate on a timescale of months. They also periodically undergo giant eruptions, violently ejecting gaseous material into space. Although scientists have long observed the variability of LBVs, the physical processes causing their behavior are still largely unknown. According to Yan-Fei Jiang, an astrophysicist at UC Santa Barbara’s Kavli Institute for Theoretical Physics, the traditional one-dimensional (1D) models of star structure are inadequate for LBVs.

“This special class of massive stars cycles between two phases: a quiescent phase when they’re not doing anything interesting, and an outburst phase when they suddenly become bigger and brighter and then eject their outer envelope,” said Jiang. “People have been seeing this for years, but 1D, spherically-symmetric models can’t determine what is going on in this complex situation.”

Instead, Jiang is leading an effort to run first-principles, 3D simulations to understand the physics behind LBV outbursts — using large-scale computing facilities provided by Lawrence Berkeley National Laboratory’s National Energy Research Scientific Computing Center (NERSC), the Argonne Leadership Computing Facility (ALCF), and NASA. NERSC and ALCF are DOE Office of Science User Facilities.

Physics Revealed by 3D

In a study published in Nature, Jiang and his colleagues from UC Santa Barbara, UC Berkeley, and Princeton University ran three 3D simulations to study three different LBV configurations. All the simulations included convection, the action when a warm gas or liquid rises while its cold counterpart sinks. For instance, convection causes hot water at the bottom of a pot on a stove to rise up to the top surface. It also causes gas from a star’s hot core to push to its outer layers.

During the outburst phase, the new 3D simulations predict that convection causes a massive star’s radius to irregularly oscillate and its brightness to vary by 10 to 30 percent on a timescale of just a few days — in agreement with current observations.

“Convection causes the star to expand significantly to a much larger size than predicted by our 1D model without convection. As the star expands, its outer layers become cooler and more opaque,” Jiang said.

Opacity describes how a gas interacts with photons. The researchers discovered that the helium opacity in the star’s outer envelope doubles during the outburst phase, making it more difficult for photons to escape. This leads the star to reach an effective temperature of about 9,000 degrees Kelvin (16,000 degrees Fahrenheit) and triggers the ejection of mass.

“The radiation force is basically a product of the opacity and the fixed luminosity coming from the star’s core. When the helium opacity doubles, this increases the radiation force that is pushing material out until it overcomes the gravitational force that is pulling the material in,” said Jiang. “The star then generates a strong stellar wind, blowing away its outer envelope.”

Massive Simulations Required

Massive stars require massive and expensive 3D simulations, according to Jiang. So he and his colleagues needed all the computing resources available to them, including about 15 million CPU hours at NERSC, 60 million CPU hours at ALCF, and 10 million CPU hours at NASA. In addition, NERSC played a special role in the project.

“The Cori supercomputer at NERSC was essential to us in the beginning because it is very flexible,” Jiang said. “We did all of the earlier exploration at NERSC, figuring out the right parameters to use and submissions to do. We also got a lot of support from the NERSC team to speed up our input/output and solve problems.”

In addition to spending about 5 million CPU hours at NERSC on the early phase of the project, Jiang’s team used another 10 million CPU hours running part of the 3D simulations.

“We used NERSC to run half of one of the 3D simulations described in the Nature paper and the other half was run at NASA. Our other two simulations were run at Argonne, which has very different machines,” said Jiang. “These are quite expensive simulations, because even half a run takes a lot of time.”

Even so, Jiang believes that 3D simulations are worth the expense because illuminating the fundamental processes behind LBV outbursts is critical to many areas of astrophysics — including understanding the evolution of these massive stars that become black holes when they die, as well as understanding how their stellar winds and supernova explosions affect galaxies.

Jiang also used NERSC for earlier studies, and his collaboration is already running follow-up 3D simulations based on their latest results. These new simulations incorporate additional parameters — including the LBV star’s rotation and metallicity — varying the value of one of these parameters per run. For example, the speed from rotation is larger at the star’s equator than at its poles. The same is true on Earth, which is one of the reasons NASA launches rockets from Florida and California near the equator.

“A massive star has a strong rotation, which is very different at the poles and the equator. So rotation is expected to affect the symmetry of the mass loss rate,” said Jiang.

The team is also exploring metallicity, which in astrophysics refers to any element heavier than helium.

“Metallicity is important because it affects opacity. In our previous simulations, we assumed a constant metallicity, but massive stars can have very different metallicities,” said Jiang. “So we need to explore the parameter space to see how the structure of the stars change with metallicity. We’re currently running a simulation with one metallicity at NERSC, another at Argonne, and a third at NASA. Each set of calculations will take about three months to run.”

Meanwhile, Jiang and his colleagues already have new 2018 data to analyze. And they have a lot more simulations planned due to their recent allocation awards from INCITE, NERSC, and NASA.

“We need to do a lot more simulations to understand the physics of these special massive stars, and I think NERSC will be very helpful for this purpose,” he said.

This is a reposting of my news feature originally published by Berkeley Lab’s Computing Sciences.

Microrobots fly, walk and jump into the future

Assembling an ionocraft microrobot in UC Berkeley’s Swarm Lab. (Photos by Adam Lau)

A tiny robot takes off and drunkenly flies several centimeters above a table in the Berkeley Sensor and Actuator Center. Roughly the size and weight of a postage stamp, the microrobot consists of a mechanical structure, propulsion system, motion-tracking sensor and multiple wires that supply power and communication signals.

This flying robot is the project of Daniel Drew, a graduate student who is working under the guidance of electrical engineering and computer sciences professor Kris Pister (M.S.’89, Ph.D.’92 EECS). The culmination of decades of research, these microrobots arose from Pister’s invention of “smart dust,” tiny chips roughly the size of rice grains packed with sensors, microprocessors, wireless radios and batteries. Pister likes to refer to his microrobots as “smart dust with legs.”

“We’re pushing back the boundaries of knowledge in the field of miniaturization, robotic actuators, micro-motors, wireless communication and many other areas,” says Pister. “Where these results will lead us is difficult to predict.”

For now, Pister and his team are aiming to make microrobots that can self-deploy, in the hopes that they could be used by first responders to search for survivors after a disaster, industrial plants to detect chemical leaks or farmers to monitor and tend their crops.

These insect-sized robots come with a unique advantage for solving problems. For example, many farmers already use large drones to monitor and spray their plants to improve crop quality and yield. Microrobots could take this to a whole new level. “A standard quadcopter gives us a bird’s eye view of the field, but a microrobot would give us a bug’s eye view,” Drew says. “We could program them to do important jobs like pollination, looking for the same visual cues on flowers as insects [see].”

But to apply this kind of technology on a mass scale, first the team has to overcome significant challenges in microtechnolgy. And as Pister says, “Making tiny robots that fly, walk or jump hasn’t been easy. Every single piece of it has been hard.”

Flying silently with ion propulsion

Most flying microrobots have flapping wings that mimic real-life insects, like bees. But the team’s flying microrobot, called an ionocraft, uses a custom ion propulsion system unlike anything in nature. There are no moving parts, so it has the potential to be very durable. And it’s completely silent when it flies, so it doesn’t make an annoying buzz like a quadcopter rotor or mosquito.

The ionocraft’s propulsion system is novel, not just a scaled down version from NASA’s spacecrafts. “We use a mechanism that’s different than the one used in space, which ejects ions out the back to propel the spacecraft forward,” Drew says. “A key difference is that we have air on Earth.”

Instead, the ionocraft thruster consists of a thin emitter wire and a collector grid. When a voltage is applied between them, a positively-charged ion cloud is created around the wire. This ion cloud zips toward the negatively-charged collector grid, colliding with neutral air molecules along the way. The air molecules are knocked out of the way, creating a wind that moves the robot.

“If you put your hand under the collector grid of the ionocraft, you’ll feel wind on your hand — that’s the air stream that propels the microrobot upwards,” explains Drew. “It’s similar to the airstream that you’d feel if you put your hand under the rotor blades of a helicopter.”

The collector grid also provides the ionocraft’s mechanical structure. Having components play more than one role is critical for these tiny robots, which need to be compact and lightweight for the propulsion system to work.

Each ionocraft has four ion thrusters that are independently controlled by adjusting their voltages. This allows the team to control the orientation of the microrobot in a similar way as standard quadcopter drones. Namely, they can control the craft’s roll, pitch and yaw. What they can’t do yet is make the microrobot hover. “So far, we can fly it bouncing around like a bug in a web, but the goal is to get it to hover steadily in the air,” Pister says.

Taking first steps and jumps

In parallel, the researchers are developing microrobots that can walk or jump. Their micro-walker is composed of three silicon chips: a body chip that plugs perpendicularly into two chips with three legs each. “The hexapod microrobot is about the size of a really big ant, but it’s boxier,” says Pister.

Not only does the body chip provide structural support, but it also routes the external power and control signals to the leg chips. These leg chips are oriented vertically, allowing the legs to move along the table in a sweeping motion. Each leg is driven by two tiny on-chip linear motors, called electrostatic inchworm motors, which were invented by Pister. One motor lifts the robot’s body and the second pushes it forward. This unique walking mechanism allows three-dimensional microrobots to be fabricated more simply and cheaply.

Pister says the design should, in theory, allow the hexapod to run. So far it can only stand up and shuffle forward. However, he believes their recent fabrication and assembly improvements will have the microrobot walking more quickly and smoothly soon.

The jumping microrobot also uses on-chip inchworm motors. Its motor assembly compresses springs to store energy, which is then released when the microrobot jumps. Currently, it can only jump several millimeters in the air, but the team’s goal is to have it to jump six meters from the floor to the table. To achieve this, they are developing more efficient springs and motors.

“Having robots that can shuffle, jump a little and fly is a major achievement,” Pister says. “They are coming together. But they’re all still tethered by wires for control, data and power signals. ”

Working toward autonomy

Currently, high voltage control signals are passed over wires that connect a computer to a robot, complicating and restricting its movement. The team is developing better ways to control the microrobots, untethering them from the external computer. But transferring the controller onto the microrobot itself is challenging. “Small robots can’t carry the same kind of increasingly powerful computer chips that a standard quadcopter drone can carry,” Drew says. “We need to do more with less.”

So the group is designing and testing a single chip platform that will act as the robots’ brains for communication and control. They plan to send control messages to this chip from a cell phone using wireless technology such as Bluetooth. Ultimately, they hope to use only high-level commands, like “go pollinate the pumpkin field,” which the self-mobilizing microrobots can follow.

The team also plans to integrate on-board sensors, including a camera and microphone to act as the robot’s eyes and ears. These sensors will be used for navigation, as well as any tasks they want the robot to perform. “As the microrobot moves around, we could use its camera and microphone to transmit live video to a cell phone,” says Pister. “This could be used for many applications, including search and rescue.”

Using the brain chip interfaced with on-board sensors will allow the team to eliminate most of the troublesome wires. The next step will be to eliminate the power wires so the robots can move freely. Pister showed early on that solar cells are strong enough to power microrobots. In fact, a microrobot prototype that has been sitting on his office shelf for about 15 years still moves using solar power.

Now, his team is developing a power chip with solar cells in collaboration with Jason Stauth (M.S.’06, Ph.D.’08 EECS), who is an associate professor of engineering at Dartmouth. They’re also working with electrical engineering and computer sciences professor Ana Arias to investigate using batteries.

Finally, the researchers are developing clever machine learning algorithms that guide a microrobot’s motion, making it as smooth as possible.

In Drew’s case, the initial algorithms are based on data from flying a small quadcopter drone. “We’re first developing the machine learning platform with a centimeter-scale, off-the-shelf quadcopter,” says Drew. “Since the control system for an ionocraft is similar to a quadcopter, we’ll be able to adapt and apply the algorithms to our ionocraft. Hopefully, we’ll be able to make it hover.”

Putting it all together

Soon, the team hopes to have autonomous microrobots wandering around the lab directed by cell phone messages. But their ambitions don’t stop there. “I think it’s beneficial to have flying robots and walking robots cooperating together,” Drew says. “Flying robots will always consume more energy than walking robots, but they can overcome obstacles and sense the world from a higher vantage point. There is promise to having both or even a mixed-mobility microrobot, like a beetle that can fly or walk.”

Mixed-mobility microrobots could do things like monitor bridges, railways and airplanes. Currently, static sensors are used to monitor infrastructure, but they are difficult and time-consuming to deploy and maintain — picture changing the batteries of 100,000 sensors across a bridge. Mixed-mobility microrobots could also search for survivors after a disaster by flying, crawling and jumping through the debris.

“Imagine you’re a first responder who comes to the base of a collapsed building. Working by flashlight, it’s hard to see much but the dust hanging in the air,” says Drew. “Now, imagine pulling out a hundred insect-sized robots from your pack, tossing them into the air and having them disperse in all directions. Infrared cameras on each robot look for signs of life. When one spots a survivor, it sends a message back to you over a wireless network. Then a swarm of robots glowing like fireflies leads you to the victim’s location, while a group ahead clears out the debris in your path.”

The applications seem almost endless given the microrobots’ potential versatility and affordability. Pister estimates they might cost as little as one dollar someday, using batch manufacturing techniques. The technology is also likely to reach beyond microrobots.

For Pister’s team, the path forward is clear; the open question is when. “All the pieces are on the table now,” Pister says, “and it’s ‘just’ a matter of integration. But system integration is a challenge in its own right, especially with packaging. We may get results in the next six months — or it may take another five years.”

This is a reposting of my news feature previously published in the fall issue of the Berkeley Engineer magazine. © Berkeley Engineering