Home — Gadgets Quest for prosthetic retinas progresses toward human trials, with a VR assist...

Quest for prosthetic retinas progresses toward human trials, with a VR assist – TechCrunch

404
0

An artificial retina would be an enormous boon to many people with visual impairments, and the possibility is creeping closer to reality year by year. One of the latest advancements takes a different and promising approach, using tiny dots that convert light to electricity. Virtual reality has helped show that it could be a viable path forward.

These photovoltaic retinal prostheses come from the École Polytechnique fédérale de Lausanne, where Diego Ghezzi has been working on the idea for several years.

Early retinal prosthetics were created decades ago, and the basic idea is as follows: A camera outside the body (on a pair of glasses, for instance) sends a signal over a wire to a tiny microelectrode array, which consists of many small electrodes that pierce the nonfunctioning retinal surface and stimulate the working cells directly.

The problems with this are mainly that powering and sending data to the array requires a wire running from outside the eye is — generally speaking, a “don’t” when it comes to prosthetics and the body in general. (The concept doesn’t translate directly because of the way the visual system works.) the array itself is also limited in the number of electrodes it can have by the size of each. For many years, the effective resolution in the best-case scenario was on the order of a few dozen or hundred “pixels.”

Ghezzi’s approach obviates both these problems using photovoltaic materials, which turn light into an electric current. It’s not so different from what happens in a digital camera, except instead of recording the charge as in the image, it sends the current into the retina as the powered electrodes did. There’s no need for a wire to relay power or data to the implant because both are provided by the light shining on it.

VR assist

In the case of the EPFL prosthesis, thousands of tiny photovoltaic dots would, in theory, be illuminated by a device outside the eye, sending light in according to what it detects from a camera. Of course, it’s still a tricky thing to engineer. The other part of the setup would be a pair of glasses or goggles that capture an image and project it through the eye onto the implant.

We first heard of this approach in 2018, and things have changed somewhat since then as a new paper document.

“We increased the number of pixels from about 2,300 to 10,500,” explained Ghezzi in an email to TechCrunch. “So now it is difficult to see them individually, and they look like a continuous film.”

Of course, when those dots are pressed right up against the retina, it’s a different story. After all, that’s only 100×100 pixels if it were a square — not exactly high definition. But the idea isn’t to replicate human vision, which may be an impossible task, let alone realistic for anyone’s first shot.

“Technically, it is possible to make pixels smaller and denser,” Ghezzi explained. “The problem is that the current generated decreases with the pixel area.”

So the more you add, the more challenging it is to make it work, and there’s also the risk (which they tested) that two adjacent dots will stimulate the same network in the retina. But too few and the image created may not be intelligible to the user. 10,500 sounds like a lot, and it may be enough — but there is no data to support that. To start on that, the team turned to what may seem like an unlikely medium: VR.

Because the team can’t exactly do a “test” installation of an experimental retinal implant on people to see if it works, they needed another way to tell whether the dimensions and resolution of the device would be sufficient for specific everyday tasks like recognizing objects and letters.

To do this, they put people in dark VR environments except for tiny simulated “phosphors,” the pinpricks of light they expect to create by stimulating the retina via the implant; Ghezzi likened what people would see to a constellation of bright, shifting stars. They varied the number of phosphors, the area they appeared over, and the length of their illumination or “tail” when the image went, asking participants how well they could perceive things like a word or scene.

The robust analysis of the visual system in the brain intuits things like edges and motion even from sparse inputs. Their primary finding was that the most critical factor was visual angle — the overall size of the area where the image appears. Even a clear picture is difficult to understand if it only takes up the center of your vision, so it’s better to have a wide field of vision even if overall clarity suffers.

This demonstration showed that the implant’s parameters are theoretically sound, and the team can start working toward human trials. That’s not something that can happen in a hurry. While this approach is promising compared with earlier, wired ones, it will still be several years, even in the best-case scenario, before it’s possible it could be made widely available. Still, the prospect of a working retinal implant of this type is exciting, and we’ll be following it closely.

LEAVE A REPLY

Please enter your comment!
Please enter your name here