¡Últimas horas! 1 año de Premium al 25% de dto ¡Lo quiero!

Podcast
On Tech & Vision With Dr. Cal Roberts
44
0
Dr. Cal Roberts, President and CEO of Lighthouse Guild, the leading provider of exceptional services that inspire people who are visually impaired to attain their goals, interviews inventors, developers and entrepreneurs who have innovative tech ideas and solutions to help improve the lives of people with vision loss.
Dr. Cal Roberts, President and CEO of Lighthouse Guild, the leading provider of exceptional services that inspire people who are visually impaired to attain their goals, interviews inventors, developers and entrepreneurs who have innovative tech ideas and solutions to help improve the lives of people with vision loss.
We Are Living In the Future of Vision Tech
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
Not long ago virtual assistants and VR gaming were the makings of science fiction. But today, as we see developers across the board incorporating these technologies into apps and other tech, including for people who are blind or have low vision, these breakthroughs aren’t just possible, they’re part of the fabric of everyday life.
In this episode Dr. Cal talks with Steve Ewell, Executive Director of the Consumer Technology Association Foundation (CTAF). Together they discuss how rapid advances in tech have made previously unimaginable products like self-driving cars, smart homes and virtual reality gaming an attainable, everyday reality for millions of people, including for those who are aging or have low vision. They talk about CES, the CTAF’s annual trade show and the increasing number of consumer companies presenting accessible tech. They also discuss Steve’s work with AARP, and how technology designed for older users is also accelerating… and how those users are keeping up with the pace of new tech.
The episode also features an interview with Walt Hickey, the executive editor at Sherwood News and the author of “You Are What You Watch - How Movies and TV Affect Everything.” Walt discusses how scientists and engineers consume the same media as everyone else, and how the ideas from science and science fiction have fed back on one another for millennia to build our modern world.
With technology for the blind and low vision community increasingly easy to build and access, companies like Amazon and XBox are making accessibility more of a priority with every new development. In other words, when it comes to accessible technology, the future may have finally arrived.
The Big Takeaways
The Simplest Tech Can Make a Big Difference: Steve discusses the rise of smart homes, and how something simple like a voice-activated assistant can transform the living space of someone who’s blind or low vision. He talks about the natural evolution from there to doorbell cams, and the possibilities that has opened for people with low vision to maintain their privacy and security while at home.
Are Self-Driving Cars Finally Here?: It can seem like self-driving cars have been on the horizon for decades. But with Waymo and other major technology companies recently unveiling breakthroughs, people who are blind or have low vision might finally be able to hop in a car by themselves and go wherever they want. Steve discusses what these breakthroughs mean, not just for activities like visiting the doctor, but for improved opportunities for social interaction.
All Fun and Games: Steve talks about his work with AbleGamers and Oak Hill, two organizations dedicated to providing an accessible video game experience for people who are blind or have low vision. Gaming can combat social isolation, he says, but it can do more than that. When paired with wearables, Steve credits accessible gaming (and a healthy dose of competitiveness) with motivating people who are aging or have disabilities to stay active.
Old School: It might be surprising to see the words “technology” and “AARP” in the same sentence. But tooday, they are indeed connected. Steve discusses the CTAF’s partnership with AARP, and how recent developments along with a savvier population of seniors has led to increased rates of development and adoption for accessible tech.
Tweetables
“One of the important things that I've seen that's changed over the last dozen years is the awareness on the company side of the needs of older adults and people with disabilities. That was one of the big challenges, companies recognizing this as a market to engage… and what we've seen is more and more companies are actively engaging this space.” — Steve Ewell, CTAF Executive Director
“Technology and movies and pop culture are ways that we express our wants and goals” — Walt Hickey, Author, “You Are What You Watch”
“I went to one senior center and a gentleman had come up to me and he said ‘I know you're going to assume I don't use technology, but I want to let you know I'm over a hundred years old and I built all the technology that your technology today is based on.’” — Steve Ewell, CTAF Executive Director
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links
Lighthouse Guild
CTA Foundation
31:53
The Paradigm Shift in Innovation: Remixing Existing Tech to Advance Accessibility
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
Technology developed specifically for people with low vision has historically been bespoke, specific, and expensive, requiring the creation of new technology for a relatively small population of users. But recently we’re seeing a paradigm shift from building new vision tech from scratch to building on top of existing technology that most people use daily.
In this episode, Dr. Cal talks with Karthik Mahadevan, the founder and CEO of Envision, about how his technology company builds on existing tech, like smartphones and smart glasses, to benefit users with low vision. Envision’s software uses AI to recognize important visual cues in a user’s environment and then relays that information to the user with computer-generated speech. Karthik’s breakthrough realization early on that he didn’t need to build hardware when he could just tap into the cameras of smartphones and smart glasses that people already owned, enabling Envision to reach a much wider audience.
The episode also features an interview with Troy Otillio, the CEO of Aira. Aira’s platform links users to a human visual interpreter, who can access the camera on their device and guide the user through the task, whether reading a letter, navigating a grocery store, or finding their way around an unfamiliar city.
Ultimately, the future of creating vision technology might not be in reinventing the wheel but in improving it and figuring out how to turn the tech people already use every day into easy and accessible solutions for people with low vision.
The Big Takeaways
Building on Existing Tech: Karthik stresses that for vision tech to be impactful, it has to be accessible. He describes his inspiration as building on top of existing consumer technology to create Envision and thus create a product that millions of people already have the hardware to operate. Later, Dr. Cal discusses the “Taco Bell Phenomenon”, which theorizes that while there may be no such thing as a new idea, the novel combinations of existing ideas are so infinite that they may be limitless.
The Importance of a Visual Interpreter: Karthik discusses the limitations of communication formats such as braille and audio announcements in public places for people with vision impairment. While well-meaning, they can never really provide complete access to spaces that are meant to be navigated visually. That’s where Envision comes in, allowing access to the camera on a smartphone or smart glasses to give the user a complete and accurate sense of their environment.
Would You Like a Human for That?: To AI or not to AI? Speech-to-text technology is advanced enough that it can excel at reading something simple like a gas bill or a chicken soup recipe. But what if you got a birthday card from your grandchildren? Troy Otillio discusses how Aira recently partnered with Envision to offer their human visual interpreters as part of Envision’s technology — providing camera access to a human agent who can fully understand the nuance of the user’s situation.
We Don’t Make Hardware: A small market company like Envision doesn’t have the capacity to build smart glasses (or smartphones) from scratch. The cost would be too high for a company that necessarily has a limited user base. Instead, the future of vision technology is likely to be in creating apps or finding novel uses for hardware that already exists. This will provide accessibility to people with vision impairments via tools they’re already quite familiar with.
Tweetables
“In the future, these AI are going to interact with other AIs… You know, someone like a Walmart will have their own AI. Someone like a Target will have their own AI. And your AI can speak to that AI and ask for specific information. And that's when it becomes super, super exciting.” — Karthik Mahadevan, Envision Founder & CEO
“So for many years, consumer electronics was evolving in one direction, and assistive technology was evolving in a different direction. But what I noticed recently is this merger of the two: that consumer electronics are becoming more accessible.” — Dr. Cal Roberts, Lighthouse Guild President & CEO
“You can look at Aira as just a layer, an accessibility layer that exists for when an automated solution, a pure software solution, doesn't work or when there is no solution, right? We're like that, call it a backstop, a more general purpose layer that you can always fall back to.” — Troy Otillio, Aira CEO
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links
Lighthouse Guild
Envision
Aira
29:24
How Well Do You Really Need to See to Drive?
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
Today, self-driving cars are all the buzz when it comes to cutting edge driving technology that could help people with vision impairmentdrive. But there are other, simpler technologies that are already helping people who are blind or visually impaired navigate behind the wheel - some of which draw on technology that’s been around for hundreds of years!
In this episode, Dr. Cal talks with Dr. Henry Greene, co-founder and president of Ocutech about their bioptic driving telescopes. These simple devices fit on top of a pair of conventional eyeglasses, and provide a wider field of vision for drivers with vision impairments. This has been shown to increase driver safety and confidence.
The episode also features an interview with Dr. Bill Seiple, Lighthouse Guild’s Chief Research Officer. He discusses how other technological advancements that are already in widespread use among all consumers - like GPS - are making it so everyone has to use their vision less while behind the wheel.
Ultimately, the ability to pilot a car without relying on our vision isn’t there yet. But with these and other technologies discussed in this episode, the day where none of us need to see to drive may be just around the corner!
The Big Takeaways
Ancient Tech, Modern Design: Dr. Greene and his team succeeded in adapting telescopic technology that’s existed since Galileo’s time into a bioptic device that fits on top of a conventional pair of eyeglasses. Coupled with other innovations like an auto-focusing device, these bioptic driving assists can vision impaired drivers to see faraway objects with the clarity required for highway driving.
Can Something You Use Only 1% of the Time Make You Safer? Ocutech’s bioptic driving telescopes aren’t designed to be used all the time while driving. In fact, Dr. Greene estimates that his patients only use them between one and two percent of the time. This means that while they’re a useful tool for very specific tasks on the road, drivers still need to be competent behind the wheel without bioptics. Per Dr. Seiple, research has shown that drivers with vision impairment tend to be intentional about where and when they drive, adding another layer of safety.
GPS Makes Everything Easier: Thanks to 5G technology that can place a user anywhere on Earth within a few feet, GPS is making it so that everyone has to use their vision less when they’re on the road .
Is Driving a Visual Thing? As Dr. Seiple points out, visual acuity isn’t of greatest importance behind the wheel. More important is attentiveness, knowing where you’re going, and reaction time once something crosses your way, regardless of how sharply you can see it. This suggests that while cars may soon be able to pilot themselves, getting from one place to another may always require the competent guidance of a human.
Tweetables
“GPS doesn't tell you that there's a box on the road or a policeman or a pedestrian or an animal or a deviation for construction. GPS won't tell you that stuff. And that's the stuff that you need to be able to anticipate problems and take the appropriate action while you're driving.” — Dr. Henry Greene, Ocutech Co-Founder & President
“So, in tongue and cheek, you can say that I don't think driving is a visual thing. Turns out field is much more important than acuity for driving safety.” — Dr. Bill Seiple, Lighthouse Guild Chief Research Officer
“These bioptics, they are a game changer for people with low vision… All my life I've actually had to get a ride or wait for someone. And I'm very, very, very independent. I don't like to wait for people..” — Crystal Davis, Ocutech Employee & Bioptic Telescope User
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links
Lighthouse Guild
Ocutech
26:18
BenVision: Navigating with Music
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
When it comes to navigation technology for people who are blind or visually impaired, many apps utilize voice commands, loud tones or beeps, or haptic feedback. In an effort to create a more natural, seamless experience, the team at BenVision has created a different type of system that allows users to navigate using musical cues instead!
For this episode, Dr. Cal spoke with BenVision’s CEO and co-founder, Patrick Burton, along with its Technology Leadd, Aaditya Vaze. They shared about the inspiration behind BenVision, how they’re able to create immersive soundscapes that double as navigation aids, and the exciting future applications this technology could offer.
The episode also features BenVision’s other co-founder and Audio Director, Soobin Ha. Soobin described her creative process for designing BenVision’s soundscapes, how she harnesses the power of AI, and her bold vision of what’s to come.
Lighthouse Guild volunteer Shanell Matos tested BenVision herself and shares her thoughts on the experience. As you’ll hear, this technology is transformative!
The Big Takeaways
Why Music? Navigation technology that uses voice, tone, or haptics can create an added distraction for some users. But the brain processes music differently. Instead of overloading the senses, for some users music works alongside them, allowing them to single out separate sound cues, or take in the entire environment as a whole. Like how the different instruments correspond to various characters in “Peter and the Wolf,” BenVision assigns unique musical cues to individual objects.
User Experience: Shanell Matos appreciated how BenVision blends in more subconsciously, allowing her to navigate a space without having to be as actively engaged with the process.
Additional Applications: BenVision began as an augmented reality program, and its creators see a potential for it to grow beyond a navigational tool to expand for use by people who are visually impaired or fully sighted. For example, it could be used to create unique soundscapes for museums, theme parks, and more, augmenting the experience in exciting new ways.
The Role of AI: Artificial Intelligence already plays a big role in how BenVision works, and its creators see it being even more important in the future. BenVision already harnesses AI for object detection and its companion app uses AI to provide instant voice support about the immediate surroundings if needed. Moving forward, AI could be used to help instantaneously generate new sound cues or to help users customize their experience at the press of a button.
Tweetables
“We thought that if the human brain can learn echolocation and we have this amazing technology that’s available to us in the modern day, then why can’t we make echolocation a little bit more intuitive and perhaps a little bit more pleasant.” — Patrick Burton, BenVision CEO & Co-Founder
“You can think of it like a bunch of virtual speakers placed at different locations around the user. So like a speaker on a door or a couch or a chair. And then there are sounds coming from all these virtual speakers at the same time.” — Aaditya Vaze, BenVision Technology Lead
“I want to gamify this idea so that the user can actually find some interest and joy by using it, rather than just find it only helpful, but also [to create ] some pleasant feeling.” — Soobin Ha, BenVision Audio Director & Co-Founder
“So if there’s a lot of people, there’s a lot of conversations happening, a lot of sounds happening, a lot of movement happening. It’s really difficult to keep up with what everything is doing. Whereas with music, it’s not as difficult to pick out layers.” – Shanell Matos, Lighthouse Guild Volunteer
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links
Lighthouse Guild
BenVision
31:33
People's Choice Podcast Awards
Episode in
On Tech & Vision With Dr. Cal Roberts
We appreciate your support for our show — and now, we need your help nominating the On Tech & Vision podcast for the People’s Choice Podcast Awards! We are participating in these awards so we can showcase On Tech & Vision to a broader audience, gain recognition within the industry, and, most importantly, help spread the message about Lighthouse Guild and the role that technology is playing in tearing down barriers for people who are blind or visually impaired. To help us nominate On Tech & Vision, please go online to www.podcastawards.com, where you can register to vote for On Tech & Vision in both the Technology and Peoples’ Choice Categories. Voting is open until July 31st. Once again, your support is greatly appreciated!
01:10
The Possibilities of Vision Restoration
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
For hundreds of years, health professionals have dreamed of restoring vision for people who are blind or visually impaired. However, doing so, either through transplanting a functioning eye or using technological aids, is an incredibly complex challenge. In fact, many considered it impossible. But thanks to cutting-edge research and programs, the ability to restore vision is getting closer than ever.
As a first for this podcast, this episode features an interview with Dr. Cal Roberts himself! Adapting audio from an interview on The Doctors Podcast, Dr. Cal describes his work as a program manager for a project on eye transplantation called Transplantation of Human Eye Allographs (THEA). Funded by a government initiative called ARPA-H, THEA is bringing some of the country’s finest minds together to tackle the complexities of connecting a person’s brain to an eye from a human donor.
This episode also features an interview with Dr. Daniel Palanker of Stanford University. Dr. Palanker is working on technology that can artificially restore sight through prosthetic replacement of photoreceptors. Having proved successful in animals, Dr. Palanker and his team are working hard to translate it to humans.
And if that can happen, then something once considered impossible could finally be accomplished!
The Big Takeaways
The Challenges of Eye Transplants: Although eyeball transplants have been done, they’ve only been cosmetic. So far, nobody has been able to successfully connect a donor eyeball to a recipient’s brain. Dr. Roberts’s work with THEA is bringing together multiple teams to tackle the challenges associated with a whole eyeball transplant, from connecting nerves and muscles to ensuring the organ isn’t rejected, and much more.
“Artificial” Vision Restoration: Dr. Palanker is working to replace the functions of photoreceptors through technological means. His photovoltaic array is placed underneath the retina and can convert light into an electrical current that activates the cells that send visual information to the brain. While it doesn’t completely restore sight for people with Age-Related Macular Degeneration, this technology shows incredible promise.
Decoding “Brain Language”: For both Dr. Roberts and Dr. Palanker, one of the biggest challenges with vision restoration is understanding how the eye and brain communicate. Dr. Roberts likens it to Morse Code — the eye speaks to the brain in “dots and dashes,” which the brain then converts into vision. Right now, the language is still foreign to us, but we’re closer than ever to decoding it.
The Evolution of the Brain-Machine Interface: Dr. Palanker imagines incredible possibilities in the interaction between the brain and technology. If we can find a way to truly translate the brain’s signals into information, Dr. Palanker envisions the possibility of direct brain-to-brain communication without verbalization. In a way, this could make people telepathic, able to understand and digest vast amounts of information in an instant.
Tweetables:
So ideally in medicine, at least the ideal therapy is the restoration of full functionality. If we can grow back photoreceptors and make them reconnect to bipolar cells, undo all the rewiring that right now underwent during degeneration, and restore the full extent of vision, that would be the ideal outcome. — Dr. Daniel Palanker, Professor of Ophthalmology, Stanford University
We can think about other aspects of brain-machine interface, which takes you maybe into the realm of capabilities that humans never had. If you enable artificial senses or enable brain-to-brain connectivity so you can communicate without verbalization that would open completely new capabilities that humanity never had. — Dr. Palanker
Forty-two years after the implantation of the first mechanical heart, there’s not a single person in the world walking around with a mechanical heart. All that work, all that research, and all that effort to come up with mechanical heart transplants are still state-of-the-art. And so, while I believe that there is a role for a bionic eye or mechanical eye, what I really believe is that everything that we learn from doing an eye transplant will just make it better and easier when we do eventually come up with a bionic or a mechanical eye. — Dr. Calvin Roberts, ARPA-H Health Science Futures Program Manager (President and CEO of Lighthouse Guild and Host of On Tech & Vision!)
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links
Lighthouse Guild
THEA Program
Prima System
People’s Choice Podcast Awards
We appreciate your support for our show — and now, we need your help nominating the On Tech & Vision podcast for the People’s Choice Podcast Awards! We are participating in these awards so we can showcase On Tech & Vision to a broader audience, gain recognition within the industry, and, most importantly, help spread the message about Lighthouse Guild and the role that technology is playing in tearing down barriers for people who are blind or visually impaired. To help us nominate On Tech & Vision, please go online to www.podcastawards.com, where you can register to vote for On Tech & Vision in both the Technology and Peoples’ Choice Categories. Voting is open until July 31st. Once again, your support is greatly appreciated!
44:09
Biosensors: The Future of Diagnostic Medicine
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
This episode is about how biosensor technology is revolutionizing the field of diagnostic and preventive medicine. Biosensors can take many forms — wearable, implantable, and even ingestible. And they can serve many different functions as well, most notably when it comes to detecting the various pressure levels in our bodies.
This episode features interviews with several luminaries working with biosensors. One of them is Doug Adams, a revolutionary entrepreneur who became inspired to create a biosensor that can assist in the treatment of glaucoma patients, initially focusing on a sensor for intraocular pressure. More recently, Doug founded a company called QURA, whose current efforts are focused on a biosensor that detects blood pressure.
To elaborate on QURA’s initiatives, this episode also includes insights from its Chief Business Officer, David Hendren. He and Dr. Cal discuss the current state of biosensor technology, the benefits of implantable biosensors, and how they work.
Finally, this episode includes a conversation with Max Ostermeier, co-founder and General Manager of Implandata Ophthalmic Products. Max was previously interviewed by Dr. Cal for the episode “Innovations in Intraocular Pressure and Closed Loop Drug Delivery Systems.” This time, Max joins Dr. Cal to discuss the possibilities of biosensor technology and his company’s Eyemate system — which includes biosensor technology for glaucoma patients.
All three guests also offer their thoughts on the future of biosensors and their endless possibilities. While it may seem like science fiction, it truly is science reality!
The Big Takeaways
What Biosensors Do: Currently, biosensors primarily sense the various pressures in the human body. QURA’s current sensor detects blood pressure and assists with hypertension. Meanwhile, Implandata’s Eyemate technology serves glaucoma patients by gathering data on intraocular pressure.
The Rapid Shrinking of Biosensors: When Doug Adams first started working on biosensors, the model he saw was the size of a microwave. Now, it’s shrunk to the size of a grain of rice! By making biosensors smaller, they are easier to implant and place in different spots within the body. And by doing so, they can gather more and more data.
The Benefits of AI: One drawback of gathering so much data is that it can sometimes be hard to analyze it. However, improvements in AI technology are making it easier to sort through all that data, giving doctors and patients valuable information for medical diagnostics and treatments.
The Future of Biosensors: As implantable biosensors become smaller and more sophisticated, all our guests see them becoming a crucial part of healthcare. In addition to gathering data on all sorts of functions within the body, biosensors could provide therapies and treatments with minimal human intervention.
Tweetables:
So, we are measuring the absolute pressure inside the eye with this kind of technology. It originates from the automotive industry. Tire pressure sensors, where you also have to measure the pressure inside the tire. And so basically we took set technology and advanced it and made it so small that you can also implant this kind of sensor in an eye. — Max Ostermeier, co-founder and General Manager of Implandata Ophthalmic Products
So I had a physical a month ago, and along with the physical, they draw blood, and they send that blood off to a lab. I have a feeling in the next decade, that goes away. Why do you have to send a vial of blood to the lab? Because if I had a sensor, not even in an artery, but on top of an artery, I could do a complete analysis of everything in that blood that you’re doing from the lab. — Doug Adams, entrepreneur and founder of QURA
The important thing is that you are automatically getting data to the care group that is taking care of these patients, where they are able to see what’s happening. They’re able to see not just a snapshot once in a while, as you’d have from an external pressure cuff, but [get] continuous data longitudinally. — David Hendren, Chief Business Officer of QURA
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links
Lighthouse Guild
QURA
Implandata Ophthalmic Products
31:50
The World in Your Hand: The Power of Generative AI
Episode in
On Tech & Vision With Dr. Cal Roberts
When it comes to emerging technology, there’s no hotter topic than artificial intelligence. Programs like ChatGPT and Midjourney are becoming more popular and are inspiring people to explore the possibilities of what AI can achieve — including when it comes to accessible technology for people who are blind or visually impaired.
One of those people is Saqib Shaikh, an engineering manager at Microsoft. Saqib leads the team that developed an app called Seeing AI, which utilizes the latest generation of artificial intelligence, known as generative AI Dr. Cal spoke with Saqib about how Generative AI works, his firsthand experience using an app like Seeing AI, and how it helped improve his daily life.
This episode also features Alice Massa, an occupational therapist at Lighthouse Guild. Alice described the many benefits of generative AI, and how it helps her clients better engage in their world.
Saqib and Alice also both agreed that the current state of AI is only the beginning of its potential. They shared their visions of what it could achieve in the future — and it doesn’t seem that far off.
The Big Takeaways:
The Power of Generative AI: Saqib discussed the present condition of artificial intelligence and why generative AI is a massive leap from what came before it. With a deep data pool to draw from, generative AI can do so much more than identify items or come up with an essay prompt. It can understand and interpret the world with startling depth and expediency.
Seeing AI: This app can truly put the world in the palm of your hand. It can perform essential tasks like reading a prescription or the sign at a bus stop — and even more than that! It can describe all the colorful details of sea life in a fish tank at the aquarium or help you order dinner off a menu. The app doesn’t just provide people who are blind or visually impaired greater access to the world — it expands it.
Embrace Change: There’s understandably a lot of uncertainty about what role AI should play in society. However, Saqib Shaikh and Alice Massa insist that there’s nothing to fear from AI, that the benefits far outweigh any potential drawbacks, and that as long as it’s handled responsibly, there’s a lot AI can do to help improve our lives.
Tweetables:
“I had a client at the Lighthouse who really was very disinterested in doing anything. The only thing he did on his phone was answer a call from his pastor and call his pastor. And I was able to put Seeing AI on his phone. And his wife said the first time in two years, she saw a smile on his face because now he could read his Bible by himself.” — Alice Massa, Occupational Therapist at Lighthouse Guild
“What if AI could understand you as a human? What are your capabilities? What are your limitations at any moment in time? Whether that's due to a disability or your preferences or something else, and understand the environment, the world you're in, or the task you’re doing on the computer or whatever. And then we can use the AI to close that gap and enable everyone to do more and realize their full potential.” — Saqib Shaikh, Engineering Manager at Microsoft
“I call my phone my sister because my phone is the person I go to when I’m on the street if I’m walking in Manhattan. The other day I was meeting someone on 47th Street. I wasn’t sure which block I was on. All I did was open Seeing AI short text, hold it up to the street sign, and it told me I was on West 46th Street.” — Alice Massa
“Some of the interesting things powered by generative AI is going from taking a photo, say from your photo gallery if you’re reliving memories from your vacation, or even just what’s in front of you right now. It can go from saying it’s a man sitting on a chair in a room to actually giving you maybe a whole paragraph describing what’s in the room, what’s on the shelf, what’s In the background, what’s through the window, even. And it’s just remarkable. I work on this every day. I understand the technology, yet as an end user, I still am surprised and delighted by what this generation of AI is capable of telling me.” — Saqib Shaikh
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links:
Lighthouse Guild
Together, Tacit
Retissa Neoviewer
State Tactile Omero Museum
Emily Metauten Artist Page (Herminia Blue)
26:46
Reimagining the Visual Arts
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
When it comes to art, a common phrase is “look, don’t touch.” Many think of art as a purely visual medium, and that can make it difficult for people who are blind or visually impaired to engage with it. But in recent years, people have begun to reimagine what it means to experience and express art.
For this episode, Dr. Cal spoke to El-Deane Naude from Sony Electronics. El-Deane discussed the Retissa NeoViewer, a project developed with QD Laser that projects images taken on a camera directly onto the photographer’s retina. This technology allows people who are visually impaired to see their work much more clearly and with greater ease.
Dr. Cal also spoke with Bonnie Collura, a sculptor and professor at Penn State University about her project, “Together, Tacit.” Bonnie and her team developed a haptic glove that allows artists who are blind or visually impaired to sculpt with virtual clay. They work in conjunction with a sighted partner wearing a VR headset, allowing both to engage with each other and gain a new understanding of the artistic process.
This episode also includes an interview with Greta Sturm, who works for the State Tactile Omero Museum in Italy. Greta described how the museum’s founders created an experience solely centered around interacting with art through touch. Not only is it accessible for people who are blind or visually impaired, but it allows everyone to engage with the museum’s collection in a fascinating new way.
Finally, a painter and makeup artist named Emily Metauten described how useful accessible technology has been for her career. But she also discussed the challenges artists who are blind or visually impaired face when it comes to gaining access to this valuable technology.
The Big Takeaways:
The Value of Versatility: Many photographers who are visually impaired require the use of large, unwieldy accessories in order to properly capture their work. Sony and QD Laser are determined to solve this problem with the Retissa NeoViewer, which can replace cumbersome accessories like screen magnifiers and optical scopes.
Sculpting Virtual Clay: The aim of Together, Tacit, is to “foster creative collaboration between blind, low-vision, and sighted individuals.” A major way this is accomplished is by using the haptic glove to sculpt virtual, rather than physical, clay. Working in VR makes it harder for the sighted partner to unintentionally influence the work of the artist who is blind or visually impaired. As a result, the experience for both users is more authentic and enriching.
Reimagining the Museum Experience: The Tactile Omero Museum is much more than an opportunity for people who are blind or visually impaired to interact with art – it’s reimagining how that art is fundamentally experienced. By giving visitors a chance to engage with pieces on a tactile level, the museum allows everyone a chance to reconnect with a vital sense that many take for granted.
Expanding Ability to Access Technology: For artists like Emily Metauten who are visually impaired, accessible technology makes it much easier to do their jobs. However, many governmental organizations don’t have the infrastructure to provide this technology to them. Emily wants to raise awareness of how valuable this technology can be, and why providing it to people is so important.
Tweetables:
“When we’re little kids, we want to touch everything … and then soon after that, we’re told, no, no, no, you shouldn’t touch. You should look and not touch. And so, it becomes the reality and it becomes what you’re supposed to do.” – Greta Sturm, Operator at State Tactile Omero Museum
“I carry a Monocular little optical scope. But it becomes extremely difficult when you’re out and about and you’re trying to take a photograph, trying to change your settings. This method, the laser projection, I can actually read, the tiniest little settings.” – El-Deane Naude, Senior Project Manager at Sony Electronics Imaging Division
“The VR glasses definitely unlock an ability to see more details more easily for me. Because peripheral vision isn’t designed to see fine details. That's what the central vision is responsible for. So that’s what I have trouble with. But it made what I was already doing easier, and also did give me inspiration. Because we’re trying to unlock the greater things in life, that aren’t just beyond the basics for people with vision loss.” – Emily Metauten, professional painter and makeup artist
“I’ve learned through teaching that if a visually impaired or blind person was to use real clay … a sighted person would inevitably start to signify it in terms of what it can be called … And already, immediately, that begins to change the power dynamic on how something is created.” – Bonnie Collura, Professor of Art, Penn State University
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links
Lighthouse Guild
Together, Tacit
Retissa Neoviewer
State Tactile Omero Museum
Emily Metauten Artist Page (Herminia Blue)
37:05
Developing Big Ideas: Product Testing and Iteration
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
When we buy a product off the shelf, we rarely think about how much work went into getting it there. Between initial conception and going to market, life-changing technology requires a rigorous testing and development process. That is especially true when it comes to accessible technology for people who are blind or visually impaired.
For this episode, Dr. Cal spoke to Jay Cormier, the President and CEO of Eyedaptic, a company that specializes in vision-enhancement technology. Their flagship product, the EYE5, provides immense benefits to people with Age-Related Macular Degeneration, Diabetic Retinopathy, and other low-vision diseases. But this product didn’t arrive by magic. It took years of planning, testing, and internal development to bring this technology to market.
This episode also features JR Rizzo, who is a professor and researcher of medicine and engineering at NYU — and a medical doctor. JR and his research team are developing a wearable “backpack” navigation system that uses sophisticated camera, computer, and sensor technology. JR discussed both the practical and technological challenges of creating such a sophisticated project, along with the importance of beta testing and feedback.
The Big Takeaways:
The importance of testing: There’s no straight line between the initial idea and the final product. It’s more of a wheel, that rolls along with the power of testing and feedback. It’s extremely important to have a wide range of beta testers engage with the product. Their experience with it can highlight unexpected blind spots and create opportunities to make something even greater than originally anticipated.
Anticipating needs: When it comes to products like the EYE5, developers need to anticipate that its users will have evolving needs as their visual acuity deteriorates. So part of the development process involves anticipating what those needs will be and finding a way to deliver new features as users need them.
Changing on the fly: Sometimes, we receive feedback we were never expecting. When JR Rizzo received some surprise reactions to his backpack device, he had to reconsider his approach and re-examine his fundamental design.
Future-Casting: When Jay Cormier and his team at Eyedaptic first started designing the EYE5 device, they were already considering what the product would look like in the future, and how it would evolve. To that end, they submitted certain patents many years ahead of when they thought they’d need them — and now, they’re finally being put to use.
Tweetables:
“I’m no Steve Jobs and I don’t know better than our users. So the best thing to do is give them a choice and see what happens.” — Jay Cormier, President & CEO of Eyedaptic
“I started to think a little bit more about … assistive technologies. … And, I thought about trying to build in and integrate other sensory inputs that we may not have natively … to augment our existing capabilities.” — JR Rizzo, NYU Professor of Medicine and Engineering
“I think the way we’ve always looked at it is the right way, which is you put the user, the end user, front and center, and they’re really your guide, if you will. And we’ve always done that even in the beginning when we start development of a project.” – Jay Cormier
“When we put a 10-pound backpack on some colleagues, they offered some fairly critical feedback that it was way too heavy and they would never wear it. … They were like … it’s a non-starter.” — JR Rizzo
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links
Lighthouse Guild
Eyedaptic
Rizzo Lab
37:33
Robotic Guidance Technology
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
The white cane and guide dogs are long-established foundational tools used by people with vision impairment to navigate. Although it would be difficult to replace the 35,000 years of bonding between humans and dogs, researchers are working on robotic technologies that can replicate many of the same functions of a guide dog.
One such project, called LYSA, is being developed by Vix Labs in Brazil. LYSA sits on two wheels and is pushed by the user. It’s capable of identifying obstacles and guiding users to saved destinations. And while hurdles such as outdoor navigation remain, LYSA could someday be a promising alternative for people who either don’t have access to guide dogs or aren’t interested in having one.
In a similar vein, Dr. Cang Ye and his team at Virginia Commonwealth University are developing a robotic white cane that augments the familiar white cane experience for people with vision loss. Like the LYSA, the robotic white cane has a sophisticated computer learning system that allows it to identify obstacles and help the user navigate around them, using a roller tip at its base. Although it faces obstacles as well, the robotic guide cane is another incredible example of how robotics can help improve the lives of people who are blind or visually impaired.
It may be a while until these technologies are widely available, and guide dogs and traditional canes will always be extremely useful for people who are blind or visually impaired. But with how fast innovations in robotics are happening, it may not be long until viable robotic alternatives are available.
The Big Takeaways:
Reliability of Biological Guide Dogs: Although guide dogs have only been around for a little over a century, humans and dogs have a relationship dating over 35,000 years. Thomas Panek, the President and CEO of Guiding Eyes for the Blind, points out that there will never be a true replacement for this timeless bond. That being said, he thinks there is a role for robotics to coexist alongside biological guide dogs, and even help augment their abilities.
LYSA the Robotic Guide Dog: LYSA may look more like a rolling suitcase than a dog, but its developers at Brazil’s Vix Systems are working on giving it many of the same functions as its biological counterpart. LYSA can identify obstacles and guide its user around them. And for indoor environments that are fully mapped out, it can bring the user to pre-selected destinations as well.
The Robotic White Cane: Dr. Cang Ye and his team at Virginia Commonwealth University are developing a Robotic White Cane that can provide more specific guidance than the traditional version. With a sophisticated camera combined with LiDAR technology, it can help its user navigate the world with increased confidence.
Challenges of Outdoor Navigation: Both LYSA and the Robotic White Cane are currently better suited for indoor navigation. A major reason for that is the unpredictability of an outdoor environment along with more fast-moving objects, such as cars on the road. Researchers are working hard on overcoming this hurdle, but it still poses a major challenge.
The Speed of Innovation: When Dr. Ye began developing the Robotic White Cane a decade ago, the camera his team used cost $500,000 and had image issues. Now, their technology can be run on a smartphone – making the technology much more affordable, and hopefully one day, more accessible if it becomes available to the public.
Tweetables:
“We’ve had a relationship with dogs for 35,000 years. And a relationship with robots for maybe, you know, 50 years. So the ability of a robot to take over that task is a way off. But technology is moving quickly.” — Thomas Panek, President and CEO of Guiding Eyes for the Blind
“Outdoor navigation is a whole new world because if you go on the streets, it could be dangerous. You have to be very careful because you are driving a person, driving a human being.” — Kaio Ribeiro, Researcher at Vix Systems
“The first … camera we used, it's close to 500 grand. … But now … the iPhone’s LiDAR … works outdoors. … And it just took … a … bit more than 10 years.” — Dr. Cang Ye, Prof. of Comp. Sci. at Virginia Commonwealth University and Program Director, National Science Foundation
“It’s not the traditional … robot … that’s stiff. … We have to move into soft robotics … to accomplish the … activity … a dog can accomplish. … It’s a way off. … If … an engineering student … wants to get into soft robotics, … that’s where it will be.” — Thomas Panek
Pertinent Links:
Lighthouse Guild
Guiding Eyes for the Blind
LYSA Robot Guide
Robotic White Cane
33:12
Smart Cities and Autonomous Driving: How Technology is Providing Greater Freedom of Movement for People with Vision Loss
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
Navigating the world can be difficult for anyone, whether or not they have vision loss. Tasks like driving safely through a city, navigating a busy airport, or finding the right bus stop all provide unique challenges. Thankfully, advances in technology are giving people more freedom of movement than ever before, allowing them to get where they want, when they want, safely.
Smart Cities are putting data collection to work in a healthy way by providing information to make busy intersections more secure, sidewalks more accessible, and navigation more accurate. They’re providing assistance for all aspects of travel, from the front door to the so-called “last hundred feet,” while using automated technology to make life easier every step of the way.
And although fully autonomous vehicles are still on the horizon, the technology being used to develop them is being applied to improve other aspects of life in incredible ways. These applications are making the world more accessible, safer, and better for everyone, including people who are blind or visually impaired.
One example of this is Dan Parker, the “World’s Fastest Blind Man,” who has developed sophisticated guidance systems for his racing vehicles, as well as a semi-autonomous bicycle that could give people with vision loss a new way to navigate the world safely and independently.
The Big Takeaways:
Smart Cities. Greg McGuire and his team at MCity in Ann Arbor, Michigan are working on the concept of Smart Cities, which focus on using data to improve the everyday lives of their citizens. That means improving traffic intersection safety, greater accessibility options, providing detailed “last hundred feet” guidance, and much more.
Autonomous Driving. In a perfect world, self-driving cars will provide ease of transportation for everyone, and create safer, less congested roads. That technology isn’t there yet – but it’s being worked on by talented researchers like John Dolan, the Principal Systems Scientist at Carnegie Mellon’s Autonomous Driving Vehicle Research Center. Sophisticated sensors and advanced robot-human interfaces are being developed to make self-driving cars possible.
Application of Technology. Even though the technologies behind Smart Cities and autonomous vehicles are still being developed, they can still be applied to everyday life in exciting ways. Things like miniature delivery robots that can deliver goods, AI-powered suitcases that can help you navigate busy airports, or semi-autonomous bicycles are already here – and there’s more on the way.
The World’s Fastest Blind Man. When professional race car driver Dan Parker lost his vision in an accident, he felt lost. But a moment of inspiration led him and his business partner Patrick Johnson to develop a sophisticated guidance system that let him continue racing without human assistance. Thanks to this revolutionary technology, Dan became the “World’s Fastest Blind Man” when he set a land-speed record of 211.043 miles an hour in his customized Corvette.
Tweetables:
“One of the key pillars of MCity is accessibility. The four areas we think about are safety, efficiency, equity, and accessibility. … Accessibility is that we can make transportation systems available to as many of us as possible.” – Greg McGuire, Managing Director of MCity
“I became the first blind man to race Bonneville, with an average speed of 55.331 mph. And I returned in 2014 and set my official FIM class record … at 62.05 mph. … I’m the only blind land speed racer … with no human assistance.”– Dan Parker, the “World’s Fastest Blind Man”
“There are chairs, there are tables. ... We know we don’t want to run into them, but we do want to walk in the walkable space. … A car wants to drive in the drivable space.” – John Dolan, Principal Systems Scientist at Carnegie Mellon’s Autonomous Driving Vehicle Research Center
“Because we know autonomous technology is increasing every day and it’s coming, you know, a hundred percent it’s coming. You know, transportation is freedom and that’s exactly what that would bring us. Freedom.” – Dan Parker
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links:
Lighthouse Guild
MCity
Carnegie Mellon Autonomous Driving Vehicle Research Center
Dan Parker
39:12
Leveling Up Accessible Video Game Features: How New Technology is Making Gaming More Immersive and Inclusive for...
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
For decades, people with vision loss had limited options when it came to accessing video games. Aside from screen magnification and text-to-voice tools, gamers who are blind or visually impaired didn’t have many ways to play their favorite titles. But in recent years, the same cutting-edge technology used to create games has been used to also make them more accessible for people with vision impairment. These advances include more visibility options, the implementation of 3D audio, haptic feedback, and customizable controllers for gamers with vision impairment. Furthermore, 3D audio technologies being developed in live sports may soon make their way to online multiplayer video games. The implementation and improvement of these technologies mean that everyone will be able to play together, regardless of their visual acuity.
The Big Takeaways:
Leap in Accessible Gaming Options. In the past, the lack of accessibility features has made it much harder for gamers like Elaine Abdalla to access her favorite titles. But as gaming technology and accessibility improve, gamers like Elaine who are visually impaired are now able to more actively participate in their favorite titles. For example, The Last of Us: Part 2 boasts over 60 accessibility options including sophisticated visibility adjustments and inclusive difficulty settings.
Participating in the Process. Robin Spinks is the Head of Inclusive Design at the UK’s Royal National Institute of Blind People (RNIB). He helps gaming companies implement accessibility options as part of the design process from the very beginning. By ensuring people with vision loss are part of the process from the jump, Robin is helping make games more immersive than ever — for everyone.
Xbox Accessibility Team. Kaitlyn Jones first became passionate about gaming when she helped her father launch a nonprofit foundation that built customized setups for gamers with disabilities. She’s now the program manager of Xbox’s Gaming Accessibility Team. They develop guidelines for all kinds of accessible gaming options, including for people with vision loss. Thanks to her team’s efforts, new titles include options like spatial audio cues, which make it easier for gamers who are blind to navigate complex dungeons and unlock achievements.
Action Audio. At the 2022 Australian Open, a new technology called Action Audio became available for tennis fans with vision loss. Through cutting-edge camera tracking technology, Action Audio creates an information-rich 3D audio experience that allows tennis fans with vision loss to experience every thrilling rally as it happened, and to share in a communal experience.
Spatial Audio in Gaming. Action Audio’s developers hope to make it a universal technology that can also be used in gaming. Whether it’s working with a team in the latest first-person shooter or dribbling around opponents on a virtual basketball court, 3D (or spatial) audio technology is positioned to help people with vision loss to more equitably participate in the gaming community.
Tweetables:
“The motivation here is to remove barriers, to form partnerships, and to collaborate so that going forward it becomes a much more possible area of life for people. ’Cause gaming's fantastic, right? And why should you be stopped from enjoying it just because you’re blind or partially sighted?” — Robin Spinks, Head of Inclusive Design at RNIB
“When something happens on the court, it’s captured quite fast with Hawkeye within hundreds of milliseconds, and we’re able to take that data relatively quickly and generate the audio and then send it to the broadcast before it’s received in the real world.” — Tim Devine, AKQA Executive Director of Innovation
“It’s like that Japanese principle of continuous improvement, kaizen, where you’re kind of constantly looking to do things better. That’s what we want to see in the gaming world. — Robin Spinks, Head of Inclusive Design at RNIB
“We’ve always really prioritized accessibility along the way. But I think in terms of the journey of, even where we started when I first joined the team a few years ago versus now, I think the bar honestly just keeps getting higher and higher.” — Kaitlyn Jones, Program Manager: Xbox Accessibility Team
“I’m just happy to see the disabled community is finally getting the assistance they need. It’s taken a while and it’s still gonna take a while, but we’re going in the right direction.” — Elaine Abdalla, creator of BlindGamerChick YouTube Channel
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links:
Lighthouse Guild
RNIB Accessible Video Games Page
Xbox Accessibility Guidelines
BlindGamerChick YouTube Channel
On Tech & Vision: Training the Brain: Sensory Substitution
31:05
A Celebration of Sound and Song: Music Tech Shines the Spotlight on Musicians with Vision Loss
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
Marcus Roberts, Stevie Wonder, Ray Charles, and even Louis Braille (who invented the Braille Music Notation system still used today) prove that musicians who are blind or visually impaired have made profound impacts on our musical landscape. However, to get their work to us, musicians who are blind have had to structure complex workarounds, like relying on sighted musicians to demonstrate complex scores; memorizing long pieces; or only performing when they can have a Braille score in front of them, shutting them out from opportunities that fall to those who can sight read, since Braille scores have often been time-consuming and expensive to produce. However, new technologies in music composition and production are making composition, nuanced scoring, and Braille printing easier than ever, bringing musicians and composers who are blind to centerstage to share their sound and song.
The Big Takeaways:
“Lullay and Lament” by James Risdon. The recorder — pivotal in music from the Renaissance and Baroque periods — only lately has emerged from a long period of obsolescence. James Risdon, a passionate player who lost his vision due to Leiber’s congenital amaurosis as a child, has written the original song for recorder “Lullay and Lament” for his album Echoes of Arcadia, which marries the early recorder with contemporary recorder music. To make this album, he relied on new musical technologies like Dancing Dots.
Dancing Dots with Bill McCann. Bill McCann is the founder and president of Dancing Dots Braille Music Notation software. Dancing Dots is a suite of software — plus educational resources and training — that helps people who are blind to read, write and record their music. McCann founded Dancing Dots in 1992.
Chris Cooke and PlayHymns.com. When Chris Cooke got the Dancing Dots software in 2016, her creativity exploded. She was able to arrange a song and bring it to church and play a duet with a member of the congregation, something she hadn’t been able to do before, given the former time and cost of translating music into Braille notation.
What is Braille Music? Louis Braille, a noted musician, created the Braille musical notation system. Being able to translate music easily between Braille Musical Notation and Western musical notation, and to easily print either of these, helps musicians who are blind share music with other musicians, both sighted and blind, and play music together with ease.
Musical Instrument Digital Interface and MusicXML. MIDI has made it possible for musicians to play music into their instruments and have it automatically translated into digital musical notation. MusicXML has made universal the file type for a score and allows musicians to share scores across popular music notation applications like Finale or Sebelius.
The question of parity. James and Chris agree that while Dancing Dots technology has enabled them to take advantage of new musical opportunities, no technology exists that offers them complete parity with sighted musicians because musicians who are blind need additional lead time to get the music scanned correctly or to memorize pieces. Chris adds that preparing the music in a timely fashion and on a budget would help.
The MIDI-to-brain connection. Bill McCann has explored using the Brainport, a technology from WICAB (which we profiled in an episode from September 2021, on “Training the Brain: Sensory Substitution”) to allow musicians who are blind to read music on their tongues. This is important if someone needs to read music live in a performance in order to play any instrument that also requires their hands. Early trials showed signs of success. He posits that maybe someday, maybe soon, people could think new music into notation.
Tweetables:
“I said, I will never put myself in this position again. If I write something and I am asking other people to play it, and they ask me questions or there’s something I am going to know before we meet exactly what I want and what I have.” — Bill McCann, founder and president of Dancing Dots
“So the duet that we played was hot off my printer and went with me and we played it. And it was great to be able to share music in that way because of the technology in the Dancing Dots program.” — Chris Cooke, musician and music arranger, creator of Playhymns.com
“As a blind person, I can say this for myself, often we end up following sighted people, or following somebody. Braille music gives blind musicians the chance to become leaders.” — Bill McCann, founder and president of Dancing Dots
“I've set aside 2023 as a year. I'd really like to kind of develop some more expertise in the area and also come to grips with some of the technology that would help the process.” — James Risdon, musician and recorder player
“Someone sitting there and getting inspired is what we call the MIDI-to-Brain connection. We’re not there yet, but you could … think the music in your head … at a computer and … music materializes in the form of a score.” — Bill McCann, founder and president of Dancing Dots
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links:
Lighthouse Guild
Dancing Dots
James Risdon
Chris Cooke and Playhymns.com
34:39
Ambient Computing and Voice Assistants: From Your Home to the Stars
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
Lots of people have voice-controlled smart home assistants like Siri, Google, or Alexa in their homes…. to listen to the news or to set timers. But they can do so much more! David Frerichs, Principal Engineer, Alexa Experience at Amazon on the aging and accessibility team, shares his design philosophy for making voice assistants more inclusive, and the preferred mode of engagement for every user. He also shares that the next stage of smart home assistants will be ambient computing, where your devices will intuit your needs without you speaking them. We talk with Lighthouse Guild client Aaron Vasquez, who has outfitted his home with smart home technology, and with Matthew Cho, a client who traveled to the Johnson Space Center in Houston to speak to the unmanned Orion Spacecraft via the Amazon Alexa on board, demonstrating that voice assistant technology can bring inclusivity and accessibility to many jobs and industries and are not just for the home anymore.
The Big Takeaways:
Alexa Onboard the Orion Spacecraft. NASA partnered with Amazon to use the Alexa voice-controlled smart assistant onboard the unmanned Orion spacecraft so that engineers could guide the spacecraft from Mission Control. This project tested the possibility of Alexa for space travel while demonstrating that voice-controlled smart assistants have uses beyond the home. Matthew Cho, an 18-year-old student and client of Lighthouse Guild had the opportunity to travel to Houston to be one of the volunteers to give voice commands to the spacecraft via an Amazon Alexa device while it hurtled through space.
Accessibility and Preferences. David Frerichs, Principal Engineer, Alexa Experience at Amazon, and someone who works on the aging and accessibility team, has spent his career developing technology that adapts to the ever-changing needs of the user and has cultivated a design philosophy that makes clear that design choices (like voice control) that enable inclusivity for people who are blind can also become the preferred way that most users engage with a device or a tool. Curb cuts are a historical example. David often thinks in terms of “Hot tub safe computing.” What can a person do to engage with the device from their hot tub?
Ambient Computing. David mentions ambient computing, the next phase of smart home technology, in which a network of devices in the built environment intuit and respond to a user's needs without the user even needing to speak a command.
Smart Homes Today, Smart Industries Tomorrow. Aaron Vasquez is a smart home user. He uses Google Echo to power two smart lamps, operate his smart TV, and control a pet camera to oversee his rambunctious kitties when he’s not at home. As a person who is visually impaired, Aaron prefers voice command for running his home in this way. This episode asks how the smart home’s tools can be integrated into offices and industries to make these more accessible and inclusive spaces.
Tweetables:
“To be able to have a rocket be dependent on an AI without anybody having to control the spacecraft is, it is really amazing, and I feel that later on that they’ll be able to use it for much more things aside from space.” – Matthew Cho, student and client of Lighthouse Guild
“So their goal was to eventually get different people to be able to go into space … They were trying to see if Alexa would work properly with all sorts of people, normal people. Not just astronauts, like regular, ordinary, everyday people.” – Matthew Cho, student and client of Lighthouse Guild
“You have … permanent need, temporary need, situational need, and preferential need that really can inform us on … how we can address [a] barrier for the particular core use case. But if we do it well, it will serve a much broader community.” – David Frerichs, Principal Engineer, Alexa Experience at Amazon
“We’re moving toward … ambient computing. That is … that the system should be able to respond to the needs of the customer, even if the customer doesn’t say anything. ... That’s … where the boundaries are and where it’s gonna continue to be pushed.” – David Frerichs, Principal Engineer, Alexa Experience at Amazon
“We had seen this pack of smart bulbs and they were relatively cheap, and I was like, huh, that’s, that’s kind of cool. So we were like, you know what? Let’s get it. So we got ‘em, we hooked them up and that kind of is what started everything.” – Aaron Vasquez, smart home user and client of Lighthouse Guild
“Honestly, it’s so much easier if I can ask Google a question and she can come up with the answer, then I’m better off that way instead of actually trying to look it up myself.” – Aaron Vasquez, smart home user and client of Lighthouse Guild
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links:
Lighthouse Guild
Alexa in Space
David Frerichs
34:52
New Approaches in Access: Smart Tools for Indoor Navigation and Information Transfer
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
Artifacts from Blackbeard’s sunken pirate ship are on display in the North Carolina Maritime Museum in Beaufort, North Carolina. But now they are also accessible to visitors who are blind, thanks to the efforts of Peter Crumley, who spearheads the Beaufort Blind Project. In this episode, we ask: How can new technology help make sites like these as accessible to people who are blind as they are to sighted people? We profile three companies applying new technologies paired with smartphone capabilities, to make strides in indoor navigation, orientation, and information transfer. Idan Meir is co-founder of RightHere, which uses Apple’s iBeacon technology to make visual signage dynamic and accessible via audio descriptions. We check in with Javier Pita, CEO of the NaviLens QR code technology which we profiled in our first season to see what they have been developing in the last two years. Rather than iBeacons or QR codes, GoodMaps uses LiDAR and geocoding to map the interior of a space. We speak with Mike May, Chief Evangelist. Thanks to Peter Crumley, the North Carolina Maritime Museum is fully outfitted with GoodMaps, and will soon have NaviLens as well. As the prices of these tools come down, the key will be getting them into all the buildings, organizations, and sites of information transfer that people who are blind need to access – which is all of them.
The Big Takeaways:
Beaufort Blind Project. Peter Crumley, a blind resident of Beaufort, North Carolina, has advocated having accessibility tools brought to various parts of his hometown. Along the way, he helped the North Carolina Maritime Museum outfit itself with GoodMaps technology for indoor navigation, and with NaviLens QR codes for information transfer. Thanks to these new technologies, the museum artifacts are now accessible to everyone.
RightHear. Idan Meir cofounded RightHear, which uses iBeacon technology paired with users’ smartphones to guide visitors who are blind through an indoor space. iBeacons send unique signatures via low Bluetooth signals to phones inside the radius. When these iBeacons are paired with areas of interest in a space (e.g. the front door, the counters, or the bathrooms) users can orient themselves within a space, and identify where they want to go and how they want to navigate to each location. RightHear translates the information embedded in each beacon into audio feedback for users. On the subject of feedback, Idan Meir is looking for beta testers to try out RightHear and provide him with constructive feedback.
NaviLens. We profiled NaviLens QR code technology in an episode from our first season. In this episode, we follow up with Javier Pita to see what has been in development in the last two years. Since we last spoke, NaviLens has launched NaviLens 360, which uses magnets to help guide users who are blind to the NaviLens codes, even if their camera is having trouble picking up the code, making the app even more user-friendly. In addition, NaviLens has launched a partnership with Kellogg’s in Europe and North America to test the effectiveness of the Navilens code on consumer product packaging.
GoodMaps. GoodMaps uses LiDAR technology to map a space. Lasers are sent out from the LiDAR sensor, and when they bounce back, they have captured distances from the point of origin. Institutions work with GoodMaps to pay for the mapping service, and then users can access the maps for free. The app uses audio to communicate navigational directions with users.
Technological advancement. Each of these tools relies on component technologies that have gotten less expensive in recent years (iBeacons, QR Codes, and LiDAR). They are also able to exist because their target markets carry smartphones in their pockets, enabling these potential users to access the tools quickly and easily, without much additional hassle or investment.
Distribution. In this episode, we profile three different approaches to broadening access to indoor navigation technology, including for orientation and information transfer, proving there are many ways to solve these problems. It is good that some of these tools can be paired, as has been done at the North Carolina Maritime Museum, and that users may be able to choose which tools work best for them. The key will be getting them into all the buildings, organizations, and sites of information transfer that people who are blind need to access – which is all of them.
Tweetables:
“The advocacy is so important; when you’re actually interfacing with the app to make the app better and make it work in a way that a blind person really needs it to work.” – Peter Crumley, Beaufort Blind Project
“Well, it's gonna be built from blind perspective philosophy. So not only will it work for me — it will work for anyone, totally blind and fully sighted to give an interactive experience.” – Peter Crumley, Beaufort Blind Project
“Imagine, if this technology will be in all the products, we will solve the problem of accessible packaging for all users.” – Javier Pita, NaviLens
“The point is we have solved the last few yards of the wayfinding problem that is super important for a blind user. And this was born in New York City with the collaboration with the MTA and the department of transportation of New York City.” – Javier Pita, NaviLens
“That camera picks up the environment and it compares it with that point cloud and says, “I see based on this particular image … that you are near the Starbucks,” or “You're near Gate 27.” –Mike May, GoodMaps
“It was important and kind of obvious for us from the very early on, you know, that nothing about us without us. It was clear to us that we have to involve users in the process. –Idan Meir, RightHear
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links:
Lighthouse Guild
RightHear
NaviLens
GoodMaps
39:12
AI Revolutionizes Vision Tech, Ophthalmology, and Medicine as We Know It
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
In 1997, Gary Kasparov lost an epic chess rematch to IBM’s supercomputer Deep Blue, but since then, artificial intelligence has become humanity’s life-saving collaborator. This episode explores how AI will revolutionize vision technology and, beyond that, all of medicine.
Karthik Kannan, co-founder of AI vision-tech company Envision, explains the difference between natural intelligence and artificial intelligence by imagining a restaurant recognizer. He describes how he would design the model and train it with positive or negative feedback through multiple “epochs” — the same process he used to build Envision. Envision uses AI to identify the world for a blind or visually-impaired user using only smartphones and smart glasses.
Beyond vision tech, AI enables faster and more effective ophthalmic diagnosis and treatment. Dr. Ranya Habash, CEO of Lifelong Vision and a world-renowned eye surgeon, and her former colleagues at Bascom Palmer, together with Microsoft, built the Multi-Disease Retinal Algorithm, which uses AI to diagnose glaucoma and diabetic retinopathy from just a photograph. She acquired for Bascom Palmer a prototype of the new Kernal device, a wearable headset that records brain wave activity. Doctors use the device to apply algorithms to brainwave activity, in order to stage glaucoma, for example, or identify the most effective treatments for pain.
Finally, AI revolutionizes drug discovery. Christina Cheddar Berk of CNBC reports that thanks to AI, Pfizer developed its COVID-19 treatment, Paxlovid, in just four months. Precision medicine, targeted to a patient’s genetic information, is one more way AI will make drugs more effective. These AI-reliant innovations will certainly lower drug costs, but the value to patients of having additional, targeted, and effective therapies will be priceless.
The Big Takeaways:
Natural vs. artificial intelligence, and the “restaurant recognizer.” Karthik Kannan, CEO and co-founder of Envision explains the difference between natural and artificial intelligence by describing how humans recognize restaurants in a foreign city and comparing that to how he’d train a “restaurant recognizing algorithm.” Here’s a hint: the algorithm needs a lot more data.
Sensor fusion AI. AI developers are interested in using different types of sensors together to give the algorithms a sense of the world closer to human intelligence. One example is the use of LiDAR in the Envision app, in addition to the phone camera.
Transhumanism. Humans don’t have LiDAR. Does that mean AI will surpass human capability? Karthik offers that some radiology AI have higher accuracy than human radiologists, but he thinks it will be much more of a partnership between the human and the machine.
Multi-Disease Retinal Algorithm. Dr. Ranya Habash and her colleagues at Bascom Palmer Eye Institute worked with Microsoft on an AI diagnostic tool. They fed the algorithm 86,000 images of eyes, labeled with relevant diseases, and taught the machine to diagnose eye disease with just a photograph, making remote diagnosis not just possible but inexpensive.
The Brain-Machine Interface. Dr. Habash wrote a grant that earned Bascom Palmer a prototype of the Kernal device, a helmet-like device that measures brainwave activity. Doctors used this device to create a “brain-machine interface” which advances brain research on glaucoma, diabetic retinopathy, Alzheimer’s, and pain management.
Bias in AI. Karthik Kannan reminds us that the biggest threat that humanity faces from AI is bias encoded in the algorithms. This is a real harm that humans have already experienced, and AI engineers need to be extremely sensitive to ensure they are not encoding their own biases.
AI for Drug Discovery. Christina Cheddar Berk, a reporter for CNBC, shares how the pace of drug discovery is set to speed up, thanks to AI algorithms and supercomputing power that can cycle through millions of possible chemical compounds per second to I.D. effective options. Pfizer used a similar process to develop Paxlovid, in a process that took only four months.
Tweetables:
“The secret sauce is always in the data.” — Karthik Kannan, CEO and Co-Founder of Envision
“Human intelligence is so holistic. We have so many sensors on our bodies. […] Whereas an AI is taught only images.” — Karthik Kannan, CEO and Co-Founder of Envision
“I know what’s going to work and what’s not going to work within thirty seconds of seeing it. […] They need to show up with a smartphone. Then I’ll take them seriously.” — Dr. Ranya Habash, CEO, Lifelong Vision
“I don’t think there’s anything more powerful in medicine than to be able to treat a patient and get rid of a problem that is plaguing them so much.” — Dr. Ranya Habash, CEO, Lifelong Vision
“If you can measure it you can control it.” — Dr. Ranya Habash, CEO, Lifelong Vision
“It strongly takes over the bias of whoever is actually feeding the data […] and I think that has much, much more potential for harm than an AI taking over humanity.” — Karthik Kannan, CEO and Co-Founder of Envision
“The value for patients of having those additional therapies available; it’s hard to put a price on.” — Christina Cheddar Berk, reporter, CNBC
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links:
Lighthouse Guild
Karthik Kannan
Dr. Ranya Habash
Zephin Livingston
Christina Cheddar Berk
35:50
Balancing Innovation and Ethics: Who is Protecting the Early Adopters?
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
Innovations in implant technology are advancing at lightning speed, profoundly impacting the lives of people who are blind or visually impaired. In On Tech And Vision, we’ve profiled some amazing new implant technologies that have the potential to restore people’s sight. But in this episode, we pump the breaks — because we need to address a critical part of the innovation process: the ethical frameworks that protect participants in early clinical trials, and the need for an updated framework that ensures patient protections without stifling innovation and development.
Discussions between doctors and participants in clinical trials almost always focus on the new technology and very rarely on the manufacturer who sponsors the clinical trial — and almost never on the long-term commitment and financial viability of the company sponsoring the technology. And while clinical trial informed consent includes whose responsibility it is to remove the implants should they fail during the trial, that responsibility usually ends once the trial is over. At that stage, who will maintain or remove the implants that are still housed in patients’ bodies?
In this episode, we talk about innovative implants such as the Argus II, which we featured in the first season of On Tech And Vision. The Argus II is a microchip implanted under the retina that, in combination with a special headset, provided some vision to people who otherwise had none. And while the technology was exciting, the company discontinued the retinal implant three years ago, and the Argus II was eventually sold to another pharmaceutical company.
Dr. Joseph Fins, Professor of Medical Ethics and Professor of Medicine at Weill Cornell Medical Center in New York, joins us to share his thoughts on today’s big idea: How do we balance the life-changing potential of electroceutical implant technology with the ethics of caring for early participants — particularly after clinical trials are over?
The Big Takeaways:
Examples of electroceutical implants. Cochlear implants, retinal implants, and deep brain stimulators are examples of scientific advances that rely on in-dwelling devices.
Regulatory framework today. The relationships between researchers and clinical trial participants are regulated by institutional review boards, which came out of the National Research Act of 1974. However, while this framework works well for drug trials, new issues specific to implants need to be addressed by new regulations. For example, who is responsible for people left with in-dwelling devices once trials are over? If the sponsoring company no longer supports their devices, are they victims of abandonment? Are the timelines for drug trial success too short to be relevant for implant device trials, since it may take the body longer to adopt a new technology than to respond to a new drug?
Ancillary care obligations. Henry Richardson, in his book Moral Entanglements: The Ancillary Care Obligations of Medical Researchers writes that historically, researchers — to avoid conflicts of interest — did not assume a clinical care role. However, that is changing, as researchers realize they have an obligation to share actionable results with patients. The result is that there is even less of a “bright line distinction,” as Dr. Fins says, “between research and therapy.”
Collective responsibility. Who is responsible for the long-term well-being of participants in electroceutical trials? Dr. Fins suggests that the sponsoring company, the medical school where the research is taking place, and the government should share responsibility. It’s a collective problem, he says.
Some solutions. Requiring researchers, sponsoring companies, and researching universities to include in the costs of development insurance to cover long-term care for participants is one potential solution that Dr. Fins imagines. He also offers that researchers and sponsoring companies that develop successful and adopted medical products could subsidize the field. Or, he suggests, a tax on gaming devices (adjacent to electroceutical implants) to sustain people who are given indwelling devices in clinical trials.
The law. The law needs to evolve to address the specific vulnerabilities of participants in electroceutical implant trials. Dr. Fins suggests that there are provisions within the Americans with Disabilities Act that account for assistive technologies that were relevant when the act was written in the nineties. According to Dr. Fins, these provisions in that could be read with a more contemporary lens, to include the assistive technologies of today (which would encompass electroceutical implants). There is room for lawyers and legal scholars to impact the legal frameworks in place now, to expand coverage from the ADA to protect participants in clinical trials for electroceutical implants.
“Victims of Our Own Success.” Electroceutical implants are a miracle, says Dr. Fins. They are human ingenuity at its best. The science is harder to solve than the bureaucracy, but the bureaucracy to sustain medical advancements like these must catch up, or, ultimately, the vulnerability of trial participants threatens to impede scientific progress.
Danger to the field. Clinical trials rely on willing participants, and when participants are not supported after trials end, it erodes participants’ trust across the field. Without a clear set of protections in place for participants in clinical trials, scientific and medical advancement in the area of electroceutical implants may be impeded.
Tweetables:
“I think this is a perfect rationale for insurance.”— Dr. Joseph Fins, Weill Cornell Medical Center
“This is a huge problem. … We’re victims of our own success.”— Dr. Joseph Fins, Weill Cornell Medical Center
“It's human ingenuity at its very best. And the fact that we can’t figure out the bureaucracy to sustain this? … The science is harder than the politics and the bureaucracy, but we’re being overmatched by the politics and the bureaucracy.” — Dr. Joseph Fins, Weill Cornell Medical Center
“When stories like this come out it makes recruitment very hard.” — Dr. Joseph Fins, Weill Cornell Medical Center
“These retinal implants, these deep brain stimulators, … they're gonna be looked upon as primitive halfway technologies 50 and 100 years from now. But we're only gonna get there if we're able to do this research.” — Dr. Joseph Fins, Weill Cornell Medical Center
“Once you understand these facts, the ethics are pristine. They’re clear.” — Dr. Joseph Fins, Weill Cornell Medical Center
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links:
Lighthouse Guild
Dr. Joseph Fins
31:00
People Choice Awards 2022
Episode in
On Tech & Vision With Dr. Cal Roberts
Listener nominations are open from July 1st - July 31, 2021. On July 1st visit: https://www.podcastawards.com, sign-up and vote
00:44
Tools for Success: Tech Convergence and Co-Designed Products Close Gaps for Children Who are Blind
Episode in
On Tech & Vision With Dr. Cal Roberts
This podcast is about big ideas on how technology is making life better for people with vision loss.
People who are blind or visually impaired know all too well the challenges of living in a sighted world. But today, the capabilities of computer vision and other tech are converging with the needs of people who are blind and low-vision and may help level the playing field for young people with all different sensory abilities. These tools can pave the way for children’s active participation and collaboration in school, in social situations, and eventually, in the workplace, facilitating the important contributions they will make to our world in their adult lives.
Access to educational materials is a consistent challenge for students and adults who are blind, but Greg Stilson, the head of Global Innovation at American Printing House for the Blind (APH), is trying to change that. Together with partner organizations Dot Inc. and Humanware, APH is on the verge of being able to deliver the “Holy Braille” of braille readers, a dynamic tactile device that delivers both Braille and tactile graphics in an instant, poised to fill a much-needed gap in the Braille textbook market. Extensive user testing means the device is as useful for people who are blind as possible. Greg sees a future in which more inclusively designed and accessible video games, augmented reality (AR), and virtual reality (VR) will help children who are blind learn with greater ease, and better engage with their sighted peers.
Enter Dr. Cecily Morrison, principal researcher at Microsoft Research in Cambridge, UK. Based on extensive research and co-designing with people who are blind, she and her team developed PeopleLens, smart glasses worn on the forehead that can identify the person whom the user is facing, giving the user a spatial map in their mind of where classmates (as one example) are in space. PeopleLens helps children who are blind overcome social inhibitions and engage with classmates and peers, a skill that will be crucial to their development, and in their lives, as they move into the cooperative workspaces of the future.
The Big Takeaways:
Robin Akselrud, an occupational therapist and assistant professor at Long Island University in Brooklyn, author of MY OT Journey Planner and The My OT Journey Podcast, explains how a baby who is born blind becomes inhibited from their first developmental milestones. She explains the stressors that these children might face upon attending school and describes the kinds of interventions that occupational therapy offers.
Bryce Weiler, disability consultant, sports enthusiast, and co-founder of the Beautiful Lives Project, emphasizes how important it is for children who are blind or low-vision to have rich sensory experiences — and life experiences — which give them a chance to flourish and socialize with peers. Beautiful Lives Project offers opportunities to do that.
Greg Stilson, Director of Global Innovation at American Printing House for the Blind, and his team are developing a dynamic tactile device (DTD) that can switch seamlessly between Braille and tactile graphics — the “Holy Braille” of braille devices. The DTD is made possible by developments in pin technology by Dot Inc, and APH. Humanware developed the software for the device. No longer using the piezoelectric effect to move pins has reduced the cost of the device significantly, and APH can funnel federal funds to reduce the price further, making the DTD a potential, viable option for institutions.
Cecily Morrison, principal researcher at Microsoft Research in Cambridge UK, and her team developed PeopleLens, a head-worn pair of smart glasses that lets the wearer know who is in their immediate vicinity. Dr. Morrison and her team tested it in classrooms for school-age children who are blind or visually impaired and found that PeopleLens reduces students’ cognitive load and helps young people overcome social anxiety and inhibitions that Robin Akselrud described at the top of the show. Wearers of PeopleLens learn to develop mental models of where people are in a room, and gain the confidence to engage others, or not, as they choose. Once social skills are built, students no longer have to wear the device but are set up for more successful social interactions at school and in their lives to come.
Tweetables:
If they have a visual impairment, it really impacts them from early on, from that first development milestone. — Robin Akselrud, occupational therapist and assistant professor at Long Island University in Brooklyn, author of MY OT Journey Planner and The My OT Journey Podcast
For children, just giving them that foundation of making friendships as they’re growing up, and the opportunity to be a part of something, sport can allow them to do that, and it also gives them the chance to do things that their peers are taking part in. —Bryce Weiler, disability consultant, sports enthusiast, and co-founder of the Beautiful Lives Project
This was what the field regards as the “Holy Braille” right? Having both [Braille and tactile graphics] on the same surface. —Greg Stilson, Director of Global Innovation at American Printing House for the Blind
With the advancements of virtual reality and augmented reality, … along with the idea of making experiences and video games and things like that more inclusive, it’s going to create a more inclusive way for blind kids to engage with their sighted peers. — Greg Stilson, Director of Global Innovation at American Printing House for the Blind
We found that “people” was the thing that was most interesting to people. And that doesn’t surprise us. We are people, and we like other people. — Dr. Cecily Morrison, principal researcher at Microsoft Research in Cambridge UK
They can go out and find someone that they want to play with. They can choose not to talk to somebody by turning their head away from them, and the moment that they understand the agency they have in those situations is when we see a significant change in their ability to place people and to engage with them. — Dr. Cecily Morrison, principal researcher at Microsoft Research in Cambridge UK
When we look at the workplaces of today, they’re often very collaborative places. So you can be the best mathematician in the world, but if you struggle to collaborate, you’re not building the AI technologies of tomorrow. By helping kids ensure that they have a strong foundation in these attentional skills, we’re giving them a significant lift. — Dr. Cecily Morrison, principal researcher at Microsoft Research in Cambridge UK
Contact Us:
Contact us at podcasts@lighthouseguild.org with your innovative new technology ideas for people with vision loss.
Pertinent Links:
Lighthouse Guild
Robin Akselrud
Bryce Weiler
Greg Stilson
Dr. Cecily Morrison
32:24
You may also like View more
Hablando Crypto
¿Te interesan las criptomonedas? A nosotros también. Somos Óscar y Cristian. Después de más de 5 años jugueteando con las criptomonedas os explicamos nuestras historias. También hablamos sobre como vemos el crypto-mundo y hacia donde creemos que irá. Updated
Loop Infinito (by Xataka)
Loop Infinito es un podcast diario de Xataka presentado por Javier Lacort.
Un nuevo episodio cada día de lunes a viernes que analiza la actualidad tecnológica dando contexto y perspectiva.. Updated
Applelianos
Podcast de tecnología, principalmente de noticias sobre el mundo de Apple, con un grupo de compañeros expertos en los temas expuestos, descubre la información adecuada con nosotros. Updated




