Arturo Estrella has a message for recession naysayers: It could hit sooner than you think.Marketsread more
Salesforce released its first earnings report since its $15.3 billion acquisition of Tableau Software, the company's largest deal ever.Technologyread more
Fed Chairman Jerome Powell faces the tough challenge of presenting a unified voice on Fed policy from the most divided Fed in years.Market Insiderread more
VMware is following through on its proposal to buy Pivotal, a fellow Dell subsidiary, and expanding into cybersecurity with the acquisition of Carbon Black.Technologyread more
Google says it shut down hundreds of YouTube channels tied to misinformation around the Hong Kong protests.Technologyread more
It is a rare scenario where long-term interest rates suddenly fall below short-term interest rates.Real Estateread more
Investors are rushing to get a piece of its privately held rival Impossible Foods before it goes public, according to the Wall Street Journal.Food & Beverageread more
Weisler has been CEO at the company since 2015 when it split from HPE.Technologyread more
Gap Inc.'s fiscal second-quarter earnings topped analysts' estimates but sales missed. Same-store sales dropped 4% during the period, worse than expected.Retailread more
The deal adds the independent studio with preschool brands such as Peppa Pig to the U.S. company known for Nerf and Power Rangers.Entertainmentread more
Dallas Fed President Robert Kaplan would like to avoid additional stimulus but is keeping an "open mind."The Fedread more
A computer that aims to translate thoughts into natural sounding speech has been hailed by its developers as an "exhilarating" breakthrough.
Researchers from the University of California, San Francisco, designed the system – a computer simulation that turns brain signals into a virtual voice – to help restore speech to people with paralysis or neurological damage. They published their paper in the scientific journal "Nature" on Wednesday.
The device works by using a brain-computer interface (BCI), which works out a person's speech intentions by matching brain signals to physical movements they would usually activate in a person's vocal tract – their larynx, jaw, lips and tongue. The data is then translated by a computer into spoken words. The same technique has been used to generate limb movement in people with paralysis.
Previous BCI systems for speech facilitation have focused on typing, generally allowing people to type a maximum of 10 words per minute – massively lagging behind the average speaking speed of around 150 words per minute.
Scientists worked with five volunteers whose brain activity was being monitored as part of a treatment for epilepsy. The researchers recorded activity in a language-producing region of the brain as the volunteers read several hundred sentences aloud.
Researchers working on the project claimed their computer system would not only restore speech, but could eventually reproduce the "musicality" of the human voice that conveys a speaker's emotions and personality.
"For the first time, this study demonstrates that we can generate entire spoken sentences based on an individual's brain activity," Edward Chang, professor of neurological surgery and the study's senior author, said in a press release. "This is an exhilarating proof of principle that with technology that is already within reach, we should be able to build a device that is clinically viable in patients with speech loss."
Gopala Anumanchipalli, a speech scientist who led the research, said the breakthrough came by linking brain activity to movements in the mouth and throat during speech, rather than associating brain signals to acoustics and sounds.
"We reasoned that if these speech centers in the brain are encoding movements rather than sounds, we should try to do the same in decoding those signals," he said in the press release.
Up to 69% of the words generated by the computer were accurately identified by people asked to transcribe the computer's voice. Researchers said this was a significantly better rate than had been achieved in previous studies.
"We still have a way to go to perfectly mimic spoken language," said Josh Chartier, a bioengineering graduate student who worked on the research. "We're quite good at synthesizing slower speech sounds like 'sh' and 'z' as well as maintaining the rhythms and intonations of speech and the speaker's gender and identity, but some of the more abrupt sounds like 'b's and 'p's get a bit fuzzy. Still, the levels of accuracy we produced here would be an amazing improvement in real-time communication compared to what's currently available."