top of page

OpenAI Announces ChatGPT's Ability to Talk, Listen, and Understand Images

OpenAI has recently announced that their ChatGPT can now "see, hear, and speak." This upgrade to the chatbot will become available to subscribers in the next fortnight. With OpenAI taking steps to push the feature, the competition concerning artificial intelligence among OpenAI, Microsoft, Google, and Anthropic is only set to get more intense. On Monday, OpenAI announced a major update to their ChatGPT system, giving it the ability to "see, hear and speak." This will include the option of five different synthetic voices on the mobile app, as well as image processing capabilities. Rolling out to paying users in the next two weeks, these features will be available on all platforms, with the exception of voice which will only be available on iOS and Android.This update comes in the midst of a heated artificial intelligence arms race between OpenAI, Microsoft, Google, and Anthropic. Google has made recent announcements for their Bard chatbot, while Microsoft has added visual search to Bing. Additionally, OpenAI was the recipient of Microsoft's massive $10 billion investment earlier this year.This has raised new concerns over AI-generated synthetic voices, as they can lead to more convincing deepfakes. OpenAI responded by saying that the voices used were from actors they worked with directly, rather than sourced from strangers. The company also emphasized that audio clips not be used to improve models and that transcriptions are considered inputs which may be utilized to improve language models. Despite this, OpenAI did not provide specifics on how it will use consumer voice inputs or how it will secure the data.

Comments


bottom of page