Add to Favourites
To login click here

OpenAI and Google are showcasing their latest AI technology, focusing on making AI models multimodal. This allows for seamless switching between different functions, such as talking, seeing, and hearing. Both companies have demonstrated their models’ abilities to process video and audio, with OpenAI’s GPT-4 Omni already rolling out to Premium users. Google’s Project Astra is still in the early stages, but shows potential for identifying objects through visual input. Multimodal AI is seen as the next step in making AI more natural and human-like in everyday life.