{"id":14445,"date":"2024-06-18T10:38:43","date_gmt":"2024-06-18T07:38:43","guid":{"rendered":"https:\/\/forklog.com\/en\/google-deepmind-develops-ai-model-for-video-soundtrack-generation\/"},"modified":"2024-06-18T10:38:43","modified_gmt":"2024-06-18T07:38:43","slug":"google-deepmind-develops-ai-model-for-video-soundtrack-generation","status":"publish","type":"post","link":"https:\/\/u1f987.com\/en\/google-deepmind-develops-ai-model-for-video-soundtrack-generation\/","title":{"rendered":"Google DeepMind Develops AI Model for Video Soundtrack Generation"},"content":{"rendered":"<p>Google DeepMind is <a href=\"https:\/\/deepmind.google\/discover\/blog\/generating-audio-for-video\/\">developing<\/a> an artificial intelligence technology to create soundtracks for videos.<\/p>\n<p>Google&#8217;s AI research division and other organizations have previously created models for video, but these have been unable to generate accompanying sound effects. To address this, DeepMind is employing V2A (video-to-audio) technology.<\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\"><p>\n<cite>\u201cVideo generation models are advancing at an incredible pace, but many current systems do not produce a soundtrack. One of the next important steps towards film generation is the creation of soundtracks for these silent videos,\u201d stated DeepMind.<\/cite><\/p><\/blockquote>\n<p>DeepMind&#8217;s V2A technology uses prompts in conjunction with video to create music, sound effects, and dialogue. For example: \u201cPulsating underwater jellyfish, marine life, ocean.\u201d The underlying diffusion AI model of V2A is trained on a database of sounds, dialogue transcripts, and video clips.<\/p>\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\">\n<div class=\"wp-block-embed__wrapper\">\n<iframe loading=\"lazy\" title=\"V2A Horror\" width=\"500\" height=\"281\" src=\"https:\/\/www.youtube.com\/embed\/b6Elcke3JMc?start=9&#038;feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe>\n<\/div>\n<\/figure>\n<p>The following prompts were used to create sound for the video: cinema, thriller, horror film, music, tension, atmosphere, footsteps on concrete.<\/p>\n<p>DeepMind acknowledges that the technology is not yet perfect, and the sound cannot be described as high-quality or convincing. Further refinements and testing are required before the full launch of V2A.<\/p>\n<p>In February, OpenAI <a href=\"https:\/\/u1f987.com\/en\/news\/openai-unveils-sora-enthusiasm-and-challenges\">introduced a new generative AI model, Sora<\/a>, which allows text to be transformed into video.<\/p>\n<p>In June, scientists from Harvard and DeepMind created a virtual rat with artificial intelligence as its brain.<\/p>\n<p>Previously, Google&#8217;s subsidiary <a href=\"https:\/\/u1f987.com\/en\/news\/deepmind-unveils-genie-an-ai-model-crafting-games-from-prompts\">unveiled the generative AI model Genie<\/a> for creating games.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Google DeepMind is developing an artificial intelligence technology to create soundtracks for videos. Google&#8217;s AI research division and other organizations have previously created models for video, but these have been unable to generate accompanying sound effects. To address this, DeepMind is employing V2A (video-to-audio) technology. \u201cVideo generation models are advancing at an incredible pace, but [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":14444,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"select":"","news_style_id":"","cryptorium_level":"","_short_excerpt_text":"","creation_source":"","_metatest_mainpost_news_update":false,"footnotes":""},"categories":[3],"tags":[438,738],"class_list":["post-14445","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news-and-analysis","tag-artificial-intelligence","tag-google"],"aioseo_notices":[],"amp_enabled":true,"views":"39","promo_type":"","layout_type":"","short_excerpt":"","is_update":"","_links":{"self":[{"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/posts\/14445","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/comments?post=14445"}],"version-history":[{"count":0,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/posts\/14445\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/media\/14444"}],"wp:attachment":[{"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/media?parent=14445"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/categories?post=14445"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/tags?post=14445"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}