Introducing AI Human Solution in NAB Show 2022

DeepBrain AI participated in the NAB Show 2022 in Las Vegas, USA, for four days from April 23 to 26, introducing AI human solutions.
NAB Show is the world's largest broadcasting exhibition organized by the National Broadcasting Association (NAB), and this year, about 1,600 broadcasting related companies from around the world participated to showcase related products and services.


DeepBrain AI has prepared booths focusing on 'AI Human' solutions at NAB 2022. We introduced cases in which AI human technology was introduced to various industries, and ran a demonstration program to see how AI human technology is applied in the broadcasting field, such as AI announcers.

In addition, 'AISTUDIOS', a SaaS AI human solution, was also exhibited. AI STUDIOS is a video composition editing platform that allows AI Human, a virtual model, to produce a video that reads the sentence naturally by entering the speech that the user wants as script.

The biggest feature of AI STUDIOS is that anyone can create images featuring AI humans without video, voice synthesis technology, or equipment. For example, if you usually need 5 to 10 staff, filming equipment, and studios to produce breaking news, AI STUDIOS can produce news videos right away with one PC and just typing in text.


At this event, officials from major broadcasters around the world from Europe and Asia, as well as U.S, were interested in discovering new business models. DeepBrain AI will discover and provide various business models that can be applied to broadcasting environments such as AI reporters, AI show hosts, and AI MC.


DeepBrain AI and KT (Korea Telecom) signed a MOU for business cooperation between GiGa Genie and AI Human.

DeepBrain AI and KT signed a MOU (Memorandum Of Understanding) for business cooperation between GiGa Genie and AI Human. Through this cooperation, the two companies will create "Giga Genie AI Human," which combines the artificial intelligence of KT GiGA Genie with the virtual human of DeepBrain AI, and develop a new GiGA Genie business using it.


While existing AI Humans could only understand existing questions, KT GiGa Genie implemented AI Human will be able to have conversations like an actual person with intelligence of GiGa Genie and appearance of AI Human. Both companies are expected to expand their services in industries such as media, education, finance, commerce and other virtual services.

News More >>

Are Virtual Humans the Next Evolution of Chatbots & Voice Assistants?

We sit down with Joe Murphy, who leads US business development for DeepBrain AI, to explore the business potential of virtual humans, examine the technologies that enable them...

They go by many names, but AI-generated virtual agents that appear to talk, move and gesture like real people appear poised to take off in coming years. Recent fundraising milestones indicate that this space is heating up and investors see significant applications for this technology across both the real world and the emerging VR and AR landscape.

To dive deeper into this space, I recently sat down with Joe Murphy, who leads US business development for DeepBrainAI, to explore the business potential of virtual humans, examine the technologies that enable them, and the many challenges ahead as this approach to human-machine interaction matures and scales.


1. Joe, thanks for your time. First, can you help orient us as to what an “AI human” is, and what are some killer use cases? 

Hi Eric, thanks for the opportunity to share. For this conversation, I will use the terms, “AI Human, Virtual Human, and Digital Twin” interchangeably. In general, they are variations on a common theme: a digital representation of a person that looks, sounds, and acts as the actual person.

On the topic of “killer use cases,” I believe that AI Humans represent the logical evolution of virtual assistants and chatbots. So, augmenting any chatbot with an AI Human is the killer use case. What I mean is that most existing and future chatbot use cases can be enhanced with a Virtual Human. For example, why when speaking to Alexa or Siri, why do I receive a disembodied voice response from a black box? This is then followed by some awkward conversational turn-taking which is haphazardly guided by flashing lights and pulsing icons.

Previous technical limitations could create a case study for the uncanny valley and the disembodied voice assistant made sense. More recently, video synthesis technology has progressed to the point where Virtual Humans can be indistinguishable from actual humans. So, we are no longer constrained to having a conversation with a faceless black box.

Not to sound overly enthusiastic, but I compare the upcoming Virtual Human chatbot evolution to several other technology shifts, where the video was preferred and overtook the audio-only solution.


Entertainment: Radio → TV

Communication: Phone Call → FaceTime Call

Business: Conference Bridge → Zoom Meeting


Each of the paradigms above was noticeably improved with the addition of video. Adding human-centric video almost always creates a more enjoyable and natural interaction. So, we fully expect that adding AI Humans to chatbots will follow this same pattern of acceptance and adoption.



DeepBrain AI, Top 20 Korean AI Startups to watch for – Best of 2022

Startups in Korea are looking to incorporate AI technology into their existing products or service. The reason is that a lot of Korean investors are very interested in investing in Korean AI startups. The Korean government said that it will invest over $330 million in the development of processing-in-memory (PIM) chips. This is in addition to their pledge of $830 million into AI semiconductors from 2020 to 2029.

The Korean government has already built 6 new AI schools in 2020/2021 to educate Korean engineers in AI technology. In addition, corporations are also looking to get into AI technology. Furthermore, Samsung, LG, Naver, Kakao, and Hyundai have shown great interest to invest in AI technologies. Therefore, it is clear that both the Korean government, investors, and companies think AI is an important technology for Korea.


Here are the top 20 AI startups in Korea to keep an eye on for 2022


1. DeepBrain AI

DeepBrain AI is a Korean AI startup that researches and develops conversational AI technology. They offer a wide range of AI-powered customer service products. However, their specialty is in synthetic humans that are able to respond to natural language questions. They use AI technology to offer video and speech synthesis and chatbot solutions to enterprise customers such as MBN, Metro News, LG, and KB Kookmin Bank. They showcased their AI human imbedded “AI Kiosks” at CES 2022. It leveraged the power of AI with its human-based AI avatars.


AI Human Avatars

In order to create AI Human avatars, DeepBrain AI first captures a video of a human model in a studio and then uses AI to analyze the model’s lip, mouth, and head movement. Therefore, it is great for creating virtual AI bankers, teachers, and even news anchors. Their revenue in 2021 was $5 million due to the rise in demand due to COVID. Moreover, DeepBrain AI has been able to raise over $50 million to date led by Korea Development Bank. The startup is currently valued at $180 million.

“There is only a very limited number of companies both locally and globally with high-quality deep learning-based voice synthesis technology. We will achieve excellent performances in diversified business domains with our competitive technology, which is at par with those of global companies,” said CEO of DeepBrain AI, Eric Jang.





DeepBrain AI, Determined as Top Ten Finalists of T-Challenge

We are very happy to announce the TOP 10 teams of our T-Challenge on XR experiences. The ideas focus on virtual shops, products and service experiences, avatars, as well as technologies and platforms that improve the XR experience.

The T-Challenge, organized by T-Labs and T-Mobile U.S., enables teams from all over the world to show and discuss their ideas for future #XR customer experiences.

The finalists are now entering the second phase of the competition. The goal is to further elaborate their ideas until May 22nd. In the “Solution Development” stream solutions should be presented as a Minimum Viable Product (MVP). In “Concept and Design Creation”, a tangible prototype is required.

Besides the prize money of up to €150,000, the teams have the chance to see their solutions integrated in selected Telekom shops in Germany, Europe and the United States.


Conversational human-alike AI avatar

DeepBrain AI creates digital twin avatars that express the same appearance, voice, accent, and behavioral movement as real ones. With this video synthesis technology, you can either easily create videos by typing text or speak to digital twins in real-time, which will enable enterprises to engage their customers more efficiently and humanly.


Virtual Humans: The Video Evolution for Metaverse-Bound Voice Assistants, Brand Ambassadors, and Media Personalities heading to the metaverse

Text-To-Speech (TTS) is the tech-du-jour for most voice assistants. It makes no difference if someone interacts with Alexa, Siri, Google, or others; the responses are typically TTS audio playing out of a smart speaker, mobile phone, or automobile speaker. The current voice assistant paradigm of speaking to a black box and receiving a disembodied voice response works with the interaction models of today, but this doesn’t translate well to the metaverse we see on the horizon.

Enter a host of new start-up companies all in a race to develop “Virtual Humans” or “Digital Twins.” They are creating what will most likely be the next generation of conversational interfaces based on more natural, authentic, and humanistic digital interactions. So why Virtual Humans, and why now? A few technology drivers and socioeconomic factors have created the perfect storm for real-time video synthesis and Virtual Humans.

When compared to conversational TTS responses, there is no doubt that video synthesis solutions require higher workloads (CPU+GPU) to generate video and higher payloads (file size) to deliver video. However, ever-increasing CPU and GPU performance and increased availability speed up the video synthesis process in the cloud and on edge. Also, advances in batch processing and smart caching have enabled real-time video synthesis that rivals TTS solutions for conversational speed. So, the bottleneck of generating ultra-realistic video on the fly has been primarily addressed. This leads to delivering video in real-time, which, thanks to broadband speeds over both Wi-Fi and 5G, is now readily available to most homes, businesses, and schools. You can see the comparison in the video below.

Businesses that require employees to engage with customers, such as hotels, banks, or quick-service restaurants, are having trouble hiring and retaining new employees. A lack of available and qualified employees can damage the customer’s perception of the brand and create a real drain on revenue. Enter the Virtual Humans that can handle basic requests quickly and consistently. In Korea, both 7-11 and KB Bank have installed AI Kiosks that rely on a Virtual Human to interact with customers. The 7-11 implementation supports a man-less (or woman-less) operation.

Another promising vertical for Virtual Humans is media, both broadcast media and social media (influencers). Whether streaming news 24 hours a day or staying relevant on TikTok, the need is the same: generate more video content and make it faster. Once again, Asia has taken the lead with Virtual Humans. Television stations such as MBN and LG HelloVision both supplement their live broadcasts with Virtual Human versions of their lead anchors that provide regular news updates throughout the day. Using either API calls or an intuitive “what you type is what you get” web interface, videos with Virtual Humans can be made in minutes without needing a camera, crew, lights, make-up, etc. A time-saving and cost-saving tool that can be intermixed throughout the day to keep content fresh.

“What is our strategy for the Metaverse?” That question is being asked within conference rooms across all sectors. It is easy to imagine how brands leveraging the 2D Virtual Humans of today for taking orders, helping, sharing news will quickly evolve to be the early pioneers of the 3D world and the metaverse. Watch throughout the year for some big announcements within this space.



DeepBrain AI participates ‘NVIDIA GTC 2022’, Presenting AI Human Technology and Research outcome

▶ DeepBrain AI CTO Kyung-Soo Chae presents time-reducing, lip sync video synthesis technology.
▶ Participates in “Digital Human and Interactive AI” session panel to introduce current status of AI Human industry.


DeepBrain AI, an artificial intelligence (AI) company, announced on the 29th that it participated in the "NVIDIA GTC 2022," the world's largest AI developer conference, and announced AI human-based research results and overall technology.

The "NVIDIA GTC (GPU Technology Conference)" is a global technology conference organized by NVIDIA, a leader in AI computing technology, and was held as an online live session from March 21 to 24. Companies from all over the world participate in more than 500 live sessions, including deep learning artificial intelligence, data science, high-performance computing, and robotics technology, to showcase technology and discuss related agendas with participants.


DeepBrain AI, Chief Technology Officer (CTO) Kyung-Soo Chae attended the session presentation under the theme of "Toward Real-Time Audiovisual Conversation with Artificial Humans," which enables real-time video and voice-based communication.

CTO Kyung-Soo Chae representatively introduced DeepBrain AI's lip-syncing video synthesis technology that enables real-time conversations between AI humans and people. In addition to the technology to generate high-resolution lip-sync videos at four times the speed of time through its own artificial neural network structure design, it also announced research results that succeeded in reducing synthesis time to 1/3 by applying NVIDIA's deep learning inference optimization SDK.


In addition, DeepBrain AI participated as a panel of "Digital Human and Convergent AI" sessions to announce AI human technology and specific business status. Through real-time questions and answers that followed, it actively communicated with industry officials and introduced DeepBrain AI's AI human technology use cases and future business plans.


DeepBrain AI CTO Kyung-Soo Chae said, "The area that takes the most time and money to implement AI humans is video and voice synthesis, and DeepBrain AI's unique technology has announced the results of reducing high-quality video synthesis time in 1/12 of real time."



Introducing AI Yoon Seok-Yeoul appearing on the campaign trail for the 20th president of South Korea.

The PPP(People Power Party) introduced AI Yoon Seok-Yeoul as part of their campaign strategy for the 20th presidential election.
DeepBrain AI developed AI Yoon Seok-Yeoul which was then utilized by PPP to actively communicate with national voters.


"AI Yoon Seok-Yeoul" first appeared at the inauguration ceremony of the presidential election, and said, "The first AI in politics symbolizes the new future and challenge of Korea," adding, "We will be visiting you all over the country."
During the actual election campaign, the PPP utilized AI Yoon Seok-Yeoul to produce personalized targets for each region into personalized videos and distributed them to party members. Which was then mounted on the large screen of the campaign car to explain local pledges.


Looking into technological advantages of AI Yoon Seok-Yeoul,

First, there are no location and time constraints.
AI Yoon Seok-Yeoul has no location and time constraints. Unlike people, he can appear in any part of the country at any given time.
Candidate Yoon Seok-Yeoul, a real person, was able to focus on campaigning in more important strategic areas.
It has also become an alternative to non-face-to-face communication factoring COVID-19.

Second, customized election strategies for each region.
Along with the familiar landmark background of each region, our neighborhood pledge and our city and provincial pledge were implemented by AI Yoon Seok-Yeoul.
AI Yoon Seok-Yeoul made a customized pledge video just by typing text and delivered it to voters through text, SNS, YouTube, and campaign vehicle screens.

Third, reduced the cost of promotion.
If a real person were to be featured in a video of a customized pledge for each region, it would have cost a lot for production such as studios, filming equipment, and staff. Or if candidates would travel all of the regions directly, it costs precious time and money.
AI Yoon Seok-Yeoul was produced in one to two weeks, and after that, it reduced the cost of promoting the election significantly.


It is expected to be used in the upcoming national local elections for local government heads.
Furthermore, we look forward to the day when AI presidents and AI mayors play a role in listening to and resolving complaints from the people and citizens.


DeepBrain Brings AI to Life at the NAB SHOW.

Visit DeepBrain AI to witness the future of Digital Twins for broadcast and social media.




Are you ready for your very Digital Twin?

News stations across Asia are already using DeepBrain AI’s technology for creating “AI Anchors.” DeepBrain AI representatives will be on-site to share information on the process: from the video shoot to deep learning to delivery. Our approach creates an ultra-realistic AI model that matches the actual person’s appearance, voice, gestures, and mannerisms.







AI Studios

DeepBrain AI’s software platform makes videos as easy as typing a script.
AI Studios eliminates investment into high-cost resources required by traditional video production. No cameras, no lights, no crew, type your script, and our video synthesis technology generates the video in seconds.






Conversational AI humans

In addition to presenter-based or informational videos, DeepBrain AI’s video synthesis can also be paired with chatbot software to create a fully interactive and conversation experience.



Booth Information.


Find us at NAB Show 2022

Booth number: C2627 in Central Hall
Check or click the floor map below to navigate your way to our booth.


DeepBrain AI will be exhibiting at #NAB at Las Vegas Convention Ceter in April 24~27.
We will be demonstrating conversational AI Humans and automatic video generator, AI Studios on site. You can come visit us at our booth and learn more about how you can benefit from AI Human technology.
To arrange a meeting contact Ahmad Elzahdan (






More Details on #ASUGSV Summit.

Meet DeepBrain AI Team at ASU+GSV Summit.

We are inviting you to witness the future of AI Human powered on-offline education and training.




Create your own AI Human

DeepBrain AI creates digital twin of a real person by utilizing deep learning and video synthesis technology. Your AI version will look and speak just like you.
The AI version of the real model will look and speak exactly the same, becoming a real digital double.










Conversational AI Humans

Speak, ask, engage with AI Humans in real-time.
Enterprises can deploy this AI solution for more personalized customer engagement as if it was a
1-on-1 in-person consultation available 24/7.











AI Studios

AI Studios eliminates the need for investment into high-cost resources required by traditional video production.
Just by typing the script, this enables AI Humans to naturally speak, use body language and gestures just like real presenters.










Find us at ASU+GSV!

Join us in Booth #213A



DeepBrain AI will be exhibiting at #ASUGSV in San Diego in April 4~6.
We will be demonstrating conversational AI Humans and automatic video generator, AI Studios on site. You can come visit us at our booth and learn more about how you can benefit from AI Human technology.
To arrange a meeting contact Ahmad Elzahdan (




News More >>


More Details on #NAB