• Friday, 10 April 2026

Fusion Of AI And Human Expression

blog

We use our hands, face and body more than words to communicate in our daily life. The skills to express our thoughts through body parts are termed as linguistic gestures. These gestures are now embedded in an AI system too. AI can now read and understand such linguistic cues and thus implement them in homes, hospitals, schools, and virtual worlds, where it is successful for limited tasks. Linguistic gesture and its embodiment in AI or AI-powered machines are an important field of study in the modern era.

According to McNeill, there are four types of gesture: iconic, metaphoric, deictic, and beat gesture. Iconic gestures are physical representations of hands that depict the things being described. For example, a 'thumbs up' means 'satisfaction' or 'approval' while a thumbs down means 'dissatisfaction' or 'disapproval.' Metaphoric gestures help in expressing abstract ideas through concrete physical actions or objects. For example, to express love, we show a rose. Deictic gesture is pointing or indicating by hand, head and eyes to something to order or draw attention. For example, pointing towards the door to close. Beat gesture is the rhythmic movements of hands to emphasise speech, draw attention, etc. For example, tapping on someone's shoulder would mean 'well done!'

Different styles 

 It is essential for AI to figure out the different styles and meanings of their gestures in context to culture. For example, in sending off a simple hand gesture in one culture, it could mean goodbye, while in another it could mean hello. Therefore, to fully comprehend these differences, the AI systems require a wide range of datasets.

 There are various technologies by which AI learns gestures. A few of them are: computer vision, machine learning, deep learning, and natural language processing. Computer vision is a field of artificial intelligence that enables machines to interpret and understand visual information from their surroundings in the same way as humans do with their own eyes. As regards education, computer vision can monitor hand and motion body details, identifying the teacher's gestures during a lesson or the student's bodily reactions to an interactive exercise. 

Machine learning enables software to learn from the data or examples instead of being explicitly programmed. According to research, machine learning in education can predict learners' efficiency and track student learning patterns or even recommend individual exercises. Deep learning is a subset of machine learning wherein neural networks are utilised in analysing videos and all other complex data. As noted by Zhou, Tang, and Koller, the competence of deep learning lies in acquiring and interpreting video data. This enables AI to acquire video lessons for class evaluations, measure the student's engagement, or even evaluating student activity through the interpretation of faces and body parts. 

Natural language processing deals with the understanding and production of human language. It can associate computer vision-based sign and gesture recognition with spoken or written words, which makes human-machine interactions smoother and more intuitive. For instance, the system can identify a student raising a hand and recognise the follow-up question. These technologies enable AI to understand and encode meaning in videos featuring motion sequences and image analysis algorithms. AI is applicable in the fields of communication, education, medical science, entertainment, and manufacturing. 

Pepper and Sophia Robots utilise gestures in order to imitate natural speaking. They build more human-like experiences using sensors to recognise the emotions of humans and return with the appropriate gesture. By converting sign language to speech or text, artificial intelligence (AI) facilitates communication for people with disabilities. Now, hearing-impaired individuals can communicate with those who do not understand sign language.  

 ChatGPT, Deepseek, Grammarly, etc. are helping teachers and students learn so many lessons. Now lessons are learned without the guidance of teachers. Students can evaluate themselves by using AI tools. For example, if students imitate a gesture about a verb or an object that was performed for them, the AI tutor can provide feedback, which will assist in the memory of their performance.

 In medical science, AI-based applications learn to read hand and bodily movements for the detection of emotional distress. Suddenly, repetitive actions may indicate an individual is anxious and no action at all may indicate the person is depressed. AI-powered aids are used for disease diagnosis, drug discovery, robotic surgery and image analysis. These applications have improved patient outcomes and support doctors and patients with high-quality service. AI is employed for content making in audio and video production. There are so many AI-created visuals for fun and entertainment. AI supports gameplay by introducing intelligent fictional characters, expert graphics and dynamic adjustment. Games like The Last of Us and Red Dead Redemption 2 use AI-powered character behaviour to make the world more immersive. 

Artificial Intelligence has revolutionised the field of manufacturing, too. It supports the industrialists by providing periodic and predictive maintenance, improving supply chain management, quality control, and smart robotics. It has provided outstanding works by maximising efficiency and lowering costs.

Even though there are so many positive aspects of Artificial Intelligence, we need to consider ethical issues related to AI gestures during its application. Privacy should be maintained when AI captures gestures through cameras. If this data is misused, it may breach the privacy of users. Systems must ensure that the data is safely stored and anonymised. Gestures from Asia, Africa, or South America may be misread by the AI if it is trained solely on Western gestures. For example, different parts of the world may read the same hand gesture differently. 

However, an over-reliance on AI can diminish human empathy and contact. Machines should facilitate communication rather than be a replacement. Misreading a friendly act, watching people’s moves all the time, and selling gesture data for profit without consent could be a threat.

Sign languages  

There are more than 100 ethnic communities in Nepal with their own sign languages and cultural gestures. We are practically applying the concept of inclusive AI by providing the following resources for Nepali sign language translation: Gesture-based education apps for mute children and a joint open-source model development initiative with international AI experts. Universities, local governments, and non-governmental organisations should work hand-in-hand to put Nepal at the front line in the development of ethical AI. This also makes sure that the diverse linguistic and cultural input of the Global South is not left out by global AI models.

Finally, artificial intelligence and linguistic gesture recognition are at the centre of a major transformation in human-machine interaction. The application field ranges from education, medical science, entertainment, and manufacturing companies to several other domains. The risks are not negligible; they range from cultural misinterpretation to surveillance and data misuse. The world should move forward with an emphasis on developing the AI systems that are ethical, inclusive, and culturally sensitive—systems that try to understand people rather than decode their gestures. When allowed to see our diversity of expression, AI brings tech closer to humans.

(The author is Vice Principal at the Little Angels' School, Hattiban, Lalitpur.)

How did you feel after reading this news?