In recent years, the exploitation of the potential of big data has resulted in significant advancements in a variety of Computer Vision and Natural Language Processing applications. However, the majority of tasks addressed thus far have been primarily visual in nature due to the unbalanced availability of labelled samples across modalities (e.g., there are numerous large labelled datasets for images but few for audio or IMU-based classification), resulting in a large performance gap when algorithms are trained separately. With its origins in audio-visual speech recognition and, more recently, in language and vision projects such as image and video captioning, multimodal machine learning is a thriving multidisciplinary research field that addresses several of artificial intelligence's (AI) original goals by integrating and modelling multiple communicative modalities, including linguistic, acoustic, and visual messages. Due to the variability of the data and the frequently observed dependency between modalities, this study subject presents some particular problems for machine learning researchers. Because the majority of this hateful content is in regional languages, they easily slip past online surveillance algorithms that are designed to target articles written in resource-rich languages like English. As a result, low-resource regional languages in Asia, Africa, Europe, and South America face a shortage of tools, benchmark datasets, and machine learning approaches. This topical collection aims to bring together members of the machine learning and multimodal data fusion fields in regional languages. We anticipate contributions that hate speech and emotional analysis in multimodality include video, audio, text, drawings, and synthetic material in regional language. This topical collection's objective is to advance scientific study in the broad field of multimodal interaction, techniques, and systems, emphasising important trends and difficulties in regional languages, with a goal of developing a roadmap for future research and commercial success.