Academic literature on the topic 'Auding scene analysis'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Auding scene analysis.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Auding scene analysis"

1

Sareya, Rosli, Zairul Anuar Md Dawam, and D. Maryama Ag. Daud. "ELEMEN BEAT DALAM BABAK DUA: ANALISIS FILEM POLIS EVO (2015)." Jurnal Gendang Alam (GA) 8 (December 20, 2018): 31. http://dx.doi.org/10.51200/ga.v8i.1567.

Full text
Abstract:
ABSTRAKKajian ini merungkai kepentingan elemen beat dalam babak dua iaitu konfrontasi dalam filem aksi "Polis Evo" (2015). Penggunaan elemen beat bertujuan meningkatkan tahap intensiti adegan-adegan yang boleh menarik perhatian audiens untuk menonton sesebuah filem aksi. Intensiti yang tinggi dalam adegan aksi contohnya akan menimbulkan emosi keterujaan dan cemas pada audiens. Analisis beat yang mempengaruhi tahap intensiti babak dua filem aksi ini dan menggunakan pendekatan Blake Snyder’s (2005) yang memperkenalkan elemen beat sheet. Elemen beat tersebut terdiri daripada opening image, them
APA, Harvard, Vancouver, ISO, and other styles
2

Rahmat, Inggi, Sugiantoro Bambang, and Prayudi Yudi. "PENERAPAN SYSTEM DEVELOPMENT LIFE CYCLE (SDLC)." semanTIK 4, no. 2 (2018): 193–200. https://doi.org/10.5281/zenodo.2528444.

Full text
Abstract:
<strong><em>Abstract</em></strong> <em>An audio recorder is one of the multimedia content that is often found at the crime scene and is used as evidence in the trial process. Audio recorders that can be made as legal evidence in court must follow the rules and audio forensic standards that apply in the process of obtaining the evidence, this is because audio digital evidence is very easy to manipulate. The forensic audio investigation process requires a framework that can be used as a reference in the investigation process. However, at present, the developing audio forensic focus more on analy
APA, Harvard, Vancouver, ISO, and other styles
3

Mahmoud, Shakir Wahhab. "EXPLORING SOUND GENERATION AND PROCESSING IN MECHATRONIC SYSTEMS THROUGH DIGITAL SIGNAL PROCESSING: A MATLAB-BASED INVESTIGATION." Engineering and Technology Journal 8, no. 05 (2023): 2272–80. https://doi.org/10.5281/zenodo.7977174.

Full text
Abstract:
Digital signal processing (DSP) encompasses various applications such as convolution, censoring, and filtering. These techniques find relevance in diverse fields, including computer games consoles, tone generation for mobile cellular devices, and theatrical performances. For computer engineers facing limited memory budgets in their devices, generating new sounds by convolving with existing sounds or noises becomes a significant challenge. The objective of this paper is to elucidate and demonstrate the generation of sounds through sound and recorded sound processing. These operations form the f
APA, Harvard, Vancouver, ISO, and other styles
4

Pawestry, Ezarine Vani Indirra. "PEMAKNAAN KEKERASAN BERBASIS GENDER ONLINE DALAM FILM LIKE & SHARE (ANALISIS RESEPSI AUDIENS FILM LIKE & SHARE)." Linimasa : Jurnal Ilmu Komunikasi 7, no. 2 (2024): 36–57. https://doi.org/10.23969/linimasa.v7i2.10798.

Full text
Abstract:
Online gender-based violence cases has recently increased statistically in Indonesia. The movie Like &amp; Share by Gina S. Noer shows how online gender-based violence happens to women. This research uses descriptive qualitative methods and uses Stuart Hall's reception theory as an approach. Reception analysis itself views at the activity of viewers and readers as recipients of a message produced by the sender of the message. The purpose of this research is to describe the audiences’ reception of messages which can later be classified into three reader positions based on Stuart Hall's receptio
APA, Harvard, Vancouver, ISO, and other styles
5

Saputri, Nur Amala. "Komodifikasi Adegan Humor dalam Akun Instagram Netflix Indonesia." Komuniti : Jurnal Komunikasi dan Teknologi Informasi 16, no. 2 (2024): 168–94. https://doi.org/10.23917/komuniti.v16i2.2118.

Full text
Abstract:
Penelitian ini bertujuan menganalisis praktik komodifikasi adegan humor dalam film dan serial TV Netflix yang dilakukan dalam akun jejaring media sosial Netflix Indonesia (@netflixid). Netflix sebagai perusahaan media yang berperan sebagai produsen sekaligus distributor film, banyak melakukan praktik komodifikasi konten untuk mencapai keuntungan lebih. Peneliti menggunakan paradigma kritis dan perspektif ekonomi politik media sebagai pisau analisisnya. Metode yang digunakan adalah analisis tekstual deskriptif kualitatif. Teknik pengumpulan data, peneliti mengurutkan dan melakukan klasifikasi k
APA, Harvard, Vancouver, ISO, and other styles
6

Ovitamaya, Eklesia. "RESEPSI PENONTON REMAJA FILM DUA GARIS BIRU TENTANG ISU PENDIDIKAN SEKS." Jurnal Audience 4, no. 01 (2021): 73–85. http://dx.doi.org/10.33633/ja.v4i01.4232.

Full text
Abstract:
AbstrakFilm Indonesia yang mengangkat isu tentang pendidikan seks masih dianggap tabu. Pada sisi yang lain, sebagai media alternatif, pesan dalam film dapat dipahami dengan mudah oleh para penontonnya. Penelitian ini membahas resepsi penonton remaja film Dua Garis Biru tentang isu pendidikan seks. Penelitian ini menggunakan kerangka encoding-decoding Stuart Hall untuk menganalisis pemakanaan audiens dari setiap adegan dalam film. Temuan data menunjukkan bahwa audiens terbagi ke dalam tiga kelompok pembacaan: dominan hegemonik, negosiasi, dan oposisi. Posisi pembacaan penonton ini dipengaruhi o
APA, Harvard, Vancouver, ISO, and other styles
7

Rifaldi, Riki. "Eksistensi Gender Perempuan Dalam Film Battle Of Sexes (Analisis Semiotika Roland Barthes)." Komunika : Jurnal Ilmiah Komunikasi 2, no. 1 (2024): 36–45. https://doi.org/10.70437/komunika.v2i1.1067.

Full text
Abstract:
Woman gettting more proof that they can living their life with their own abilities. With the growth of mass media this time, woman existention become more exist. One of those following mass media that have huge influence was film. This film is also a film that bring up gender existention theme which woman fight for. This research purpose to know about gender existention using Roland Barthes semiotic analysis through denotation, conotation and myth indicator. This research using descriptive qualitative methode. Data collection technique is using observation and documentation to this film. Unit
APA, Harvard, Vancouver, ISO, and other styles
8

Kovsh, Oleksandr, and Oleksii Kopachinskyi. "Features of Editing in Modern Audiovisual Production: Special Effects and Transitions." Bulletin of Kyiv National University of Culture and Arts. Series in Audiovisual Art and Production 6, no. 1 (2023): 105–17. https://doi.org/10.31866/2617-2674.6.1.2023.279255.

Full text
Abstract:
The purpose of the research&nbsp;is to analyze the phenomenon of human perception of the combination of film images in varying space and time and with different actions. To determine the role of audio and video combination: soundtrack, counterpoint; to trace the parallels between the structure of musical works and music notation with composition, as well as the system of editing audiovisual works; to outline the main editing transitions and special effects, to note the feasibility of their use in modern audiovisual production.&nbsp;The research methodology&nbsp;is based on the following scient
APA, Harvard, Vancouver, ISO, and other styles
9

Rejeki, Amanda Sri, and Juwita Oktaviani Pramudya. "Analisis Poster Film “Saya Nikahkan” Menggunakan Pendekatan Semiotika Roland Barthes." Magenta | Official Journal STMK Trisakti 8, no. 1 (2024): 1188–97. https://doi.org/10.61344/magenta.v8i1.141.

Full text
Abstract:
Movie posters are often used as promotional media to communicate the content or conflict of the story contained in a movie. Various visual elements contained in the poster contain signs. These signs will later have a meaning that will try to convey to the audience. The science that studies signs is semiotics. Semiotics studies the theory of signs in the context of scenarios, images, texts or scenes in movies that can be interpreted. This research will discuss the analysis of the movie poster "Saya Nikahkan" using Roland Barthes' semiotic approach. The result of this research is known that the
APA, Harvard, Vancouver, ISO, and other styles
10

Harsono, Ari, and Khalif Anant Pangastono. "Film sebagai medium untuk menyampaikan pendidikan seks: analisis isi serial produksi netflix “sex education”." Penelitian Ilmu Pengetahuan Sosial 1, no. 2 (2024): 140–57. https://doi.org/10.61511/pips.v1i2.2024.1650.

Full text
Abstract:
ABSTRACT Background: This paper analyzes how the Netflix series “Sex Education” as a form of mass communication media deliver sex education to the audiences. The concept analyzed in this paper is the implication of the five categories of comprehensive sexuality education (CSE) topics promoted by UNFPA. In addition, this paper also looks for the root causes of problems related to sex education that this series tries to overcome. Methods: This study uses content analysis method by observing the scenes in the series as well as analyzing its relationship with the concept. Findings: The results of
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Auding scene analysis"

1

Bando, Yoshiaki. "Robust Audio Scene Analysis for Rescue Robots." Kyoto University, 2018. http://hdl.handle.net/2433/232410.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Carlo, Diego Di. "Echo-aware signal processing for audio scene analysis." Thesis, Rennes 1, 2020. http://www.theses.fr/2020REN1S075.

Full text
Abstract:
La plupart des méthodes de traitement du signal audio considèrent la réverbération et en particulier les échos acoustiques comme une nuisance. Cependant, ceux-ci transmettent des informations spatiales et sémantiques importantes sur les sources sonores et des méthodes essayant de les prendre en compte ont donc récemment émergé.. Dans ce travail, nous nous concentrons sur deux directions. Tout d’abord, nous étudions la manière d’estimer les échos acoustiques à l’aveugle à partir d’enregistrements microphoniques. Deux approches sont proposées, l’une s’appuyant sur le cadre des dictionnaires cont
APA, Harvard, Vancouver, ISO, and other styles
3

Parekh, Sanjeel. "Learning representations for robust audio-visual scene analysis." Thesis, Université Paris-Saclay (ComUE), 2019. http://www.theses.fr/2019SACLT015/document.

Full text
Abstract:
L'objectif de cette thèse est de concevoir des algorithmes qui permettent la détection robuste d’objets et d’événements dans des vidéos en s’appuyant sur une analyse conjointe de données audio et visuelle. Ceci est inspiré par la capacité remarquable des humains à intégrer les caractéristiques auditives et visuelles pour améliorer leur compréhension de scénarios bruités. À cette fin, nous nous appuyons sur deux types d'associations naturelles entre les modalités d'enregistrements audiovisuels (réalisés à l'aide d'un seul microphone et d'une seule caméra), à savoir la corrélation mouvement/audi
APA, Harvard, Vancouver, ISO, and other styles
4

Parekh, Sanjeel. "Learning representations for robust audio-visual scene analysis." Electronic Thesis or Diss., Université Paris-Saclay (ComUE), 2019. http://www.theses.fr/2019SACLT015.

Full text
Abstract:
L'objectif de cette thèse est de concevoir des algorithmes qui permettent la détection robuste d’objets et d’événements dans des vidéos en s’appuyant sur une analyse conjointe de données audio et visuelle. Ceci est inspiré par la capacité remarquable des humains à intégrer les caractéristiques auditives et visuelles pour améliorer leur compréhension de scénarios bruités. À cette fin, nous nous appuyons sur deux types d'associations naturelles entre les modalités d'enregistrements audiovisuels (réalisés à l'aide d'un seul microphone et d'une seule caméra), à savoir la corrélation mouvement/audi
APA, Harvard, Vancouver, ISO, and other styles
5

Butler, Darren Edward. "Scene analysis & compression." Thesis, Queensland University of Technology, 2004.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
6

Phillips, Nicola Jane. "Audio-visual scene analysis : attending to music in film." Thesis, University of Cambridge, 2000. https://www.repository.cam.ac.uk/handle/1810/251745.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Unnikrishnan, Harikrishnan. "AUDIO SCENE SEGEMENTATION USING A MICROPHONE ARRAY AND AUDITORY FEATURES." UKnowledge, 2010. http://uknowledge.uky.edu/gradschool_theses/622.

Full text
Abstract:
Auditory stream denotes the abstract effect a source creates in the mind of the listener. An auditory scene consists of many streams, which the listener uses to analyze and understand the environment. Computer analyses that attempt to mimic human analysis of a scene must first perform Audio Scene Segmentation (ASS). ASS find applications in surveillance, automatic speech recognition and human computer interfaces. Microphone arrays can be employed for extracting streams corresponding to spatially separated sources. However, when a source moves to a new location during a period of silence, such
APA, Harvard, Vancouver, ISO, and other styles
8

Alameda-Pineda, Xavier. "Egocentric Audio-Visual Scene Analysis : a machine learning and signal processing approach." Thesis, Grenoble, 2013. http://www.theses.fr/2013GRENM024/document.

Full text
Abstract:
Depuis les vingt dernières années, l'industrie a développé plusieurs produits commerciaux dotés de capacités auditives et visuelles. La grand majorité de ces produits est composée d'un caméscope et d'un microphone embarqué (téléphones portables, tablettes, etc). D'autres, comme la Kinect, sont équipés de capteurs de profondeur et/ou de petits réseaux de microphones. On trouve également des téléphones portables dotés d'un système de vision stéréo. En même temps, plusieurs systèmes orientés recherche sont apparus (par exemple, le robot humanoïde NAO). Du fait que ces systèmes sont compacts, leur
APA, Harvard, Vancouver, ISO, and other styles
9

Baque, Mathieu. "Analyse de scène sonore multi-capteurs : un front-end temps-réel pour la manipulation de scène." Thesis, Le Mans, 2017. http://www.theses.fr/2017LEMA1013/document.

Full text
Abstract:
La thèse s’inscrit dans un contexte d’essor de l’audio spatialisé (5.1, Dolby Atmos...). Parmi les formats audio 3D existants, l’ambisonie permet une représentation spatiale homogène du champ sonore et se prête naturellement à des manipulations : rotations, distorsion du champ sonore. L’objectif de cette thèse est de fournir un outil d’analyse et de manipulation de contenus audio (essentiellement vocaux) au format ambisonique. Un fonctionnement temps-réel et en conditions acoustiques réelles sont les principales contraintes à respecter. L’algorithme mis au point est basé sur une analyse en com
APA, Harvard, Vancouver, ISO, and other styles
10

Melih, Kathy, and n/a. "Audio Source Separation Using Perceptual Principles for Content-Based Coding and Information Management." Griffith University. School of Information Technology, 2004. http://www4.gu.edu.au:8080/adt-root/public/adt-QGU20050114.081327.

Full text
Abstract:
The information age has brought with it a dual problem. In the first place, the ready access to mechanisms to capture and store vast amounts of data in all forms (text, audio, image and video), has resulted in a continued demand for ever more efficient means to store and transmit this data. In the second, the rapidly increasing store demands effective means to structure and access the data in an efficient and meaningful manner. In terms of audio data, the first challenge has traditionally been the realm of audio compression research that has focused on statistical, unstructured audio represent
APA, Harvard, Vancouver, ISO, and other styles
More sources

Books on the topic "Auding scene analysis"

1

Computational Auditory Scene Analysis: Proceedings of the Ijcai-95 Workshop. CRC, 1998.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
2

Rosenthal, David, Hiroshi Okuno, Hiroshi G. Okuno, and David F. Rosenthal. Computational Auditory Scene Analysis: Proceedings of the Ijcai-95 Workshop. Taylor & Francis Group, 2021.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
3

Rosenthal, David, Hiroshi Okuno, Hiroshi G. Okuno, and David F. Rosenthal. Computational Auditory Scene Analysis: Proceedings of the Ijcai-95 Workshop. Taylor & Francis Group, 2021.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
4

Rosenthal, David, Hiroshi Okuno, Hiroshi G. Okuno, and David F. Rosenthal. Computational Auditory Scene Analysis: Proceedings of the Ijcai-95 Workshop. Taylor & Francis Group, 2021.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
5

Rosenthal, David, Hiroshi Okuno, Hiroshi G. Okuno, and David F. Rosenthal. Computational Auditory Scene Analysis: Proceedings of the Ijcai-95 Workshop. Taylor & Francis Group, 2021.

Find full text
APA, Harvard, Vancouver, ISO, and other styles

Book chapters on the topic "Auding scene analysis"

1

Saraceno, Caterina, and Riccardo Leonardi. "Audio-visual processing for scene change detection." In Image Analysis and Processing. Springer Berlin Heidelberg, 1997. http://dx.doi.org/10.1007/3-540-63508-4_114.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Tsekeridou, Sofia, Stelios Krinidis, and Ioannis Pitas. "Scene Change Detection Based on Audio-Visual Analysis and Interaction." In Multi-Image Analysis. Springer Berlin Heidelberg, 2001. http://dx.doi.org/10.1007/3-540-45134-x_16.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Krstulović, Sacha. "Audio Event Recognition in the Smart Home." In Computational Analysis of Sound Scenes and Events. Springer International Publishing, 2017. http://dx.doi.org/10.1007/978-3-319-63450-0_12.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Pham, Lam, Alexander Schindler, Mina Schutz, Jasmin Lampert, Sven Schlarb, and Ross King. "Deep Learning Frameworks Applied For Audio-Visual Scene Classification." In Data Science – Analytics and Applications. Springer Fachmedien Wiesbaden, 2022. http://dx.doi.org/10.1007/978-3-658-36295-9_6.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Owens, Andrew, and Alexei A. Efros. "Audio-Visual Scene Analysis with Self-Supervised Multisensory Features." In Computer Vision – ECCV 2018. Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-030-01231-1_39.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Kang, Chan-Mi, and Joong-Hwan Baek. "Audio Content Analysis for Understanding Structures of Scene in Video." In Lecture Notes in Computer Science. Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/11816157_151.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Li, Qi, Huadong Ma, and Dong Zhao. "A Neural Network Based Framework for Audio Scene Analysis in Audio Sensor Networks." In Advances in Multimedia Information Processing - PCM 2009. Springer Berlin Heidelberg, 2009. http://dx.doi.org/10.1007/978-3-642-10467-1_42.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Gupta, Vaibhavi, Vinay Detani, Vivek Khokar, and Chiranjoy Chattopadhyay. "C2VNet: A Deep Learning Framework Towards Comic Strip to Audio-Visual Scene Synthesis." In Document Analysis and Recognition – ICDAR 2021. Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-86331-9_11.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Ganesh, Attigodu Chandrashekara, Frédéric Berthommier, and Jean-Luc Schwartz. "Audio Visual Integration with Competing Sources in the Framework of Audio Visual Speech Scene Analysis." In Advances in Experimental Medicine and Biology. Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-25474-6_42.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Vuolo, Mike, Christopher Uggen, and Sarah Lageson. "To Match or Not to Match? Statistical and Substantive Considerations in Audit Design and Analysis." In Audit Studies: Behind the Scenes with Theory, Method, and Nuance. Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-319-71153-9_6.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Auding scene analysis"

1

Nigro, Michael, Binh Nguyen, and Sridhar Krishnan. "Comparison of Hand-Crafted and Deep Features Towards Explainable AI at the Edge for Analysis of Audio Scenes." In 2024 IEEE 34th International Workshop on Machine Learning for Signal Processing (MLSP). IEEE, 2024. http://dx.doi.org/10.1109/mlsp58920.2024.10734782.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Wang, Shanshan, Annamaria Mesaros, Toni Heittola, and Tuomas Virtanen. "A Curated Dataset of Urban Scenes for Audio-Visual Scene Analysis." In ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021. http://dx.doi.org/10.1109/icassp39728.2021.9415085.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Asefi, Hamid, Behnaz Ghoraani, Andy Ye, and Sridhar Krishnan. "Audio scene analysis using parametric signal features." In 2011 24th IEEE Canadian Conference on Electrical and Computer Engineering (CCECE). IEEE, 2011. http://dx.doi.org/10.1109/ccece.2011.6030593.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Venugopal, S., K. R. Ramakrishnan, S. H. Srinivas, and N. Balakrishnan. "Audio scene analysis and scene change detection in the MPEG compressed domain." In 1999 IEEE Third Workshop on Multimedia Signal Processing (Cat. No.99TH8451). IEEE, 1999. http://dx.doi.org/10.1109/mmsp.1999.793819.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Alves, Diego Cardoso, and Paula Dornhofer Paro Costa. "Multimodal social scenario perception model for initial human-robot interaction." In XXXII Conference on Graphics, Patterns and Images. Sociedade Brasileira de Computação - SBC, 2019. http://dx.doi.org/10.5753/sibgrapi.est.2019.8309.

Full text
Abstract:
Human-robot interaction imposes many challenges and artificial intelligence researchers are demanded to improve scene perception, social navigation and engagement. Great attention is being dedicated to the development of computer vision and multimodal sensing approaches that are focused on the evolution of social robotic systems and the improvement of social model accuracy. Most recent works related to social robotics rely on the engagement process with a focus on maintaining a previously established conversation. This work brings up the study of initial human-robot interaction contexts, propo
APA, Harvard, Vancouver, ISO, and other styles
6

Bregman, Albert S. "Progress in the Study of Auditory Scene Analysis." In 2007 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics. IEEE, 2007. http://dx.doi.org/10.1109/aspaa.2007.4393058.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Li, Qi, Bin Tian, and Miao Zhang. "An event sequence based method for audio scene analysis." In Multimedia Technology (IC-BNMT 2011). IEEE, 2011. http://dx.doi.org/10.1109/icbnmt.2011.6155936.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Cho, Youngmin, and Lawrence K. Saul. "Learning dictionaries of stable autoregressive models for audio scene analysis." In the 26th Annual International Conference. ACM Press, 2009. http://dx.doi.org/10.1145/1553374.1553396.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Moncrieff, S., S. Venkatesh, and C. Dorai. "Horror film genre typing and scene labeling via audio analysis." In 2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698). IEEE, 2003. http://dx.doi.org/10.1109/icme.2003.1221586.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Fayek, Haytham M., and Anurag Kumar. "Large Scale Audiovisual Learning of Sounds with Weakly Labeled Data." In Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20}. International Joint Conferences on Artificial Intelligence Organization, 2020. http://dx.doi.org/10.24963/ijcai.2020/78.

Full text
Abstract:
Recognizing sounds is a key aspect of computational audio scene analysis and machine perception. In this paper, we advocate that sound recognition is inherently a multi-modal audiovisual task in that it is easier to differentiate sounds using both the audio and visual modalities as opposed to one or the other. We present an audiovisual fusion model that learns to recognize sounds from weakly labeled video recordings. The proposed fusion model utilizes an attention mechanism to dynamically combine the outputs of the individual audio and visual models. Experiments on the large scale sound events
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!