Search results

Filters

  • Journals
  • Authors
  • Keywords
  • Date
  • Type

Search results

Number of results: 5
items per page: 25 50 75
Sort by:
Download PDF Download RIS Download Bibtex

Abstract

A hybrid method is presented for the integration of low-, mid-, and high-frequency driver filters in loud-speaker crossovers. The Pascal matrix is exploited to calculate denominators; the locations of minimum values in frequency magnitude responses are associated with the forms of numerators; the maximum values are used to compute gain factors. The forms of the resulting filters are based on the physical meanings of low-pass, band-pass, and high-pass filters, an intuitive idea which is easy to be understood. Moreover, each coefficient is believed to be simply calculated, an advantage which keeps the software-implemented crossover running smoothly even if crossover frequencies are being changed in real time. This characteristic allows users to efficiently adjust the bandwidths of the driver filters by subjective listening tests if objective measurements of loudspeaker parameters are unavailable. Instead of designing separate structures for a low-, mid-, and high-frequency driver filter, by using the proposed techniques we can implement one structure which merges three types of digital filters. Not only does the integration architecture operate with low computational cost, but its size is also compact. Design examples are included to illustrate the effectiveness of the presented methodology
Go to article

Authors and Affiliations

Shu-Nung Yao
Download PDF Download RIS Download Bibtex

Abstract

As the virtual reality (VR) market is growing at a fast pace, numerous users and producers are emerging with the hope to navigate VR towards mainstream adoption. Although most solutions focus on providing highresolution and high-quality videos, the acoustics in VR is as important as visual cues for maintaining consistency with the natural world. We therefore investigate one of the most important audio solutions for VR applications: ambisonics. Several VR producers such as Google, HTC, and Facebook support the ambisonic audio format. Binaural ambisonics builds a virtual loudspeaker array over a VR headset, providing immersive sound. The configuration of the virtual loudspeaker influences the listening perception, as has been widely discussed in the literature. However, few studies have investigated the influence of the orientation of the virtual loudspeaker array. That is, the same loudspeaker arrays with different orientations can produce different spatial effects. This paper introduces a VR audio technique with optimal design and proposes a dual-mode audio solution. Both an objective measurement and a subjective listening test show that the proposed solution effectively enhances spatial audio quality.
Go to article

Authors and Affiliations

Shu-Nung Yao
1

  1. Department of Electrical Engineering, National Taipei University, No. 151, University Rd., Sanxia Dist., New Taipei City 237303, Taiwan
Download PDF Download RIS Download Bibtex

Abstract

There are an increasing number of binaural systems embedded with head-related transfer functions (HRTFs), so listeners can experience virtual environments via conventional stereo loudspeakers or head- phones. As HRTFs vary from person to person, it is difficult to select appropriated HRTFs from already existing databases for users. Once the HRTFs in a binaural audio device hardly match the real ones of the users, poor localization happens especially on the cone of confusion. The most accurate way to obtain personalized HRTFs might be doing practical measurements. It is, however, expensive and time consuming. Modifying non-individualized HRTFs may be an effort-saving way, though the modifications are always accompanied by undesired audio distortion. This paper proposes a flexible HRTF adjustment system for users to define their own HRTFs. Also, the system can keep sounds from suffering intolerable distortion based on an objective measurement tool for evaluating the quality of processed audio.
Go to article

Authors and Affiliations

Shu-Nung Yao
Li Jen Chen
Download PDF Download RIS Download Bibtex

Abstract

This research determines an identification system for the types of Beiguan music – a historical, nonclassical music genre – by combining artificial neural network (ANN), social tagging, and music information retrieval (MIR). Based on the strategy of social tagging, the procedure of this research includes: evaluating the qualifying features of 48 Beiguan music recordings, quantifying 11 music indexes representing tempo and instrumental features, feeding these sets of quantized data into a three-layered ANN, and executing three rounds of testing, with each round containing 30 times of identification. The result of ANN testing reaches a satisfying correctness (97% overall) on classifying three types of Beiguan music. The purpose of this research is to provide a general attesting method, which can identify diversities within the selected non-classical music genre, Beiguan. The research also quantifies significant musical indexes, which can be effectively identified. The advantages of this method include improving data processing efficiency, fast MIR, and evoking possible musical connections from the high-relation result of statistical analyses.
Go to article

Bibliography

1. Briot J.-P., Hadjeres G., Pachet F.-D. (2019), Deep Learning Techniques for Music Generation, Computational Synthesis and Creative Systems, Springer, arXiv: 1709.01620.
2. Hagan M.T., Demuth H.B., Beale M. (2002), Neural Network Design, CITIC Publishing House, Beijing.
3. Lamere P. (2008), Social tagging and music information retrieval, Journal of New Music Research, 37(2): 101–114, doi: 10.1080/09298210802479284.
4. Lu C.-K. (2011), Beiguan Music, Taichung, Taiwan: Morningstar.
5. Pan J.-T. (2019), The transmission of Beiguan in higher education in Taiwan: A case study of the teaching of Beiguan in the department of traditional music of Taipei National University of the Arts [in Chinese], Journal of Chinese Ritual, Theatre and Folklore, 2019.3(203): 111–162.
6. Rosner A., Schuller B., Kostek B. (2014), Classification of music genres based on music separation into harmonic and drum components, Archives of Acoustics, 39(4): 629–638, doi: 10.2478/aoa-2014-0068.
7. Tzanetakis G., Kapur A., Scholoss W.A., Wright M. (2007), Computational ethnomusicology, Journal of Interdisciplinary Music Studies, 1(2): 1–24.
8. Wiering F., de Nooijer J., Volk A., Tabachneck- Schijf H.J.M. (2009), Cognition-based segmentation for music information retrieval systems, Journal of New Music Research, 38(2): 139–154, doi: 10.1080/09298210903171145.
9. Yao S.-N., Collins T., Liang C. (2017), Head-related transfer function selection using neural networks, Archives of Acoustics, 42(3): 365–373, doi: 10.1515/aoa-2017-0038.
10. Yeh N. (1988), Nanguan music repertoire: categories, notation, and performance practice, Asian Music, 19(2): 31–70, doi: 10.2307/833866.
Go to article

Authors and Affiliations

Yu-Hsin Chang
1
Shu-Nung Yao
2

  1. Department of Music, Tainan National University of the Arts, No. 66, Daqi, Guantian Dist., Tainan City 72045, Taiwan
  2. Department of Electrical Engineering, National Taipei University, No. 151, University Rd., Sanxia District, New Taipei City 237303, Taiwan

This page uses 'cookies'. Learn more