2DEQ: An Intuitive Audio Equalizer Andrew T. Sabin Northwestern University [email protected] ABSTRACT

The complexity of music production tools can be a significant bottleneck in the creative process. Here we describe the development of a simple, intuitive audio equalizer with the idea that our approach could also be applied to other types of music production tools. First, users generate a large set of equalization curves representative of the most common types of modifications. Next, we represent the entire set of curves in 2-dimensional space and determine the spatial location of common auditory adjectives. Finally we create an interface, called 2DEQ, where the user can drag a single dot to control equalization in this adjective-labeled space. Author Keywords

adaptive filters, music, user interfaces ACM Classification Keywords

H.5.2 [User Interfaces]: Auditory (non-speech) feedback General Terms

Algorithms, Experimentation, Human Factors. INTRODUCTION

Software tools are widely used in audio recording and production. While these tools are powerful, they are also complex. The result of this complexity is that musicians without technical expertise can spend a great deal of time stumbling through a large range of parameter settings, thereby disrupting the creative process by directing attention away from the music itself. We wish to rethink interface design for audio production tools. Musical creators should not need to have technical expertise and should not reconceptualize their ideas in terms of a fixed interface with esoteric parameters. Instead, we propose to build a software tool based on user data with a simple interface designed to provide easy access to common audio modifications. In this paper, we apply this approach to the audio equalizer, but we anticipate that it can be applied to a wider range of audio processing techniques. Audio equalizers are perhaps the most common signal processing tools used in audio production. Equalizers selectively boost or cut restricted portions of the frequency spectrum, and in doing so dramatically alter the timbre of a

Copyright is held by the author/owner(s). C&C’09, October 26–30, 2009, Berkeley, California, USA. ACM 978-1-60558-403-4/09/10.

Bryan Pardo Northwestern University [email protected] sound. The effect of an equalizer can be described by the equalization curve, which describes the extent of energy boost or cut at each frequency. In general, frequency is on the x-axis and gain (the amount of boost or cut) is on the yaxis. The most similar previous work is that of Mecklenburg and Loviscach [1] who also designed a simple equalizer (subjEQt) where a user drags a single dot in 2 dimensional space labeled with natural language. However, in this work the spatial layout of the equalization curves on the control structure was determined using a self organizing map [2] governed by rules that the researchers felt were intuitive, but were not based on any explicit perceptual data. Similarly the researchers’ intuition was used to create the equalization curves themselves. The present work distinguishes itself from this approach because the potential equalization curves, the spatial layout, and the mapping to descriptors are all derived through an analysis of usergenerated behavioral data. EQUALIZER CONSTRUCTION

First we collected a large set of user-generated equalization curves that were intended to represent that user’s mapping to a particular descriptor. For a more detailed description of this process see [3]. Briefly, a series of 25 equalization curves are applied to a single recording. After each curve is presented, the user rates how well that curve captured their concept of a particular descriptor (e.g., how “tinny” was that piano sound?). After rating a line is fit to gain vs. rating data for each frequency band. We call the array of slopes across frequency the weighting function, which represents the direction and extent to which each frequency band influences the perception of the descriptor. In total we gathered a total 279 weighting functions from nineteen normal hearing listeners using 5 different recordings (piano, drums, saxophone, vocals, guitar) and 4 different descriptors (tinny, warm, bright, dark). Next, we represented the entire set of curves in low dimensional space. Each weighting function was comprised of 40 separate frequency bands. Assigning a parameter to each band would create a complex interface. To address this problem, we performed a dimension reducing procedure called Principal Component Analysis (PCA) to determine how well the learned set of equalization weighting functions could be described as a linear combination of a small number of component weighting functions. Interestingly, two components (depicted in Fig 1 A and B) can account for 87% of the variance in weighting functions, and therefore could be used to create a useful interface.

two dimensional principal component space. Each point in the space has a particular color (same as above) that reflects the probability given our dataset that the current equalization curve (the bottom of the interface) will be perceived as a particular descriptor.

Figure 1. (A-B) The two principal components shown as weighting functions. (C) All weighting functions in the space of principal components. We then computed the location of each weighting function in this two-dimensional principal component space. Each weighting function can be represented by a pair of multipliers where each multiplier is applied to one of the components before the two are summed together. In Figure 1C the multipliers summarizing each weighting function are represented in the horizontal and vertical position of each point. The color of the symbol indicates the descriptor corresponding to the weighting function (red = warm, yellow = bright, black = dark, green = tinny), and the shape indicates which recording was used (diamond = saxophone, vocals = square, drums = star, piano = triangle, acoustic guitar = circle). The size of each symbol indicates the reliability of the learned equalization weighting function (larger = more reliable, see [3] for details). Despite a fair amount of variability, there appears to be some consensus in the mapping of the descriptors to location in this 2dimensional space. Finally, we used this space and the corresponding mapping to descriptors to construct a novel equalizer interface, which we call 2DEQ (Figure 2). The user can simultaneously manipulate all 40 frequency bands of the equalization curve by moving a single point around in the

Figure 2. A screenshot of 2DEQ CONCLUSIONS

We feel that the current approach could be applied to a wide array of creative signal processing applications with some limitations. A representative set of the most common parameter settings must be gathered and then analyzed by a dimension reduction technique. Those dimensions are then presented to the user with a simplified interface, labeled with descriptors where applicable. ACKNOWLEDGMENTS

This research was supported by National Science Foundation grant IIS-0757544 and benefitted from conversations with John Woodruff. REFERENCES

1.Mecklenburg, S. and J. Loviscach. “subjEQt: Controlling an equalizer through subjective terms.” Computer Human Interaction Conference. 2006: Montreal, Quebec. 2.Kohonen, T. (2001) Self-Organizing Maps, Springer. 3.Sabin, A.T. and B. Pardo, “A Method for Rapid Personalization of Audio Equalization Parameters,” ACM Multimedia. 2009: Beijing, China.

2DEQ: An Intuitive Audio Equalizer

we propose to build a software tool based on user data with a simple interface ... equalization curves, the spatial layout, and the mapping to descriptors are all ...

266KB Sizes 6 Downloads 189 Views

Recommend Documents

2DEQ: An Intuitive Audio Equalizer
Software tools are widely used in audio recording and production. While these ... can account for 87% of the variance in weighting functions, and therefore could ...

equalizer
V g. 113. FREQUENCY OFSET. ESTIMATION CIRCUIT. 114. V {. FREQUENCY. 6,408,330 B1. 6/2002 DeLaHuerga. FOREIGN PATENT DOCUMENTS. EP. 0 851 602 A2 * 7/1998. JP. 3-254255. * 11/1991. JP. 6-252966 ... mm? a. w mm,. mOhéug .rKOEm A mm? EDGE-0 ZObwPmD wOZm

pdf-136\the-options-edge-an-intuitive-approach-to-generating ...
About the Author. Page 3 of 11. pdf-136\the-options-edge-an-intuitive-approach-to-gener ... ice-to-the-experienced-practitioner-wiley-trading-b.pdf.

The Equalizer: First Series.pdf
Macgregor Card. Laura Carter. Adam Clay. Shanna Compton. Joshua Corey. John Cotter. Matt Cozart. Cynthia Cruz. Barbara Cully. Alana Dagen. Peter Davis ... Kristi Maxwell. James Meetze. Jason Myers. Alexis Orgera. Cate Peebles. Craig Santos Perez. Chr

deep-time-series-forecasting-with-python-an-intuitive-introduction-to ...
deep-time-series-forecasting-with-python-an-intuitive-introduction-to-de.pdf. deep-time-series-forecasting-with-python-an-intuitive-introduction-to-de.pdf. Open.

FAB: An Intuitive Consensus Protocol using Raft and ...
Mar 23, 2015 - also provides weighted leader election to influence ... highest weighted member to become the Leader. ...... I would also like to thank my team.

People's Intuitions About Intuitive Insight and Intuitive ...
ings can be overcome through the effortful exercise of conscious will (Devine, 1989; Fiske, ... This research was supported by National Science Foundation Research ..... holistic) independently predicted the degree to which people thought those .....

Intuitive Decision Making
Of course, business is not a game, and much ... business decisions I ultimately rely on my intuition. .... (9 a.m.-5 p.m. ET) at the phone numbers listed below.

The Great Equalizer: How Main Street Capitalism Can ...
deal-making has smothered America s innovative spirit. ... In response, Smick calls for the great equalizer, a Main Street Capitalism of mass small-business ...

FREE Download Intuitive Biostatistics: A ...
perspectives of the previous editions: a focus on how to interpret statistical results rather than on how to analyze data, minimal use of equations, and a detailed.