Over a million developers have joined DZone.

Controlling Web Audio With React and Redux Middleware

DZone's Guide to

Controlling Web Audio With React and Redux Middleware

We look at how to use the popular React framework and Redux middleware together to create a web-based phone application.

· Web Dev Zone ·
Free Resource

A true open source, API-first CMS — giving you the power to think outside the webpage. Try it for free.

Classy pic of nutty old phone by Joe Haupt on Flickr

Let’s Build a TouchTone Keypad!

If you’ve built React/Redux applications before, you know there is a standard pattern of uni-directional data flow. The UI dispatches an action. A reducer handles the action, returning a new application state. The UI reorganizes itself accordingly.

But what if you need a Redux action to trigger interaction with a complex system? Say, a collection of Web Audio components used to create or analyze sound. Those are not serializable objects. They shouldn’t be managed by a reducer. Nor should a UI component manage them because it could be subject to unmounting at runtime, causing a loss of the audio system.

Instead, we use middleware, and that is the focus of this article.

Middleware is long-lived, and can therefore be expected to maintain whatever audio system we construct for the life of the application. Furthermore, it has access to the store, so it can inspect the application state, dispatch actions in response to events from its charge, e.g., a Web Audio system, and can respond to actions that direct it to interact with such a system.

In this demo, we will keep it simple and just trigger some sounds in response to an action. The goal is simply to demonstrate how to use middleware to adapt the Web Audio API system to a Redux application.

To celebrate the fact that I just this month ditched my landline after 12 years, (porting the number to the awesome openphone.co online service), we will simulate a TouchTone telephone keypad.

Researching the Problem Domain

DTMF – The Magic FrequenciesTouchTone telephones use a system called DTMF (Dual-tone multi-frequency) signaling, which triggers two separate frequencies when a key is pressed.

The linked Wikipedia article contains lots more trivia, like the fact that the ‘#’ symbol was called an “octothorpe” by the original engineers. But, for the most part, all we need to know is contained in the one table you see here.

In our application, two oscillators in a small Web Audio API system can generate those frequencies, and we’ll use a Redux middleware function to act as the go-between.

Building the App

First, we just need to encode this table of magic frequencies in such a way that we can easily create a keypad that uses that information.

The brute force method would be to declare each button separately and hard-code the appropriate frequencies in each button’s click handler. The optimal way, however, would be to arrange the data so that we can actually generate the keypad from it. Needless to say, the latter is the choice we’ll use here.

Representing the Domain

We define the row frequencies and column frequencies first, then create an array for each key with the frequency constants for its row and column position. Finally, an array of arrays representing keypad rows is built, with each key being represented by an array containing a label and the array of tones for that key.


const ROW_1 = 697;
const ROW_2 = 770;
const ROW_3 = 852;
const ROW_4 = 941;

const COL_1 = 1209;
const COL_2 = 1336;
const COL_3 = 1477;

const KEY_1 = [ROW_1, COL_1];
const KEY_2 = [ROW_1, COL_2];
const KEY_3 = [ROW_1, COL_3];

const KEY_4 = [ROW_2, COL_1];
const KEY_5 = [ROW_2, COL_2];
const KEY_6 = [ROW_2, COL_3];

const KEY_7 = [ROW_3, COL_1];
const KEY_8 = [ROW_3, COL_2];
const KEY_9 = [ROW_3, COL_3];

const KEY_STAR  = [ROW_4, COL_1];
const KEY_0     = [ROW_4, COL_2];
const KEY_POUND = [ROW_4, COL_3];

export const KEYPAD = [
    [ ['1', KEY_1],    ['2', KEY_2], ['3', KEY_3], ],   // KEYPAD ROW 1
    [ ['4', KEY_4],    ['5', KEY_5], ['6', KEY_6], ],   // KEYPAD ROW 2
    [ ['7', KEY_7],    ['8', KEY_8], ['9', KEY_9], ],   // KEYPAD ROW 3
    [ ['*', KEY_STAR], ['0', KEY_0], ['#', KEY_POUND] ] // KEYPAD ROW 4

UI to Middleware Messaging

Before we get into either the UI for representing the keypad or the middleware for playing the DTMF tones, let’s have a quick look at the message that will be sent between the two.

The action creator playDTMFPair  will accept a pair of tones as defined in the KEY_  constants above, and return an action of type PLAY_DTMF_PAIR via which the tones can be dispatched from the UI to the middleware each time a key is pressed.


// Audio related actions
export const PLAY_DTMF_PAIR = 'audio/play-dtmf';

// Play a DTMF tone pair
export const playDTMFPair = tones => {
    return {
        type: PLAY_DTMF_PAIR,

Creating the UI

Image title

The demo is a standard React/Redux setup. Additionally, it uses React Bootstrap and Styled Components to achieve the look and feel of a typical TouchTone keypad with big, square, shaded buttons arranged in tight rows. You can review the project code for the styling aspects, but below are the two main React components used to render the keypad.

The App component is, as usual, the main container. In its render method, it creates a StyledKeypad which is basically a columnar flexbox with content centering and some upper margin. Inside that, it renders a StyledKeypadRow container for each row of the keypad (you guessed it, a row-oriented flexbox). Finally, inside each of those, it renders a KeypadKey component for every key in the row, passing a label, the tones that the key needs to trigger, and a dispatcher for the action.


import React, {Component} from 'react';
import {connect} from 'react-redux';

import {KEYPAD} from '../constants/dtmf';
import {playDTMFPair} from '../store/audio/actions';
import {StyledKeypad} from '../styles/StyledKeypad';
import {StyledKeypadRow} from '../styles/StyledKeypadRow';
import KeypadKey from './KeypadKey';

class App extends Component {

    render() {

        const {playTones} = this.props;

        return <StyledKeypad>
                KEYPAD.map( (row, rindex) =>
                <StyledKeypadRow key={rindex}>
                    {row.map( key => <KeypadKey

const mapDispatchToProps = (dispatch) => ({
    playTones: tones => dispatch(playDTMFPair(tones))

export default connect(null, mapDispatchToProps)(App);

The KeypadKey component is a simple functional component which accepts the labeltones, and handleClick function we pass as props. It returns a StyledKeypadButton, which is just a big square Bootstrap button with no outline, a readable font size, and an onClick handler that calls the handleClick function, passing the tones array.


import React from 'react';

import {StyledKeypadButton} from '../styles/StyledKeypadButton';

export default function KeypadKey(props) {

    const {label, tones, handleClick} = props;

    return <StyledKeypadButton onClick={() => handleClick(tones)}>{label}</StyledKeypadButton>


The Web Audio API

Complete polyphonic synthesizers have been built using the Web Audio API, such is the awesome breadth of its implementation. And choosing React/Redux for your overall application framework is a great way to start such a project. While our demo will be piddling in comparison, the architecture could easily be adapted to such a grand purpose.

If you’re new to it, a fantastic introduction to the Web Audio API is available at css-tricks. In fact, a section of it forms the basis for our TouchTone class. The main differences are that we get the audio context in the constructor rather than accepting it as an argument, we create two oscillators instead of just one, we start the sound immediately rather than accepting a time, and we turn the sound off immediately after half a second rather than ramping off exponentially.

Since our focus is on adapting an audio system to Redux, I’ll let the css-tricks article describe the particulars of the Web Audio API touched on here.


export default class TouchTone {

    constructor() {
        // Get the audio context
        this.context = new (window.AudioContext || window.webkitAudioContext)();

    init() {

        // Create, amplify, and connect row oscillator
        this.rowOscillator = this.context.createOscillator();
        this.rowOscillator.type = 'sine';
        this.rowGain = this.context.createGain();

        // Create, amplify, and connect column oscillator
        this.colOscillator = this.context.createOscillator();
        this.colOscillator.type = 'sine';
        this.colGain= this.context.createGain();


    play(tones) {

        // initialize

        // Get the current time from the audio context
        const time = this.context.currentTime;

        // Load tones into oscillators
        this.rowOscillator.frequency.value = tones[0];
        this.colOscillator.frequency.value = tones[1];

        // Set gain and start oscillators
        this.rowGain.gain.setValueAtTime(1, time);
        this.colGain.gain.setValueAtTime(1, time);

        // Set the stop time
        this.stop(time + .5);


    stop(time) {

        // Ramp down gain and stop oscillators
        this.rowGain.gain.setValueAtTime(0, time);
        this.colGain.gain.setValueAtTime(0, time);


The Redux Middleware

When you first read the official introduction to Redux middleware it can easily twist your melon. This is because they go through a whole bunch of “wrong ways” to do things before arriving at their solution. That’s why I thought it would be nice to have a dead simple example to get you started.

It’s really not that complicated at all. A simple function accepts the Redux store and returns a function that accepts the next piece of middleware in the chain’s callback. That callback accepts an action. Inside that function we can handle or ignore any action passing through, but, when we’re done, we need to call the next piece of middleware’s callback, passing it the action.

The key things to remember are that it’s going to be around for the life of the application, and it will be given a chance to handle any dispatched action. This makes it a perfect mediator for interaction with non-serializable parts of our applications like sockets or audio components. It will have access to the store, so you can refer to the state and you can dispatch actions from it if you need to. In our case, we’re only going to respond to a single action dispatched by the keypad.


import TouchTone from './TouchTone';
import {PLAY_DTMF_PAIR} from "./actions";

export const audioMiddleware = store => {

    const touchTone = new TouchTone();

    return next => action => {

        switch (action.type) {

            case PLAY_DTMF_PAIR:






Obviously, with a different interface and audio system, this application could implement a musical device, triggering much more pleasing tones than DTMF. The main focus was on how to adapt such a system.

Using middleware to control the Web Audio API is pretty easy and architecturally the right way to go in a React/Redux application. So don’t let the convoluted documentation on the Redux site put you off if you’ve never built a piece of middleware.

More ambitious things that can happen in the middleware are that it could respond to events from an AudioListener or ScriptProcessorNode and dispatch actions by calling store.dispatch(). This would allow the app to perform audio spatialization or visualization.

You can download the project from GitHub: https://github.com/cliffhall/react-dtmf-dialer

The New Standard for a Hybrid CMS: GraphQL Support, Scripting as a Service, SPA Support. Watch on-demand now.

middleware ,javascript ,redux tutorial ,react.js tutorial ,web dev

Published at DZone with permission of

Opinions expressed by DZone contributors are their own.

{{ parent.title || parent.header.title}}

{{ parent.tldr }}

{{ parent.urlSource.name }}