1Rhythm Game sample
2==================
3
4This sample demonstrates how to build a simple musical game. The objective of the game is to clap in time to a song by copying what you hear. You do this by listening to the clap sounds, then tapping on the screen to copy those claps.
5
6For a step-by-step guide on how this game works and how to build it check out this codelab: [Build a Musical Game using Oboe](https://codelabs.developers.google.com/codelabs/musicalgame-using-oboe/index.html).
7
8
9Screenshots
10-----------
11The UI is deliberately very simple - just tap anywhere in the grey area after hearing the claps. The UI will change color to indicate the game state. The colors are:
12
13- Yellow: Game is loading (assets are being decompressed)
14- Grey: Game is being played
15- Orange: You tapped too early
16- Green: You tapped on time
17- Purple: You tapped too late
18- Red: There was a problem loading the game (check logcat output)
19
20![RhythmGame Screenshot](images/RhythmGame-screenshot.png)
21
22
23### Audio timeline
24![Game timeline](images/1-timeline.png "Game timeline")
25
26The game plays the clap sounds on the first 3 beats of the bar. These are played in time with the backing track.
27
28 When the user taps on the screen, a clap sound is played and the game checks whether the tap occurred within an acceptable time window.
29
30### Architecture
31
32![Game architecture](images/2-architecture.png "Game architecture")
33
34Oboe provides the [`AudioStream`](https://github.com/google/oboe/blob/master/include/oboe/AudioStream.h) class and associated objects to allow the sample to output audio data to the audio device. All other objects are provided by the sample.
35
36Each time the `AudioStream` needs more audio data it calls [`AudioDataCallback::onAudioReady`](https://github.com/google/oboe/blob/master/include/oboe/AudioStreamCallback.h). This passes a container array named `audioData` to the `Game` object which must then fill the array with `numFrames` of audio frames.
37
38
39![onAudioReady signature](images/3-audioData.png "onAudioReady signature")
40
41### Latency optimizations
42The sample uses the following optimizations to obtain a low latency audio stream:
43
44- Performance mode set to [Low Latency](https://github.com/google/oboe/blob/master/FullGuide.md#setting-performance-mode)
45- Sharing mode set to [Exclusive](https://github.com/google/oboe/blob/master/FullGuide.md#sharing-mode)
46- Buffer size set to twice the number of frames in a burst (double buffering)
47
48### Audio rendering
49
50The `IRenderableAudio` interface (abstract class) represents objects which can produce frames of audio data. The `Player` and `Mixer` objects both implement this interface.
51
52Both the clap sound and backing tracks are represented by `Player` objects which are then mixed together using a `Mixer`.
53
54![Audio rendering](images/4-audio-rendering.png "Audio rendering")
55
56### Sharing objects with the audio thread
57
58It is very important that the audio thread (which calls the `onAudioReady` method) is never blocked. Blocking can cause underruns and audio glitches. To avoid blocking we use a `LockFreeQueue` to share information between the audio thread and other threads. The following diagram shows how claps are enqueued by pushing the clap times (in milliseconds) onto the queue, then dequeuing the clap time when the clap is played.
59
60![Lock free queue](images/5-lockfreequeue.png "Lock free queue")
61
62We also use [atomics](http://en.cppreference.com/w/cpp/atomic/atomic) to ensure that threads see a consistent view of any shared primitives.
63
64### Keeping UI events and audio in sync
65
66When a tap event arrives on the UI thread it only contains the time (milliseconds since boot) that the event occurred. We need to figure out what the song position was when the tap occurred.
67
68To do this we keep track of the song position and the time it was last updated. These values are updated each time the `onAudioReady` method is called. This enables us to keep the UI in sync with the audio timeline.
69
70![Audio/UI synchronization](images/6-audio-ui-sync.png "Audio/UI synchronization")
71
72### Calculating whether a tap was successful
73Once we know when the user tapped in the song, we can calculate whether that tap was successful i.e whether it fell within an acceptable time range. This range is known as the "tap window".
74
75![Tap window calculation](images/7-tap-window.png "Tap window calculation")
76
77Once we know the result of the tap the UI is updated with a color to give the user visual feedback. This is done in `getTapResult`.
78
79Note that once a tap has been received the tap window is removed from the queue - the user only gets one chance to get their tap right!
80
81### Use of compressed audio assets
82In order to reduce APK size this game uses MP3 files for its audio assets. These are extracted on game startup in `AAssetDataSource::newFromCompressedAsset`. A yellow screen will be shown during this process.
83
84By default the game uses `NDKExtractor` for asset extraction and decoding. Under the hood this uses the [NDK Media APIs](https://developer.android.com/ndk/reference/group/media).
85
86There are some limitations with this approach:
87
88- Only available on API 21 and above
89- No resampling: The extracted output format will match the input format of the MP3. In this case a sample rate of 48000. If your audio stream's sample rate doesn't match the assets will not be extracted and an error will be displayed in logcat.
90- 16-bit output only.
91
92A faster, more versatile solution is to use [FFmpeg](https://www.ffmpeg.org/). To do this follow [the instructions here](https://medium.com/@donturner/using-ffmpeg-for-faster-audio-decoding-967894e94e71) and use the `ffmpegExtractor` build variant found in `app.gradle`. The extraction will then be done by `FFmpegExtractor`.
93