public DetectThread(CaptureThread rec) { notes.put("C4", 261.6); notes.put("C4#", 277.2); notes.put("D4", 293.6); notes.put("D4#", 311.1); notes.put("E4", 329.6); notes.put("F4", 349.2); notes.put("F4#", 369.9); notes.put("G4", 391.9); notes.put("G4#", 415.3); notes.put("A4", 440.0); notes.put("A4#", 466.1); notes.put("B4", 493.8); notes.put("C5", 523.25); notes.put("C5#", 554.36); notes.put("D5", 587.3); notes.put("D5#", 622.254); notes.put("E5", 659.25); notes.put("F5", 698.45); notes.put("F5#", 739.98); notes.put("G5", 783.9); ct = rec; AudioRecord ar = rec.getAudioRecord(); int bps = 0; if (ar.getAudioFormat() == AudioFormat.ENCODING_PCM_16BIT) { bps = 16; } else if (ar.getAudioFormat() == AudioFormat.ENCODING_PCM_8BIT) { bps = 8; } int channel = 0; if (ar.getChannelConfiguration() == AudioFormat.CHANNEL_IN_MONO) { channel = 1; } wh = new WaveHeader(); wh.setChannels(channel); wh.setBitsPerSample(bps); wh.setSampleRate(ar.getSampleRate()); analyzer = new Analyse(wh); }
public void testAudioRecordProperties() throws Exception { if (!hasMicrophone()) { return; } assertEquals(AudioFormat.ENCODING_PCM_16BIT, mAudioRecord.getAudioFormat()); assertEquals(MediaRecorder.AudioSource.DEFAULT, mAudioRecord.getAudioSource()); assertEquals(1, mAudioRecord.getChannelCount()); assertEquals(AudioFormat.CHANNEL_CONFIGURATION_MONO, mAudioRecord.getChannelConfiguration()); assertEquals(AudioRecord.STATE_INITIALIZED, mAudioRecord.getState()); assertEquals(mHz, mAudioRecord.getSampleRate()); assertEquals(AudioRecord.RECORDSTATE_STOPPED, mAudioRecord.getRecordingState()); int bufferSize = AudioRecord.getMinBufferSize( mHz, AudioFormat.CHANNEL_CONFIGURATION_DEFAULT, AudioFormat.ENCODING_PCM_16BIT); assertTrue(bufferSize > 0); }
private int initRecording(int sampleRate, int channels) { Logging.d(TAG, "initRecording(sampleRate=" + sampleRate + ", channels=" + channels + ")"); if (!WebRtcAudioUtils.hasPermission(context, android.Manifest.permission.RECORD_AUDIO)) { Logging.e(TAG, "RECORD_AUDIO permission is missing"); return -1; } if (audioRecord != null) { Logging.e(TAG, "InitRecording() called twice without StopRecording()"); return -1; } final int bytesPerFrame = channels * (BITS_PER_SAMPLE / 8); final int framesPerBuffer = sampleRate / BUFFERS_PER_SECOND; byteBuffer = ByteBuffer.allocateDirect(bytesPerFrame * framesPerBuffer); Logging.d(TAG, "byteBuffer.capacity: " + byteBuffer.capacity()); emptyBytes = new byte[byteBuffer.capacity()]; // Rather than passing the ByteBuffer with every callback (requiring // the potentially expensive GetDirectBufferAddress) we simply have the // the native class cache the address to the memory once. nativeCacheDirectBufferAddress(byteBuffer, nativeAudioRecord); // Get the minimum buffer size required for the successful creation of // an AudioRecord object, in byte units. // Note that this size doesn't guarantee a smooth recording under load. int minBufferSize = AudioRecord.getMinBufferSize( sampleRate, AudioFormat.CHANNEL_IN_MONO, AudioFormat.ENCODING_PCM_16BIT); if (minBufferSize == AudioRecord.ERROR || minBufferSize == AudioRecord.ERROR_BAD_VALUE) { Logging.e(TAG, "AudioRecord.getMinBufferSize failed: " + minBufferSize); return -1; } Logging.d(TAG, "AudioRecord.getMinBufferSize: " + minBufferSize); // Use a larger buffer size than the minimum required when creating the // AudioRecord instance to ensure smooth recording under load. It has been // verified that it does not increase the actual recording latency. int bufferSizeInBytes = Math.max(BUFFER_SIZE_FACTOR * minBufferSize, byteBuffer.capacity()); Logging.d(TAG, "bufferSizeInBytes: " + bufferSizeInBytes); try { audioRecord = new AudioRecord( AudioSource.VOICE_COMMUNICATION, sampleRate, AudioFormat.CHANNEL_IN_MONO, AudioFormat.ENCODING_PCM_16BIT, bufferSizeInBytes); } catch (IllegalArgumentException e) { Logging.e(TAG, e.getMessage()); return -1; } if (audioRecord == null || audioRecord.getState() != AudioRecord.STATE_INITIALIZED) { Logging.e(TAG, "Failed to create a new AudioRecord instance"); return -1; } Logging.d( TAG, "AudioRecord " + "session ID: " + audioRecord.getAudioSessionId() + ", " + "audio format: " + audioRecord.getAudioFormat() + ", " + "channels: " + audioRecord.getChannelCount() + ", " + "sample rate: " + audioRecord.getSampleRate()); if (effects != null) { effects.enable(audioRecord.getAudioSessionId()); } // TODO(phoglund): put back audioRecord.getBufferSizeInFrames when // all known downstream users supports M. // if (WebRtcAudioUtils.runningOnMOrHigher()) { // Returns the frame count of the native AudioRecord buffer. This is // greater than or equal to the bufferSizeInBytes converted to frame // units. The native frame count may be enlarged to accommodate the // requirements of the source on creation or if the AudioRecord is // subsequently rerouted. // Logging.d(TAG, "bufferSizeInFrames: " // + audioRecord.getBufferSizeInFrames()); // } return framesPerBuffer; }