1
0
mirror of https://github.com/godotengine/godot synced 2024-07-05 15:23:35 +00:00

Fix member names of AudioFrame to match extension

This commit is contained in:
A Thousand Ships 2024-01-09 15:01:12 +01:00
parent dfe226b933
commit d8b29efe66
No known key found for this signature in database
GPG Key ID: 2033189A662F8BD7
20 changed files with 154 additions and 136 deletions

View File

@ -51,105 +51,123 @@ static const float AUDIO_PEAK_OFFSET = 0.0000000001f;
static const float AUDIO_MIN_PEAK_DB = -200.0f; // linear_to_db(AUDIO_PEAK_OFFSET)
struct AudioFrame {
//left and right samples
float l = 0.f, r = 0.f;
// Left and right samples.
union {
struct {
float left;
float right;
};
#ifndef DISABLE_DEPRECATED
struct {
float l;
float r;
};
#endif
float levels[2] = { 0.0 };
};
_ALWAYS_INLINE_ const float &operator[](int idx) const { return idx == 0 ? l : r; }
_ALWAYS_INLINE_ float &operator[](int idx) { return idx == 0 ? l : r; }
_ALWAYS_INLINE_ const float &operator[](int p_idx) const {
DEV_ASSERT((unsigned int)p_idx < 2);
return levels[p_idx];
}
_ALWAYS_INLINE_ float &operator[](int p_idx) {
DEV_ASSERT((unsigned int)p_idx < 2);
return levels[p_idx];
}
_ALWAYS_INLINE_ AudioFrame operator+(const AudioFrame &p_frame) const { return AudioFrame(l + p_frame.l, r + p_frame.r); }
_ALWAYS_INLINE_ AudioFrame operator-(const AudioFrame &p_frame) const { return AudioFrame(l - p_frame.l, r - p_frame.r); }
_ALWAYS_INLINE_ AudioFrame operator*(const AudioFrame &p_frame) const { return AudioFrame(l * p_frame.l, r * p_frame.r); }
_ALWAYS_INLINE_ AudioFrame operator/(const AudioFrame &p_frame) const { return AudioFrame(l / p_frame.l, r / p_frame.r); }
_ALWAYS_INLINE_ AudioFrame operator+(const AudioFrame &p_frame) const { return AudioFrame(left + p_frame.left, right + p_frame.right); }
_ALWAYS_INLINE_ AudioFrame operator-(const AudioFrame &p_frame) const { return AudioFrame(left - p_frame.left, right - p_frame.right); }
_ALWAYS_INLINE_ AudioFrame operator*(const AudioFrame &p_frame) const { return AudioFrame(left * p_frame.left, right * p_frame.right); }
_ALWAYS_INLINE_ AudioFrame operator/(const AudioFrame &p_frame) const { return AudioFrame(left / p_frame.left, right / p_frame.right); }
_ALWAYS_INLINE_ AudioFrame operator+(float p_sample) const { return AudioFrame(l + p_sample, r + p_sample); }
_ALWAYS_INLINE_ AudioFrame operator-(float p_sample) const { return AudioFrame(l - p_sample, r - p_sample); }
_ALWAYS_INLINE_ AudioFrame operator*(float p_sample) const { return AudioFrame(l * p_sample, r * p_sample); }
_ALWAYS_INLINE_ AudioFrame operator/(float p_sample) const { return AudioFrame(l / p_sample, r / p_sample); }
_ALWAYS_INLINE_ AudioFrame operator+(float p_sample) const { return AudioFrame(left + p_sample, right + p_sample); }
_ALWAYS_INLINE_ AudioFrame operator-(float p_sample) const { return AudioFrame(left - p_sample, right - p_sample); }
_ALWAYS_INLINE_ AudioFrame operator*(float p_sample) const { return AudioFrame(left * p_sample, right * p_sample); }
_ALWAYS_INLINE_ AudioFrame operator/(float p_sample) const { return AudioFrame(left / p_sample, right / p_sample); }
_ALWAYS_INLINE_ void operator+=(const AudioFrame &p_frame) {
l += p_frame.l;
r += p_frame.r;
left += p_frame.left;
right += p_frame.right;
}
_ALWAYS_INLINE_ void operator-=(const AudioFrame &p_frame) {
l -= p_frame.l;
r -= p_frame.r;
left -= p_frame.left;
right -= p_frame.right;
}
_ALWAYS_INLINE_ void operator*=(const AudioFrame &p_frame) {
l *= p_frame.l;
r *= p_frame.r;
left *= p_frame.left;
right *= p_frame.right;
}
_ALWAYS_INLINE_ void operator/=(const AudioFrame &p_frame) {
l /= p_frame.l;
r /= p_frame.r;
left /= p_frame.left;
right /= p_frame.right;
}
_ALWAYS_INLINE_ void operator+=(float p_sample) {
l += p_sample;
r += p_sample;
left += p_sample;
right += p_sample;
}
_ALWAYS_INLINE_ void operator-=(float p_sample) {
l -= p_sample;
r -= p_sample;
left -= p_sample;
right -= p_sample;
}
_ALWAYS_INLINE_ void operator*=(float p_sample) {
l *= p_sample;
r *= p_sample;
left *= p_sample;
right *= p_sample;
}
_ALWAYS_INLINE_ void operator/=(float p_sample) {
l /= p_sample;
r /= p_sample;
left /= p_sample;
right /= p_sample;
}
_ALWAYS_INLINE_ void undenormalize() {
l = ::undenormalize(l);
r = ::undenormalize(r);
left = ::undenormalize(left);
right = ::undenormalize(right);
}
_FORCE_INLINE_ AudioFrame lerp(const AudioFrame &p_b, float p_t) const {
AudioFrame res = *this;
res.l += (p_t * (p_b.l - l));
res.r += (p_t * (p_b.r - r));
res.left += (p_t * (p_b.left - left));
res.right += (p_t * (p_b.right - right));
return res;
}
_ALWAYS_INLINE_ AudioFrame(float p_l, float p_r) {
l = p_l;
r = p_r;
_ALWAYS_INLINE_ AudioFrame(float p_left, float p_right) {
left = p_left;
right = p_right;
}
_ALWAYS_INLINE_ AudioFrame(const AudioFrame &p_frame) {
l = p_frame.l;
r = p_frame.r;
left = p_frame.left;
right = p_frame.right;
}
_ALWAYS_INLINE_ void operator=(const AudioFrame &p_frame) {
l = p_frame.l;
r = p_frame.r;
left = p_frame.left;
right = p_frame.right;
}
_ALWAYS_INLINE_ operator Vector2() const {
return Vector2(l, r);
return Vector2(left, right);
}
_ALWAYS_INLINE_ AudioFrame(const Vector2 &p_v2) {
l = p_v2.x;
r = p_v2.y;
left = p_v2.x;
right = p_v2.y;
}
_ALWAYS_INLINE_ AudioFrame() {}
};
_ALWAYS_INLINE_ AudioFrame operator*(float p_scalar, const AudioFrame &p_frame) {
return AudioFrame(p_frame.l * p_scalar, p_frame.r * p_scalar);
return AudioFrame(p_frame.left * p_scalar, p_frame.right * p_scalar);
}
_ALWAYS_INLINE_ AudioFrame operator*(int32_t p_scalar, const AudioFrame &p_frame) {
return AudioFrame(p_frame.l * p_scalar, p_frame.r * p_scalar);
return AudioFrame(p_frame.left * p_scalar, p_frame.right * p_scalar);
}
_ALWAYS_INLINE_ AudioFrame operator*(int64_t p_scalar, const AudioFrame &p_frame) {
return AudioFrame(p_frame.l * p_scalar, p_frame.r * p_scalar);
return AudioFrame(p_frame.left * p_scalar, p_frame.right * p_scalar);
}
#endif // AUDIO_FRAME_H

View File

@ -143,11 +143,11 @@ void AudioStreamPreviewGenerator::_preview_thread(void *p_preview) {
}
for (int j = from; j < to; j++) {
max = MAX(max, mix_chunk[j].l);
max = MAX(max, mix_chunk[j].r);
max = MAX(max, mix_chunk[j].left);
max = MAX(max, mix_chunk[j].right);
min = MIN(min, mix_chunk[j].l);
min = MIN(min, mix_chunk[j].r);
min = MIN(min, mix_chunk[j].left);
min = MIN(min, mix_chunk[j].right);
}
uint8_t pfrom = CLAMP((min * 0.5 + 0.5) * 255, 0, 255);

View File

@ -666,11 +666,11 @@ Ref<Texture2D> EditorAudioStreamPreviewPlugin::generate(const Ref<Resource> &p_f
}
for (int j = from; j < to; j++) {
max = MAX(max, frames[j].l);
max = MAX(max, frames[j].r);
max = MAX(max, frames[j].left);
max = MAX(max, frames[j].right);
min = MIN(min, frames[j].l);
min = MIN(min, frames[j].r);
min = MIN(min, frames[j].left);
min = MIN(min, frames[j].right);
}
int pfrom = CLAMP((min * 0.5 + 0.5) * h / 2, 0, h / 2) + h / 4;

View File

@ -177,13 +177,13 @@ int AudioStreamPlaybackOggVorbis::_mix_frames_vorbis(AudioFrame *p_buffer, int p
if (info.channels > 1) {
for (int frame = 0; frame < frames; frame++) {
p_buffer[frame].l = pcm[0][frame];
p_buffer[frame].r = pcm[1][frame];
p_buffer[frame].left = pcm[0][frame];
p_buffer[frame].right = pcm[1][frame];
}
} else {
for (int frame = 0; frame < frames; frame++) {
p_buffer[frame].l = pcm[0][frame];
p_buffer[frame].r = pcm[0][frame];
p_buffer[frame].left = pcm[0][frame];
p_buffer[frame].right = pcm[0][frame];
}
}
vorbis_synthesis_read(&dsp_state, frames);

View File

@ -123,20 +123,20 @@ void AudioStreamPlayer3D::_calc_output_vol(const Vector3 &source_dir, real_t tig
switch (AudioServer::get_singleton()->get_speaker_mode()) {
case AudioServer::SPEAKER_SURROUND_71:
output.write[3].l = volumes[5]; // side-left
output.write[3].r = volumes[6]; // side-right
output.write[3].left = volumes[5]; // side-left
output.write[3].right = volumes[6]; // side-right
[[fallthrough]];
case AudioServer::SPEAKER_SURROUND_51:
output.write[2].l = volumes[3]; // rear-left
output.write[2].r = volumes[4]; // rear-right
output.write[2].left = volumes[3]; // rear-left
output.write[2].right = volumes[4]; // rear-right
[[fallthrough]];
case AudioServer::SPEAKER_SURROUND_31:
output.write[1].r = 1.0; // LFE - always full power
output.write[1].l = volumes[2]; // center
output.write[1].right = 1.0; // LFE - always full power
output.write[1].left = volumes[2]; // center
[[fallthrough]];
case AudioServer::SPEAKER_MODE_STEREO:
output.write[0].r = volumes[1]; // front-right
output.write[0].l = volumes[0]; // front-left
output.write[0].right = volumes[1]; // front-right
output.write[0].left = volumes[0]; // front-left
break;
}
}
@ -168,25 +168,25 @@ void AudioStreamPlayer3D::_calc_reverb_vol(Area3D *area, Vector3 listener_area_p
// Stereo pair.
float c = rev_pos.x * 0.5 + 0.5;
reverb_vol.write[0].l = 1.0 - c;
reverb_vol.write[0].r = c;
reverb_vol.write[0].left = 1.0 - c;
reverb_vol.write[0].right = c;
if (channel_count >= 3) {
// Center pair + Side pair
float xl = Vector3(-1, 0, -1).normalized().dot(rev_pos) * 0.5 + 0.5;
float xr = Vector3(1, 0, -1).normalized().dot(rev_pos) * 0.5 + 0.5;
reverb_vol.write[1].l = xl;
reverb_vol.write[1].r = xr;
reverb_vol.write[2].l = 1.0 - xr;
reverb_vol.write[2].r = 1.0 - xl;
reverb_vol.write[1].left = xl;
reverb_vol.write[1].right = xr;
reverb_vol.write[2].left = 1.0 - xr;
reverb_vol.write[2].right = 1.0 - xl;
}
if (channel_count >= 4) {
// Rear pair
// FIXME: Not sure what math should be done here
reverb_vol.write[3].l = 1.0 - c;
reverb_vol.write[3].r = c;
reverb_vol.write[3].left = 1.0 - c;
reverb_vol.write[3].right = c;
}
for (int i = 0; i < channel_count; i++) {

View File

@ -213,8 +213,8 @@ void AudioStreamPlaybackWAV::do_resample(const Depth *p_src, AudioFrame *p_dst,
final_r = final; //copy to right channel if stereo
}
p_dst->l = final / 32767.0;
p_dst->r = final_r / 32767.0;
p_dst->left = final / 32767.0;
p_dst->right = final_r / 32767.0;
p_dst++;
p_offset += p_increment;

View File

@ -49,7 +49,7 @@ PackedVector2Array AudioEffectCapture::get_buffer(int p_frames) {
streaming_data.resize(p_frames);
buffer.read(streaming_data.ptrw(), p_frames);
for (int32_t i = 0; i < p_frames; i++) {
ret.write[i] = Vector2(streaming_data[i].l, streaming_data[i].r);
ret.write[i] = Vector2(streaming_data[i].left, streaming_data[i].right);
}
return ret;
}

View File

@ -96,8 +96,8 @@ void AudioEffectChorusInstance::_process_chunk(const AudioFrame *p_src_frames, A
//vol modifier
AudioFrame vol_modifier = AudioFrame(base->wet, base->wet) * Math::db_to_linear(v.level);
vol_modifier.l *= CLAMP(1.0 - v.pan, 0, 1);
vol_modifier.r *= CLAMP(1.0 + v.pan, 0, 1);
vol_modifier.left *= CLAMP(1.0 - v.pan, 0, 1);
vol_modifier.right *= CLAMP(1.0 + v.pan, 0, 1);
for (int i = 0; i < p_frame_count; i++) {
/** COMPUTE WAVEFORM **/

View File

@ -59,10 +59,10 @@ void AudioEffectCompressorInstance::process(const AudioFrame *p_src_frames, Audi
for (int i = 0; i < p_frame_count; i++) {
AudioFrame s = src[i];
//convert to positive
s.l = Math::abs(s.l);
s.r = Math::abs(s.r);
s.left = Math::abs(s.left);
s.right = Math::abs(s.right);
float peak = MAX(s.l, s.r);
float peak = MAX(s.left, s.right);
float overdb = 2.08136898f * Math::linear_to_db(peak / threshold);

View File

@ -64,13 +64,13 @@ void AudioEffectDelayInstance::_process_chunk(const AudioFrame *p_src_frames, Au
AudioFrame tap1_vol = AudioFrame(tap_1_level_f, tap_1_level_f);
tap1_vol.l *= CLAMP(1.0 - base->tap_1_pan, 0, 1);
tap1_vol.r *= CLAMP(1.0 + base->tap_1_pan, 0, 1);
tap1_vol.left *= CLAMP(1.0 - base->tap_1_pan, 0, 1);
tap1_vol.right *= CLAMP(1.0 + base->tap_1_pan, 0, 1);
AudioFrame tap2_vol = AudioFrame(tap_2_level_f, tap_2_level_f);
tap2_vol.l *= CLAMP(1.0 - base->tap_2_pan, 0, 1);
tap2_vol.r *= CLAMP(1.0 + base->tap_2_pan, 0, 1);
tap2_vol.left *= CLAMP(1.0 - base->tap_2_pan, 0, 1);
tap2_vol.right *= CLAMP(1.0 + base->tap_2_pan, 0, 1);
// feedback lowpass here
float lpf_c = expf(-Math_TAU * base->feedback_lowpass / mix_rate); // 0 .. 10khz

View File

@ -46,14 +46,14 @@ void AudioEffectEQInstance::process(const AudioFrame *p_src_frames, AudioFrame *
AudioFrame dst = AudioFrame(0, 0);
for (int j = 0; j < band_count; j++) {
float l = src.l;
float r = src.r;
float l = src.left;
float r = src.right;
proc_l[j].process_one(l);
proc_r[j].process_one(r);
dst.l += l * bgain[j];
dst.r += r * bgain[j];
dst.left += l * bgain[j];
dst.right += r * bgain[j];
}
p_dst_frames[i] = dst;

View File

@ -34,7 +34,7 @@
template <int S>
void AudioEffectFilterInstance::_process_filter(const AudioFrame *p_src_frames, AudioFrame *p_dst_frames, int p_frame_count) {
for (int i = 0; i < p_frame_count; i++) {
float f = p_src_frames[i].l;
float f = p_src_frames[i].left;
filter_process[0][0].process_one(f);
if constexpr (S > 1) {
filter_process[0][1].process_one(f);
@ -46,11 +46,11 @@ void AudioEffectFilterInstance::_process_filter(const AudioFrame *p_src_frames,
filter_process[0][3].process_one(f);
}
p_dst_frames[i].l = f;
p_dst_frames[i].left = f;
}
for (int i = 0; i < p_frame_count; i++) {
float f = p_src_frames[i].r;
float f = p_src_frames[i].right;
filter_process[1][0].process_one(f);
if constexpr (S > 1) {
filter_process[1][1].process_one(f);
@ -62,7 +62,7 @@ void AudioEffectFilterInstance::_process_filter(const AudioFrame *p_src_frames,
filter_process[1][3].process_one(f);
}
p_dst_frames[i].r = f;
p_dst_frames[i].right = f;
}
}

View File

@ -41,8 +41,8 @@ void AudioEffectLimiterInstance::process(const AudioFrame *p_src_frames, AudioFr
float scmult = Math::abs((ceildb - sc) / (peakdb - sc));
for (int i = 0; i < p_frame_count; i++) {
float spl0 = p_src_frames[i].l;
float spl1 = p_src_frames[i].r;
float spl0 = p_src_frames[i].left;
float spl1 = p_src_frames[i].right;
spl0 = spl0 * makeup;
spl1 = spl1 * makeup;
float sign0 = (spl0 < 0.0 ? -1.0 : 1.0);
@ -62,8 +62,8 @@ void AudioEffectLimiterInstance::process(const AudioFrame *p_src_frames, AudioFr
spl0 = MIN(ceiling, Math::abs(spl0)) * (spl0 < 0.0 ? -1.0 : 1.0);
spl1 = MIN(ceiling, Math::abs(spl1)) * (spl1 < 0.0 ? -1.0 : 1.0);
p_dst_frames[i].l = spl0;
p_dst_frames[i].r = spl1;
p_dst_frames[i].left = spl0;
p_dst_frames[i].right = spl1;
}
}

View File

@ -35,8 +35,8 @@ void AudioEffectPannerInstance::process(const AudioFrame *p_src_frames, AudioFra
float rvol = CLAMP(1.0 + base->pan, 0, 1);
for (int i = 0; i < p_frame_count; i++) {
p_dst_frames[i].l = p_src_frames[i].l * lvol + p_src_frames[i].r * (1.0 - rvol);
p_dst_frames[i].r = p_src_frames[i].r * rvol + p_src_frames[i].l * (1.0 - lvol);
p_dst_frames[i].left = p_src_frames[i].left * lvol + p_src_frames[i].right * (1.0 - rvol);
p_dst_frames[i].right = p_src_frames[i].right * rvol + p_src_frames[i].left * (1.0 - lvol);
}
}

View File

@ -61,20 +61,20 @@ void AudioEffectPhaserInstance::process(const AudioFrame *p_src_frames, AudioFra
allpass[0][2].update(
allpass[0][3].update(
allpass[0][4].update(
allpass[0][5].update(p_src_frames[i].l + h.l * base->feedback))))));
h.l = y;
allpass[0][5].update(p_src_frames[i].left + h.left * base->feedback))))));
h.left = y;
p_dst_frames[i].l = p_src_frames[i].l + y * base->depth;
p_dst_frames[i].left = p_src_frames[i].left + y * base->depth;
y = allpass[1][0].update(
allpass[1][1].update(
allpass[1][2].update(
allpass[1][3].update(
allpass[1][4].update(
allpass[1][5].update(p_src_frames[i].r + h.r * base->feedback))))));
h.r = y;
allpass[1][5].update(p_src_frames[i].right + h.right * base->feedback))))));
h.right = y;
p_dst_frames[i].r = p_src_frames[i].r + y * base->depth;
p_dst_frames[i].right = p_src_frames[i].right + y * base->depth;
}
}

View File

@ -87,8 +87,8 @@ void AudioEffectRecordInstance::_io_store_buffer() {
while (to_read) {
AudioFrame buffered_frame = rb_buf[ring_buffer_read_pos & ring_buffer_mask];
recording_data.push_back(buffered_frame.l);
recording_data.push_back(buffered_frame.r);
recording_data.push_back(buffered_frame.left);
recording_data.push_back(buffered_frame.right);
ring_buffer_read_pos++;
to_read--;

View File

@ -51,20 +51,20 @@ void AudioEffectReverbInstance::process(const AudioFrame *p_src_frames, AudioFra
int to_mix = MIN(todo, Reverb::INPUT_BUFFER_MAX_SIZE);
for (int j = 0; j < to_mix; j++) {
tmp_src[j] = p_src_frames[offset + j].l;
tmp_src[j] = p_src_frames[offset + j].left;
}
reverb[0].process(tmp_src, tmp_dst, to_mix);
for (int j = 0; j < to_mix; j++) {
p_dst_frames[offset + j].l = tmp_dst[j];
tmp_src[j] = p_src_frames[offset + j].r;
p_dst_frames[offset + j].left = tmp_dst[j];
tmp_src[j] = p_src_frames[offset + j].right;
}
reverb[1].process(tmp_src, tmp_dst, to_mix);
for (int j = 0; j < to_mix; j++) {
p_dst_frames[offset + j].r = tmp_dst[j];
p_dst_frames[offset + j].right = tmp_dst[j];
}
offset += to_mix;

View File

@ -115,9 +115,9 @@ void AudioEffectSpectrumAnalyzerInstance::process(const AudioFrame *p_src_frames
float *fftw = temporal_fft.ptrw();
for (int i = 0; i < to_fill; i++) { //left and right buffers
float window = -0.5 * Math::cos(to_fill_step * (double)temporal_fft_pos) + 0.5;
fftw[temporal_fft_pos * 2] = window * p_src_frames->l;
fftw[temporal_fft_pos * 2] = window * p_src_frames->left;
fftw[temporal_fft_pos * 2 + 1] = 0;
fftw[(temporal_fft_pos + fft_size * 2) * 2] = window * p_src_frames->r;
fftw[(temporal_fft_pos + fft_size * 2) * 2] = window * p_src_frames->right;
fftw[(temporal_fft_pos + fft_size * 2) * 2 + 1] = 0;
++p_src_frames;
++temporal_fft_pos;
@ -135,8 +135,8 @@ void AudioEffectSpectrumAnalyzerInstance::process(const AudioFrame *p_src_frames
for (int i = 0; i < fft_size; i++) {
//abs(vec)/fft_size normalizes each frequency
hw[i].l = Vector2(fftw[i * 2], fftw[i * 2 + 1]).length() / float(fft_size);
hw[i].r = Vector2(fftw[fft_size * 4 + i * 2], fftw[fft_size * 4 + i * 2 + 1]).length() / float(fft_size);
hw[i].left = Vector2(fftw[i * 2], fftw[i * 2 + 1]).length() / float(fft_size);
hw[i].right = Vector2(fftw[fft_size * 4 + i * 2], fftw[fft_size * 4 + i * 2 + 1]).length() / float(fft_size);
}
fft_pos = next; //swap
@ -199,8 +199,8 @@ Vector2 AudioEffectSpectrumAnalyzerInstance::get_magnitude_for_frequency_range(f
Vector2 max;
for (int i = begin_pos; i <= end_pos; i++) {
max.x = MAX(max.x, r[i].l);
max.y = MAX(max.y, r[i].r);
max.x = MAX(max.x, r[i].left);
max.y = MAX(max.y, r[i].right);
}
return max;

View File

@ -39,8 +39,8 @@ void AudioEffectStereoEnhanceInstance::process(const AudioFrame *p_src_frames, A
unsigned int delay_frames = (base->time_pullout / 1000.0) * AudioServer::get_singleton()->get_mix_rate();
for (int i = 0; i < p_frame_count; i++) {
float l = p_src_frames[i].l;
float r = p_src_frames[i].r;
float l = p_src_frames[i].left;
float r = p_src_frames[i].right;
float center = (l + r) / 2.0f;
@ -65,8 +65,8 @@ void AudioEffectStereoEnhanceInstance::process(const AudioFrame *p_src_frames, A
r = delay_ringbuff[(ringbuff_pos - delay_frames) & ringbuff_mask];
}
p_dst_frames[i].l = l;
p_dst_frames[i].r = r;
p_dst_frames[i].left = l;
p_dst_frames[i].right = r;
ringbuff_pos++;
}
}

View File

@ -285,13 +285,13 @@ void AudioServer::_driver_process(int p_frames, int32_t *p_buffer) {
const AudioFrame *buf = master->channels[k].buffer.ptr();
for (int j = 0; j < to_copy; j++) {
float l = CLAMP(buf[from + j].l, -1.0, 1.0);
float l = CLAMP(buf[from + j].left, -1.0, 1.0);
int32_t vl = l * ((1 << 20) - 1);
int32_t vl2 = (vl < 0 ? -1 : 1) * (ABS(vl) << 11);
*dest = vl2;
dest++;
float r = CLAMP(buf[from + j].r, -1.0, 1.0);
float r = CLAMP(buf[from + j].right, -1.0, 1.0);
int32_t vr = r * ((1 << 20) - 1);
int32_t vr2 = (vr < 0 ? -1 : 1) * (ABS(vr) << 11);
*dest = vr2;
@ -588,22 +588,22 @@ void AudioServer::_mix_step() {
for (uint32_t j = 0; j < buffer_size; j++) {
buf[j] *= volume;
float l = ABS(buf[j].l);
if (l > peak.l) {
peak.l = l;
float l = ABS(buf[j].left);
if (l > peak.left) {
peak.left = l;
}
float r = ABS(buf[j].r);
if (r > peak.r) {
peak.r = r;
float r = ABS(buf[j].right);
if (r > peak.right) {
peak.right = r;
}
}
bus->channels.write[k].peak_volume = AudioFrame(Math::linear_to_db(peak.l + AUDIO_PEAK_OFFSET), Math::linear_to_db(peak.r + AUDIO_PEAK_OFFSET));
bus->channels.write[k].peak_volume = AudioFrame(Math::linear_to_db(peak.left + AUDIO_PEAK_OFFSET), Math::linear_to_db(peak.right + AUDIO_PEAK_OFFSET));
if (!bus->channels[k].used) {
//see if any audio is contained, because channel was not used
if (MAX(peak.r, peak.l) > Math::db_to_linear(channel_disable_threshold_db)) {
if (MAX(peak.right, peak.left) > Math::db_to_linear(channel_disable_threshold_db)) {
bus->channels.write[k].last_mix_with_audio = mix_frames;
} else if (mix_frames - bus->channels[k].last_mix_with_audio > channel_disable_frames) {
bus->channels.write[k].active = false;
@ -639,7 +639,7 @@ void AudioServer::_mix_step_for_channel(AudioFrame *p_out_buf, AudioFrame *p_sou
ERR_FAIL_NULL(p_processor_l);
ERR_FAIL_NULL(p_processor_r);
bool is_just_started = p_vol_start.l == 0 && p_vol_start.r == 0;
bool is_just_started = p_vol_start.left == 0 && p_vol_start.right == 0;
p_processor_l->set_filter(&filter, /* clear_history= */ is_just_started);
p_processor_l->update_coeffs(buffer_size);
p_processor_r->set_filter(&filter, /* clear_history= */ is_just_started);
@ -650,8 +650,8 @@ void AudioServer::_mix_step_for_channel(AudioFrame *p_out_buf, AudioFrame *p_sou
float lerp_param = (float)frame_idx / buffer_size;
AudioFrame vol = p_vol_final * lerp_param + (1 - lerp_param) * p_vol_start;
AudioFrame mixed = vol * p_source_buf[frame_idx];
p_processor_l->process_one_interp(mixed.l);
p_processor_r->process_one_interp(mixed.r);
p_processor_l->process_one_interp(mixed.left);
p_processor_r->process_one_interp(mixed.right);
p_out_buf[frame_idx] += mixed;
}
@ -1107,14 +1107,14 @@ float AudioServer::get_bus_peak_volume_left_db(int p_bus, int p_channel) const {
ERR_FAIL_INDEX_V(p_bus, buses.size(), 0);
ERR_FAIL_INDEX_V(p_channel, buses[p_bus]->channels.size(), 0);
return buses[p_bus]->channels[p_channel].peak_volume.l;
return buses[p_bus]->channels[p_channel].peak_volume.left;
}
float AudioServer::get_bus_peak_volume_right_db(int p_bus, int p_channel) const {
ERR_FAIL_INDEX_V(p_bus, buses.size(), 0);
ERR_FAIL_INDEX_V(p_channel, buses[p_bus]->channels.size(), 0);
return buses[p_bus]->channels[p_channel].peak_volume.r;
return buses[p_bus]->channels[p_channel].peak_volume.right;
}
bool AudioServer::is_bus_channel_active(int p_bus, int p_channel) const {