summaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
Diffstat (limited to 'src')
-rw-r--r--src/gpu/effect.cc16
-rw-r--r--src/gpu/effect.h13
-rw-r--r--src/gpu/effects/flash_effect.cc2
-rw-r--r--src/gpu/effects/post_process_helper.h13
-rw-r--r--src/gpu/gpu.cc6
-rw-r--r--src/gpu/gpu.h3
-rw-r--r--src/main.cc22
-rw-r--r--src/test_demo.cc26
-rw-r--r--src/tests/assets/test_sequence.cc10
9 files changed, 62 insertions, 49 deletions
diff --git a/src/gpu/effect.cc b/src/gpu/effect.cc
index 58e011c..e0a9c24 100644
--- a/src/gpu/effect.cc
+++ b/src/gpu/effect.cc
@@ -226,7 +226,8 @@ void MainSequence::resize(int width, int height) {
}
}
-void MainSequence::render_frame(float global_time, float beat, float peak,
+void MainSequence::render_frame(float global_time, float beat_time,
+ float beat_phase, float peak,
float aspect_ratio, WGPUSurface surface) {
WGPUCommandEncoder encoder =
wgpuDeviceCreateCommandEncoder(gpu_ctx.device, nullptr);
@@ -260,11 +261,12 @@ void MainSequence::render_frame(float global_time, float beat, float peak,
// Construct common uniforms once (reused for all effects)
CommonPostProcessUniforms base_uniforms = {
.resolution = {static_cast<float>(width_), static_cast<float>(height_)},
- ._pad = {0.0f, 0.0f},
.aspect_ratio = aspect_ratio,
.time = 0.0f, // Will be set per-effect
- .beat = beat,
+ .beat_time = beat_time,
+ .beat_phase = beat_phase,
.audio_intensity = peak,
+ ._pad = 0.0f,
};
for (const SequenceItem* item : scene_effects) {
@@ -564,7 +566,8 @@ void MainSequence::simulate_until(float target_time, float step_rate,
for (float t = 0.0f; t < target_time; t += step_rate) {
WGPUCommandEncoder encoder =
wgpuDeviceCreateCommandEncoder(gpu_ctx.device, nullptr);
- float beat = fmodf(t * bpm / 60.0f, 1.0f);
+ float absolute_beat_time = t * bpm / 60.0f;
+ float beat_phase = fmodf(absolute_beat_time, 1.0f);
std::vector<SequenceItem*> scene_effects, post_effects;
for (ActiveSequence& entry : sequences_) {
if (t >= entry.start_time) {
@@ -575,11 +578,12 @@ void MainSequence::simulate_until(float target_time, float step_rate,
for (const SequenceItem* item : scene_effects) {
CommonPostProcessUniforms test_uniforms = {
.resolution = {static_cast<float>(width_), static_cast<float>(height_)},
- ._pad = {0.0f, 0.0f},
.aspect_ratio = aspect_ratio,
.time = t - item->start_time,
- .beat = beat,
+ .beat_time = absolute_beat_time,
+ .beat_phase = beat_phase,
.audio_intensity = 0.0f,
+ ._pad = 0.0f,
};
item->effect->compute(encoder, test_uniforms);
}
diff --git a/src/gpu/effect.h b/src/gpu/effect.h
index ed90ac7..b9709a4 100644
--- a/src/gpu/effect.h
+++ b/src/gpu/effect.h
@@ -49,16 +49,19 @@ class Effect {
// Helper: get initialized CommonPostProcessUniforms based on current dimensions
// If aspect_ratio < 0, computes from width_/height_
- CommonPostProcessUniforms get_common_uniforms(float time = 0.0f, float beat = 0.0f,
+ CommonPostProcessUniforms get_common_uniforms(float time = 0.0f,
+ float beat_time = 0.0f,
+ float beat_phase = 0.0f,
float intensity = 0.0f,
float aspect_ratio = -1.0f) const {
return {
.resolution = {static_cast<float>(width_), static_cast<float>(height_)},
- ._pad = {0.0f, 0.0f},
.aspect_ratio = aspect_ratio < 0.0f ? static_cast<float>(width_) / static_cast<float>(height_) : aspect_ratio,
.time = time,
- .beat = beat,
+ .beat_time = beat_time,
+ .beat_phase = beat_phase,
.audio_intensity = intensity,
+ ._pad = 0.0f,
};
}
@@ -130,8 +133,8 @@ class MainSequence {
void init_test(const GpuContext& ctx);
void add_sequence(std::shared_ptr<Sequence> seq, float start_time,
int priority = 0);
- void render_frame(float global_time, float beat, float peak,
- float aspect_ratio, WGPUSurface surface);
+ void render_frame(float global_time, float beat_time, float beat_phase,
+ float peak, float aspect_ratio, WGPUSurface surface);
void resize(int width, int height);
void shutdown();
diff --git a/src/gpu/effects/flash_effect.cc b/src/gpu/effects/flash_effect.cc
index 4357c34..e53cbce 100644
--- a/src/gpu/effects/flash_effect.cc
+++ b/src/gpu/effects/flash_effect.cc
@@ -77,7 +77,7 @@ void FlashEffect::render(WGPURenderPassEncoder pass,
// Animate color based on time and beat
const float r = params_.color[0] * (0.5f + 0.5f * sinf(uniforms.time * 0.5f));
const float g = params_.color[1] * (0.5f + 0.5f * cosf(uniforms.time * 0.7f));
- const float b = params_.color[2] * (1.0f + 0.3f * uniforms.beat);
+ const float b = params_.color[2] * (1.0f + 0.3f * uniforms.beat_phase);
// Update uniforms with computed (animated) values
const FlashUniforms u = {
diff --git a/src/gpu/effects/post_process_helper.h b/src/gpu/effects/post_process_helper.h
index 23cde0e..1c649e7 100644
--- a/src/gpu/effects/post_process_helper.h
+++ b/src/gpu/effects/post_process_helper.h
@@ -8,12 +8,13 @@
// Uniform data common to all post-processing effects
struct CommonPostProcessUniforms {
- vec2 resolution;
- float _pad[2]; // Padding for 16-byte alignment
- float aspect_ratio;
- float time;
- float beat;
- float audio_intensity;
+ vec2 resolution; // Screen dimensions
+ float aspect_ratio; // Width/height ratio
+ float time; // Physical time in seconds (unaffected by tempo)
+ float beat_time; // Musical time in beats (absolute, tempo-scaled)
+ float beat_phase; // Fractional beat (0.0-1.0 within current beat)
+ float audio_intensity;// Audio peak for beat sync
+ float _pad; // Padding for 16-byte alignment
};
static_assert(sizeof(CommonPostProcessUniforms) == 32,
"CommonPostProcessUniforms must be 32 bytes for WGSL alignment");
diff --git a/src/gpu/gpu.cc b/src/gpu/gpu.cc
index e89a2f0..41f5bcf 100644
--- a/src/gpu/gpu.cc
+++ b/src/gpu/gpu.cc
@@ -381,8 +381,10 @@ void gpu_init(PlatformState* platform_state) {
platform_state->height);
}
-void gpu_draw(float audio_peak, float aspect_ratio, float time, float beat) {
- g_main_sequence.render_frame(time, beat, audio_peak, aspect_ratio, g_surface);
+void gpu_draw(float audio_peak, float aspect_ratio, float time,
+ float beat_time, float beat_phase) {
+ g_main_sequence.render_frame(time, beat_time, beat_phase, audio_peak,
+ aspect_ratio, g_surface);
}
void gpu_resize(int width, int height) {
diff --git a/src/gpu/gpu.h b/src/gpu/gpu.h
index 8c59aee..c7ee89f 100644
--- a/src/gpu/gpu.h
+++ b/src/gpu/gpu.h
@@ -42,7 +42,8 @@ struct RenderPass {
class MainSequence; // Forward declaration
void gpu_init(PlatformState* platform_state);
-void gpu_draw(float audio_peak, float aspect_ratio, float time, float beat);
+void gpu_draw(float audio_peak, float aspect_ratio, float time,
+ float beat_time, float beat_phase);
void gpu_resize(int width, int height);
void gpu_shutdown();
diff --git a/src/main.cc b/src/main.cc
index 6132841..41c881b 100644
--- a/src/main.cc
+++ b/src/main.cc
@@ -325,11 +325,10 @@ int main(int argc, char** argv) {
const float raw_peak = audio_get_realtime_peak();
const float visual_peak = fminf(raw_peak * 8.0f, 1.0f);
- // Beat calculation should use audio time to align with audio events.
- // The graphics loop time (current_physical_time) is used for frame rate.
- const float beat_time = current_audio_time * g_tracker_score.bpm / 60.0f;
- const int beat_number = (int)beat_time;
- const float beat = fmodf(beat_time, 1.0f); // Fractional part (0.0 to 1.0)
+ // Beat calculation: convert audio time to musical beats
+ const float absolute_beat_time = current_audio_time * g_tracker_score.bpm / 60.0f;
+ const int beat_number = (int)absolute_beat_time;
+ const float beat_phase = fmodf(absolute_beat_time, 1.0f); // Fractional part (0.0 to 1.0)
// Print beat/time info periodically for identifying sync points
// Use graphics time for the print interval to avoid excessive output if
@@ -339,20 +338,21 @@ int main(int argc, char** argv) {
0.5f) { // Print every 0.5 seconds
if (tempo_test_enabled) {
printf(
- "[GraphicsT=%.2f, AudioT=%.2f, MusicT=%.2f, Beat=%d, Frac=%.2f, "
+ "[GraphicsT=%.2f, AudioT=%.2f, MusicT=%.2f, Beat=%d, Phase=%.2f, "
"Peak=%.2f, Tempo=%.2fx]\n",
current_physical_time, current_audio_time, g_music_time,
- beat_number, beat, visual_peak, g_tempo_scale);
+ beat_number, beat_phase, visual_peak, g_tempo_scale);
} else {
- printf("[GraphicsT=%.2f, AudioT=%.2f, Beat=%d, Frac=%.2f, Peak=%.2f]\n",
- current_physical_time, current_audio_time, beat_number, beat,
+ printf("[GraphicsT=%.2f, AudioT=%.2f, Beat=%d, Phase=%.2f, Peak=%.2f]\n",
+ current_physical_time, current_audio_time, beat_number, beat_phase,
visual_peak);
}
last_graphics_print_time = current_physical_time;
}
- // Draw graphics using the graphics frame time and synchronized audio events
- gpu_draw(visual_peak, aspect_ratio, (float)current_physical_time, beat);
+ // Draw graphics using physical time and musical beat time
+ gpu_draw(visual_peak, aspect_ratio, (float)current_physical_time,
+ absolute_beat_time, beat_phase);
last_frame_time = current_physical_time;
// Update audio systems (tracker, synth, etc.) based on audio time
diff --git a/src/test_demo.cc b/src/test_demo.cc
index b8e9381..edbcae0 100644
--- a/src/test_demo.cc
+++ b/src/test_demo.cc
@@ -104,8 +104,10 @@ class PeakMeterEffect : public PostProcessEffect {
.resolution = {(float)width_, (float)height_},
.aspect_ratio = aspect_ratio,
.time = time,
- .beat = beat,
+ .beat_time = beat,
+ .beat_phase = beat,
.audio_intensity = peak_value,
+ ._pad = 0.0f,
};
uniforms_.update(ctx_.queue, u);
@@ -347,11 +349,10 @@ int main(int argc, char** argv) {
const float raw_peak = audio_get_realtime_peak();
const float visual_peak = fminf(raw_peak * 8.0f, 1.0f);
- // Beat calculation should use audio time to align with audio events.
- // The graphics loop time (current_physical_time) is used for frame rate.
- const float beat_time = current_audio_time * g_tracker_score.bpm / 60.0f;
- const int beat_number = (int)beat_time;
- const float beat = fmodf(beat_time, 1.0f); // Fractional part (0.0 to 1.0)
+ // Beat calculation: convert audio time to musical beats
+ const float absolute_beat_time = current_audio_time * g_tracker_score.bpm / 60.0f;
+ const int beat_number = (int)absolute_beat_time;
+ const float beat_phase = fmodf(absolute_beat_time, 1.0f); // Fractional part (0.0 to 1.0)
#if !defined(STRIP_ALL)
// Log peak (either per-frame or per-beat)
@@ -377,22 +378,23 @@ int main(int argc, char** argv) {
if (current_physical_time - last_graphics_print_time >= 0.5f) {
if (tempo_test_enabled) {
printf(
- "[GraphicsT=%.2f, AudioT=%.2f, MusicT=%.2f, Beat=%d, Frac=%.2f, "
+ "[GraphicsT=%.2f, AudioT=%.2f, MusicT=%.2f, Beat=%d, Phase=%.2f, "
"Peak=%.2f, Tempo=%.2fx]\n",
current_physical_time, current_audio_time, g_music_time,
- beat_number, beat, visual_peak, g_tempo_scale);
+ beat_number, beat_phase, visual_peak, g_tempo_scale);
} else {
- printf("[GraphicsT=%.2f, AudioT=%.2f, Beat=%d, Frac=%.2f, Peak=%.2f]\n",
- current_physical_time, current_audio_time, beat_number, beat,
+ printf("[GraphicsT=%.2f, AudioT=%.2f, Beat=%d, Phase=%.2f, Peak=%.2f]\n",
+ current_physical_time, current_audio_time, beat_number, beat_phase,
visual_peak);
}
last_graphics_print_time = current_physical_time;
}
#endif
- // Draw graphics using the graphics frame time and synchronized audio events
+ // Draw graphics using physical time and musical beat time
const float graphics_frame_time = (float)current_physical_time;
- gpu_draw(visual_peak, aspect_ratio, graphics_frame_time, beat);
+ gpu_draw(visual_peak, aspect_ratio, graphics_frame_time,
+ absolute_beat_time, beat_phase);
// Update audio systems (tracker, synth, etc.) based on audio time
// progression
diff --git a/src/tests/assets/test_sequence.cc b/src/tests/assets/test_sequence.cc
index 44aac46..157b462 100644
--- a/src/tests/assets/test_sequence.cc
+++ b/src/tests/assets/test_sequence.cc
@@ -96,7 +96,7 @@ void test_effect_lifecycle() {
main_seq.add_sequence(seq1, 0.0f, 0);
// Before effect starts
- main_seq.render_frame(0.5f, 0, 0, 1.0f,
+ main_seq.render_frame(0.5f, 0, 0, 0, 1.0f,
dummy_surface); // This will still call real render, but
// test counts only init
assert(effect1->init_calls == 1);
@@ -105,26 +105,26 @@ void test_effect_lifecycle() {
assert(effect1->end_calls == 0);
// Effect starts
- main_seq.render_frame(1.0f, 0, 0, 1.0f, dummy_surface);
+ main_seq.render_frame(1.0f, 0, 0, 0, 1.0f, dummy_surface);
assert(effect1->start_calls == 1);
// assert(effect1->render_calls == 1); // No longer checking render calls
// directly from here
assert(effect1->end_calls == 0);
// During effect
- main_seq.render_frame(2.0f, 0, 0, 1.0f, dummy_surface);
+ main_seq.render_frame(2.0f, 0, 0, 0, 1.0f, dummy_surface);
assert(effect1->start_calls == 1);
// assert(effect1->render_calls == 2);
assert(effect1->end_calls == 0);
// Effect ends
- main_seq.render_frame(3.0f, 0, 0, 1.0f, dummy_surface);
+ main_seq.render_frame(3.0f, 0, 0, 0, 1.0f, dummy_surface);
assert(effect1->start_calls == 1);
// assert(effect1->render_calls == 2); // Render not called on end frame
assert(effect1->end_calls == 1);
// After effect ends
- main_seq.render_frame(3.5f, 0, 0, 1.0f, dummy_surface);
+ main_seq.render_frame(3.5f, 0, 0, 0, 1.0f, dummy_surface);
assert(effect1->start_calls == 1);
// assert(effect1->render_calls == 2);
assert(effect1->end_calls == 1);