Upload files to "Source"

This commit is contained in:
ed
2025-10-22 16:58:14 +00:00
parent 8ecb5e774d
commit 1c23e3a1d0
2 changed files with 366 additions and 409 deletions

View File

@@ -1,308 +1,276 @@
/* #include "PluginProcessor.h"
============================================================================== #include "PluginEditor.h"
This file contains the basic framework code for a JUCE plugin processor. //==============================================================================
NeuralSynthAudioProcessor::NeuralSynthAudioProcessor()
============================================================================== : parameters(*this, nullptr, "PARAMETERS", createParameterLayout())
*/ , AudioProcessor(BusesProperties().withOutput("Output", juce::AudioChannelSet::stereo(), true))
, audioEngine(sp)
#include "PluginProcessor.h" {
#include "PluginEditor.h" parameters.addParameterListener("waveform", this);
// Wavetable params
//============================================================================== sp.wtOn = parameters.getRawParameterValue("wt_on");
NeuralSynthAudioProcessor::NeuralSynthAudioProcessor() : parameters(*this, nullptr, "PARAMETERS", createParameterLayout()) sp.wtMorph = parameters.getRawParameterValue("wt_morph");
, AudioProcessor(BusesProperties().withOutput("Output", juce::AudioChannelSet::stereo(), true))
, audioEngine(sp) // === Per-panel bypass (default OFF) ===
{ sp.chorusOn = parameters.getRawParameterValue("chorus_on");
parameters.addParameterListener("waveform", this); sp.delayOn = parameters.getRawParameterValue("delay_on");
//sp.waveform = parameters.getRawParameterValue("waveform"); sp.reverbOn = parameters.getRawParameterValue("reverb_on");
sp.flangerOn = parameters.getRawParameterValue("flanger_on");
// === Chorus === sp.distortionOn = parameters.getRawParameterValue("distortion_on");
parameters.addParameterListener("chorus_rate", this); sp.filterOn = parameters.getRawParameterValue("filter_on");
parameters.addParameterListener("chorus_depth", this); sp.eqOn = parameters.getRawParameterValue("eq_on");
parameters.addParameterListener("chorus_centre", this);
parameters.addParameterListener("chorus_feedback", this); // === Chorus ===
parameters.addParameterListener("chorus_mix", this); parameters.addParameterListener("chorus_rate", this);
parameters.addParameterListener("chorus_depth", this);
sp.chorusRate = parameters.getRawParameterValue("chorus_rate"); parameters.addParameterListener("chorus_centre", this);
sp.chorusDepth = parameters.getRawParameterValue("chorus_depth"); parameters.addParameterListener("chorus_feedback", this);
sp.chorusCentre = parameters.getRawParameterValue("chorus_centre"); parameters.addParameterListener("chorus_mix", this);
sp.chorusFeedback = parameters.getRawParameterValue("chorus_feedback");
sp.chorusMix = parameters.getRawParameterValue("chorus_mix"); sp.chorusRate = parameters.getRawParameterValue("chorus_rate");
sp.chorusDepth = parameters.getRawParameterValue("chorus_depth");
// === Delay === sp.chorusCentre = parameters.getRawParameterValue("chorus_centre");
parameters.addParameterListener("delay_delay", this); sp.chorusFeedback = parameters.getRawParameterValue("chorus_feedback");
sp.delayTime = parameters.getRawParameterValue("delay_delay"); sp.chorusMix = parameters.getRawParameterValue("chorus_mix");
// === Reverb === // === Delay ===
parameters.addParameterListener("reverb_roomSize", this); parameters.addParameterListener("delay_delay", this);
parameters.addParameterListener("reverb_damping", this); sp.delayTime = parameters.getRawParameterValue("delay_delay");
parameters.addParameterListener("reverb_wetLevel", this);
parameters.addParameterListener("reverb_dryLevel", this); // === Reverb ===
parameters.addParameterListener("reverb_width", this); parameters.addParameterListener("reverb_roomSize", this);
parameters.addParameterListener("reverb_freezeMode", this); parameters.addParameterListener("reverb_damping", this);
parameters.addParameterListener("reverb_wetLevel", this);
sp.reverbRoomSize = parameters.getRawParameterValue("reverb_roomSize"); parameters.addParameterListener("reverb_dryLevel", this);
sp.reverbDamping = parameters.getRawParameterValue("reverb_damping"); parameters.addParameterListener("reverb_width", this);
sp.reverbWetLevel = parameters.getRawParameterValue("reverb_wetLevel"); parameters.addParameterListener("reverb_freezeMode", this);
sp.reverbDryLevel = parameters.getRawParameterValue("reverb_dryLevel");
sp.reverbWidth = parameters.getRawParameterValue("reverb_width"); sp.reverbRoomSize = parameters.getRawParameterValue("reverb_roomSize");
sp.reverbFreezeMode = parameters.getRawParameterValue("reverb_freezeMode"); sp.reverbDamping = parameters.getRawParameterValue("reverb_damping");
sp.reverbWetLevel = parameters.getRawParameterValue("reverb_wetLevel");
// === ADSR === sp.reverbDryLevel = parameters.getRawParameterValue("reverb_dryLevel");
parameters.addParameterListener("adsr_attack", this); sp.reverbWidth = parameters.getRawParameterValue("reverb_width");
parameters.addParameterListener("adsr_decay", this); sp.reverbFreezeMode= parameters.getRawParameterValue("reverb_freezeMode");
parameters.addParameterListener("adsr_sustain", this);
parameters.addParameterListener("adsr_release", this); // === Amp ADSR ===
parameters.addParameterListener("adsr_attack", this);
sp.adsrAttack = parameters.getRawParameterValue("adsr_attack"); parameters.addParameterListener("adsr_decay", this);
sp.adsrDecay = parameters.getRawParameterValue("adsr_decay"); parameters.addParameterListener("adsr_sustain", this);
sp.adsrSustain = parameters.getRawParameterValue("adsr_sustain"); parameters.addParameterListener("adsr_release", this);
sp.adsrRelease = parameters.getRawParameterValue("adsr_release");
sp.adsrAttack = parameters.getRawParameterValue("adsr_attack");
// === Flanger === sp.adsrDecay = parameters.getRawParameterValue("adsr_decay");
parameters.addParameterListener("flanger_rate", this); sp.adsrSustain = parameters.getRawParameterValue("adsr_sustain");
parameters.addParameterListener("flanger_depth", this); sp.adsrRelease = parameters.getRawParameterValue("adsr_release");
parameters.addParameterListener("flanger_feedback", this);
parameters.addParameterListener("flanger_dryMix", this); // === Filter Env ===
parameters.addParameterListener("flanger_phase", this); parameters.addParameterListener("fenv_attack", this);
parameters.addParameterListener("flanger_delay", this); parameters.addParameterListener("fenv_decay", this);
parameters.addParameterListener("fenv_sustain", this);
sp.flangerRate = parameters.getRawParameterValue("flanger_rate"); parameters.addParameterListener("fenv_release", this);
sp.flangerDepth = parameters.getRawParameterValue("flanger_depth"); parameters.addParameterListener("fenv_amount", this);
sp.flangerFeedback = parameters.getRawParameterValue("flanger_feedback");
sp.flangerDryMix = parameters.getRawParameterValue("flanger_dryMix"); sp.fenvAttack = parameters.getRawParameterValue("fenv_attack");
sp.flangerPhase = parameters.getRawParameterValue("flanger_phase"); sp.fenvDecay = parameters.getRawParameterValue("fenv_decay");
sp.flangerDelay = parameters.getRawParameterValue("flanger_delay"); sp.fenvSustain = parameters.getRawParameterValue("fenv_sustain");
sp.fenvRelease = parameters.getRawParameterValue("fenv_release");
// === Filter === sp.fenvAmount = parameters.getRawParameterValue("fenv_amount");
parameters.addParameterListener("filter_cutoff", this);
parameters.addParameterListener("filter_resonance", this); // === Filter base ===
parameters.addParameterListener("filter_type", this); parameters.addParameterListener("filter_cutoff", this);
parameters.addParameterListener("filter_drive", this); parameters.addParameterListener("filter_resonance", this);
parameters.addParameterListener("filter_mod", this); parameters.addParameterListener("filter_type", this);
parameters.addParameterListener("filter_key", this); parameters.addParameterListener("filter_drive", this);
parameters.addParameterListener("filter_mod", this);
sp.filterCutoff = parameters.getRawParameterValue("filter_cutoff"); parameters.addParameterListener("filter_key", this);
sp.filterResonance = parameters.getRawParameterValue("filter_resonance");
sp.filterType = parameters.getRawParameterValue("filter_type"); sp.filterCutoff = parameters.getRawParameterValue("filter_cutoff");
sp.filterDrive = parameters.getRawParameterValue("filter_drive"); sp.filterResonance = parameters.getRawParameterValue("filter_resonance");
sp.filterMod = parameters.getRawParameterValue("filter_mod"); sp.filterType = parameters.getRawParameterValue("filter_type");
sp.filterKey = parameters.getRawParameterValue("filter_key"); sp.filterDrive = parameters.getRawParameterValue("filter_drive");
sp.filterMod = parameters.getRawParameterValue("filter_mod");
// === Distortion === sp.filterKey = parameters.getRawParameterValue("filter_key");
parameters.addParameterListener("distortion_drive", this);
parameters.addParameterListener("distortion_mix", this); // === Distortion ===
parameters.addParameterListener("distortion_bias", this); parameters.addParameterListener("distortion_drive", this);
parameters.addParameterListener("distortion_tone", this); parameters.addParameterListener("distortion_mix", this);
parameters.addParameterListener("distortion_shape", this); parameters.addParameterListener("distortion_bias", this);
parameters.addParameterListener("distortion_tone", this);
sp.distortionDrive = parameters.getRawParameterValue("distortion_drive"); parameters.addParameterListener("distortion_shape", this);
sp.distortionMix = parameters.getRawParameterValue("distortion_mix");
sp.distortionBias = parameters.getRawParameterValue("distortion_bias"); sp.distortionDrive = parameters.getRawParameterValue("distortion_drive");
sp.distortionTone = parameters.getRawParameterValue("distortion_tone"); sp.distortionMix = parameters.getRawParameterValue("distortion_mix");
sp.distortionShape = parameters.getRawParameterValue("distortion_shape"); sp.distortionBias = parameters.getRawParameterValue("distortion_bias");
sp.distortionTone = parameters.getRawParameterValue("distortion_tone");
sp.distortionShape = parameters.getRawParameterValue("distortion_shape");
parameters.addParameterListener("master", this); // === Master / EQ ===
parameters.addParameterListener("lowEQ", this); parameters.addParameterListener("master", this);
parameters.addParameterListener("midEQ", this); parameters.addParameterListener("lowEQ", this);
parameters.addParameterListener("highEQ", this); parameters.addParameterListener("midEQ", this);
parameters.addParameterListener("highEQ", this);
sp.masterDbls = parameters.getRawParameterValue("master");
sp.lowGainDbls = parameters.getRawParameterValue("lowEQ"); sp.masterDbls = parameters.getRawParameterValue("master");
sp.midGainDbls = parameters.getRawParameterValue("midEQ"); sp.lowGainDbls = parameters.getRawParameterValue("lowEQ");
sp.highGainDbls = parameters.getRawParameterValue("highEQ"); sp.midGainDbls = parameters.getRawParameterValue("midEQ");
} sp.highGainDbls = parameters.getRawParameterValue("highEQ");
}
NeuralSynthAudioProcessor::~NeuralSynthAudioProcessor()
{ NeuralSynthAudioProcessor::~NeuralSynthAudioProcessor() = default;
}
//==============================================================================
//============================================================================== const juce::String NeuralSynthAudioProcessor::getName() const { return JucePlugin_Name; }
const juce::String NeuralSynthAudioProcessor::getName() const
{ bool NeuralSynthAudioProcessor::acceptsMidi() const
return JucePlugin_Name; {
} #if JucePlugin_WantsMidiInput
return true;
bool NeuralSynthAudioProcessor::acceptsMidi() const #else
{ return false;
#if JucePlugin_WantsMidiInput #endif
return true; }
#else
return false; bool NeuralSynthAudioProcessor::producesMidi() const
#endif {
} #if JucePlugin_ProducesMidiOutput
return true;
bool NeuralSynthAudioProcessor::producesMidi() const #else
{ return false;
#if JucePlugin_ProducesMidiOutput #endif
return true; }
#else
return false; bool NeuralSynthAudioProcessor::isMidiEffect() const
#endif {
} #if JucePlugin_IsMidiEffect
return true;
bool NeuralSynthAudioProcessor::isMidiEffect() const #else
{ return false;
#if JucePlugin_IsMidiEffect #endif
return true; }
#else
return false; double NeuralSynthAudioProcessor::getTailLengthSeconds() const { return 0.0; }
#endif
} int NeuralSynthAudioProcessor::getNumPrograms() { return 1; }
int NeuralSynthAudioProcessor::getCurrentProgram() { return 0; }
double NeuralSynthAudioProcessor::getTailLengthSeconds() const void NeuralSynthAudioProcessor::setCurrentProgram (int) {}
{ const juce::String NeuralSynthAudioProcessor::getProgramName (int) { return {}; }
return 0.0; void NeuralSynthAudioProcessor::changeProgramName (int, const juce::String&) {}
}
//==============================================================================
int NeuralSynthAudioProcessor::getNumPrograms() void NeuralSynthAudioProcessor::prepareToPlay (double sampleRate, int samplesPerBlock)
{ {
return 1; // NB: some hosts don't cope very well if you tell them there are 0 programs, audioEngine.prepare({ sampleRate, (juce::uint32)samplesPerBlock, 2 });
// so this should be at least 1, even if you're not really implementing programs. }
}
void NeuralSynthAudioProcessor::releaseResources() {}
int NeuralSynthAudioProcessor::getCurrentProgram()
{ bool NeuralSynthAudioProcessor::isBusesLayoutSupported (const BusesLayout& layouts) const
return 0; {
} if (layouts.getMainOutputChannelSet() != juce::AudioChannelSet::mono()
&& layouts.getMainOutputChannelSet() != juce::AudioChannelSet::stereo())
void NeuralSynthAudioProcessor::setCurrentProgram (int index) return false;
{
} return true;
}
const juce::String NeuralSynthAudioProcessor::getProgramName (int index)
{ void NeuralSynthAudioProcessor::processBlock(juce::AudioSampleBuffer& buffer, juce::MidiBuffer& midiMessages)
return {}; {
} const int newWaveform = sp.waveform.exchange(-1);
void NeuralSynthAudioProcessor::changeProgramName (int index, const juce::String& newName) if (newWaveform != -1) {
{ audioEngine.applyToVoices([newWaveform](NeuralSynthVoice* v)
} {
v->changeWaveform(newWaveform);
//============================================================================== });
void NeuralSynthAudioProcessor::prepareToPlay (double sampleRate, int samplesPerBlock) }
{
audioEngine.prepare({ sampleRate, (juce::uint32)samplesPerBlock, 2 }); juce::ScopedNoDenormals noDenormals;
midiMessageCollector.reset(sampleRate); auto totalNumInputChannels = getTotalNumInputChannels();
} auto totalNumOutputChannels = getTotalNumOutputChannels();
void NeuralSynthAudioProcessor::releaseResources() for (int i = totalNumInputChannels; i < totalNumOutputChannels; ++i)
{ buffer.clear(i, 0, buffer.getNumSamples());
// When playback stops, you can use this as an opportunity to free up any
// spare memory, etc. audioEngine.renderNextBlock(buffer, midiMessages, 0, buffer.getNumSamples());
} scopeDataCollector.process(buffer.getReadPointer(0), (size_t)buffer.getNumSamples());
}
bool NeuralSynthAudioProcessor::isBusesLayoutSupported (const BusesLayout& layouts) const
{ //==============================================================================
// This is the place where you check if the layout is supported. bool NeuralSynthAudioProcessor::hasEditor() const { return true; }
// In this template code we only support mono or stereo.
if (layouts.getMainOutputChannelSet() != juce::AudioChannelSet::mono() juce::AudioProcessorEditor* NeuralSynthAudioProcessor::createEditor()
&& layouts.getMainOutputChannelSet() != juce::AudioChannelSet::stereo()) {
return false; return new NeuralSynthAudioProcessorEditor (*this);
}
return true;
} //==============================================================================
void NeuralSynthAudioProcessor::getStateInformation (juce::MemoryBlock& destData) { juce::ignoreUnused(destData); }
void NeuralSynthAudioProcessor::processBlock(juce::AudioSampleBuffer& buffer, juce::MidiBuffer& midiMessages) void NeuralSynthAudioProcessor::setStateInformation (const void* data, int sizeInBytes) { juce::ignoreUnused(data, sizeInBytes); }
{
const int newWaveform = sp.waveform.exchange(-1); void NeuralSynthAudioProcessor::parameterChanged(const juce::String& id, float newValue)
{
if (newWaveform != -1) { juce::ignoreUnused(newValue);
audioEngine.applyToVoices([newWaveform](NeuralSynthVoice* v) if (id == "waveform")
{ sp.waveform.store((int)newValue, std::memory_order_release);
v->changeWaveform(newWaveform); }
});
} //==============================================================================
// This creates new instances of the plugin..
juce::ScopedNoDenormals noDenormals; juce::AudioProcessor* JUCE_CALLTYPE createPluginFilter() { return new NeuralSynthAudioProcessor(); }
auto totalNumInputChannels = getTotalNumInputChannels();
auto totalNumOutputChannels = getTotalNumOutputChannels(); void NeuralSynthAudioProcessor::buildParams(std::vector<std::unique_ptr<juce::RangedAudioParameter>>& params, const std::string& paramGroup) {
const auto& paramGroupSettings = PARAM_SETTINGS.at(paramGroup);
midiMessageCollector.removeNextBlockOfMessages(midiMessages, buffer.getNumSamples());
for (const auto& [name, s] : paramGroupSettings) {
for (int i = totalNumInputChannels; i < totalNumOutputChannels; ++i) params.push_back(std::make_unique<juce::AudioParameterFloat>(
buffer.clear(i, 0, buffer.getNumSamples()); paramGroup + "_" + name, s.label,
juce::NormalisableRange<float>(s.min, s.max, s.interval),
audioEngine.renderNextBlock(buffer, midiMessages, 0, buffer.getNumSamples()); s.defValue));
scopeDataCollector.process(buffer.getReadPointer(0), (size_t)buffer.getNumSamples()); }
} }
//============================================================================== juce::AudioProcessorValueTreeState::ParameterLayout NeuralSynthAudioProcessor::createParameterLayout()
bool NeuralSynthAudioProcessor::hasEditor() const {
{ std::vector<std::unique_ptr<juce::RangedAudioParameter>> params;
return true; // (change this to false if you choose to not supply an editor)
} params.push_back(std::make_unique<juce::AudioParameterChoice>(
"waveform", "Waveform",
juce::AudioProcessorEditor* NeuralSynthAudioProcessor::createEditor() juce::StringArray{ "Sine", "Saw", "Square", "Triangle" }, 0));
{ // --- Wavetable on/off + morph position (0..15) ---
return new NeuralSynthAudioProcessorEditor (*this); params.push_back(std::make_unique<juce::AudioParameterBool>("wt_on", "Wavetable On", true));
} params.push_back(std::make_unique<juce::AudioParameterFloat>(
"wt_morph", "WT Morph",
//============================================================================== juce::NormalisableRange<float>(0.0f, 15.0f, 0.001f), 0.0f));
void NeuralSynthAudioProcessor::getStateInformation (juce::MemoryBlock& destData)
{
// You should use this method to store your parameters in the memory block. // Per-panel bypass toggles (default OFF)
// You could do that either as raw data, or use the XML or ValueTree classes params.push_back(std::make_unique<juce::AudioParameterBool>("chorus_on", "Chorus On", false));
// as intermediaries to make it easy to save and load complex data. params.push_back(std::make_unique<juce::AudioParameterBool>("delay_on", "Delay On", false));
} params.push_back(std::make_unique<juce::AudioParameterBool>("reverb_on", "Reverb On", false));
params.push_back(std::make_unique<juce::AudioParameterBool>("flanger_on", "Flanger On", false));
void NeuralSynthAudioProcessor::setStateInformation (const void* data, int sizeInBytes) params.push_back(std::make_unique<juce::AudioParameterBool>("distortion_on", "Distortion On", false));
{ params.push_back(std::make_unique<juce::AudioParameterBool>("filter_on", "Filter On", false));
// You should use this method to restore your parameters from this memory block, params.push_back(std::make_unique<juce::AudioParameterBool>("eq_on", "EQ On", false));
// whose contents will have been created by the getStateInformation() call.
} buildParams(params, "adsr");
buildParams(params, "fenv");
void NeuralSynthAudioProcessor::parameterChanged(const juce::String& id, float newValue) buildParams(params, "chorus");
{ buildParams(params, "delay");
if (id == "waveform") buildParams(params, "reverb");
sp.waveform.store((int)newValue, std::memory_order_release); buildParams(params, "flanger");
} buildParams(params, "distortion");
buildParams(params, "filter");
//==============================================================================
// This creates new instances of the plugin.. params.push_back(std::make_unique<juce::AudioParameterFloat>("master", "Master",
juce::AudioProcessor* JUCE_CALLTYPE createPluginFilter() juce::NormalisableRange<float>(-24.0f, 24.0f, 0.1f), 0.1f));
{
return new NeuralSynthAudioProcessor(); params.push_back(std::make_unique<juce::AudioParameterFloat>("lowEQ", "Low Gain",
} juce::NormalisableRange<float>(-24.0f, 24.0f, 0.1f), 0.5f));
params.push_back(std::make_unique<juce::AudioParameterFloat>("midEQ", "Mid EQ",
void NeuralSynthAudioProcessor::buildParams(std::vector<std::unique_ptr<juce::RangedAudioParameter>> &params, const std::string& paramGroup) { juce::NormalisableRange<float>(-24.0f, 24.0f, 0.1f), 0.8f));
const auto& paramGroupSettings = PARAM_SETTINGS.at(paramGroup); params.push_back(std::make_unique<juce::AudioParameterFloat>("highEQ", "High EQ",
juce::NormalisableRange<float>(-24.0f, 24.0f, 0.1f), 1.0f));
for (const auto& [name, paramSettings] : paramGroupSettings) {
params.push_back(std::make_unique<juce::AudioParameterFloat>(paramGroup + "_" + name, paramSettings.label, return { params.begin(), params.end() };
juce::NormalisableRange<float>(paramSettings.min, paramSettings.max, paramSettings.interval), }
paramSettings.defValue));
}
}
juce::AudioProcessorValueTreeState::ParameterLayout NeuralSynthAudioProcessor::createParameterLayout()
{
std::vector<std::unique_ptr<juce::RangedAudioParameter>> params;
params.push_back(std::make_unique<juce::AudioParameterChoice>(
"waveform", "Waveform",
juce::StringArray{ "Sine", "Saw", "Square", "Triangle" }, 0));
buildParams(params, "adsr");
buildParams(params, "chorus");
buildParams(params, "delay");
buildParams(params, "reverb");
buildParams(params, "flanger");
buildParams(params, "distortion");
buildParams(params, "filter");
params.push_back(std::make_unique<juce::AudioParameterFloat>("master", "Master",
juce::NormalisableRange<float>(-24.0f, 24.0f, 0.1f), 0.1f));
params.push_back(std::make_unique<juce::AudioParameterFloat>("lowEQ", "Low Gain",
juce::NormalisableRange<float>(-24.0f, 24.0f, 0.1f), 0.5f));
params.push_back(std::make_unique<juce::AudioParameterFloat>("midEQ", "Mid EQ",
juce::NormalisableRange<float>(-24.0f, 24.0f, 0.1f), 0.8f));
params.push_back(std::make_unique<juce::AudioParameterFloat>("highEQ", "High EQ",
juce::NormalisableRange<float>(-24.0f, 24.0f, 0.1f), 1.0f));
return { params.begin(), params.end() };
}

View File

@@ -1,101 +1,90 @@
/* #pragma once
==============================================================================
#include <JuceHeader.h>
This file contains the basic framework code for a JUCE plugin processor. #include "AudioBufferQueue.h"
#include "AudioEngine.h"
============================================================================== #include "ScopeDataCollector.h"
*/ #include "NeuralSharedParams.h"
#pragma once //==============================================================================
// Processor
#include <JuceHeader.h> class NeuralSynthAudioProcessor : public juce::AudioProcessor,
#include "AudioBufferQueue.h" private juce::AudioProcessorValueTreeState::Listener
#include "AudioEngine.h" {
#include "ScopeDataCollector.h" public:
#include "NeuralSharedParams.h" NeuralSynthAudioProcessor();
~NeuralSynthAudioProcessor() override;
//==============================================================================
/** // AudioProcessor overrides
*/ void prepareToPlay(double sampleRate, int samplesPerBlock) override;
class NeuralSynthAudioProcessor : public juce::AudioProcessor, void releaseResources() override;
private juce::AudioProcessorValueTreeState::Listener
{ #ifndef JucePlugin_PreferredChannelConfigurations
public: bool isBusesLayoutSupported(const BusesLayout& layouts) const override;
//============================================================================== #endif
NeuralSynthAudioProcessor();
~NeuralSynthAudioProcessor() override; void processBlock(juce::AudioBuffer<float>&, juce::MidiBuffer&) override;
//============================================================================== // Editor
void prepareToPlay(double sampleRate, int samplesPerBlock) override; juce::AudioProcessorEditor* createEditor() override;
void releaseResources() override; bool hasEditor() const override;
bool isBusesLayoutSupported(const BusesLayout& layouts) const;
// Info
#ifndef JucePlugin_PreferredChannelConfigurations const juce::String getName() const override;
bool isBusesLayoutSupported(const BusesLayout& layouts) const override; bool acceptsMidi() const override;
#endif bool producesMidi() const override;
bool isMidiEffect() const override;
void processBlock(juce::AudioBuffer<float>&, juce::MidiBuffer&) override; double getTailLengthSeconds() const override;
//============================================================================== // Programs
juce::AudioProcessorEditor* createEditor() override; int getNumPrograms() override;
bool hasEditor() const override; int getCurrentProgram() override;
void setCurrentProgram(int index) override;
//============================================================================== const juce::String getProgramName(int index) override;
const juce::String getName() const override; void changeProgramName(int index, const juce::String& newName) override;
bool acceptsMidi() const override; // State
bool producesMidi() const override; void getStateInformation(juce::MemoryBlock& destData) override;
bool isMidiEffect() const override; void setStateInformation(const void* data, int sizeInBytes) override;
double getTailLengthSeconds() const override;
// Parameters
//============================================================================== void parameterChanged(const juce::String& id, float newValue) override;
int getNumPrograms() override; void buildParams(std::vector<std::unique_ptr<juce::RangedAudioParameter>>& params,
int getCurrentProgram() override; const std::string& paramGroup);
void setCurrentProgram(int index) override; juce::AudioProcessorValueTreeState::ParameterLayout createParameterLayout();
const juce::String getProgramName(int index) override;
void changeProgramName(int index, const juce::String& newName) override; // Utilities
juce::MidiMessageCollector& getMidiMessageCollector() noexcept { return midiMessageCollector; }
//============================================================================== AudioBufferQueue<float>& getAudioBufferQueue() noexcept { return audioBufferQueue; }
void getStateInformation(juce::MemoryBlock& destData) override;
void setStateInformation(const void* data, int sizeInBytes) override; AudioBufferQueue<float>& getChorusAudioBufferQueue() noexcept { return chorusBufferQueue; }
AudioBufferQueue<float>& getDelayAudioBufferQueue() noexcept { return delayBufferQueue; }
//============================================================================== AudioBufferQueue<float>& getReverbAudioBufferQueue() noexcept { return reverbBufferQueue; }
void parameterChanged(const juce::String& id, float newValue) override; AudioBufferQueue<float>& getFlangerAudioBufferQueue() noexcept { return flangerBufferQueue; }
AudioBufferQueue<float>& getDistortionAudioBufferQueue() noexcept { return distortionBufferQueue; }
void buildParams(std::vector<std::unique_ptr<juce::RangedAudioParameter>>& params, const std::string& paramGroup); AudioBufferQueue<float>& getFilterAudioBufferQueue() noexcept { return filterBufferQueue; }
juce::MidiMessageCollector& getMidiMessageCollector() noexcept { return midiMessageCollector; } // Public members (by JUCE convention)
juce::MidiMessageCollector midiMessageCollector;
juce::MidiMessageCollector midiMessageCollector; juce::AudioProcessorValueTreeState parameters;
juce::AudioProcessorValueTreeState parameters;
juce::AudioProcessorValueTreeState::ParameterLayout createParameterLayout(); private:
JUCE_DECLARE_NON_COPYABLE_WITH_LEAK_DETECTOR (NeuralSynthAudioProcessor)
AudioBufferQueue<float>& getAudioBufferQueue() noexcept { return audioBufferQueue; }
// ---- IMPORTANT ORDER FIX ----
AudioBufferQueue<float>& getChorusAudioBufferQueue() noexcept { return chorusBufferQueue; } // Objects are constructed in THIS order. 'sp' must come BEFORE audioEngine.
AudioBufferQueue<float>& getDelayAudioBufferQueue() noexcept { return delayBufferQueue; } NeuralSharedParams sp; // <— construct first
AudioBufferQueue<float>& getReverbAudioBufferQueue() noexcept { return reverbBufferQueue; } NeuralAudioEngine audioEngine; // needs a valid reference to 'sp'
AudioBufferQueue<float>& getFlangerAudioBufferQueue() noexcept { return flangerBufferQueue; } // Meter/scope queues
AudioBufferQueue<float>& getDistortionAudioBufferQueue() noexcept { return distortionBufferQueue; } AudioBufferQueue<float> audioBufferQueue;
AudioBufferQueue<float>& getFilterAudioBufferQueue() noexcept { return filterBufferQueue; } AudioBufferQueue<float> chorusBufferQueue;
AudioBufferQueue<float> delayBufferQueue;
private: AudioBufferQueue<float> reverbBufferQueue;
//============================================================================== AudioBufferQueue<float> flangerBufferQueue;
JUCE_DECLARE_NON_COPYABLE_WITH_LEAK_DETECTOR(NeuralSynthAudioProcessor) AudioBufferQueue<float> distortionBufferQueue;
AudioBufferQueue<float> filterBufferQueue;
NeuralAudioEngine audioEngine;
AudioBufferQueue<float> audioBufferQueue; // Scope collector (uses audioBufferQueue, so declare after it)
ScopeDataCollector<float> scopeDataCollector { audioBufferQueue };
AudioBufferQueue<float> chorusBufferQueue; };
AudioBufferQueue<float> delayBufferQueue;
AudioBufferQueue<float> reverbBufferQueue;
AudioBufferQueue<float> flangerBufferQueue;
AudioBufferQueue<float> distortionBufferQueue;
AudioBufferQueue<float> filterBufferQueue;
ScopeDataCollector<float> scopeDataCollector{ audioBufferQueue };
NeuralSharedParams sp;
};