SpeechConfiguration
@objc
public class SpeechConfiguration : NSObject
Configuration properties for Spokestack modules.
-
A comma-separated list of wakeword keywords
Remark
ex: “up,dog”Warning
cannot contain spacesSee also
AppleWakewordRecognizer
Declaration
Swift
@objc public var wakewords: String
-
The name of the window function to apply to each audio frame before calculating the STFT.
Remark
Currently the “hann” window is supported.See also
TFLiteWakewordRecognizer
Declaration
Swift
public var fftWindowType: SignalProcessing.FFTWindowType
-
The desired linear Root Mean Squared (RMS) signal energy, which is used for signal normalization and should be tuned to the RMS target used during wakeword model training.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@available(*, deprecated, message: "RMS normalization is no longer used during wakeword recognition.") @objc public var rmsTarget: Float
-
The Exponentially Weighted Moving Average (EWMA) update rate for the current Root Mean Squared (RMS) signal energy (0 for no RMS normalization).
See also
TFLiteWakewordRecognizer
Declaration
Swift
@available(*, deprecated, message: "RMS normalization is no longer used during wakeword recognition.") @objc public var rmsAlpha: Float
-
The size of the signal window used to calculate the STFT, in number of samples - should be a power of 2 for maximum efficiency.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var fftWindowSize: Int
-
The length of time to skip each time the overlapping STFT is calculated, in milliseconds.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var fftHopLength: Int
-
The length of a frame in the mel spectrogram used as an input to the wakeword recognizer encoder, in milliseconds.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var melFrameLength: Int
-
The number of filterbank components in each mel spectrogram frame sent to the wakeword recognizer.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var melFrameWidth: Int
-
The size of the wakeword recognizer’s encoder state output.
Remarks
Defaults to matching theencodeWidth
value.Declaration
Swift
@objc public var stateWidth: Int
-
The size of the wakeword recognizer’s encoder window output.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var encodeWidth: Int
-
The length of the sliding window of encoder output used as an input to the wakeword recognizer classifier, in milliseconds.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var encodeLength: Int
-
The threshold of the wakeword recognizer classifier’s posterior output, above which the wakeword recognizer activates the pipeline, in the range [0, 1].
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var wakeThreshold: Float
-
The minimum length of an activation, in milliseconds. Used to ignore a Voice Activity Detector (VAD) deactivation after the wakeword.
See also
TFLiteWakewordRecognizer`Declaration
Swift
@objc public var wakeActiveMin: Int
-
The maximum length of an activation, in milliseconds. Used to time out the speech pipeline activation.
Remarks
Defaults to 5 seconds to improve perceived responsiveness, although most NLUs use a longer timeout (eg 7s).Declaration
Swift
@objc public var wakeActiveMax: Int
-
Indicate to the VAD the level of permissiveness to non-speech activation.
Declaration
Swift
public var vadMode: VADMode
-
Delay between a VAD deactivation and the delivery of the recognition results.
See also
AppleSpeechRecognizer
Remark
unique to iOSDeclaration
Swift
@objc public var vadFallDelay: Int
-
Audio sampling rate, in Hz.
Declaration
Swift
public var sampleRate: Int
-
Audio frame width, in milliseconds.
To do
Should be renamed wakeFrameWidth.Declaration
Swift
@objc public var frameWidth: Int
-
Length of time to allow an Apple ASR request to run, in milliseconds.
See also
AppleWakewordRecognizer
Remark
Apple has an undocumented limit of 60000ms per request. Unique to iOS.Declaration
Swift
@objc public var wakewordRequestTimeout: Int
-
The pre-emphasis filter weight to apply to the normalized audio signal, in a range of [0, 1].
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var preEmphasis: Float
-
The filename of the machine learning model used for the filtering step.
Remarks
Both the file name and the file path are configurable to allow for flexibility in constructing the path that the recognizer will attempt to load the model from.Declaration
Swift
@objc public var filterModelName: String
-
The filename of the machine learning model used for the encoding step.
Remarks
Both the file name and the file path are configurable to allow for flexibility in constructing the path that the recognizer will attempt to load the model from.Declaration
Swift
@objc public var encodeModelName: String
-
The filename of the machine learning model used for the detect step.
Remarks
Both the file name and the file path are configurable to allow for flexibility in constructing the path that the recognizer will attempt to load the model from.Declaration
Swift
@objc public var detectModelName: String
-
The filesystem path to the machine learning model for the filtering step.
Declaration
Swift
@objc public var filterModelPath: String
-
The filesystem path to the machine learning model for the encoding step.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var encodeModelPath: String
-
The filesystem path to the machine learning model for the detect step.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var detectModelPath: String
-
Text To Speech API client identifier key.
See also
TextToSpeech
Declaration
Swift
@objc public var apiId: String
-
Text To Speech API client secret key.
See also
TextToSpeech
Declaration
Swift
@objc public var apiSecret: String
-
The filesystem path to the vocabulary used for tokenizer encoding.
See also
Tokenizer
Declaration
Swift
@objc public var nluVocabularyPath: String
-
The index in the vocabulary of the terminator token. Determined by the NLU vocabulary.
See also
BertTokenizer
Declaration
Swift
@objc public var nluTerminatorTokenIndex: Int
-
The index in the vocabulary of the terminator token. Determined by the NLU vocabulary.
See also
BertTokenizer
Declaration
Swift
@objc public var nluPaddingTokenIndex: Int
-
The filesystem path to the machine learning model for Natural Language Understanding processing.
See also
TensorflowNLU
Declaration
Swift
@objc public var nluModelPath: String
-
The filesystem path to the model metadata for Natural Language Understanding processing.
See also
TensorflowNLU
Declaration
Swift
@objc public var nluModelMetadataPath: String
-
The maximum utterance length the NLU can process. Determined by the NLU model.
See also
BertTokenizer
Declaration
Swift
@objc public var nluMaxTokenLength: Int
-
Debugging trace levels, for simple filtering.
Declaration
Swift
@objc public var tracing: Trace.Level
-
Delegate events will be sent using the specified dispatch queue.
Declaration
Swift
@objc public var delegateDispatchQueue: DispatchQueue
-
Automatically run Spokestack’s NLU classification on ASR transcripts for clients that use the
Spokestack
facade.Note
RequiresNLUTensorflow
to be correctly configured, notably withnluModelPath
,nluModelMetadataPath
, andnluVocabularyPath
.Declaration
Swift
@objc public var automaticallyClassifyTranscript: Bool
-
The filename of the machine learning model used for the filtering step of the keyword recognizer.
Remarks
Both the file name and the file path are configurable to allow for flexibility in constructing the path that the recognizer will attempt to load the model from.See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywordFilterModelName: String
-
The filename of the machine learning model used for the encoding step of the keyword recognizer.
Remarks
Both the file name and the file path are configurable to allow for flexibility in constructing the path that the recognizer will attempt to load the model from.See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywordEncodeModelName: String
-
The filename of the machine learning model used for the detect step of the keyword recognizer.
Remarks
Both the file name and the file path are configurable to allow for flexibility in constructing the path that the recognizer will attempt to load the model from.See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywordDetectModelName: String
-
The filename of the model metadata for keyword recognition
Remarks
Both the file name and the file path are configurable to allow for flexibility in constructing the path that the recognizer will attempt to load the model from.See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywordMetadataName: String
-
The filesystem path to the machine learning model for the filtering step of the keyword recognizer.
See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywordFilterModelPath: String
-
The filesystem path to the machine learning model for the encoding step of the keyword recognizer.
See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywordEncodeModelPath: String
-
The filesystem path to the machine learning model for the detect step of the keyword recognizer.
See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywordDetectModelPath: String
-
The threshold of the keyword recognizer’s posterior output, above which the keyword recognizer emits a recognition event for the most probable keyword.
See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywordThreshold: Float
-
The filesystem path to the model metadata for keyword recognition
See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywordMetadataPath: String
-
A comma-separated list of keywords to recognize.
Remark
ex: “yes,no”Warning
Cannot contain spaces. Will be ignored in favor ofkeywordMetadataPath
if available.See also
TFLiteKeywordRecognizer
Declaration
Swift
@objc public var keywords: String
-
The name of the window function to apply to each audio frame before calculating the STFT.
Remark
Currently the “hann” window is supported.See also
TFLiteWakewordRecognizer
Declaration
Swift
public var keywordFFTWindowType: SignalProcessing.FFTWindowType
-
The size of the signal window used to calculate the STFT, in number of samples - should be a power of 2 for maximum efficiency.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var keywordFFTWindowSize: Int
-
The length of time to skip each time the overlapping STFT is calculated, in milliseconds.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var keywordFFTHopLength: Int
-
The length of a frame in the mel spectrogram used as an input to the wakeword recognizer encoder, in milliseconds.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var keywordMelFrameLength: Int
-
The number of filterbank components in each mel spectrogram frame sent to the wakeword recognizer.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var keywordMelFrameWidth: Int
-
The size of the wakeword recognizer’s encoder window output.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var keywordEncodeWidth: Int
-
The length of the sliding window of encoder output used as an input to the wakeword recognizer classifier, in milliseconds.
See also
TFLiteWakewordRecognizer
Declaration
Swift
@objc public var keywordEncodeLength: Int
-
Timeout in seconds used for semaphore waits in the speech pipeline
Warning
There is not normally a need to change this value.Declaration
Swift
@objc public var semaphoreTimeout: Double