The figures were updated on April 3rd at 1pm PT.

Geometric Trajectory Matrix for Voice Modeling

MFCCs, while historically central to audio representation, primarily encode short-window spectral features and discard higher-order temporal structure. We propose the geometric trajectory matrix, a representation that models speech as a trajectory through feature space, preserving sequential dynamics that may better support downstream inference of context and prediction of future utterances.

At Sanguine, we’ve been developing a set of voice features for speaker verification and other identity-based tasks, inspired by audio intelligence systems such as Shazam.

Walk into any room and sound is everywhere, but machines rarely hear it the way we do. Voices, footsteps, even the hum of electronics—they carry patterns that can tell you who’s speaking, what’s happening, and when it matters. Our lab is using Geometric Trajectory Matrices to help machines make sense of these patterns in real time.

Traditional features like Mel-Frequency Cepstral Coefficients capture slices of sound in time. They’re snapshots—useful, but incomplete. Modern embedding models go further, but often compress entire utterances into a single vector, losing the temporal structure of speech.

Geometric Trajectory Matrices take a different approach. We start by mapping audio into a low-dimensional trajectory:

\[ \mathbf{z}_1, \mathbf{z}_2, \dots, \mathbf{z}_T \in \mathbb{R}^k \]

Each \(\mathbf{z}_t\) represents the voice at a moment in time. Together, they form a path through acoustic space—the vocal trajectory.

Instead of only tracking local motion, we explicitly model the geometry of this trajectory by capturing relationships between all pairs of time steps.

Distance geometry:

\[ D_{ij} = \|\mathbf{z}_i - \mathbf{z}_j\| \]

This matrix encodes the global shape of the trajectory—how similar or different any two moments in time are.

Inner-product structure:

\[ G_{ij} = \mathbf{z}_i^\top \mathbf{z}_j \]

This captures angular relationships and correlations across time, revealing how the trajectory is oriented in space.

Temporal dynamics:

\[ \Delta \mathbf{z}_t = \mathbf{z}_{t+1} - \mathbf{z}_t \]

These local motion vectors describe how the voice evolves from frame to frame—capturing articulation patterns and speaking style.

The vocal path is the journey of your articulators—the tongue, lips, jaw, and vocal cords—through acoustic space. Every frame in the matrix is a snapshot along that journey. The difference between frames tells you the local step. Geometric invariants, like curvature:

\[ \kappa_t = \frac{\|\Delta \mathbf{z}_t \times \Delta \mathbf{z}_{t+1}\|}{\|\Delta \mathbf{z}_t\|^3} \]

or alignment between two trajectories:

\[ S(i,j) = \frac{\operatorname{Tr}(\mathbf{Z}_i^\top \mathbf{Z}_j)}{\|\mathbf{Z}_i\|_F \|\mathbf{Z}_j\|_F} \]

give models a way to measure similarity and structure in a way that is both interpretable and robust.

Together, these components form the Geometric Trajectory Matrix: a structured representation of both where the voice is and how it moves.

Visualizing GMT Trajectories

To make this idea more tangible, we can visualize a 3D trajectory of a single speaker’s voice. As time moves forward, the trajectory traces how the features change over the course of their speech.

GMT Trajectory of a Single Speaker

Figure: GMT trajectory of a single speaker.

Trajectory Colored by Time

Figure: Trajectory colored by time to show the temporal progression of features along the vocal path. Early frames are purple, late frames are yellow.

By examining the trajectory, patterns of motion—curves, twists, and transitions—become visible, providing insight into temporal dynamics that are not captured by static representations.

Building Foundations for What Comes Next

Today, most systems that rely on voice for identity reduce speech to static representations. These approaches perform well under controlled conditions, but often degrade in real-world settings where noise, variability, and temporal structure matter.

This limitation shows up across voice authentication systems and autonomous agents—anywhere identity must be inferred under uncertainty. In warehouse robotics, for example, operator authentication often happens in noisy, fast-moving environments where reliability is critical. The issue is not just model capacity, but how the signal itself is represented. When temporal structure is compressed away, important information about consistency and evolution is lost.

We think the problem is upstream. Not just in the models, but in the representation itself.

Building on prior work in voice biometrics, we introduce a data construct we call the vocal path. Instead of collapsing time, it preserves it. Instead of summarizing, it traces. It captures how a voice moves—not just what it looks like.

We believe this shift enables a range of high-impact applications, from operator authentication in warehouse robotics to persistent identity for agents operating in irreversible decision contexts.

Reach out to us at team@sanguineai.xyz to learn more.