He started manually. For the first verse, he drew a flat, almost robotic delivery. The lyrics were about waiting—the numb, dissociative kind. He wanted Hana to sound like she’d forgotten why she was even at the station. He set the Dynamics to a low, steady 32. Breathiness at 18. A faint, constant hiss of air, like a radiator.
VOCALOid 6’s new "Expressive Control" feature was supposed to allow for this. It let you import an audio reference, and the AI would analyze the timbre, the portamento, the raw, ugly gasps for air. But when Kenji hit "apply," Hana’s voice emerged polished. The crack was there, but it was a diamond crack—symmetrical, beautiful, meaningless. vocaloid 6 tuning
The screen glowed a soft, sterile white. Kenji stared at the grid of parameters—Dynamics, Pitch Deviation, Growl, Breathiness—each one a tiny lever he could pull to bend reality, or at least, to bend the ghost in the machine. He started manually
For the next three hours, Kenji became a micro-surgeon of silence. He inserted a tiny, 0.2-second dip in the Pitch Deviation right before the chorus—a moment of doubt, a slight downward glance before the leap upward. He manually painted a "Growl" parameter on the long, held note of "yo-ake" (dawn), not a full rasp, just a granular flutter, like sand slipping through fingers. He took the AI’s perfect, buttery portamento between two notes and replaced it with a jagged, stair-stepped curve, making Hana sound like she was choking on the word. He wanted Hana to sound like she’d forgotten
The chorus needed lift. He selected the four bars and switched back to the AI "Dynamic Mode." He sang into his laptop’s cheap mic: "Kaze ga fuitara…" with a swelling, desperate rise in pitch. The AI parsed it. For a moment, Hana’s voice bloomed—rich, powerful, heartbreaking. But the transition from the flat, robotic verse to the AI-generated chorus was a cliff. A hard, digital step.
At 2:47 AM, he played it back.
"Damn it," he muttered, zooming into the Pitch Rendering graph.