Tuco Benedicto Pacifico
Arcane
Yep.Sounds like shit tbhThey should have gone with Vivec Voice from the in game file.
Yep.Sounds like shit tbhThey should have gone with Vivec Voice from the in game file.
Just since it's vaguely pertinent... I've crossed a guy on the Larian forum that was experimenting a bit with this stuff and voiced several lines of Viconia in BG2 using AI:
Eh, it still sounds leagues better than a lot of "amateur voice acting" seen in plenty of fan mods (the recent Final Fantasy VII abomination comes to mind) and of a lot of bad voice acting seen in games, in general.Still doesn't sound right tbh.
They should have gone with Vivec Voice from the in game file.
Just since it's vaguely pertinent... I've crossed a guy on the Larian forum that was experimenting a bit with this stuff and voiced several lines of Viconia in BG2 using AI:
Still doesn't sound right tbh. Grey DeLisle has a certain coquettishness to her voice when she's acting seductive. Especially apparent in number 48 after the "lend me your ear" line.
"That is too bad..." and onward sounds flat, certain words that the actress would draw out are neglected by the A.I. The tech isn't quite there yet.
At this point in time in this specific field there are SEVERAL competing solutions that can vary drastically in ease of use, amount of data required to "train" them and verisimilitude/"naturalness" of the result..Its all about how much text you feed the labs. Too much, and as you say the delivery goes flat. Too little, you don't hear enough context, and neither does the algorithm.
Are they, though? It's dirt cheap all things considered. You can make an audibook for an average book for as much as 25$. Pennies.levenlabs is a highly impressive piece of tech but they're taking the absolute piss with their character limits and payment plans.
For unimportant NPCs, AI would probably be preferrable, because otherwise every single unimportant NPC sounds the same unless you have an enormous pool of voice actors to do unimportant roles. With AI-generated, every NPC can sound different when being unimportant.Whatever the case, it's clear to me that in a single character the twain should never meet. Switching between the real actress and the A.I. for subsequent lines is somehow more distracting than going from authentic voiceover to text. For unimportant NPCs like innkeepers, merchants, random soldiers etc. you can probably get away with A.I., but I'd always choose a real actor, even a bad one, for the important roles. At the very least, I'll have something to chuckle at (in Oblivion's case especially).
At this point in time in this specific field there are SEVERAL competing solutions that can vary drastically in ease of use, amount of data required to "train" them and verisimilitude/"naturalness" of the result..Its all about how much text you feed the labs. Too much, and as you say the delivery goes flat. Too little, you don't hear enough context, and neither does the algorithm.
For instance this Microsoft version can clone a vocal timbre with 3 seconds of audio feed:
But generally speaking, the more training data the better. There's no such a thing as "too much".
Sometimes the errors are funny, like Joe Rogan mispronouncing Automaton https://www.youtube.com/watch?v=-0QxuL7StOMI'd be willing to pay a monthly fee for unlimited access, but the problem with their quotas is that you sometimes get results that just go wrong - I've had a couple where certain segments of the prompt got pronounced in totally bizarre ways, meaning the entire prompt needs to be redone, and you thus get "charged" twice for it, since there's no way to refund a bad result.
And then factor in how, if you're looking for serious high-quality results, you'll be messing around with the very sensitive intensity sliders to try and get the exact result you want, meaning you might end up re-entering a prompt four or five times. The quotas which seem generous at first glance end up being pretty tight.
I wasted a good chunk of my free monthly quota just trying to fine-tune a very short joke message for my friends.
This reply shows how you basically desire the degenerate future where machines are pushed in front of humans. Kinda like how this retard below instantly betrays itEh, it still sounds leagues better than a lot of "amateur voice acting" seen in plenty of fan mods (the recent Final Fantasy VII abomination comes to mind) and of a lot of bad voice acting seen in games, in general.Still doesn't sound right tbh.
Oblivion, for one, is filled to the brim of lines recited with worse delivery than a lot of AI-based stuff in this thread.
I can't wait until this is good enough to replace voice actors - insufferable f*gs.
I'm pleased that the technology has reached a point where it sounds realistic enough for games to actually use. I've been complaining for 30 years about game devs using voice actors instead of speech synthesis. (The fucking Commodore 64 SID chip could synthesize voices with all kinds of inflections and parameters.) It's exciting as hell to see this happen in my lifetime and I'm looking forward to populated game worlds filled with characters who each have their own voice. Once AI conversation engines can generate content the insanity will be complete.
It's hard for me to exactly point out the absurdity of going "yeah I wish more games used shitty sound chip voice synthesis instead of voice acting", you know, if you don't like voice acting then how about just... read the text?
Are you morons using some kind of voice synthesizer to read these web pages or something?
Yours shows me you are a retarded with a bias that makes a load of moronic assumptions.This reply shows how you basically desire the degenerate future where machines are pushed in front of humans. Kinda like how this retard below instantly betrays it
The problem with humans is that most of them are bad at voice acting. So if you can get the technology to where a synthetized voice sounds exactly like a human, and has the added benefit of voice acting well, then this is something to be excited for, considering how weak voice acting for games is in general. But I find that to be a tad optimistic.So you've been complaining for 30 years about devs using voice actors instead of speech synthesis, but now that you have supposedly synthesis that can mimic humans perfectly that is exciting? Isn't this redundant, if the speech synthesizer sounds exactly like a human?