DMCA, complaints, and other inquiries:
the opie model came out way better than the ant one. The Ant one was more so a proof of concept, so I just used one of his long gun rants from a random 2012 show to train it because it was easy. As soon as I heard his voice in OPs audio, I knew for sure that was my ant model because of the artifacts, like how it sounds kind of robotic or auto-tuned when transitioning to certain sounds. I spent more time on the opie model and used more audio from different shows with a wider range of vocal inflections and stuff, so it sounds much more natural.It's actually scary how accurate Opie's voice is. I honestly wouldn't be able to tell it apart from the actual opester if I didn't know this was AI.
I vote bobo or intern Davidthe opie model came out way better than the ant one. The Ant one was more so a proof of concept, so I just used one of his long gun rants from a random 2012 show to train it because it was easy. As soon as I heard his voice in OPs audio, I knew for sure that was my ant model because of the artifacts, like how it sounds kind of robotic or auto-tuned when transitioning to certain sounds. I spent more time on the opie model and used more audio from different shows with a wider range of vocal inflections and stuff, so it sounds much more natural.
On my to-do list is cleaning up and adding more stuff to the Ant dataset and re-training it so it sounds as natural as opie's, then making one for jimmy. After those are done and they start getting used (like in this post), I'll throw up a poll for who should be next.
I vote Patrick Tomlinson. Imagine how mad he'd be if there was a fake Patrick Tomlinson podcast where he comes clean about being a pedophile and explains in detail all the crimes he's committed.the opie model came out way better than the ant one. The Ant one was more so a proof of concept, so I just used one of his long gun rants from a random 2012 show to train it because it was easy. As soon as I heard his voice in OPs audio, I knew for sure that was my ant model because of the artifacts, like how it sounds kind of robotic or auto-tuned when transitioning to certain sounds. I spent more time on the opie model and used more audio from different shows with a wider range of vocal inflections and stuff, so it sounds much more natural.
On my to-do list is cleaning up and adding more stuff to the Ant dataset and re-training it so it sounds as natural as opie's, then making one for jimmy. After those are done and they start getting used (like in this post), I'll throw up a poll for who should be next.
I thought about that a while ago. It's possible, but it's a lot of work, especially if you want their dialogue to actually sound somewhat natural and accurate. Here's the high level flow I came up with:
The only problem here would be spacing out the audio files when creating the combined audio track. I have no idea how you could automate natural pauses and interruptions, so you'd probably be better off manually editing it all together, but that's still not super easy to do.
- The program starts by calling GPT4 for a list of current events, with a focus on topics relevant to what O&A would discuss, and output it to some format.
- Now you could use GPT-4 to create the actual script, but it won't have enough data to output something that would sound like them actually conversing. To get that, you could use a local LLM and train a LoRA on a dataset full of quotes and conversations from O&A and certain facts that might not be there already, like nana's love for a boy named sue.
- Send the list of events and stuff to the local model with a prompt for creating a script that will output to a standard format. Depending on how long you want it to be, the token limit of the model, and the hardware used to run the model, you may have to break it up into separate prompts.
- Then you have the app break up the conversation and send each instance of each speaker talking to a text-to-speech program (this can also be done locally).
- Most TTS models aren't very accurate, so you'd probably then want to export the audio files to RVC (which was used for OP's audio) to replace the TTS voices with O&A's. The files should output with some sequenced naming convention.
- The program could then re-compile the conversation into one single audio file and throw Street Fighting Man at the beginning.
There's a lot of manual stuff that would make it better, like recording yourself reading the script and using RVC on that instead of the TTS output, but this is just a theoretical automated end to end solution.
so good luck to to whoever wants to make that
Yes, and Anthony agrees with him while Opie is disgusted. Yimmy would be a bit on the fenceI vote Patrick Tomlinson. Imagine how mad he'd be if there was a fake Patrick Tomlinson podcast where he comes clean about being a pedophile and explains in detail all the crimes he's committed.
I never got the obsession with the Patrick shit. He's a fat nobody midwest faggot with no connection to O&A. He had some shitty take on Norm, which made some autists on the O&A sub decide to fuck with him to the point of getting the sub shut down. I don't understand why he's still relevant here.I vote Patrick Tomlinson. Imagine how mad he'd be if there was a fake Patrick Tomlinson podcast where he comes clean about being a pedophile and explains in detail all the crimes he's committed.
He’s still on his phone trawling for shemales to plug up his egg white omelette fartsIt's the first five minutes of the show. Of course he's not speaking yet.
Are you ready to SPREAD THE VIRUS in 2024?!
Well now you can, because the long-awaited reunion of Opie and Anthony has happened! (as AI)
Has anyone got this to Opie?
These are good ideas.Redbar needs to hear this as well!
They'd need a separate server farm just to process all of the blinking.The AI has a tough time processing long pontificating breaths through the noses, the thousand different inflections of “It’s a tough one man.”, and the complexities of AIDS puddle jokes.
I'm shocked at how accurate their individual cadences and inflections are. Is this something that is programmed in some way in the voice model, or does it just pick up the inflection from sourced samples?Assuming he's using the RVC models I made a few months ago, I only did Opie and Ant. I was waiting for someone to actually do something with them before I made new ones and refined the existing ones.
This is seriously fascinating.I thought about that a while ago. It's possible, but it's a lot of work, especially if you want their dialogue to actually sound somewhat natural and accurate. Here's the high level flow I came up with:
The only problem here would be spacing out the audio files when creating the combined audio track. I have no idea how you could automate natural pauses and interruptions, so you'd probably be better off manually editing it all together, but that's still not super easy to do.
- The program starts by calling GPT4 for a list of current events, with a focus on topics relevant to what O&A would discuss, and output it to some format.
- Now you could use GPT-4 to create the actual script, but it won't have enough data to output something that would sound like them actually conversing. To get that, you could use a local LLM and train a LoRA on a dataset full of quotes and conversations from O&A and certain facts that might not be there already, like nana's love for a boy named sue.
- Send the list of events and stuff to the local model with a prompt for creating a script that will output to a standard format. Depending on how long you want it to be, the token limit of the model, and the hardware used to run the model, you may have to break it up into separate prompts.
- Then you have the app break up the conversation and send each instance of each speaker talking to a text-to-speech program (this can also be done locally).
- Most TTS models aren't very accurate, so you'd probably then want to export the audio files to RVC (which was used for OP's audio) to replace the TTS voices with O&A's. The files should output with some sequenced naming convention.
- The program could then re-compile the conversation into one single audio file and throw Street Fighting Man at the beginning.
There's a lot of manual stuff that would make it better, like recording yourself reading the script and using RVC on that instead of the TTS output, but this is just a theoretical automated end to end solution.
so good luck to to whoever wants to make that
I'm guessing OP recorded himself doing the voices and then used RVC. Cadence and inflection come from the source performance, tone and timbre come from the model. Using TTS --> RVC works, but having tried that using different TTS methods, it never sounded natural to me. It's a lot easier to record a half-assed impression, since you just need to mimic the speaking style and not the actual sound of their voice.I'm shocked at how accurate their individual cadences and inflections are. Is this something that is programmed in some way in the voice model, or does it just pick up the inflection from sourced samples?
This forum is dedicated exclusively to parody, comedy, and satirical content. None of the statements, opinions, or depictions shared on this platform should be considered or treated as factual information under any circumstances. All content is intended for entertainment purposes only and should be regarded as fictional, exaggerated, or purely the result of personal opinions and creative expression.
Please be aware that this forum may feature discussions and content related to taboo, controversial, or potentially offensive subjects. The purpose of this content is not to incite harm but to engage in satire and explore the boundaries of humor. If you are sensitive to such subjects or are easily offended, we kindly advise that you leave the forum.
Any similarities to real people, events, or situations are either coincidental or based on real-life inspirations but used within the context of fair use satire. By accepting this disclaimer, you acknowledge and understand that the content found within this forum is strictly meant for parody, satire, and entertainment. You agree not to hold the forum, its administrators, moderators, or users responsible for any content that may be perceived as offensive or inappropriate. You enter and participate in this forum at your own risk, with full awareness that everything on this platform is purely comedic, satirical, or opinion-based, and should never be taken as factual information.
If any information or discussion on this platform triggers distressing emotions or thoughts, please leave immediately and consider seeking assistance.
National Suicide Prevention Lifeline (USA): Phone: 1-800-273-TALK (1-800-273-8255) Website: https://suicidepreventionlifeline.org/