this post was submitted on 25 Jul 2024
1 points (100.0% liked)
Technology
59587 readers
3037 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Same, I've automated alot of my tasks with AI. No way 77% is "hampered" by it.
77% are smart enough not to admit their job is being done by a bot.
What have you actually replaced/automated with AI?
Voiceover recording, noise reduction, rotoscoping, motion tracking, matte painting, transcription - and there's a clear path forward to automate rough cuts and integrate all that with digital asset management. I used to do all of those things manually/practically.
All the models I've used that do TTS/RVC and rotoscoping have definitely not produced professional results.
What are you using? Cause if you're a professional, and this is your experience, I'd think you'd want to ask me what I'm using.
Coqui for TTS, RVC UI for matching the TTS to the actor's intonation, and DWPose -> controlnet applied to SDXL for rotoscoping
Full open source, nice! I respect the effort that went into that implementation. I pretty much exclusively use 11 Labs for TTS/RVC, turn up the style, turn down the stability, generate a few, and pick the best. I do find that longer generations tend to lose the thread, so it's better to batch smaller script segments.
Unless I misunderstand ya, your controlnet setup is for what would be rigging and animation rather than roto. I do agree that while I enjoy the outputs of pretty much all the automated animators, they're not ready for prime time yet. Although I'm about to dive into KREA's new key framing feature and see if that's any better for that use case.
I was never able to get appreciably better results from 11 labs than using some (minorly) trained RVC model :/ The long scripts problem is something pretty much any text-to-something model suffers from. The longer the context the lower the cohesion ends up.
I do rotoscoping with SDXL i2i and controlnet posing together. Without I found it tends to smear. Do you just do image2image?
They're right IMO. Practical effects still look and age better than (IMO very obvious) digital effects. Oh and digital deaging IMO looks like crap.
But, this will always remain an opinion battle anyway, because quantifying "artistry" is in and of itself a fool's errand.
Digital video, not digital effects - I mean the guys I went to film school with that refused to touch digital videography.
This may come as a shock to you, but the vast majority of the world does not work in tech.
I'm not working in tech either. Everyone relying on a computer can use this.
Also, medicin and radiology are two areas that will benefit from this - especially the patients.
I dunno, mishandling of AI can be worse than avoiding it entirely. There's a middle manager here that runs everything her direct-report copywriter sends through ChatGPT, then sends the response back as a revision. She doesn't add any context to the prompt, say who the audience is, or use the custom GPT that I made and shared. That copywriter is definitely hampered, but it's not by AI, really, just run-of-the-mill manager PEBKAC.
I'm infuriated on their behalf.
E-fucking-xactly. I hate reading long winded bullshit AI stories with a passion. Drivel all of it.
A lot of people are keen to hear that AI is bad, though, so the clicks go through on articles like this anyway.