Metahuman Animator Tutorial
Фильм және анимация
Metahuman Animator Tutorial
JSFILZM Mocap Helmet: • Cheap Mocap Helmet for...
Mocap Waitlist: JSFILMZMOCAP AT GMAIL DOT COM
Grab my new Unreal Engine 5.1 Course here! Be sure to share it with everyone!
Link to lighting course: www.artstation.com/a/25961360
Link to How to make a movie in UE5.1 www.artstation.com/a/22299532
jsfilmz.gumroad.com/l/lmaqam
My Realistic Warehouse VR Demo: www.artstation.com/a/27325570
My Fortnite Map: 3705-9661-2941
Join this channel if you want to support it!
/ @jsfilmz
Sign up with Artlist and get two extra months free when using my link below.
Artlist
artlist.io/artlist-70446/?art...
Artgrid
artgrid.io/Artgrid-114820/?ar...
@UnrealEngine #unrealengine5 #metahumananimator #metahuman
Metahuman Animator Tutorial,unreal engine metahuman animator tutorial,How to Use MetaHuman Animator in Unreal Engine,unreal engine 5.2 metahuman animator,unreal engine metahuman animator app,unreal engine 5 metahuman animator tutorial,metahuman animator tutorial,metahuman animator release date,unreal engine 5 metahuman animator release date,metahuman animator unreal engine 5,metahuman animator unreal engine,unreal engine 5.2 metahuman animator tutorial,jsfilmz
Пікірлер: 294
ITS ON LINKEY DONKEYYYYY! Here is video of the animator app studio.kzread.infoiXH79mrKADM/edit First rap test kzread.info/dash/bejne/n3WIk7ptqMWrlKQ.html
@jimmwagner
Жыл бұрын
Do you need the depth capture in Live Link or can you just use video?
Yes so glad to hear you’ll be doing more tests with this. Bro just realised I’ve been on this unreal journey with you since your channel was a baby. You’re literally the don in this field. Keep em coming
Bruh! It is nuts how accurate this thing is.
Incredible! You're so ON IT. Thanks for the walk-thru. Very helpful!!!
Thanks for the tutorial, I can't wait to try it, although it will take hours since the shorts that I'm doing have a lot of dialogue, but it will look a lot more realistic. Best
Bro, you are a master! And the way that Unreal it's improving this technology day by day it's amazing.
@Jsfilmz
Жыл бұрын
thx dude!
BIG! I was wanting to explore this, thanks for the tutorials!
Knew this was coming the second I saw the Unreal post about this. Excited to try this out today.
just done my first quick test following your tuts to guide me through the process, got to say wowzer animator is amazing. so much cleaner then i expected. well worth the iphone rental :}. keep the vids coming :}
@Jsfilmz
Жыл бұрын
oof someones bout to buy an iphone 😂
You're amazing. Thank you for staying on the pulse!
@Jsfilmz
Жыл бұрын
thanks for being here!
Great job!
Good and fast tutorial, nice man
@Jsfilmz
Жыл бұрын
tx
You the man bro
This was jammed packed with so much information. I have to pause this video and slowly follow all of the steps. This is amazing! I'd like to purchase the headset for my iPhone. Did you make it yourself? or did you get it from a website?
@Jsfilmz
Жыл бұрын
home made bro with my sweat and blood its not $129 anymore though its $169 now
Great stuff!
@Jsfilmz
Жыл бұрын
Thanks!
Awesome.. Randomly woke up at 4am knew there was a reason... thank you!
@Jsfilmz
Жыл бұрын
hahaha
Thank you so much, its amazing
Some day i think you will make tutorials before release. haha. So quick!!!!!!!!!!!
@Jsfilmz
Жыл бұрын
hahaha yea they didnt select me for beta access unfortunately i know some did so im gonna have to catch up to them
sooo quick , Dude ! you are fast as hell .... 😁
@Jsfilmz
Жыл бұрын
i woke up 5 am today haha
Hey! I just wanted to say thank you for your videos! If it wasn't for you I couldn't have created animation for my metahumans. ❤ Thank you
@Jsfilmz
Жыл бұрын
That is awesome!
@HQLNOH
Жыл бұрын
Omg yes! My video will never make it if it wasn't for you! I even mentioned you in appreciation in the description ❤️
@Jsfilmz
Жыл бұрын
@@HQLNOHthanks man not manu give credit back to me i appreciate it
The accuracy looks insane
@Jsfilmz
Жыл бұрын
yea
Awesome video! Thanks for the quick setup explanation! But I have to point out that MetaHuman Animator already uses AI for its solves. When you hit the "Prepare for performance" button, it trains a model on your face to later mimic the way it moves so it can animate other metahuman characters to that likeness. Thats why this step took 8-10 minutes : )
Thanks Lord Helmet!!!
LUV'd your last vid with your little girl (all of your vids ❤ ) 🏆😁👍!
@Jsfilmz
Жыл бұрын
hahah thx man it got 20k views on twitter hahah
yeeeeaaaaaaa 🔥
@Jsfilmz
Жыл бұрын
hey J can i borrow 10k subs so i can hit 100k? thanks mang
7:00 I'm listening to this video driving around, windows down, level 20 on Bluetooth and at a Bus Stop with little old ladies 🤦♂️
@Jsfilmz
Жыл бұрын
lol
This really is such a big step forward. Very exciting. The only problem is the iPhone only requirement. That makes no sense for something like this, but let's hope it changes sooner or later.
@Jsfilmz
Жыл бұрын
beats buying a real mocap system 😂
@oldmatttv
Жыл бұрын
@@Jsfilmz True I suppose :)
BOSS!
I need to buy iPhone 12 + first hahaha
@Jsfilmz
Жыл бұрын
looks like 11 works
@aaagaming2023
Жыл бұрын
@@Jsfilmz In their docs they group the X, 11 and 12 together and then 13 and 14 together. Then they give a caveat about the X, saying that its not capable of capturing more than a few sec, but they dont say that about the 11, so yeah, I think the 11 is capable of capturing longer form content like the 12.
@andrewwelch5017
Жыл бұрын
@@aaagaming2023I tested the iPhone 11 Pro last night and it works fine, zero issues.
A lost detail in Hideo Kojima's DS2 trailer, was the end saying the performance-capture is powered by Metahuman. That game is going to be insane levels of detail.
@Jsfilmz
Жыл бұрын
yea hideo was lurking around on my channel when ue5 first came out i was one of the first ones to cover it
You're absolutely smashing it man 👏🏽 Thank you so much for the awesome content! I do have one slight issue though!.. For the life of me, I cannot get the Livelink Face mocap to work with separate body mocap. The head/chest just detaches itself and they are both independent. It's driving me insane. I've tried following some advice on the UE forums to no avail 😭 Have you experienced this yet? any tips? Thank you
Great showcase! In case you plan to make more test video's, can you show expressions that are hard to do with Apple Arkit? I'm curious how it compares. For example a sad face with hanging lower lip, asymmetric brow movement🤨, worried face 😟 or any interaction between teeth, lips and tongue.
@Jsfilmz
Жыл бұрын
man im a terrible actor but ill try
@Pauliotoshi
Жыл бұрын
@@Jsfilmz Thanks a lot!
Super sick! Do you feel like the results are significantly better than the live stream app?
@Jsfilmz
Жыл бұрын
bro go watch my rap with it it will answer your question
I Hope that behind the scenes they are working on full body mocap
This is awesome! Would I be able to take the metahuman/face mocap and export it into blender? I don't need any of the textures or materials. Thanks!
Amazing content as always! Could you please make a video on troubleshooting these three issues: - “Promote Frame” randomly jumping to a different frame than the selected one. - Metahuman Identity Solve not accurate result. - “Add Teeth Pose” breaking the Identity Solve even more. Thanks a lot!!!!!
Thank you so much for your tutorial, so timely. In addition, I am trying to import from the mesh body in UE5.2, and it seems that the tracking mark link can no longer be carried out. Have you encountered it?
Great video - you mentioned about when MHA will later use AI. Just a heads up it is AI driven currently. Pixel tracking is only a small part of the foundation. Great video!
@Jsfilmz
Жыл бұрын
wait like its doing ai pose estimations already?
@AllanMcKay
Жыл бұрын
@@Jsfilmz yeah there’s a lot under the hood that’s ML driven already. Facial tracking doesn’t account for wrinkles or much else other than eye and mouth shapes, everything else is interpolated with AI. there’s more coming but the foundation is already utilizing AI in a lot of areas.The second pass animation is still being improved on, so it’ll continue to get better. But it’s a training model based on a lot of human facial animation, to know what to do when cheeks are raised, nostrils flared, eyebrow wrinkles etc Night and day different to something like live face or other tools which purely track eye and mouth shapes and don’t leverage any AI to them interpolate wrinkles and pseudo face muscles into the animation
@Jsfilmz
Жыл бұрын
@@AllanMcKayoh wow hahaha crazy stuff being 1.0 its not bad oh btw for iphone it can only output 30 even when recording 60 right? Thanks i love knowing about the tech
...and off we go!
Have u done any vids how to take that head with animation and apply to another metahuman ?
Hope my XS works. Don't want any more Apple pradux
@IamSH1VA
Жыл бұрын
Please try it & please report if it works, I have same iPhone XS. I am not gonna have access to windows system for at least 15 days, but I am dying to test this feature.
Thank you for sharing a fun and essential tutorial!! Anyway, is there a way to use the neck animation recorded by Unreal Live link facial capture? When I imported the facial anim with neck animation, and apply to my metahuman skeleton, the body and face is broken because of the neck anim. It's also possible to just facial anim without neck movement, is there a way to use neck anim...? Is there just one solution using the mocap data (with neck anim) + facial capture anim just face movement(without neck anim)...?
epic insane , looks much better then faceware or livelink .like you say the curves look smooth and no jitter . and to think i spent 4 hours last night doing 30 sec of manual facial animation that looks rubbish. so if i get a friend who has iphone they can send me clips ? .
@Jsfilmz
Жыл бұрын
yes
@kool-movies
Жыл бұрын
@@Jsfilmz awesome ,, is there any reason not to get the iphone 12 mini ?
@Jsfilmz
Жыл бұрын
@@kool-movies i havent tested longer takes with 12 mini yet but it used to overheat on me alot hahaha
@kool-movies
Жыл бұрын
@@Jsfilmz your rap video is a long take so hopefully will be good , i just ordered/renting a cheap refurbished one. hopefully arrives tomorrow. :} the 3lateral video that just released is mind blowing.assume they used a stereo camera ?,
Great video man! I'm getting an error when I hit the process button in the Metahuman Performance. It says "The Processing Pipeline failed with an error." Any ideas on how to fix this would be appreciated. Thanks!
So how do we connect this to a body so we can add animations to it? 😅
4:43 "Iden-TITTIES" hahahahahaha Great video man! Thank you for the content.
crazy times!
Thanks for the tutorial! Do you have a recorded take file I can use as a test? I dont have an iphone. Thanks
it´s here!!!!! downloading plugin!!!!!
Great vid!! In the Showcase ,didnt they show a way you could use this app to generate textures for your metahuman? Will you be showing us how as well?
@Jsfilmz
Жыл бұрын
can you send me that video? i dont think i saw that
@ielohim2423
Жыл бұрын
@@Jsfilmz I think I may be mistaken. I thought the HellBlade II showcase did it ,but I think they may have had a premade metahuman.
@ielohim2423
Жыл бұрын
@@HellMunky What's your workflow? What do you use to generate the mesh/textures that you import into UE to use in the plug-in?
If I'm not wrong, you should be turning off the neck solving on a headmounted camera. And use it only for static cameras.
@Jsfilmz
Жыл бұрын
in my case ill use neck movements from my mocap the exported sequence doesnt come with neck check out new rap video i uploaded
does someone got it working good with iphone 11? the new epic post says it needs aleast iphone 12..
@Jsfilmz
Жыл бұрын
try it broski
I wonder how long a given take can be. I have to make some EDU products with this, and they might need to be on the long-ish side.
hey bro! awesome! I have a little issue, when I track my face with the animation sequence it's like nothing happend in my sequencer, but it exported correctly because when I open it alone the animation is fine. any idea? Thank you!
@Jsfilmz
Жыл бұрын
i dont understand :( maybe join unreal discord and post pics and issue there
I receive software release and your notifications at the same time, how is it even possible :D
@Jsfilmz
Жыл бұрын
i woke up 5 am today to make this video
Epic is great but they gotta give us some options with the markers, like maybe don't make the ones that go on teeth straight up yellow / green lol? Great video though.
Did you see issues of floating head? Face animates and body remains still.. if neck rotation is disabled it's fixed but the neck rotation part of the capture is lost.. Do you know how to avoid that? Thanks! 🙂🙏🏻
Hello bro.I just went back to this video ,becasue for me when update to 5.2.1 its says Preparation for Performance Fialed,any tips on it bro.
So, whenever I want to make a facial animation (of myself doing it with live link) I need to go through these steps? But they can be used on every metahuman?
when I play the level in a new window (PIE) the Metahuman character is moving his face with my motions very good, but when I click in any window other than the PIE window it starts to lag. and when I click again on the PIE window, the character moves normally ! what should I do ?
Hi! Thanks for the cool video. Listen to the question, will it work if you upload a regular video shot on a camera?) just no iphone))
@Jsfilmz
Жыл бұрын
dont think so iphones are precalibrated like i showed in the performance editor
@andrewwelch5017
Жыл бұрын
No, the software requires depth information which a regular camera can’t provide. Borrow an iPhone 11 or newer.
my head keeps detaching from the body when i attach the animation to the face and play it in sequencer
Maybe in a year or 2 Epic will give us a way to import our videos for mocap, I’ve tried so many apps and don’t like the results, only one left to try is supposedly the best one, move ai.
@Jsfilmz
Жыл бұрын
yea its good man
Incredible! Is there a way to export Animation data to Maya to have more freedom for further tweaks?
@Jsfilmz
Жыл бұрын
that would be amazing but i dont think thats possible yet
@Luke_wp
Жыл бұрын
This guy made a script to do it kzread.info/dash/bejne/l5eNsZdmZcmyYLg.html
Having a huge problem when I try to add the animation to my metahuman in the sequencer it becomes detached from the body from around the shoulders area. Was it because I may have moved too much in the capture? Can’t seem to get the body and the head attached
JS can that be done with DAZ characters too?
Hi! It's amazing video. I try to import my recording video using iPhone 11 but, in the capture manager the video can be reading. Any suggest for it?
@Jsfilmz
Жыл бұрын
firewall?
after ubdate iphone app It gives a warning for the animation part of the application live link face app saying your device model is unsupported you may continue but your results could be affcted"for meta human animator capture ." and its work but i wonder My economic situation is not good and I will develop games. How much performance difference does the iPhone 11 make?
@andrewwelch5017
Жыл бұрын
The software requires an iPhone with a “True Depth” (LiDAR) sensor because it needs depth data to accurately track your face. You can always borrow an iPhone 11 (or newer) to do the tracking and then transfer the file to your computer.
@UnrealEnginecode
Жыл бұрын
thank you for your answer, what are they doing this warning for, do you think there will be a noticeable quality difference?
@andrewwelch5017
Жыл бұрын
@@UnrealEnginecode The quality I got was excellent so I'm not worried about it.
Hold up! Did it just create a Metahuman face with your facial proportions just based on the iPhone calibration!? that's crazy!
@Jsfilmz
Жыл бұрын
yes
@Impaczus43
Жыл бұрын
@@Jsfilmz that's awesome! Glad you were able to share this! Awesome video
@Jsfilmz
Жыл бұрын
@@Impaczus43 it made a fatter version of me which is effed up lol
@soncho.editz1
Жыл бұрын
@@JsfilmzI mean honestly it does have your facial features but do u think you can adjust that manually ???
Do you think there will be a marketplace to buy mesh to meta human data scans eventually? For example I don't know any korean girls to run this on with an iphone like that company did but i'd be willing to pay people who know attractive people from every race if they did
What about body movement and hands?
I get the "assertion failed" crash every time I promote the first frame... what should I do?
can i move the mocap data to another software, like blender?
Can you do a video of how to warp a metahuman to look like a custom character? RS3D Zwrap is a good wrapper. I have this model of NAS I want to put to the test, as well as Michael Jordan and the Rock.
@Jsfilmz
Жыл бұрын
just mesh to metahuman it mang save u the headache hee hee
@SkyHandOneTen
Жыл бұрын
@@Jsfilmz Hmmm, I guess I'm a bit behind. Not sure what that is or where it is. I'll look it up. Thanks
@Jsfilmz
Жыл бұрын
@@SkyHandOneTenoh yea man mesh to metahuman jsfilmz look it up its easy
@SkyHandOneTen
Жыл бұрын
Seems that info is everywhere I was just looking up the wrong terminology all this time. Thanks.
Hi, i asked this to a lot person but couldn't get appropriate answer. After creating identity and perform with it, my Metahuman Character's default lips and teeth are changing, it's because of getting my identity. Is it possible to keep animation exactly same but with character's default facial features?
His next epic: "Get Shorty"
Was about to say the same thing about your teeth JS 😂
add get hangs when I try to bake the animation....???
Did everything to the t and my animation sequence doesn't show up in the Face animation menu. Any tips?
3:52 it's says iPhone 11 here But my phone is also 11 and over there on live link for metahuman animator I get that your device model is not supported you can continue to use but results won't be that good
@Jsfilmz
Жыл бұрын
yea ive had people here try it
@reubencf
Жыл бұрын
@@Jsfilmz I just went back to the site and it seems that iPhone X and 11 are removed
@Jsfilmz
Жыл бұрын
wtfff on the docs?
@reubencf
Жыл бұрын
@@Jsfilmz yes
when doing facial mocap, what do you do for a mic? do you have a little one that you attach to your helmet? what kind of mic is goof for that?
@Jsfilmz
Жыл бұрын
just my good ole senheisser g2
@binyaminbass
Жыл бұрын
@@Jsfilmz do you attach it to the arm of your helmet?
is there a way to create the captures from a webcam?
I don't have an iPhone. Will it work with my 2021 iPad pro ?
The Performance audio track shows 'Unresolved Binding' though I can hear the audio. When I export the animation no sound is exported......the internet has failed me. Anyone? UE 5.3 and 5.4
Do you think it's worth shelling out a bit more money for the iPhone mini 13 over 12?
@Jsfilmz
Жыл бұрын
im broke so ur askin wrong person
Were you able to figure out how to connect the iPhone to wirelessly be triggered to record from Unreal?
@Jsfilmz
Жыл бұрын
just the regular livelink way?
@tomhalpin8
Жыл бұрын
@@Jsfilmz With the new MetaHuman Animator version of LiveLinkFace, can't seem to connect. Trying to match my body mocap with the face mocap.
@Jsfilmz
Жыл бұрын
cant do it live animator is offline
@tomhalpin8
Жыл бұрын
@@Jsfilmz I wonder if I can I can use OSC to sync up the recording process on the phone with the mocap
Isn't it necessary to have an iPhone with this? I don't have an iphone sorry
I thought when you're calibrating you need a minimum of 3 the front, left and right, and that the teeth only need one in front?
@Jsfilmz
Жыл бұрын
turn ur head with helmet on
@ComanderJTC
Жыл бұрын
@@Jsfilmz ah I see your point but wouldn't it be better for calibration to do it without the helmet to get the three angles then so the depth part of the camera recordings with the helmet on make it more accurate for certain things like creases on face and stuff similar to that when making facial expressions?
@Jsfilmz
Жыл бұрын
@@ComanderJTCar as i know u gotta calibrate like how ur gonna record
@ComanderJTC
Жыл бұрын
@@Jsfilmz ok makes sense I just figured it would read your face better if it had an initial depth analysis from multiple angles was my thinking
how to add my facial animation to different character
can it be used not with my face but with the one created in metahuman creator? and how to do it?
@Jsfilmz
Жыл бұрын
stay tuned
I have 16 gb ram. I can't use prepare for performance, help.
Hey J, for some reason the result of your test is not very good, compared to other tests I have seen online. Do you reckon there was something in the configuration/shooting conditions that interfeered? Or perhaps the other tests used other types of cameras, such as stereo cams, rather than iPhone? Thanks for the tut though.
@Jsfilmz
Жыл бұрын
i think the demo videos that came out were done with stereo cams not sure
@matteo.grossi
Жыл бұрын
@@Jsfilmz The rig used when they announced MHA wasn't a metahuman rig, in fact they only showcased how the new system works on metahuman rigs for like, 5" at the end of the presentation.
@Jsfilmz
Жыл бұрын
@@matteo.grossi hahaha thats cheating then right lol
I have a problem that the character shows more bottom teeth than top. I dont speak like that andclive link doesnt do it. I tried re-tracking face markers but still had the same problem.
I may be a little late to the party, but im confused why you choose in CAPTURE SOURCE LivelLink Archives (which is uploading footage etc from PC drive?) then when you go to import you do it from the iPhone in CAPTURE MANAGER?
@Jsfilmz
9 ай бұрын
i made tutorial both ways my usb transfers faster for bigger files
@Blairjones3d
9 ай бұрын
@@Jsfilmz oh sweet as no worries bro. Any tips on uploading from USB then using archives? Seems more complicated to setup the Performance that way...
Idk what's going on its says that's its no longer available 😕 I been trying to this for weeks now
Also hey I'd like one of those helmets, but I do have an iphone 11, will that work? It's bigger than the mini, of course.
@Jsfilmz
10 ай бұрын
i know some people whos tried 11 with MHA and they said it works i havent tested it myself
@JoshuaLundquist
10 ай бұрын
Gonna buy an iphone 12 mini like you have, can you tell me which mount you use so I can buy that and yr helmet? @@Jsfilmz
On the metahuman video they use 4 calibration images, is there a reason why you did not use side views?
@Jsfilmz
Жыл бұрын
i have helmet on bro hahahaha they on a tripod if i move my head the camera wil move too
@chipcode5538
Жыл бұрын
@@Jsfilmz You could make the calibration video with a tripod and use the helmet for the capture. I don’t know if it will improve the calibration, the metahuman seems a little off especially the nose.
Jae, are from the future? I think you are a time traveler, you had been creating the tutorials before the release of the tool. 👀😃
@Jsfilmz
Жыл бұрын
hahahaha
Damned... I got an I Phone X... Time to go to the 13 mini after all.
@Jsfilmz
Жыл бұрын
looks like x is unofficially supported
@CoreGamingProject
Жыл бұрын
@@Jsfilmz Yep, but only for few second.. I need 3 or 5 minutes recording for all my Project. I will test tonight if it's ok or not, but... There is no reason for them to lie. XD
@CoreGamingProject
Жыл бұрын
It seems to work without Problem, for the one Minutes I tested.
@Jsfilmz
Жыл бұрын
@@CoreGamingProject dope
I've done it but how can I render it... help me pls