Do all these steps work when performed on the phone?
@nnablaАй бұрын
Thanks for watching our video. The answer depends on how you want to use this. You can access to Google Colaboratory and run this demo via smartphone, but that means it is not performed on the phone, but on their server. I hope this answers your question.
3:12 結果パート「GANベースのADM」--> 「Diffusion ModelのADM」です 4:25 右側の論文タイトルは「Pre-training Vision Transformers with Very Limited Synthesized Images」-->「SegRCDB: Semantic Segmentation via Formula-Driven Supervised Learning」です
@白葛 Жыл бұрын
分かりやすい解説ありがとうございます!
@net_stack4176 Жыл бұрын
Hi, thanks for the video. This really good. How do I run an inference for a set of images and get the prediction results saved?
@nnabla Жыл бұрын
Hi, thanks for using our Colab demo. Since it's a bit hard to show you how to do that here, we've opened an issue and answered your question. Please refer to github.com/sony/nnabla-examples/issues/394.
Thanks for this! Does the png image and mp4 video have to be the same pixel ratio? I tried uploading an mp4 from my files and ffmpeg gave me a warning, then tried to play and it gave this error: FileNotFoundError: [Errno 2] No such file or directory: 'result/arbitrary/input_image.png_by_input_video.mp4'
@nnabla Жыл бұрын
Hi, sorry for a bit late reply. Thanks for using our demo! > Does the png image and mp4 video have to be the same pixel ratio? No. We confirmed it works even when they have different pixel ratio. Note that it would affect the generation quality. Judging from the error message, it seems inference code failed to generate the resulting video. I don't think it was because of the different ratio. There might be something wrong in input video such as some special codec or non-ascii filename? Can you try again if you're still interested? Thanks.
@Nightspire1 Жыл бұрын
Thanks for replying! I've tried several different ai lipsyncing models, they all seem to warp the face. I feel ai lipsyncing is still in its infancy, and I don't know python too well so that is the biggest hurdle. SO many dependencies with so many different versions makes using any program in python very difficult, again if you don't really know what you're doing. I appreciate the reply, and the collab demo, as collabs are the only thing i can get to work!@@nnabla
@moyamoyamoyamoya Жыл бұрын
毎度勉強させてもらっています!
@mkii5095 Жыл бұрын
いつも論文紹介お世話になってます!
@andrewshin8704 Жыл бұрын
1万人おめでとうございます!🎵
@どこかのだれか-s8g Жыл бұрын
すごくわかりやすかったです! これを機に自分でも最新の論文を追っていきたいと思います!!
@toritometo Жыл бұрын
お腹すいたカバ!可愛い名前
@mariosalamanca7743 Жыл бұрын
Currently not working. Code stops at from generate import *
@nnabla Жыл бұрын
Hi, thanks for using our demo. We confirmed that the problem has been solved and it works as expected. Can you try that if you're still interested? Thanks!
@demaxism Жыл бұрын
今GPT人気爆発ですね もっと前からこれを見るべきだった~
@YukioHatoyama114511 Жыл бұрын
やろうと思ったけどグラボ積んでるpcなかった
@truth2 Жыл бұрын
Brother I am doing in mobile And the site is Demo for paper "First Order Motion Model for Image Animation" And I was playing cells, those cells were being played well, after mounting Google drive, the next cell was Creat a model and load checkpoints When i pressed, it showed me error of ImportError Traceback (most recent call last) <ipython-input-6-dbd18151b569> in <module> ----> 1 from demo import load_checkpoints 2 generator, kp_detector = load_checkpoints(config_path='config/vox-256.yaml', 3 checkpoint_path='/content/gdrive/My Drive/first-order-motion-model/vox-cpk.pth.tar') 3 frames /content/first-order-model/augmentation.py in <module> 10 11 from skimage.transform import resize, rotate ---> 12 from skimage.util import pad 13 import torchvision 14 ImportError: cannot import name 'pad' from 'skimage.util' (/usr/local/lib/python3.9/dist-packages/skimage/util/__init__.py
@hungah Жыл бұрын
11:08 わかりやすい
@googIe.com.2 жыл бұрын
ファインチューニングなしでUniTuneできるZeroTuneがほしいですね
@googIe.com.2 жыл бұрын
拡散モデル完全に理解した
@dazof76712 жыл бұрын
python初心者なので質問の意図が動画の主旨から外れている場合は無視してください。コラボラトリ上のRun the trainingの再生ボタン(実行ボタン?)をクリックするとenv.close()の行でNameError:name 'env' is definedとなりました。githubにアップロードしていただいている内容に問題ないでしょうか。
@nnabla2 жыл бұрын
ご質問ありがとうございます。おそらくですが、最後のRun the trainingの再生/実行ボタンだけ押していないでしょうか?Google Colaboratory上での実行では上から順番に全部の再生/実行ボタンを押して準備をしていくので、上にある他のブロックの再生ボタンを押した後で、最後にRun the trainingを押せば、学習が開始されると思います。
@nnabla2 жыл бұрын
Thanks for watching this video and trying the Colab demo! Unfortunately, as of now (Nov. 8, 2022), we observe an error when running the Colab demo. We will fix this issue very soon and let you know here.
@walidflux2 жыл бұрын
hey man can you do update for this?
@nnabla2 жыл бұрын
Yes, we observe there'll be an error if you run the current Colab demo as is. A simple solution is to install the latest nnabla ("!pip install nnabla-ext-cuda114" in the first cell). Anyway we will update the demo soon.
Quality of generation results depends on the input images or videos. One recommendation is to use well-aligned (same scale and same orientation) images as input videos (or vice versa). Or, there could be an issue when downloading the required pretrained weights file, and the AI model failed to use it. If the model was forced to run without proper weights file, generation results would be totally collapsed.
@zimma43352 жыл бұрын
When i try to play the video it says "FileNotFoundError: [Errno 2] No such file or directory: 'result/arbitrary/input_image.png_by_input_video.mp4':
@nnabla2 жыл бұрын
One possible cause is that you executed the cell before the previous cells had not finished. Some cells would take long until the process ends.
@amathpati91072 жыл бұрын
Can we give the video as input instead of images to mimic the expression ?