AviSynth

From SDA Knowledge Base

Revision as of 23:50, 10 September 2013 by Manocheese (Talk | contribs)

(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search

Introduction

IMPORTANT NOTE: AviSynth scripting is an advanced topic. All beginners are recommended to use Anri-chan instead.

To put it briefly, AviSynth is a video editor like VirtualDub except everything is done with scripts. For details, check Wikipedia. The most important thing to learn right now is the concept of AviSynth.

Let's say you're still using VirtualDub:

  • You go through the menu or drag and drop your source video inside the program.
  • You load the audio.
  • You use the brackets to cut off frames you don't need.
  • You go to the filters section and use resize.
  • Resizing has made the picture blurry, so you add the sharpen filter.

You did all that by moving your mouse, going through menus, etc. With AviSynth you're creating a text-based file (.avs) to tell it what to do with text commands. The above would be, as an example:

  • avisource("myvideo.avi")
  • wavsource("myaudio.wav")
  • Trim(4000,7000)
  • Lanczos4Resize(320,240)
  • XSharpen(30,40)

You save the avs file, and load that into MeGUI or VirtualDub for final compression. The advantages are that you don't have go through as many menus, you don't have to remember which frames you want to cut out, you have access to more advanced deinterlacing filters like mvbob, and you can keep your scripts forever so that you don't have to start from scratch in case you want to re-encode them later.

How to use this guide.

Yes, AviSynth can be confusing and hard to learn, but it is very rewarding once you get the hang of it. I suggest you look at the sample scripts at the bottom of the page to get an idea of what a final script looks like. Then go through each section putting whatever you need into your own script. If you are having trouble, and you probably will >:), do not hesitate to ask for help in the Tech Support forum.


Installation / plugins

Go to http://www.avisynth.org/ or http://sourceforge.net/project/showfiles.php?group_id=57023 to download AviSynth 2.5.7. Note that Part 11: SDA StatID cannot be completed with version 2.5.6 or lower.

Download these necessary avisynth plugins.

With AviSynth installed, go to Start menu -> [All] Programs -> AviSynth -> Plugin Directory. This will open the directory where AviSynth stores its plugins. Copy the files from inside the AviSynth plugins zip file to the Avisynth plugins directory window you just opened. Note that you can't copy the directory you unzipped in; you have to copy the files inside the directory you unzipped into the AviSynth plugins directory. This is because AviSynth won't recognize the plugins if they are inside a directory inside the AviSynth plugins directory.

If AviSynth complains that the MSVCR71.dll file is missing, you can download it at http://www.dll-files.com/. Place it in your "C:\WINDOWS\SYSTEM32\" directory.

You'll also want to have VirtualDub to test your script as you go along.

The AviSynth script

Create a .txt document with a name of your choice in the same folder as your source files. Rename the extension from .txt to .avs. If you can't see the extension and are running Windows, open Windows Explorer, go to Tools -> Folder Options -> View and uncheck "Hide extensions for known filetypes".

Open the avs file in Notepad or any text editor.


Tip: You should create separate AviSynth scripts for each quality version that has differences between them. There's no sense in creating one AviSynth script, encoding it, editing the script, encoding again, etc. With separate AviSynth scripts you'll get less script errors and less headaches, and you'll be able to queue many encodes.


Part 1: Loading the plugins

AviSynth will automatically load any dll and avsi files located in the plugins directory. If you're working with DVD source material, the only plugin that I recommend you load manually is DGDecode.dll since it's in its own folder. You can of course just copy and paste it into the plugins folder, but be sure to re-copy it if you update DGMPGDec. Change the file path if needed.

Loadplugin("C:\Program Files\DGMPGDec\DGDecode.dll") # For DVD source

#### For MvBob deinterlacing. Remove the # to load it.
#import("C:\Program Files\AviSynth 2.5\plugins\mvbob.avs")


Part 2: Loading the source files

There are different ways to load the source files and it all depends on what it is you're working with. Here is a list of available commands.

  • avisource(video.avi)
  • directshowsource(video)
  • MPEG2source(video.d2v)
  • Ac3source(video, sound.ac3)
  • Wavsource(sound.wav)
  • Mpasource(sound.mpa)
  • AudioDub(video, sound)

If you used a capture card or screen capture software (e.g., Fraps, Camtasia) then it is quite simple to load the files. If avisource does not work, try directshowsource. If you are using Camtasia and chose to use its own Techsmith codec, then use directshowsource.

### Video and audio are already combined
avisource("video.avi")

### Video and audio are split
#AudioDub(avisource("video.avi"), wavsource("audio.wav"))

Notice the # before AudioDub. This is telling AviSynth to skip over the line. If you need to use this line, remove the # and add one before the first avisource command.

Tip: If your avs script file is in a different folder than your source files, just use relative or absolute paths, like "c:\my work folder\video.avi".

Tip: If your video is separated in multiple parts, which is usually the case when recording with Fraps, then be sure to look at part 4 since it is linked to part 2.


If you used a DVD recorder then your video and audio is most likely split. Make sure you've gone over the DVD page before continuing.

AC3source(MPEG2source("vob.d2v",upConv=1),"vob T01 2_0ch 192Kbps DELAY -66ms.ac3")
#AudioDub(MPEG2source("vob.d2v",upConv=1),MPASource("vob T01 2_0ch 192Kbps DELAY -66ms.mpa")) 

I hope you haven't removed the delay information from the sound files. Not that it's the end of the world if you did remove it, you'll just have to listen by ear until you get a close value with DelayAudio().


Part 3: Fixing audio delay

Constant audio desync

Constant desync is when the desync at the beginning of the video is the same as the desync at the end of the video.

The DelayAudio command is straightforward, but there is an extra concept worth learning. Look at the following two scripts:

Ac3Source(MPEG2source("vob.d2v",upConv=1),"vob DELAY -66ms.ac3")
DelayAudio(-0.066)
Ac3Source(MPEG2source("vob.d2v",upConv=1),"vob DELAY -66ms.ac3").DelayAudio(-0.066)

Is there a difference between them? Yes and no. They both get the same result, however the script with one line makes it easier for projects where you append files together. I suggest using the format of the second script.

Progressive audio desync

Progressive (gradual) desync is when the desync at the beginning of the video is not the same as the desync at the end of the video. For example, the audio seems fine but then later on someone fires his gun and you hear the bang a second later. This type of desync can be hard to fix since you usually have to figure out the correct desync value manually. THE DESYNC MUST BE GRADUAL! Do not apply to audio that has an abrupt desync somewhere along the timeline.

#Positive value stretches audio, negative value shrinks audio. Value in milliseconds.
stretch=-300
#Do not change the line below.
TimeStretch(tempo=(last.AudioLengthF/(last.AudioLengthF+(stretch*(last.audiorate/1000)))*100))

Tip: Preview avisynth scripts in Media Player Classic


Part 4: Appending

One of the best features of AviSynth is its ability to do an aligned splice when appending video. There is usually a mismatch between the length of the video and the length of the audio, typically ranging from -50 ms to +50 ms. This means that appending files in VirtualDub(Mod) will in almost all cases cause a desync because the audio of clip2 will be appended right after clip1. VirtualDub(Mod) is unable to do an aligned splice. Here is an illustration:

Append.png

You will probably never use UnalignedSplice. Here are different methods for using AlignedSplice:

### Method 1
AlignedSplice(avisource("clip1.avi"), avisource("clip2.avi"), avisource("clip3.avi"))

### Method 2
a = avisource("clip1.avi")
b = avisource("clip2.avi")
c = avisource("clip3.avi")
AlignedSplice(a,b,c)

### Method 3
avisource("clip1.avi")++avisource("clip2.avi")++avisource("clip3.avi")

Ac3Source(MPEG2source("vob1.d2v",upConv=1),"vob1 DELAY -66ms.ac3").DelayAudio(-0.066)++Ac3Source(MPEG2source("vob2.d2v",upConv=1),"vob2 DELAY -30ms.ac3").DelayAudio(-0.030)

### Method 4
import("script1.avs")++import("script2.avs")++import("script3.avs")

Method 2 is good for when you're joining a lot of clips since it's easier to edit. Notice the double plus signs in method 3; this is the same as AlignedSplice. One plus sign would indicate UnalignedSplice. Method 4 is to join independent AviSynth scripts.


Part 5: Deinterlacing / Full framerate video

Deinterlacing isn't needed for video that is not interlaced such as computer games recorded with Fraps/Camtasia, or if your console game outputs in progressive mode and you captured in progressive mode as well. Otherwise, if your footage looks anything like the picture below with the horizontal lines, then you definitely need to deinterlace.

Tip: The SelectEven() command used in this section is still useful for those with beefy computers who recorded their computer game at 60 fps. The command will get you half framerate (30 fps) needed for LQ/MQ.


Dmc3interlaced.jpg


What is deinterlacing?

It's important that you understand why you'll be using whichever deinterlacing method is needed for the video you recorded. Since I'm lazy, I'll briefly explain what deinterlacing does.

An interlaced frame has two fields. Think of it as a single frame with two smaller frames inside. Those two fields (the smaller frames) represent two different moments in time. Take the picture as an example, with four different fields, A,B,C,D:

Abcd interlaced.png

When deinterlaced by separating the fields, we get:

Abcd deinterlaced.png

This is how video that has been recorded at 29.97 fps interlaced can be deinterlaced to achieve 59.94 fps progressive.

You may be wondering: "doesn't splitting the fields squish the video?". Well, yes, but it's not a problem for low resolution video since we can just resize it horizontally. For high resolution, experts out there have created filters like telecide and mvbob to reconstruct the height.

Here's where it gets a little more complicated... Some games only run at half framerate (or even lower) but are still interlaced. So, instead of A/B,C/D becoming A,B,C,D for full framerate you get A/A,B/B becoming A,A,B,B. Obviously, there's no point in keeping the duplicate frames. These games are usually very easy to deinterlace with good results, but you have to learn how to determine the framerate of your game.


Determining the framerate of the game

First, check to see if your game is listed on this page. If it's not there, read on.

By this point in the guide, you should have a script similar to this:

Loadplugin("C:\Program Files\DGMPGDec\DGDecode.dll")

Ac3Source(MPEG2source("segment1.d2v",upConv=1),"segment1 192Kbps DELAY -48ms.ac3").DelayAudio(-0.048)

Add SeparateFields() at the end and save your script. Load this script into VirtualDub(Mod). You'll notice that the video has been squished vertically, don't worry about this.

To make it easier for you, move to a part of your video that has a lot of motion. Also make sure you're looking at actual gameplay, not a cutscene, since gameplay is much more important for speed runs. Use the buttons at the bottom of VirtualDub(Mod) to step through your video frame by frame. What you're looking for is difference in motion. If you see unique motion every single frame, then your game is full framerate. If you see unique motion every two frames, the game is half framerate. If it's every three frames, you've got one-third framerate.

For example, if U=unique and D=duplicate, you can write full framerate as a sequence of frames U,U,U,U,etc, and you can write half framerate as a sequence of frames U,D,U,D,etc.

There are some games that have a strange sequence of frames. We recently saw one in Tech Support for a PAL game that had a sequence of frames U,D,U,U,D,etc. If you get something strange like that, make a topic in Tech Support about it so we can look at it.

Before leaving this section, make sure you've determined whether your game is F1 (full framerate), F2 (half framerate) or F3 (third framerate).

Remove SeparateFields() from the script.


Determining the resolution of the game

If you're unfortunate enough to have recorded your game at a low resolution, then this section probably doesn't matter.

First, check to see if your game is listed on this page. If it's not there, read on.

This section requires a bit of your own research. Wikipedia is a good place to start. Older consoles like the NES and Sega Genesis can only output in low resolution, however the SNES, N64 and PS1 are capable of high resolution. From the PS2 and onwards, high resolution for games is more probable.

http://www.google.com


Deinterlacing

Refer to the Glossary of terms if needed.

Now that you know the framerate and resolution of your game, you can do some actual deinterlacing. Note that there are limitations on the quality versions that SDA produces.

  • LQ for Xvid/H.264 must be half or third framerate, low resolution.
  • MQ for Xvid must be half or third framerate, low resolution.
  • MQ for H.264 can be any framerate you want, low resolution.
  • HQ/IQ for Xvid/H.264 must be full framerate (if available), high resolution (if available).

Ask yourself which one you want to do, then follow the appropriate link/section:

F1 - Full framerate - high resolution
F1 - Full framerate - low resolution
F2 - Half framerate - high resolution
F2 - Half framerate - low resolution
F3 - One third framerate


F1 - Full framerate - high resolution

For full resolution deinterlacing leakkerneldeint or mvbob come into play. If you notice too many "lines," or interlacing artifacts, as we like to call them, then lower the threshold value. The negative effect to lowering this is that you end up with more jaggedy edges, and loss of details. Use what you think is best.

### You may have to set order to 0. Try both to see which one works.
LeakKernelBob(order=1,threshold=10,sharp=true,twoway=true,map=false)

Those who want more quality at the cost of encoding speed can use mvbob instead of leakkernelbob. Nate uses this himself, so if you want to go the SDA way, go with mvbob. Beware; this thing is very, very slow. But it comes with a free Frogurt! You may also want to read this tip from General Advice to save time.

### Use complementparity if the motion of the video seems to go back and forth.
#ComplementParity()
ConvertToYUY2()
mvbob()

Here are two pictures that show the difference between mvbob and LeakKernelBob. Loading the pages in two tabs and switching back and forth is recommended.

mvbob

LeakKernelBob

Note how LeakKernelBob tends to create artifacts near edges.


F1 - Full framerate - low resolution

Method 1 - Pros: Easy and fast. Cons: You'll notice a bobbing effect which is unpleasant to the eye. Harder for the encoder, bitrate usage is higher.

### Use complementparity if the motion of the video seems to go back and forth.
#ComplementParity()
separatefields()

Method 2 - Pros: Zero bobbing. Great for the encoder, bitrate usage is lower. Cons: Slightly blurry image from resizing. Slow because of the mvbob filter (don't even think about using leakkernelbob).

 
### Use complementparity if the motion of the video seems to go back and forth.
#ComplementParity()
mvbob()
(RESIZE FILTER GOES HERE, check part 7)
(SHARPENING FILTER GOES HERE, check part 8)

Method 2 is commonly used for Game Boy footage. Sharpening is probably a good idea, too, but don't go overboard.


Method 3 - This method eliminates bobbing by shifting fields. It is much faster than method 2, but may not work as well for some types of video. Try both on a sample and if the difference is unnoticeable, use this method.

separatefields()
converttorgb32
clip1=SelectEven.Crop(0,0,0,last.height).AddBorders(0,0,0,0)
clip2=SelectOdd.Crop(0,0,0,last.height-1).AddBorders(0,1,0,0)
Interleave(clip1,clip2)
converttoyv12


F2 - Half framerate - high resolution

In certain cases gameplay will be in 30 fps(25 for PAL), but you still end up with interlaced video. This is because the fields shifted out of order, and in this case it's called combing. All you need to do is shift the fields back with their proper frame and it'll be deinterlaced, or decombed. Sounds simple, but usually these games just seem to be progressive at one point, and combed at another; almost as if it were random. This is fixed by using the Decomb filter with the Telecide() command.

### Use complementparity if the motion of the video seems to go back and forth.
#ComplementParity()
Telecide()

With games that run at half framerate, there might be certain small parts in the game that are actually at full framerate such as menu screens. Since they are usually just a small part of gameplay, you can just ignore it and let the Decomb filter deinterlace it.

You can also manually tell the Decomb filter to decomb certain parts. You can do so by creating a file called decomb.tel in the same directory as the avs file. First change your Telecide() command.

Telecide(ovr="decomb.tel")

Now open up decomb.tel and tell it what frames you want matched. For example:

100,200 n
500,600 n

This says frames 100-200 and 500-600 will be matched with the next. Use p for the previous frame and c for the current frame. If you're not sure, just try them all out.


F2 - Half framerate - low resolution

This is the easiest kind of deinterlacing. You barely even have to think about it. Looking at the picture above, notice all the individual horizontal lines. These show the fields. Every even numbered line is one field, while every odd numbered line is the other field. With half framerate we simply remove one of the fields. This reduces the height of the video by half, we'll take care of the width later on when resizing.

separatefields()
selecteven()
F3 - One third framerate

Used for games that have problems with flickering sprites, typically what you see in older games. This is probably the trickiest kind of deinterlacing, so you may want to consult with others in the Tech Support forum.

### Use complementparity if the motion of the video seems to go back and forth.
#ComplementParity()
separatefields()
selectevery(3)

http://www.avisynth.org/SelectEvery


Part 6: Trimming

Trim is used to cut out frames. The numbers inside the brackets represent the range of frames you want to keep and are inclusive. To make it easier to find the frame range numbers, load the avs file into VirtualDub(Mod) since it displays the current frame number.

Note that deinterlacing methods such as leakkernelbob and mvbob will double the amount of frames in your video. If you figure out the frame range before deinterlacing, but then mvbob it, you'll have to double the numbers. For example, the following two are equivalent:

avisource("dmc3_1.avi")
trim(106,31200)
mvbob()
avisource("dmc3_1.avi")
mvbob()
trim(212,62400)

Selecteven() will grab all even numbered frames, and discard the odd numbered. The following are equivalent, as well:

avisource("dmc3_1.avi")
trim(106,31200)
selecteven()
avisource("dmc3_1.avi")
selecteven()
trim(53,15600)

You can also combine trim commands.

Trim(100,35000)++Trim(36000,50000)

http://www.avisynth.org/Trim


Part 7: Resizing

Rule 1: Never resize to a resolution greater than that of the original source video. This is called "stretching" and does nothing to increase the quality of your video.

Rule 2: Never resize to a resolution greater than that of the original game! You may have captured that NES game at full resolution, but it truly plays at low resolution.

Correcting the aspect ratio

The NTSC DVD format is 720x480, this is an aspect ratio of 1.5:1 while your game probably play at a ratio of 1.33:1, or 4:3. The correct resolution is 640x480.

Resizing for HQ/IQ full resolution:

The only time you should be resizing for HQ/IQ is if the aspect ratio is incorrect. People with NTSC DVD recorders will end up with video at 720x480 resolution, an aspect ratio of 1.5:1. This is a problem since the game you're recording probably plays at an aspect ratio of 1.33:1, or more commonly reffered to as 4:3. In this case you would resize the video to 640x480. PAL DVD video will come at 720x576 and needs to be resized to 704x576.

### NTSC
lanczos4resize(640,480)

### PAL
lanczos4resize(704,576)

Resizing for MQ/LQ low resolution:

This step is required to meet SDA standards.

### NTSC
lanczos4resize(320,240)

### PAL
lanczos4resize(352,288)

Be sure to read Part 8: Sharpening since resizing can possibly make the picture too blurry.


http://www.avisynth.org/Resize
http://www.avisynth.org/ReduceBy2


Part 8: Sharpening

When a picture or video has gone through significant resizing you usually end up with a blurry image. This is where sharpening comes in. Those who recorded their computer game with screen capture software are pretty much guaranteed a blurry image when resizing to the MQ resolution. Same with those who chose to deinterlace their video to the MQ resolution with the mvbob + resize method. Do not go overboard with the sharpening, play with the values until it looks right.

Since XSharpen works in the YV12 colorspace, you will have to convert it first.

ConvertToYV12()
### Defaults are 30,40
XSharpen(30,40)


Part 9: Cropping / Adding borders

You may be tempted to crop out the black border of console recorded footage. It's best to leave them in, otherwise you're going to make more work for yourself dealing with strange resolutions, encoders not accepting them, and possibly using the resize command later only to get an incorrect aspect ratio. Cropping should be used sparingly.

The following code corresponds to the illustration.

Crop(10,8,-14,-16)

Crop.jpg


You will probably not use the crop and add border commands. A good example of its combined use is if you want to get rid of some noise at the bottom of the video. You would crop it away, then add the border back. Only do this if the noise is along the black border of the video, since you don't want to crop away gameplay footage.

Crop(0,0,0,-10)
Addborders(0,0,0,10)


Part 10: Color / Brightness

Be very careful when playing around with color and brightness. If your video is too bright and looks greyish it will be rejected. Feel free to ask others in the Tech Support forum for their opinions about your video.

http://www.avisynth.org/Tweak

http://www.avisynth.org/Levels

http://forum.doom9.org/showthread.php?t=93571


Part 11: SDA StatID

Tip: This should be the last thing you do in your script. Make sure the script works without the StatID before continuing.

SDA uses Station ID's to protect the runner's and the site's identities. StatID's are placed at the beginning of a video and shown for a few seconds. They are the next best thing to watermarks. Below is an example:

StatIDexample.png

SDA realizes that those who encode their own runs and need manual timing can't show the time in the StatID, since final encodes are submitted to SDA and then timed. Until a solution is found, just make one without the time. A partial StatID is better than none.

You will need to install Avisynth 2.5.7 or later in order for the script to work. You also need to download the StatID header and place it in the same folder as your avisynth script, or make sure the paths are correct in the script.

The script is designed to work with any source file at any resolution and at any framerate. The only thing you need to change is the subtitle. The \n indicates a new line. Place the code at the end of your script.

###StatID version 1.5
mySubtitle="""Besmir ‘Zoid’ Sheqi\n\nMetroid Prime: Hard [1:15]\nPart 1"""

template=last
StatIDheader = (template.width < 640 ? ImageSource("statidheader.png").lanczos4resize(template.width,round(172.0*template.width/640.0)) : ImageSource("statidheader.png"))
FontSize = round(template.width * 0.05 + 4)
StatID = Blankclip(template,length=round(template.framerate*5))
StatID = Overlay(StatID,StatIDheader,x=(template.width-StatIDheader.width)/2,y=2)
StatID = Subtitle(StatID,mySubtitle,font="Verdana",size=FontSize,text_color=$FFFFFF,align=5,lsp=40)
#StatID = Subtitle(StatID,"Audio commentary on track 2",font="Verdana",size=round(FontSize*0.7),text_color=$E1CE8B,align=2)
StatID = StatID.ConvertToYV12()

StatID++last++StatID

Try to understand the concept behind the line "template=last" in the StatID. When creating the StatID from scratch, how do we know what resolution and framerate to use? We can't just put anything, otherwise we won't be able to join the StatID with the actual gameplay. One way is to manually input the exact values in the whole script. A better way is to make AviSynth look at what you've already set up (the gameplay), and extract the information needed. So, the "last" thing being worked on goes into a "template" and the script uses that.

Now that you understand the concept, note that you might not use "last" in "template=last". Same goes for the "last" in "StatID++last++StatID" (StatID is repeated at the end to try to fool pirates who cut off the StatID and steal the video). Take sample script #2 as an example.


Audio Commentary

If you are adding audio commentary to your run you may wish to add a note at the bottom of the StatID stating this. You will notice in the script that one line has been commented out with a single # sign. Removing it will enable the new subtitle.


Part 12: QuickTime compatibility

You wouldn't expect that something in AviSynth could cause problems with QuickTime; usually it's a result of incorrect encoding. Mp4 files store a value called 'timescale' that defines the rate at which the video should play. Sometimes AviSynth can cause the timescale to go as high as 10000000 instead of a more normal 2997 and breaks playback in QuickTime. Thankfully, there's an easy way to ensure that this won't happen. Add this piece of code at the end of your script.

changefps(last.framerate)


Sample scripts

1. DVD source, one segment, for HQ and IQ, deinterlacer: leakkerneldeint, gamma correction, no statid

Loadplugin("C:\Program Files\DGMPGDec\DGDecode.dll")

Ac3Source(MPEG2source("segment3.d2v",upConv=1),"segment3 192Kbps DELAY -66ms.ac3").DelayAudio(-0.066)
LeakKernelBob(order=1,threshold=10,sharp=true,twoway=true,map=false)
Trim(588,37648)
Lanczos4Resize(640,480)
Levels(0, 1.2, 255, 16, 235)

ConvertToYV12()

2. DVD source, two segments appended, for LQ and MQ DivX/Xvid, statid

Loadplugin("C:\Program Files\DGMPGDec\DGDecode.dll")

seg1 = Ac3Source(MPEG2source("segment1.d2v",upConv=1),"segment1 192Kbps DELAY -66ms.ac3").DelayAudio(-0.066)
seg1 = seg1.separatefields().selecteven()
seg1 = seg1.Trim(76,109763)
seg1 = seg1.lanczos4resize(320,240)

seg2 = Ac3Source(MPEG2source("segment2.d2v",upConv=1),"segment2 192Kbps DELAY -32ms.ac3").DelayAudio(-0.032)
seg2 = seg2.separatefields().selecteven()
seg2 = seg2.Trim(143,76875)
seg2 = seg2.lanczos4resize(320,240)

#StatID version 1.5
template=seg1
mySubtitle="""Besmir ‘Zoid’ Sheqi\n\nMetroid Prime: Hard [1:15]\nPart 1"""

StatIDheader = (template.width < 640 ? ImageSource("statidheader.png").lanczos4resize(template.width,round(172.0*template.width/640.0)) : ImageSource("statidheader.png"))
FontSize = round(template.width * 0.05 + 4)
StatID = Blankclip(template,length=round(template.framerate*5))
StatID = Overlay(StatID,StatIDheader,x=(template.width-StatIDheader.width)/2,y=2)
StatID = Subtitle(StatID,mySubtitle,font="Verdana",size=FontSize,text_color=$FFFFFF,align=5,lsp=40)
#StatID = Subtitle(StatID,"Audio commentary on track 2",font="Verdana",size=round(FontSize*0.7),text_color=$E1CE8B,align=2)
StatID = StatID.ConvertToYV12()

StatID++seg1++seg2++StatID
ConvertToYV12()

2a - Alternative code, same functionality. DVD source, two segments appended, for LQ and MQ DivX/Xvid, statid

Loadplugin("C:\Program Files\DGMPGDec\DGDecode.dll")

Ac3Source(MPEG2source("segment1.d2v",upConv=1),"segment1 192Kbps DELAY -66ms.ac3").DelayAudio(-0.066)
separatefields()
selecteven()
Trim(76,109763)
lanczos4resize(320,240)

seg1=last

Ac3Source(MPEG2source("segment2.d2v",upConv=1),"segment2 192Kbps DELAY -32ms.ac3").DelayAudio(-0.032)
separatefields()
selecteven()
Trim(143,76875)
lanczos4resize(320,240)

seg2=last

AlignedSplice(seg1,seg2)

#StatID version 1.5
mySubtitle="""Besmir ‘Zoid’ Sheqi\n\nMetroid Prime: Hard [1:15]\nPart 1"""

template=last
StatIDheader = (template.width < 640 ? ImageSource("statidheader.png").lanczos4resize(template.width,round(172.0*template.width/640.0)) : ImageSource("statidheader.png"))
FontSize = round(template.width * 0.05 + 4)
StatID = Blankclip(template,length=round(template.framerate*5))
StatID = Overlay(StatID,StatIDheader,x=(template.width-StatIDheader.width)/2,y=2)
StatID = Subtitle(StatID,mySubtitle,font="Verdana",size=FontSize,text_color=$FFFFFF,align=5,lsp=40)
#StatID = Subtitle(StatID,"Audio commentary on track 2",font="Verdana",size=round(FontSize*0.7),text_color=$E1CE8B,align=2)
StatID = StatID.ConvertToYV12()

StatID++last++StatID
ConvertToYV12()

3. Fraps source, split parts, one segment, for LQ and MQ, statid

avisource("seg7_part1.avi")++avisource("seg7_part2.avi")++avisource("seg7_part3.avi")
Trim(439,60938)
lanczos4resize(320,240)

ConvertToYV12()
XSharpen(35,40)  # A little sharper than default 30,40

#StatID version 1.5
mySubtitle="""Besmir ‘Zoid’ Sheqi\n\nMetroid Prime: Hard [1:15]\nPart 1"""

template=last
StatIDheader = (template.width < 640 ? ImageSource("statidheader.png").lanczos4resize(template.width,round(172.0*template.width/640.0)) : ImageSource("statidheader.png"))
FontSize = round(template.width * 0.05 + 4)
StatID = Blankclip(template,length=round(template.framerate*5))
StatID = Overlay(StatID,StatIDheader,x=(template.width-StatIDheader.width)/2,y=2)
StatID = Subtitle(StatID,mySubtitle,font="Verdana",size=FontSize,text_color=$FFFFFF,align=5,lsp=40)
#StatID = Subtitle(StatID,"Audio commentary on track 2",font="Verdana",size=round(FontSize*0.7),text_color=$E1CE8B,align=2)
StatID = StatID.ConvertToYV12()

StatID++last++StatID
ConvertToYV12()

4. Fraps source, split parts, segmented run, all appended into one video, for HQ and IQ, no statid

a = avisource("seg1_1.avi")++avisource("seg1_2.avi")++avisource("seg1_2.avi")
b = avisource("seg2_1.avi")++avisource("seg2_2.avi")
c = avisource("seg3_1.avi")++avisource("seg3_2.avi")

a = a.trim(100,20877) ++ a.trim(21000,30000)
b = b.trim(16,10988)
c = c.trim(875,13000)

AlignedSplice(a,b,c)  # Same as a++b++c

ConvertToYV12()

5a. DVD source, one segment, for HQ and IQ, deinterlacer: mvbob, brightness tweak, no statid

Loadplugin("C:\Program Files\DGMPGDec\DGDecode.dll")

Ac3Source(MPEG2source("segment1.d2v",upConv=1),"segment1 192Kbps DELAY -48ms.ac3").DelayAudio(-0.048)
mvbob()
Tweak(bright=10,cont=1.0)
Trim(600,38000)
Lanczos4Resize(640,480)

ConvertToYV12()

5b. DVD source, one segment, for MQ and LQ, brightness tweak, no statid

Loadplugin("C:\Program Files\DGMPGDec\DGDecode.dll")

Ac3Source(MPEG2source("segment1.d2v",upConv=1),"segment1 192Kbps DELAY -48ms.ac3").DelayAudio(-0.048)
separatefields()
selecteven()
Tweak(bright=10,cont=1.0)
Trim(300,19000)
Lanczos4Resize(320,240)

ConvertToYV12()

6. DVD source, one segment, for HQ and IQ, deinterlacers: mvbob and telecide, brightness correction, no statid

Loadplugin("C:\Program Files\DGMPGDec\DGDecode.dll")

Ac3Source(MPEG2source("SEG7_5155.d2v",upConv=1),"SEG7_5155_DELAY-73ms.ac3").DelayAudio(-0.073)
Tweak(bright=10,cont=1.0)

assumetff()

# MvBob is used for gameplay moments where the game runs at 60 fps. 
# Telecide is used for cutscenes running at 30 fps. Framerate is 
# doubled to match video "a". There are three cutscenes in this segment.
# "last" refers to the ac3source line above.

a=last.Trim(0,1720).mvbob()
b=last.Trim(1721,4080).telecide().changefps(a)
c=last.Trim(4081,11480).mvbob()
d=last.Trim(11481,12455).telecide().changefps(a)
e=last.Trim(12456,13485).mvbob()
f=last.Trim(13486,16680).telecide().changefps(a)
g=last.Trim(16681,17200).mvbob()

AlignedSplice(a,b,c,d,e,f,g)

Lanczos4Resize(640,480)

ConvertToYV12()

Script verification

When you think you're done

If you're reading this guide, you're probably a beginner at AviSynth scripting and should check if your scripts are done properly. Make a topic on the Tech Support forum, tell us what game you're running, the console, whether it's NTSC or PAL, the framerate of the game, and of course the AviSynth scripts.


Note: When you go through the MeGUI guide and post video samples in the forum, they may reveal mistakes in your scripts that we simply couldn't catch from the scripts alone. The goal here is to avoid the more obvious mistakes. So be prepared to make changes to your scripts later on.


When you know you're done

If you are sure that your AviSynth script is done properly, you are ready to compress to H.264 or XviD.

Personal tools