{"id":27333,"date":"2025-01-22T09:13:32","date_gmt":"2025-01-22T01:13:32","guid":{"rendered":"https:\/\/www.1ai.net\/?p=27333"},"modified":"2025-01-18T15:32:42","modified_gmt":"2025-01-18T07:32:42","slug":"%e6%89%8b%e6%8a%8a%e6%89%8b%e6%95%99%e4%bd%a0%e7%94%a8%e5%8f%af%e7%81%b5ai%ef%bc%8c%e5%8f%af%e7%81%b5ai%e4%bd%bf%e7%94%a8%e6%8c%87%e5%8d%97%e8%a7%86%e9%a2%91%e7%94%9f%e6%88%90%e8%bf%9b%e9%98%b6","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/27333.html","title":{"rendered":"Hands-on teaching you to use the Keyline AI, Keyline AI User Guide Video Generation Advanced Tutorials"},"content":{"rendered":"<p><strong>Want to become a master of AI video production? Master the<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%8f%af%e7%81%b5ai\" title=\"[SEE ARTICLES WITH [CHIN AI] LABEL]\" target=\"_blank\" >KeLing AI<\/a>s universal generation function to make your video creation more handy, this post will introduce the Kerin AI<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e4%bd%bf%e7%94%a8%e6%8c%87%e5%8d%97\" title=\"[See articles with [Use Guide] labels]\" target=\"_blank\" >Guidelines for use<\/a>Video generation advanced tutorial.<\/strong><\/p>\n<p><strong>Standard versus high quality models<\/strong><\/p>\n<p>\"Standard Mode\" is a model with faster video generation speed and lower reasoning cost, which can be used to quickly verify the effect of the model and satisfy users' creative realization needs; [High Quality Mode] is a model with richer video generation details and higher reasoning cost, which can be used to generate high-quality videos and satisfy the creators' needs for higher-order works.<\/p>\n<p>For the standard model and high quality model, respectively, there are the following advantages, we can choose the model generation according to the actual situation; the<\/p>\n<p><strong>Standard model:<\/strong>Video generation is faster and less costly to reason. It is good at generating portraits, animals, and scenes with large dynamic amplitude, and generates more intimate animals with a soft color palette, and is also a model that received good reviews when Kerin was first released;.<\/p>\n<p><strong>High quality model:<\/strong>Video generation is richer in details, higher inference cost, good at generating portraits, animals, buildings, landscapes and other videos, richer in details, more advanced composition and tonal atmosphere, is a model most used by Kerin at this stage for fine video creation.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27338\" title=\"66d2605bj00sq9w6z009xd000qg00blp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/66d2605bj00sq9w6z009xd000qg00blp.jpg\" alt=\"66d2605bj00sq9w6z009xd000qg00blp\" width=\"952\" height=\"417\" \/><\/p>\n<p><strong>Camera Control<\/strong><\/p>\n<p>The high quality modes of Kerin 1.0 and Kerin 1.5 both support camera control, which now supports 6 basic camera movements including \"Horizontal camera movement, Vertical camera movement, Advancement\/Distraction, Vertical panning, Horizontal panning\", \"Rotary panning, Horizontal panning\", \"Rotary panning\", and \"Horizontal panning\". It now supports 6 basic frames including \"horizontal frame, vertical frame, push\/pull, vertical pan, rotate frame, and horizontal frame\". Koring 1.0 model Vince video, but also additional support for \"left rotation advance, right rotation advance, advance upward, downward shift and pull farther\" 4 masters of the lens, to help creators generate video images with a clear effect of the lens.<\/p>\n<p>The lens control belongs to a kind of lens language, in order to meet the diversity of video creation, so that the model better respond to the creator's control of the lens, the platform to increase the lens control function, to control the lens behavior of the video screen with absolute commands, you can adjust the displacement parameter for the amplitude of the lens to choose, the following is a different example of the \"a big cat playing the piano by the lake\". The following is an example of the different camera movements for \"A big cat playing piano by the lake\".<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27337\" title=\"87a62220j00sq9w8k00mzd000ln00n0p\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/87a62220j00sq9w8k00mzd000ln00n0p.jpg\" alt=\"87a62220j00sq9w8k00mzd000ln00n0p\" width=\"779\" height=\"828\" \/><\/p>\n<p><strong>head-to-tail frame capability<\/strong><\/p>\n<p>First and last frame function, that is, upload two pictures, the model will use these two pictures as the first and last frames to generate a video, by clicking on \"Add Last Frame\" in the upper right corner of the \"Graphic Video\" function to experience it.<\/p>\n<p>The first and last frame function can realize a finer control of the video, at this stage is mainly used in the video creation of the first frame of the last frame of the control requirements of the video generation, can better achieve the expected dynamic transition of the generated video, but it should be noted that the first frame of the last frame of the video content needs to be as similar as possible, if the difference is large will cause the camera to switch.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27335\" title=\"0324653cj00sq9w9z00n1d000lj00lap\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/0324653cj00sq9w9z00n1d000lj00lap.jpg\" alt=\"0324653cj00sq9w9z00n1d000lj00lap\" width=\"775\" height=\"766\" \/><\/p>\n<p><strong>Some tips:<\/strong><\/p>\n<p>Try to choose two pictures with the same theme and close together, so that the model is easy to articulate smoothly within 5s, and if the two pictures are quite different, it may trigger a lens switch.<\/p>\n<p>Many creators will perform similar image selection through image generation and subsequently utilize the first and last frame capabilities for video generation, e.g.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27334\" title=\"903bd64aj00sq9wb9007ld000q8007op\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/903bd64aj00sq9wb9007ld000q8007op.jpg\" alt=\"903bd64aj00sq9wb9007ld000q8007op\" width=\"944\" height=\"276\" \/><\/p>\n<p><strong>Motion Brush Capability<\/strong><\/p>\n<p>Motion brush function, that is, upload any picture, the user can in the picture through the \"automatic selection\" or \"paint\" on a region or subject to select, add motion trajectory, at the same time, enter the expected motion Prompt (subject +), click to generate the model will generate for the user to add the specified motion after the graphic video results to control the performance of a particular subject to complement the advanced graphic video can be generated. At the same time, enter the expected motion Prompt (Subject + Motion), click Generate and the model will generate the video result for the user after adding the specified motion, so as to control the motion performance of a specific subject, complementing the advanced controllable generation of graphic video.<\/p>\n<p>The exercise brush function, which serves as a more manageable production capability for graphic videos, can further generate the desired area or subject-specific sports, such as the \"ball sport\" where the graphic video is more difficult to achieve, as well as the \"person 1 animal's turn and walk path\" generation, which supports the simultaneous setting of six subjects and tracks. In addition, this time, the \"static brush\" function is supported, after which the model will fix the pixel point of the area and avoid a mirroring event, and if it is not desirable that the motion track may cause, it is recommended to add a static brush to the bottom of the picture\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27336\" title=\"0660b486j00sq9wdh00khd000pu00j3p\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/0660b486j00sq9wdh00khd000pu00j3p.jpg\" alt=\"0660b486j00sq9wdh00khd000pu00j3p\" width=\"930\" height=\"687\" \/><\/p>\n<p><strong>Some tips<\/strong><\/p>\n<p>It is recommended that, when using the exercise brush function, the Promot description be added to the extent possible, and that the monthly Promot cat account be consistent with the area, the main movement, such as \u201cdog running on the road\u201d, and also follow the formulation of the \"subject + movement\" hint\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27342\" title=\"a6b945e1j00sq9wfk005ud000lg00e4p\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/a6b945e1j00sq9wfk005ud000lg00e4p.jpg\" alt=\"a6b945e1j00sq9wfk005ud000lg00e4p\" width=\"772\" height=\"508\" \/><\/p>\n<p>Selecting critical localizations of an object (such as an animal's head) enables more accurate motion control.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27343\" title=\"1dd77529j00sq9wgh00apd000lk00dhp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/1dd77529j00sq9wgh00apd000lk00dhp.jpg\" alt=\"1dd77529j00sq9wgh00apd000lk00dhp\" width=\"776\" height=\"485\" \/><\/p>\n<p>For objects that can't move in the physical world, if we give a motion trajectory setting, the model understands the picture and motion commands and generates a running mirror effect.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27341\" title=\"1c8a4539j00sq9whf003jd000na006yp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/1c8a4539j00sq9whf003jd000na006yp.jpg\" alt=\"1c8a4539j00sq9whf003jd000na006yp\" width=\"838\" height=\"250\" \/><\/p>\n<p>If you want to avoid a mirror effect on the model, you need to use the \"Static Brush\" function, which fixes the pixels in the area after painting.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27340\" title=\"c9552efaj00sq9wj400kod000r300bjp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/c9552efaj00sq9wj400kod000r300bjp.jpg\" alt=\"c9552efaj00sq9wj400kod000r300bjp\" width=\"975\" height=\"415\" \/><\/p>\n<p><strong>Suggestions for image selection areas.<\/strong><\/p>\n<p>Individual dynamic brushes only select single objects with consistent categories [Recommended].<\/p>\n<p>A single dynamic brush paints only one interconnected area, not multiple areas separated from each other [Recommended].<\/p>\n<p>The State Brush can select multiple areas that are not linked to each other, but it is still recommended that each individual selection be within the same category.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27339\" title=\"2f4461c8j00sq9ws30011d000gy00bxp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/2f4461c8j00sq9ws30011d000gy00bxp.jpg\" alt=\"2f4461c8j00sq9ws30011d000gy00bxp\" width=\"610\" height=\"429\" \/><\/p>\n<p><strong>Recommendations on movement trajectories<\/strong><\/p>\n<p>Both the direction and length of the trace curve come into play, and assuming that the start of the trace curve is inside the selection, the end of the trace is expected to be where the object rests at the end of the video.<\/p>\n<p>The intermediate motion process of the selected object will move strictly in accordance with the drawn trajectory.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27344\" title=\"1c553ea4j00sq9wl9009zd000h600bwp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/1c553ea4j00sq9wl9009zd000h600bwp.jpg\" alt=\"1c553ea4j00sq9wl9009zd000h600bwp\" width=\"618\" height=\"428\" \/><\/p>\n<p><strong>Lip Sync<\/strong><\/p>\n<p>The [lip-sync] function supports you to upload local dubbing\/singing files after generating character videos in Keyline A1, or generate dubbing online through the text reading function, and Keyline A1 will be able to synchronize your video character's lip-sync with the audio perfectly as if a real person is talking\/singing, and the video is instantly as if it were born!<\/p>\n<p><strong>Usage.<\/strong><\/p>\n<p>1) Using Kerin AI, generate a video that contains the full face of the character: Click on \"lip-sync\" under Preview Video.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27345\" title=\"6c3e299dj00sq9wmo004fd000bv006vp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/6c3e299dj00sq9wmo004fd000bv006vp.jpg\" alt=\"6c3e299dj00sq9wmo004fd000bv006vp\" width=\"427\" height=\"247\" \/><\/p>\n<p>2) Within the open lip-sync pop-up window, use the text to read aloud to generate a dub, or upload a local dub\/singing file.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27346\" title=\"d71cbb43j00sq9wnf003kd000qx007np\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/d71cbb43j00sq9wnf003kd000qx007np.jpg\" alt=\"d71cbb43j00sq9wnf003kd000qx007np\" width=\"969\" height=\"275\" \/><\/p>\n<p>The tones of the text reading are super realistic large model tones, and support to adjust the speaking speed in the range of 0.8-2x speed.<\/p>\n<p>3) Click the lip-sync button and wait for the video to be generated to get the result of perfect synchronization between the video character's lip-sync and the audio.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27347\" title=\"5600da35j00sq9woy001zd000bd006lp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/5600da35j00sq9woy001zd000bd006lp.jpg\" alt=\"5600da35j00sq9woy001zd000bd006lp\" width=\"409\" height=\"237\" \/><\/p>\n<p>Note: lip-syncing is a paid feature, the price is related to the length of the character video, a 5s video lip-syncing costs 5 Inspiration Value, a 10s video lip-syncing costs 10 Inspiration Value.<\/p>\n<p>If you upload the audio, or if the text is read out, the audio is more than the long library, it will also provide you with the function of cutting the length of the audio\u3002<\/p>\n<p><strong>Some tips:<\/strong><\/p>\n<p>The videos generated by Keyline 1.0 model and Keyline 1.5 model support lip-synching as long as the face conditions of the video screen are met.<\/p>\n<p>Currently, Keling AI supports lip-syncing for character-based characters (real\/3D\/2D), but not for animal-based characters.<\/p>","protected":false},"excerpt":{"rendered":"<p>Want to become a master of AI video creation? Master the universal generation function of Ke Ling AI, so that your video creation more comfortable, this article will introduce the Ke Ling AI User's Guide video generation advanced tutorial. Standard Mode and High Quality Mode \"Standard Mode\" is a model for faster video generation and lower inference cost, you can quickly verify the model effect through the standard mode to meet the user's creative realization needs;.<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[149,144],"tags":[5592,3676,5593],"collection":[],"class_list":["post-27333","post","type-post","status-publish","format-standard","hentry","category-jiaocheng","category-baike","tag-5592","tag-ai","tag-5593"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/27333","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=27333"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/27333\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=27333"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=27333"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=27333"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=27333"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}