{"id":47633,"date":"2025-12-20T19:00:34","date_gmt":"2025-12-20T11:00:34","guid":{"rendered":"https:\/\/www.1ai.net\/?p=47633"},"modified":"2025-12-20T19:00:34","modified_gmt":"2025-12-20T11:00:34","slug":"ai%e8%87%aa%e5%8a%a8%e5%8c%96%e8%a7%86%e9%a2%91%e5%89%aa%e8%be%91%e5%b7%a5%e5%85%b7%ef%bc%8c%e7%94%a8ai%e6%90%ad%e5%bb%ba%e7%9a%84%e5%ae%89%e8%a3%85%e4%bd%bf%e7%94%a8%e6%95%99%e7%a8%8b","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/47633.html","title":{"rendered":"AI AUTOMATED VIDEO CLIP TOOL, USING THE AI-BUILT INSTALLATION TUTORIAL"},"content":{"rendered":"<p>I wonder if you've had any of these problems when you've compiled some of the videos:<strong>There's always a lot of good footage in the long video, but they're scattered everywhere\u3002<\/strong><\/p>\n<p>It takes a little bit of time to find the progress strips, to remember the time, to cut them manually in the editing software, and it takes time and effort\u3002<\/p>\n<p>NOW THAT AI CAN READ AND READ THE VIDEO, IS IT POSSIBLE FOR IT TO AUTOMATICALLY PICK OUT THE KEY FOOTAGE FOR ME<\/p>\n<p>The logic of the whole tool is not complex and the process is roughly three steps:<\/p>\n<p><strong>UPLOAD THE VIDEO FILE, LET AI ANALYZE THE VIDEO CONTENT, FIND THE CLIPS THAT ARE WORTH CUTTING, THEN I SELECT THE PART I WANT IN THE INTERFACE, AND THEN I AUTOMATICALLY EDIT AND EXPORT IT IN THE BACK\u3002<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47636\" title=\"24b42e65g00t7ke0a01rvd000ty00gam\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/24b42e65g00t7ke0a01rvd000ty00gam.gif\" alt=\"24b42e65g00t7ke0a01rvd000ty00gam\" width=\"1078\" height=\"586\" \/><\/p>\n<p>But when it's actually done, it's in more detail than I thought, and there's a lot of grounding<strong>Here's how I built this application\u3002<\/strong><\/p>\n<p>THE BIGGEST PROBLEM IN THE IMPLEMENTATION OF THE TOOL LIES IN `AAI SEEING THE VIDEO' AND THEN ANALYSING DIFFERENT TIME SEGMENTS AND WHY, WHICH IS CRITICAL TO THE BASIC ABILITY OF THE VISUAL MODEL\u3002<\/p>\n<p>Here's what I need<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e6%99%ba%e8%b0%b1\" title=\"[View articles tagged with [Smart Spectrum]]\" target=\"_blank\" >Zhipu<\/a>The latest open-source GLM-4.6V visual model, which shows in some promotional parameters that GLM-4.6V performed well in the MMBench, MathVista and others 30+ assessment, 128k context (approximately 150 pages of document or 1 hour video), supports \"overblind\" long video understanding and multi-document analysis\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47634\" title=\"65ff734aj00t7ke0b00c2d000ulwm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/65ff734aj00t7ke0b00c2d000u000lwm.jpg\" alt=\"65ff734aj00t7ke0b00c2d000ulwm\" width=\"1080\" height=\"788\" \/><\/p>\n<p>Before entering the theme or thesis directly into the presentation video, automatically retrieve the graph, verify the quality of the picture, generate the graphics and the popular\/research content, or make me more curious, so choose it as the player\u3002<\/p>\n<p>The GMCD Plan, which was purchased at a low price in the previous period, coincided with this GM4.6V, the best partner for this tool\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47635\" title=\"cf93b508j00t7ke0b00a2d000uhzm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/cf93b508j00t7ke0b00a2d000u000hzm.jpg\" alt=\"cf93b508j00t7ke0b00a2d000uhzm\" width=\"1080\" height=\"647\" \/><\/p>\n<p>The use of GM 4.6 in Claude Code, Cursor, Cline is not gradually configured here, and interested friends can read the official document note at the textbook level: https:\/\/docs.bigmodel.cn\/cn\/coding-plan\/overview<\/p>\n<p>Now that everything's ready, we'll start talking about how to do it\u3002<\/p>\n<p>YOU CAN ENTER A DEMAND DIRECTLY AT THE BEGINNING, YOU CAN BE A VAGUE ONE, THEN YOU CAN EXPRESS THE WHOLE LOGIC OF REALIZATION, I ENTER IT DIRECTLY BY VOICE, AND EVEN IF THERE IS SOME CONFUSION, THEN GLM4 4.6 OF THE SPECTRA IS STILL ABLE TO UNDERSTAND MY CLAIM WELL, AND THEN HELP ME DESIGN THE CHART:<\/p>\n<blockquote>\n<ul>\n<li>I'd like to make a good video that can be uploaded, recognize some of the good footage, and make an automatic clip. I will provide a large-language model with multimodular video understanding API. I need you to do this for me on the basis of this identification, and then do this clip of the video clip, and then I need a front-end page, and then back-end processing the footage. Please help me design an entire structure information, and then I like to use a simplistic design, so you can use the tailwindcs CSS style as this foreground\u3002<\/li>\n<\/ul>\n<\/blockquote>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47637\" title=\"8c394390j00t7ke0b0051d000mo012sm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/8c394390j00t7ke0b0051d000mo012sm.jpg\" alt=\"8c394390j00t7ke0b0051d000mo012sm\" width=\"816\" height=\"1396\" \/><\/p>\n<p>The front end is Tailwind CSS, with a layout reference to the cut:<strong>ON THE LEFT IS THE LIST OF CLIPS, WITH THE VIDEO PREVIEW AREA IN THE MIDDLE, AND ON THE RIGHT IS THE MESSAGE OF THE FRAGMENTS ANALYSED BY AI, WHICH IS VERY SIMPLE AND VIOLENT\u3002<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47638\" title=\"5e577607j00t7ke0a000td000n09am\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/5e574607j00t7ke0a000td000n0009am.jpg\" alt=\"5e577607j00t7ke0a000td000n09am\" width=\"828\" height=\"334\" \/><\/p>\n<p><strong>Because the Code Plan that we chose to use has the ability to access tools with the GLM 4.6V model, and if I want to adjust the UI, I'll just have to cut it in\u3002<\/strong><\/p>\n<p>The picture I'm bringing in is actually this one, and then a simple one, because it's a good way to understand the layout you want\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47639\" title=\"c8954549j00t7ke0b001nd000u000imm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/c8954549j00t7ke0b001nd000u000imm.jpg\" alt=\"c8954549j00t7ke0b001nd000u000imm\" width=\"1080\" height=\"670\" \/><\/p>\n<p>For me, dark themes look more comfortable\u3002<\/p>\n<p>The backend uses Python, which is whether it uses the FFmpeg package at the bottom or not. And the two other back-end services that we need to use are the API capability for visual analysis (GLM 4.6V) and the service capacity to convert the uploaded video into a public web-based OSS address, which will simply give GLM 4.6 access to the relevant interface documents that we provide\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47640\" title=\"3ab811e2j00t7ke0b0047d000my00om\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/3ab811e2j00t7ke0b0047d000my00oom.jpg\" alt=\"3ab811e2j00t7ke0b0047d000my00om\" width=\"826\" height=\"888\" \/><\/p>\n<p>YOU CAN GO STRAIGHT TO THE NETWORK OF INTELLECTUALS AND POST THE GLM 4.6V FILES DIRECTLY TO THE DIALOGUE WINDOW, AND IT'S STILL SIMPLE AND ROUGH\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47641\" title=\"9c78151bj00t7ke0b004rd000ugbm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/9c78151bj00t7ke0b004rd000u000gbm.jpg\" alt=\"9c78151bj00t7ke0b004rd000ugbm\" width=\"1080\" height=\"587\" \/><\/p>\n<p>reference document address: https:\/\/docs.bigmodel.cn\/api-reference\/<\/p>\n<p>WE FOUND OUT IN OUR TESTING THAT BECAUSE OUR VIDEO IS LOCAL, WE NEED AN ADDRESS THAT CAN BE ACCESSED ON THE PUBLIC WEB IN ORDER TO MAKE OUR VIDEO AVAILABLE TO GLM-4.6V\u3002<\/p>\n<p>Finally:<strong>THE USER UPLOADS THE BACKEND AUTOMATICALLY TO OSS \u2192 TO GET ACCESSABLE ADDRESSES ON THE INTERNET \u2192 TO SEND GLM-4.6V ANALYSIS \u2192 TO RETURN THE RESULT OF THE RENDERING\u3002<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47642\" title=\"6b815c63j00t7ke0b0035d000me00ngm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/6b815c63j00t7ke0b0035d000me00ngm.jpg\" alt=\"6b815c63j00t7ke0b0035d000me00ngm\" width=\"806\" height=\"844\" \/><\/p>\n<p>THE OSC OBJECT OF ARIYUN WAS THEREFORE STORED AND THE VIDEO LINK RETURNED TO THE GLM-4.6V WAS THEN USED\u3002<\/p>\n<p>And as for how to connect and get the corresponding key information, it's done by asking GLM-4.6, and it's in its own training data, there's a docking information on the product, and it's all we need to do throughout the coding process\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47643\" title=\"47f71afdj00t7ke0a0021d000uaym\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/47f71afdj00t7ke0a0021d000u000aym.jpg\" alt=\"47f71afdj00t7ke0a0021d000uaym\" width=\"1080\" height=\"394\" \/><\/p>\n<p>EVEN IN THE PROCESS OF ASKING QUESTIONS, YOU CAN GIVE SOME ERRORS DIRECTLY TO THE GLM-4.6, WHICH WILL TEACH YOU VERY CAREFULLY HOW TO SET UP THE RIGHT INFORMATION\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47644\" title=\"4f7c64fj00t7ke0b0047d000m00qum\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/4af7c64fj00t7ke0b0047d000mo00qum.jpg\" alt=\"4f7c64fj00t7ke0b0047d000m00qum\" width=\"816\" height=\"966\" \/><\/p>\n<p>Throughout the process, it was not the code that really took the most time, but rather the hints, which are shared for information\u3002<\/p>\n<ul class=\"code-snippet__line-index code-snippet__js\">\n<li><\/li>\n<li><\/li>\n<li><\/li>\n<li><\/li>\n<li><\/li>\n<\/ul>\n<pre class=\"code-snippet__js\" data-lang=\"markdown\"><code>You're a top-level short video operator and a viral content planner. Your mission is to pinpoint from the video high-light footage with a \"explosive potential\" \u2014 something that can stop the audience, react strongly, and share\u3002<\/code><code><span class=\"code-snippet__section\">## \ud83d\udd25 CORE SCREENING CRITERIA: WHAT IS A \"BLOWPOINT\"<\/span><\/code><code><span class=\"code-snippet__section\">#1T# \ud83d\udca5 FIRST PRIORITY: EMOTIONAL BOMB<\/span><\/code><code>I'm not sure if I'm going to be able to do this<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** Alarmed\/unforeseen**<\/span>Unforeseen reverses, divine operations, unexpected results<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** laughter**<\/span>: funny mistakes, God's response, abstract behavior, living scenes<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** moved \/ tears**<\/span>: true love, warm moments, tears<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** Anger\/controversial**<\/span>: point of view that triggers the discussion, disproportionate operation, full slots<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">**Stunny\/shock**<\/span>Visual impact, high-energy images, technology presentations<\/code><code><span class=\"code-snippet__section\">#1T#\ud83c\udfaf SECOND PRIORITY: SPREADING HOOKS<\/span><\/code><code>Identification of content that has a viral transmission characteristic:<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** Gold sentence\/noun**<\/span>: A sentence that captures the fine expression that spreads<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** Emoticons package**<\/span>: Exaggerated expression, magic action, reaction lens<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** imitation material**<\/span>: Interesting footage that can be replayed in two<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** Topic Detonation Point**<\/span>: elements that could trigger discussion in the comment area<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** Suspended \/ hook**<\/span>: Makes you want to see the \"over and over.\"<\/code><code><span class=\"code-snippet__section\">#1T1T# \u26a1 THIRD PRIORITY: RHYTHM HIGH ENERGY<\/span><\/code><code>To identify high-density, high-paced clips:<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** High-energy early warning**<\/span>: intensive laughter\/explosion continuous output<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** beat turn**<\/span>: sudden acceleration, wind mutation, contrast<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** Climax hour**<\/span>: Top of the story, duel, answer<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** BGM POINT**<\/span>: The light of music and the image is perfect<\/code><code><span class=\"code-snippet__section\">#1T# \ud83c\udfac FOURTH PRIORITY: CONTENT PRECISION<\/span><\/code><code>Video core value enrichment:<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">**Dry goods **<\/span>: most valuable point of knowledge\/skills<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** Product highlights**<\/span>: most attractive feature\/characterization<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">** High profile**<\/span>: the decisive moment of glamour\/power<\/code><code><span class=\"code-snippet__bullet\">-<\/span>\u00a0<span class=\"code-snippet__strong\">**Story core**<\/span>: I can't read the key to the whole video<\/code><code><span class=\"code-snippet__section\">## \ud83d\udcca POPPOINT SCORING CRITERIA<\/span><\/code><code>| fractional range | explosive point level | standard |<\/code><code>|---------|---------|------|<\/code><code>| 0.9-1.0 | S | ABSOLUTE EXPLOSIVE: IRRESISTIBLE SHARING, CUT-OFF, |2<\/code><code>| 0.8-0.89 | A-CLASS | HIGH-ENERGY TIME: A VISIBLE EMOTIONAL PEAK OR MEMORY POINT<\/code><code>|0.7-0.79 |B| \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69 \u7cbe\u5f69<\/code><code>|0.6-0.69 |C | GENERAL CONTENT: OF VALUE BUT LACK OF COMMUNICATION<\/code><code>&lt; 0.6 | D | TRANSITION CONTENT: NO CUT RECOMMENDED<\/code><code><span class=\"code-snippet__strong\">**Note: Only fractions of 0.7 minutes or more are exported, but do not output below this fraction! **<\/span><\/code><code><span class=\"code-snippet__section\">## \ud83d\udcdd OUTPUT FORMAT<\/span><\/code><code><span class=\"code-snippet__code\">``json<\/span><\/code><code>{<\/code><code>\u00a0 \u00a0 \"highlights\":<\/code><code>\u00a0 \u00a0 \u00a0 \u00a0 {<\/code><code>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \"start_time\": 10.5,<\/code><code>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \"end_time\": 25.0,<\/code><code>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \"description\": \"detailed description of [explosive point type]: 1) what happened (2) why is the suitable use of the scene at the blast point (emotional point\/transmission point) (3) (seismic beginning\/climax\/end\/separate)\" i don't know,<\/code><code>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \"score\": 0.92,<\/code><code>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \"category\": \"category.\"<\/code><code>\u00a0 \u00a0 \u00a0 \u00a0 }<\/code><code>\u00a0 \u00a0 ]<\/code><code>}<\/code><code>```<\/code><code><span class=\"code-snippet__section\">##\ud83c\udff7\ufe0f CLASS LABEL (MUST CHOOSE FROM THE FOLLOWING):<\/span><\/code><code>| Category | Applicable scene |<\/code><code>|-----|---------|<\/code><code>\u266a Bang, bang, bang \u266a<\/code><code>\u266a Bang, bang, bang \u266a<\/code><code>\u266a Bang, bang, bang, bang \u266a<\/code><code>\u266a \u266a In the back of God \u266a<\/code><code>| \ud83d\udcac \ud83d\udcac | | | | |  \u53ef  \u53ef  \u53ef  \u53ef \u7cbe \u7cbe \u7cbe \u7cbe |<\/code><code>\u266a Bang, bang, bang \u266a<\/code><code>It's a high-power early warning, a high-intensity output<\/code><code>It's the most valuable core content<\/code><code>\u266a The light of the man \u266a<\/code><code>\u266a Controversy \u266a<\/code><code>It's a very powerful image<\/code><code>\u266a Bang, bang, bang, bang \u266a<\/code><code><span class=\"code-snippet__section\">## \u26a0\ufe0f IMPORTANT REQUIREMENTS<\/span><\/code><code><span class=\"code-snippet__bullet\">1.<\/span>\u00a0<span class=\"code-snippet__strong\">** I'd rather have nothing to lose**<\/span>: Select only the real blast, not the count. A 10-minute video could only have two to five real flashpoints<\/code><code><span class=\"code-snippet__bullet\">2.<\/span>\u00a0<span class=\"code-snippet__strong\">** Precision card points**<\/span>Time is stuck to the full wave of emotion\/content, do not interrupt<\/code><code><span class=\"code-snippet__bullet\">3.<\/span>\u00a0<span class=\"code-snippet__strong\">** Detailed information**<\/span>: The description needs to be clear: \"Why is this a flashpoint\" instead of a simple picture<\/code><code><span class=\"code-snippet__bullet\">4.<\/span>\u00a0<span class=\"code-snippet__strong\">** Moderate duration**<\/span>: Recommendations for short video blast spot 8-30 seconds, maximum 60 seconds<\/code><code><span class=\"code-snippet__bullet\">5.<\/span>\u00a0<span class=\"code-snippet__strong\">** To avoid flatness**<\/span>: Transition, mating, daily dialogue, etc<\/code><code>PLEASE WATCH THE VIDEO CAREFULLY AND SCREEN IT AS THE CORE CRITERION. RETURNS ONLY THE JSON FORMAT RESULT, NO OTHER TEXT\u3002<\/code><\/pre>\n<p>AT THE BEGINNING, I ASKED AI TO \"GET A GOOD CLIP,\" AND IT GAVE AN ANALYSIS THAT WAS COMPLETELY BROAD, AND WHAT WAS THE \"SCIENTIFIC\" \"RICH\" JUDGEMENT OF NO VALUE\u3002<\/p>\n<p>I then changed the term, instead of \u201cexcellent\u201d, to \u201cexplosive point\u201d, the kind of footage that people would stop and share\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47646\" title=\"55988d91j00t7ke0b0060d000ugbm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/55988d91j00t7ke0b0060d000u000gbm.jpg\" alt=\"55988d91j00t7ke0b0060d000ugbm\" width=\"1080\" height=\"587\" \/><\/p>\n<p>A number of specific criteria have been added to the hints, such as whether there is an emotional outbreak, whether there is a golden sentence, whether there is a reversal, whether it is not appropriate to start short videos, and, in particular, to emphasize that there is no need to give me points below 0.7\u3002<\/p>\n<p>WITH THIS ADJUSTMENT, AI HAS CLEARLY RELIED ON MUCH OF ITS JUDGEMENT\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47645\" title=\"cbae6ac4j00t7ke0b004ed000un9m\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/cbae6ac4j00t7ke0b004ed000u000n9m.jpg\" alt=\"cbae6ac4j00t7ke0b004ed000un9m\" width=\"1080\" height=\"837\" \/><\/p>\n<p>Not only does it tell me which clips are worth cutting, but it also gives reasons for changes in rhythm, emotional orgasms, or is the word particularly contagious, all of which make the selection process more relevant\u3002<\/p>\n<p><strong>I'm comfortable with this tool\u3002<\/strong><\/p>\n<p>A VIDEO OF SEVERAL MINUTES, UPLOADED AND ANALYSED BY AI, WILL MARK ALL THE CLIPS WORTH CUTTING IN A FEW MINUTES AND PRODUCE THE FINISHED PRODUCT IN A FEW MINUTES. IT'S MUCH MORE EFFICIENT THAN IT USED TO BE<strong>And the last clip I've edited, I expect it to be better than bad, with only high-quality output\u3002<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-47647\" title=\"19e7cf73j00t7ke0b007nd000ugbm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/12\/19e7cf73j00t7ke0b007nd000u000gbm.jpg\" alt=\"19e7cf73j00t7ke0b007nd000ugbm\" width=\"1080\" height=\"587\" \/><\/p>\n<p>IF YOU'RE DOING SOMETHING LIKE THIS, OR IF YOU WANT TO LEARN HOW TO INTEGRATE AI INTO YOUR OWN TOOLS, THIS PROJECT MIGHT GIVE YOU SOME INFORMATION\u3002<\/p>\n<p>And then if you haven't found the right programming tools and basic models, you go straight to Coding Plan and it's done. A whole set of solutions is open, and it's very expensive\u3002<\/p>\n<p>In addition, if your GLM-4.6V model analysis resources are insufficient, you can go directly to the Specialized Zone to buy this $5.9 GLM-4.6V preferential package, and 10 million tokens are used more than enough on a daily basis\u3002<\/p>\n<p>AI ' S CAPACITY IS GROWING, BUT IT REMAINS A TOOL, AND THE KEY IS TO THINK CLEARLY ABOUT ITS NEEDS BEFORE DECIDING WHICH STEP TO TAKE\u3002<\/p>\n<p>For me, the most time-consuming part of this tool is not \u201cclips\u201d, but \u201cstrips\u201d\u3002<\/p>\n<p>AI HELPED ME AUTOMATE THE MOST MECHANICAL, TIME-CONSUMING PART, AND I STILL HAVE THE FINAL DECISION\u3002<\/p>\n<p><strong>I THINK IT'S PROBABLY THE WAY I THINK I'M REALLY USING AI\u3002<\/strong><\/p>\n<p>That's all about sharing this. I'm Silver. I'll see you next time<\/p>","protected":false},"excerpt":{"rendered":"<p>I WONDER IF YOU'VE HAD SOME OF THE VIDEO MATERIAL THAT YOU'VE BEEN ABLE TO PUT TOGETHER, BUT THERE'S ALWAYS A LOT OF GOOD FOOTAGE, BUT THEY'RE SCATTERED AROUND. IT TAKES A LITTLE BIT OF TIME TO FIND THE PROGRESS STRIPS, TO REMEMBER THE TIME, TO CUT THEM MANUALLY IN THE EDITING SOFTWARE, AND IT TAKES TIME AND EFFORT. NOW THAT AI CAN READ THE VIDEO, IS IT POSSIBLE FOR IT TO AUTOMATICALLY PICK OUT THE KEY FOOTAGE FOR ME? THE LOGIC OF THE WHOLE TOOL IS NOT COMPLICATED AND THE PROCESS IS ROUGHLY THREE-STEP: UPLOAD THE VIDEO FILE, LET AI ANALYZE THE VIDEO CONTENT, FIND THE CLIPS THAT ARE WORTH CUTTING, THEN I SELECT THE PART IN THE INTERFACE, AND THEN AUTOMATICALLY EDIT AND EXPORT IT IN THE BACKSTAGE. BUT WHEN IT'S REALLY DONE, IT'S MORE DETAILED THAN I THOUGHT, AND THERE'S MORE TO IT<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[149,144],"tags":[4466,2680,2174],"collection":[],"class_list":["post-47633","post","type-post","status-publish","format-standard","hentry","category-jiaocheng","category-baike","tag-ai","tag-2680","tag-2174"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/47633","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=47633"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/47633\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=47633"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=47633"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=47633"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=47633"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}