open sora
open soraxueyoo/Open-Sora-Plan.gitee-modal {width: 500px !important; }-rn
Apache License Version 2.0, January 2004 http://www.apache.org/licenses/ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION 1. Definitions. "License" shall mean the terms and conditions for use, reproduction, and distribution as defined by Sections 1 through 9 of this document. "Licensor" shall mean the copyright owner or entity authorized by the copyright owner that is granting the License. "Legal Entity" shall mean the union of the acting entity and all other entities that control, are controlled by, or are under common control with that entity. For the purposes of this definition, "control" means (i) the power, direct or indirect, to cause the direction or management of such entity, whether by contract or otherwise, or (ii) ownership of fifty percent (50%) or more of the outstanding shares, or (iii) beneficial ownership of such entity. "You" (or "Your") shall mean an individual or Legal Entity exercising permissions granted by this License. "Source" form shall mean the preferred form for making modifications, including but not limited to software source code, documentation source, and configuration files. "Object" form shall mean any form resulting from mechanical transformation or translation of a Source form, including but not limited to compiled object code, generated documentation, and conversions to other media types. "Work" shall mean the work of authorship, whether in Source or Object form, made available under the License, as indicated by a copyright notice that is included in or attached to the work (an example is provided in the Appendix below). "Derivative Works" shall mean any work, whether in Source or Object form, that is based on (or derived from) the Work and for which the editorial revisions, annotations, elaborations, or other modifications represent, as a whole, an original work of authorship. For the purposes of this License, Derivative Works shall not include works t……open soraWhat is Sora? A new generative AI tool could transform video production and amplify disinformation risks-rn
Late last week, OpenAI announced a new generative AI system named Sora, which produces short videos from text prompts. While Sora is not yet available to the public, the high quality of the sample outputs published so far has provoked both excited and concerned reactions.The sample videos published by OpenAI, which the company says were created directly by Sora without modification, show outputs from prompts like “photorealistic closeup video of two pirate ships battling each other as they sail inside a cup of coffee” and “historical footage of California during the gold rush”. At first glance, it is often hard to tell they are generated by AI, due to the high quality of the videos, textures, dynamics of scenes, camera movements, and a good level of consistency. OpenAI chief executive Sam Altman also posted some videos to X (formerly Twitter) generated in response to user-suggested prompts, to demonstrate Sora’s capabilities.Sora combines features of text and image generating tools in what is called a “diffusion transformer model”. Transformers are a type of neural network first introduced by Google in 2017. They are best known for their use in large language models such as ChatGPT and Google Gemini.Diffusion models, on the other hand, are the foundation of many AI image generators. They work by starting with random noise and iterating towards a “clean” image that fits an input prompt.A video can be made from a sequence of such images. However, in a video, coherence and consistency between frames are essential. Sora uses the transformer architecture to handle how frames relate to one another. While transformers were initially designed to find patterns in tokens representing text, Sora instead uses tokens representing small patches of space and time.Sora is not the first text-to-video model. Earlier models include Emu by Meta, Gen-2 by Runway, Stable Video Diffusion by Stability AI, and recently Lumiere by Google. Lumiere, released just a few weeks ago, claimed to produce better video than its predecessors. But Sora appears to be more powerful than Lumiere in at least some respects.Sora can generate videos with a resolution of up open sora to 1920 × 1080 pixels, and i……
open soraTesting locally run Open Sora AI Video Generation-rn
As I mentioned in my recent post about locally run portrait animation, the thing I am most excited about in the generative AI space is the ability to generate uncensored videos locally, without subscriptions or censorship. The difference between my current self and the version of myself who wrote that article 11 days ago is that I have discovered a locally run AI video generation repository—Open Sora.When I first learned about this repo from a post on the LocalLlama subreddit, I eagerly browsed through it and was excited by the high-quality 720p video samples present in the repository. Unfortunately for me, I then realized that these samples were generated using an Nvidia H100, an 80 GB GPU with a price point comparable to a new Japanese sedan. Since my localllm machine uses two Nvidia 3090ti’s, I searched for a way to see if I could run any sort of generation on my setup and came across a page that mentioned the ability to generate a 4-second-long 240p video using a 3090. While this may not sound enticing to most, those who share my fantasy of uncensored offline video generation will appreciate my excitement at the prospect of being able to run this with my local setup.open soraI followed the GitHub instructions to install everything and had a bit of trouble, though that was related to some obscure issues regarding my use of an SSH session into the localllm machine, and not inherent to the repo itself. Once everything was installed, open sora I attempted to run the script, but for some reason, I was not able to get the command line inference script to work with only one card. Fortunately, the option to utilize multiple GPUs was available and through this, I was able to get my first generation. I chose to initially attempt to recreate the results from the backprop page, utilizing the prompts listed there for my examples.The open sora first prompt was for a bustling city transitioning from day to night, and my output was definitely impressive. As of the conclusion of my testing, I would say that I found this one to be the most impressive in terms of “realism.” Next, I tried the art museum tour, which began in a very “trippy” manner, ending in a reasonable walk through t……