AIGC Chain Documents
  • ๐ŸŒIntroduction
    • Introduction
    • Solutions
    • Who are AIGC Chainโ€™s contributors?
    • AIGC Chain Roadmap
  • ๐ŸŽ‡diffusion models
    • Diffusion Models
    • Unique architecture Modules
    • Comparison with other Diffusion Models
    • Use case for infrastructure
  • ๐ŸงฒIncentive models
    • Why are users deciding to build here?
    • Domain knowledge contribution
    • GPU and storage contribution
  • ๐ŸงฎCapabilities
    • 2D Profile Picture (PFP, DID and NFT)
    • 2D Utilities
      • Editing Skill
      • Merging subjects for new creation
      • Combining artistic styles to create something new
      • Optimizing graphics and industrial design
  • Text to Video
  • 3D/metaverse capabilities
  • ๐Ÿช™FINANCIAL BASED SECTIONS
    • Tokenomics
    • Metanaunt NFT
  • ๐Ÿ’ฟResources
    • Social Media
    • TERMS OF SERVICE
Powered by GitBook
On this page
  • The forward diffusion process
  • The reverse diffusion process
  • Accuracy control for coherent visual generation
  1. diffusion models

Unique architecture Modules

PreviousDiffusion ModelsNextComparison with other Diffusion Models

Last updated 2 years ago

The forward diffusion process

Our raw-feature based autoencoder uses a raw-feature initialization mechanism, which has been proven to be more effective than red-green-blue (or RGB-based) diffusion methods. This raw-feature initialization allows the model to capture more detailed and nuanced information about the input data, resulting in higher-quality generation results.

The reverse diffusion process

We use a low-rank approximation denoising method to achieve dimension reduction in the latent space for the large matrix's semantic features. This low-rank approximation method significantly reduces the computation cost compared to other methods such as stable diffusion. This allows the model to be more efficient and cost-effective.

Accuracy control for coherent visual generation

For better controllability and consistency of the generated contents (especially human characters), we provide an integrated solution that includes our efforts in diffusion models and experience in past video/3D avatar/multiview stereo/segmentation projects. Besides modifying the diffusion methods, the autoencoder, and inserting new CLIP, a variety of proprietary algorithms are coded to provide extra conditioning beyond plain images and texts. This yields results that are not only visually appealing and diverse, but coherent and consistent with real-world perception.

๐ŸŽ‡