AIResearch AIResearch
Back to articles
AI

AI Generates Realistic Images Without Traditional Neural Networks

A new method using cellular automata creates and fuses images while preserving privacy, offering a simpler path to AI-generated content with fewer data risks.

AI Research
November 14, 2025
3 min read
AI Generates Realistic Images Without Traditional Neural Networks

In a world where AI image generation often relies on vast datasets and complex neural networks, a new approach offers a simpler, more transparent alternative. Researchers have developed a system that uses cellular automata—simple grid-based models inspired by biology—to generate and manipulate images without the heavy computational demands of traditional methods. This innovation matters because it could make AI tools more accessible and reduce privacy concerns, as the process doesn't require storing sensitive original data.

The key finding is that cellular automata, when combined with variational autoencoders (a type of AI model), can reconstruct and fuse images effectively. For example, the system takes an input image and uses a grid of cells, each holding information like color or class, to gradually build up a new image through iterative updates. This method successfully reconstructed images from datasets such as MNIST (handwritten digits) and CIFAR (small objects), and even fused styles from multiple sources, like combining an umbrella with a face from the celebA dataset.

Methodology involves training the system to map images into a hidden high-dimensional space using an encoder, then using a decoder to set parameters for the cellular automata. The automata start with most cells set to zero, except a central cell initialized to one, and evolve over time based on local rules—similar to how patterns grow in nature. For instance, in MNIST examples, the system used 64 iterations with an 8-dimensional state vector per cell, while for CIFAR, it used 128 iterations with a 16-dimensional vector. This process avoids the need for complex upsampling in many AI models, making it more efficient.

Results analysis shows that the approach performs well in tasks like image restoration and style fusion. As illustrated in Figure 3, the model's outputs closely match ground truth images, with the left side showing original test images and the right showing reconstructions. In style fusion, such as blending elements from different images, the system produces coherent combinations without direct access to the original data. Figure 4 demonstrates applications like repairing defaced pictures, where the algorithm calculates and fills in missing parts based on training, highlighting its robustness.

Contextually, this method has real-world implications for everyday users by enabling faster, cheaper image generation for creative projects, data augmentation in research, or privacy-sensitive applications. Unlike generative adversarial networks (GANs), which have raised concerns about misuse like deepfakes, this cellular automata-based approach is simpler and could be easier to regulate, reducing risks of disinformation or unethical use. It builds on the idea that technology should be encouraged morally, with trust restored through better detection models and legislation.

Limitations include the current lack of in-depth ethical analysis for neural cellular automata, as noted in the paper, and the potential for future abuse if refined for difficult tasks. The method is not yet explored for video generation, though it holds promise, and its performance depends on specific training setups, which may not generalize to all image types without further development.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn