Researchers at Stanford University have developed an artificial intelligence (AI) model, EG3D, that can generate random, high-resolution images of faces and other objects with basic geometric structures.

by time news

Artificial intelligence models have developed recently, and users will soon be able to use these models to create and transform semi-realistic 3D scenes from their comfortable laptops. As these technologies make it easy to create ultra-realistic avatars, they will revolutionize the way artists work in video games and CGI for movies. For some time, AIs have been able to create realistic 2D images. However, 3D visualizations have proven more difficult due to the massive computing power required. It uses the EG3D model, an AI model developed by a team of academics at Stanford University, to create random, high-resolution images of faces and other objects with basic geometric structures. This model is one of the first 3D models currently used to achieve near photo-realistic display quality.

EG3D and its predecessors use a very popular machine learning method called the Generative Adversarial Network to generate graphics. Using one neural network to generate the images and the other to assess their accuracy, these systems pit two neural networks against each other. This process is repeated several times until the result is possible. Researchers have developed a component that can transform these images into 3D space by combining features of existing high-resolution 2D GANs. This two-piece building accomplishes two goals simultaneously. In addition, it is fast enough to run in real time on a laptop and can be used to create complex 3D designs. It is compatible with existing architectures and has efficient computing performance.

Although it is possible to create nearly reality-like 3D images using tools like EG3D, there is still the problem of how difficult it can be to convert them into design software. This is because although the result is a visible image, it is not clear how GANs produced it. A machine learning model called GiraffeHD developed by researchers at the University of Wisconsin-Madison could help in this situation. This model is useful for extracting manipulable features from 3D images. It allows the user to choose many elements including the shape, color, and width of the image or background. Giraffe HD is trained using countless images. The model looks for the factors inherent in the image to generate these images so that these multiple features act as controllable variables. By editing these controllable features of future 3D images, users will be able to precisely change the properties of the desired scenes. A more prominent trend is the use of artificial intelligence to create 3D images, including EG3D and Giraffe HD. However, there is still much work to be done regarding the algorithm’s bias and broad application. The nutrition training data type still restricts these models. More research is being done to address these issues.

Although still in its infancy, this research opens possibilities for more realistic 3D images and models. It will be interesting to see where this research goes and how it can be used in the future. I’d love to hear your thoughts on this new approach to our ML subreddit.

This Article is written as a summary article by Marktechpost Staff based on the paper 'Efficient Geometry-aware 3D Generative Adversarial Networks'. All Credit For This Research Goes To Researchers on This Project. Checkout the paper, github and reference article.

Please Don't Forget To Join Our ML Subreddit

You may also like

Leave a Comment