Latest In


AI Build AI - Developing AI Models To Build AI

We have worked hard to create real machine intelligence. Maybe we should have let them get on with it. The majority of artificial intelligence is a numbers game.

Author:Suleman Shah
Reviewer:Han Ju
May 28, 202333.4K Shares445.4K Views
AI Build AI- Machine Learning is a subset of Artificial Intelligencein which a machine learns from previous experience, i.e. data. Unlike traditional programming, which requires the developer to foresee and write every possible scenario, a machine learning solution changes the output based on the data. A Machine Learning algorithm does not write code directly but constructs a computer model of the environment, which it then updates based on how it is educated.
We have worked hard to create machine intelligence. Maybe we should have let them get on with it. The majority of artificial intelligence is a numbers game. Deep neural networks, a type of AI that learns to recognize patterns in data, began outperforming standard algorithms 10 years ago because we finally had enough data and processing capacity to utilize them fully. Today's neural networks are even more data and power-hungry. Training them necessitates fine-tuning the values of millions, if not billions, of parameters that define these networks and indicate the strength of connections between artificial neurons. The objective is to obtain near-ideal settings for them, a process called optimization, but training the networks to get there is difficult. "Training might take days, weeks, or even months," said Petar Velickovic, a DeepMind staff research scientist in London. That might soon change.

AutoML By Google

The invention of AutoML, an artificial intelligence (AI) capable of producing its artificial intelligence, was disclosed in May 2017 by Google Brain researchers. Google scientists used a method known as reinforcement learning to automate the building of machine learning models for use in artificial intelligence. AutoML functions as a neural controller network that generates a child artificial intelligence network tailored to a given purpose. The aim for this specific kid artificial intelligence, which the researchers dubbed NASNet, was to recognise items in a video in real-time, such as people, automobiles, traffic lights, purses, backpacks, etc. According to the researchers, researchers found that NASNet was 82.7% accurate in predicting pictures on ImageNet's validation set.

Can An AI Create An AI?

Boris Knyazev, a professor at the University of Guelph, and his colleagues created and trained a "hypernetwork," a form of overlord for other neural networks and can speed up the training process. When given a new, untrained deep neural network built for a specific purpose, the hypernetwork predicts its parameters in fractions of a second, potentially eliminating the requirement for entirely training the deep neural network. This is because the hyper network has been trained to recognise highly subtle patterns in deep neural network design, and this study may also have more enormous theoretical implications. For the time being, the hypernetwork performs well in some situations, but there is still room for improvement – which is only natural given the scale of the problem. If they succeed in addressing it, "this will have a tremendous impact across the board for machine learning," according to Velikovi.

Hypernetwork GHN-2

Ren, his former University of Toronto colleague Chris Zhang, and their mentor Raquel Urtasun took a new strategy in 2018. Given a series of potential architectures, they created a graph hypernetwork (GHN) to discover the optimum deep neural network architecture to accomplish a job. When Knyazev and his colleagues discovered the graph hypernetwork concept, they recognized they could expand on it. The team demonstrates how to utilize GHNs to determine the best architecture from a set of samples and forecast the parameters for the best network. It performs well in an absolute sense in their new study. In cases where the best is insufficient, the network can be trained further via gradient descent.
Knyazev and his colleagues dubbed their hyper network GHN-2, improving on two critical characteristics of Ren and colleagues' graph hyper network. First, they relied on Ren's representation of a neural network's architecture as a graph. Each node in the graph represents a subset of neurons that do a particular computation. The graph's edges show how information moves from node to node, from input to output.
The process of training the hyper network to produce predictions for new candidate designs was the second notion they drew on. This necessitates the use of two more neural networks. The first allows computations on the original candidate graph, resulting in changes to information associated with each node, while the second takes the updated nodes as input and predicts the parameters for the candidate neural network's corresponding computational units. These two networks each have their own set of parameters that must be improved before the hyper network can forecast parameter values appropriately.
However, Velikovi raises a potentially significant issue if hypernetworks like GHN-2 become the standard way for improving neural networks. "You have a neural network — effectively a black box — anticipating the parameters of another neural network," he explained of graph hypernetworks. As a result, if it makes a mistake, you have no method of explaining it. Deep neural networks typically detect patterns in photos, text, or audio signals, all of which are very organized sources of information. GHN-2 searches for patterns in the graphs of entirely random neural network topologies. That's a lot of information. GHN-2, on the other hand, can generalize, which means it can generate good predictions of parameters for unknown and even out-of-distribution network structures. This work demonstrates that many patterns are similar in many architectures and that a model may learn how to transfer information from one architecture to another.
Jump to
Suleman Shah

Suleman Shah

Suleman Shah is a researcher and freelance writer. As a researcher, he has worked with MNS University of Agriculture, Multan (Pakistan) and Texas A & M University (USA). He regularly writes science articles and blogs for science news website and open access publishers OA Publishing London and Scientific Times. He loves to keep himself updated on scientific developments and convert these developments into everyday language to update the readers about the developments in the scientific era. His primary research focus is Plant sciences, and he contributed to this field by publishing his research in scientific journals and presenting his work at many Conferences. Shah graduated from the University of Agriculture Faisalabad (Pakistan) and started his professional carrier with Jaffer Agro Services and later with the Agriculture Department of the Government of Pakistan. His research interest compelled and attracted him to proceed with his carrier in Plant sciences research. So, he started his Ph.D. in Soil Science at MNS University of Agriculture Multan (Pakistan). Later, he started working as a visiting scholar with Texas A&M University (USA). Shah’s experience with big Open Excess publishers like Springers, Frontiers, MDPI, etc., testified to his belief in Open Access as a barrier-removing mechanism between researchers and the readers of their research. Shah believes that Open Access is revolutionizing the publication process and benefitting research in all fields.
Han Ju

Han Ju

Hello! I'm Han Ju, the heart behind World Wide Journals. My life is a unique tapestry woven from the threads of news, spirituality, and science, enriched by melodies from my guitar. Raised amidst tales of the ancient and the arcane, I developed a keen eye for the stories that truly matter. Through my work, I seek to bridge the seen with the unseen, marrying the rigor of science with the depth of spirituality. Each article at World Wide Journals is a piece of this ongoing quest, blending analysis with personal reflection. Whether exploring quantum frontiers or strumming chords under the stars, my aim is to inspire and provoke thought, inviting you into a world where every discovery is a note in the grand symphony of existence. Welcome aboard this journey of insight and exploration, where curiosity leads and music guides.
Latest Articles
Popular Articles