Science topic

Neural Networks (Computer) - Science topic

Explore the latest questions and answers in Neural Networks (Computer), and find Neural Networks (Computer) experts.
Questions related to Neural Networks (Computer)
  • asked a question related to Neural Networks (Computer)
Question
3 answers
Our study goal is to develop and validate questionnaire for patient satisfaction and then using this questionnaire tool to evaluate patient satisfaction then building a model to predict patient satisfaction using artificial neural network. need help to calculate sample size for this study.
Relevant answer
Answer
Of course. What a single layer neural network can do, that regression can't, is to provide insight into what kind of associations can be classified (if they're linesrly separable, they can be) and into the fact that very few actually can be-there's a critical capacity, where the number of associations, that can be classified at all, scales linearly with the number of inputs.
  • asked a question related to Neural Networks (Computer)
Question
2 answers
In my opinion, the perceptron and other machine learning algorithms can evaluate quite complex functional dependencies of time series, if you have any ideas for further research in this vein, welcome to a private or public discussion.
The study presents a bio-inspired chaos sensor model based on the perceptron neural network for the estimation of entropy of spike train in neurodynamic systems.
Relevant answer
Answer
Dear Dr. Yazen Alawaide thanks for discussion, looks like a good plane for future research, and application to funding. If you interested in particular some joint research direction we can discuss later. Now I am filling out an application for funding in Russia, but it can be expanded and submitted to another fund based on your ideas.
  • asked a question related to Neural Networks (Computer)
Question
2 answers
how to implement neural network for 2D planar robotic manipulator to estimate joint angles for a commanded position in a circular path? and how to estimate its error for defined mathematical model and neural network model in a circular path??
Relevant answer
Answer
Christian Schmidt Thank You
  • asked a question related to Neural Networks (Computer)
Question
4 answers
What type of deep learning architectures should we prefer while working on CNN models, Standards models such as AlexNET, VGGNET, or Customised models (with user-defined layers in the neural network architecture)?
Relevant answer
Answer
Hello, Jyoti Mishra ,
The choice between standard or customized neural network models, specifically for convolutional neural networks (CNNs), depends on several factors and the specific requirements of your task. Let's explore both options:
  1. Standard Models (Pretrained Networks): Standard models, such as AlexNet, VGGNet, ResNet, and Inception, are well-established architectures that have been extensively studied and validated on large-scale datasets. These models often serve as a starting point for many computer vision tasks due to their strong performance and generalizability.
Advantages of Standard Models:
  • Proven Performance: Standard models have achieved impressive results on various benchmark datasets, making them reliable choices.
  • Transfer Learning: Pretrained models can be fine-tuned on your specific task with smaller amounts of data, saving training time and resources.
  • Community Support: Standard models have extensive documentation, pre-trained weights, and community support, facilitating easier implementation and troubleshooting.
  1. Customized Models: Customized models involve designing neural network architectures tailored to your specific problem domain. This approach provides flexibility and allows you to incorporate domain-specific knowledge or experimental ideas into the network design.
Advantages of Customized Models:
  • Task-specific Adaptation: Customized models can be designed to capture specific characteristics or constraints of your dataset, potentially leading to improved performance.
  • Model Compactness: Customized models can be more lightweight and efficient if you have constraints on computational resources or deployment scenarios.
  • Innovative Research: Customized models provide the opportunity for innovative exploration of novel architectures, activation functions, or layer connections.
When to Choose Standard Models:
  • Limited Data: If you have limited labeled data, starting with a pretrained model and fine-tuning it can be a viable option to leverage knowledge learned from larger datasets.
  • General Computer Vision Tasks: Standard models work well for common computer vision tasks such as image classification, object detection, and semantic segmentation.
When to Choose Customized Models:
  • Domain-specific Challenges: If your task has specific requirements or unique characteristics that are not well-addressed by standard models, customization can be beneficial.
  • Research or Innovation: If you are conducting research or exploring new ideas, designing custom models allows you to test novel architectures or incorporate domain-specific knowledge.
In practice, it is often beneficial to consider a hybrid approach. You can start with a standard model as a baseline and then customize it by adding or modifying specific layers to suit your needs.
Ultimately, the choice between standard or customized models should be based on factors such as available data, task requirements, computational resources, and the level of innovation or customization desired for your project.
  • asked a question related to Neural Networks (Computer)
Question
1 answer
كيف يتم نمذجة الشبكات العصبية؟
Relevant answer
Answer
Hi, what do you mean by “modeled”? If you mean how to implement, where area you want to apply?
  • asked a question related to Neural Networks (Computer)
Question
3 answers
Latest ANN Ideas
Relevant answer
Answer
Artificial neural networks (ANNs) are a form of machine learning that have been around for several decades. Over the years, researchers have proposed new ideas and improvements to ANNs to make them more powerful and efficient. Here are some of the latest ideas on ANNs:
  1. Deep learning architectures: Deep learning is a type of neural network that is capable of learning and extracting features from large datasets. These architectures consist of multiple layers of interconnected neurons, allowing them to learn complex representations of data.
  2. Attention mechanisms: Attention mechanisms are a type of neural network architecture that allows the model to focus on specific parts of the input. This is particularly useful in image recognition, natural language processing, and other applications where the input can be complex and diverse.
  3. Reinforcement learning: Reinforcement learning is a type of machine learning that involves training an agent to take actions in an environment in order to maximize a reward. ANNs can be used to learn the policy for the agent, allowing it to make better decisions over time.
  4. Transfer learning: Transfer learning is a technique that involves using a pre-trained neural network on one task and applying it to another related task. This can save time and resources in training new models and can improve performance on the target task.
  5. Bayesian neural networks: Bayesian neural networks are a type of neural network that incorporates Bayesian probability theory to make predictions. This allows the model to make probabilistic predictions and provide a measure of uncertainty in its predictions.
These new ideas and improvements on ANNs are constantly being developed and refined, with the goal of making machine learning models more powerful and effective.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
In my research, I need to compare the neural networks I have built, consisting mainly of perceptron and normalization layers, with networks from other publications that have convolution, pulling, normalization and perceptron layers in terms of computational complexity. I have the ability to calculate the number of parameters a given neural network has on a given layer, but I don't know how I should compare it.
Should I take only convolution layers as the most taxing, or sum the number of parameters from all of layers?
How should I compare neural networks that have computationally stressful convolution layers with others that do not have them, but perform feature extraction in a different way?
Relevant answer
Answer
Calculating the computational complexity of convolutional neural networks (CNNs) can be a challenging task, but it is crucial for comparing different network architectures. There are several approaches that can be taken to calculate the complexity of CNNs, each with its own advantages and limitations.
One common approach is to count the number of floating-point operations (FLOPs) required to process a single input image. This approach takes into account the number of operations required for each layer, including convolutions, pooling, normalization, and fully connected layers. The FLOPs can be calculated using the number of parameters in each layer and the size of the input feature maps.
Another approach is to count the total number of parameters in the network. This approach is simpler than counting FLOPs, but it may not accurately reflect the computational demands of the network. For example, a network with many small convolutional filters may have a large number of parameters, but may be more computationally efficient than a network with fewer, larger filters.
When comparing networks with different architectures, it is important to consider the specific task the network is designed for. For example, if two networks have similar accuracy on a given task, but one uses convolutional layers while the other uses a different feature extraction method, it may be more appropriate to compare their computational efficiency rather than their total complexity.
In summary, when calculating the complexity of CNNs, it is important to consider both the number of parameters and the computational demands of each layer, and to take into account the specific task the network is designed for.
  • asked a question related to Neural Networks (Computer)
Question
5 answers
I noticed that in some very bad models of neural networks, the value of R² (coefficient of determination) can be negative. That is, the model is so bad that the mean of the data is better than the model.
In linear regression models, the multiple correlation coefficient (R) can be calculated using the root of R². However, this is not possible for a model of neural networks that presents a negative R². In that case, is R mathematically undefined?
I tried calculating the correlation y and y_pred (Pearson), but it is mathematically undefined (division by zero). I am attaching the values.
Obs.: The question is about artificial neural networks.
Relevant answer
Answer
Raid, apologies here's the attachment. David Booth
  • asked a question related to Neural Networks (Computer)
Question
6 answers
I am building a Neural Network Model for approximating SINR value at a specified location in 2D. The scenario consists of a central gNB placed at (1500, 1500), and other 6 gNBs are placed surrounding it in a hexagonal manner with ISD equal to 1000 m. The UEs are placed around the central gNB at distances from 30m to 500m in steps of 20m, along X and Y. I thus have 2500 points at which I am getting SINR.
For training, X and Y coordinates of UE and their values obtained from NetSim simulation have been used. Thus, the features are coordinates (X, Y), while the Label is SINR.
The NN model architectures I have tried using (i) 2-5 layers (ii) Activation functions: ReLU, LeakyReLU, tanh, sigmoid
(iii) Different units size: 32 to 128, (iv) Dropout at various layers from 0.4 to 0.5,
However, the minimum MSE obtained was around 37 while training, which is very high.
Relevant answer
Answer
Maitri Saraf You might attempt the following to lower the MSE and enhance the R2 measure:
1. Use additional data: You might try gathering more data points or training the model with more data. This may aid the model in better learning the link between the characteristics and the target variable.
2. Feature engineering: You might try to create new features that are more relevant to your problem by combining current ones. This may aid the model in better learning the link between the characteristics and the target variable.
3. Experiment with alternative model architectures: You can use a Convolutional Neural Network (CNN), a Recurrent Neural Network (RNN), or a combination of both.
4. Fine-tune the hyperparameters: You may experiment with changing the learning rate, number of hidden layers, and number of neurons in each layer.
5. Regularization: To avoid overfitting, try using regularization techniques such as L1 or L2 regularization.
6. Cross-validation: To acquire a better estimate of how well your model will perform on unseen data, consider employing cross-validation approaches.
7. Early stopping: To minimize overfitting and enhance model performance, attempt early halting.
8. Ensemble methods: To boost overall performance, you might try mixing numerous models.
It is crucial to remember that finding the ideal model and hyperparameters is an iterative process, and you may need to test several combinations before you discover the one that works best for you.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
I have trained a deep neural network model. Now I want to use information from this model to predict. One way is to load the model and then predict. But is there any way we predict without loading or training the model?
I want to write a simple python code in which I manually give data from the trained model and do predictions without using any library. But I don't know how and what type of information should I extract from the model?
Relevant answer
Answer
Muhammad Adnan Anwar I used to do this easily in MATLAB. The task I was performing is called cross-modal transfer learning. In this case, I'm training a model in a specific source domain. I take the parameters of the trained model (weight, bias, activation, etc.) and use them again to transfer additional knowledge to a new model in a target domain. If you are interested, I will post a code soon on how to access these parameters BUT in MATLAB, just follow my researchgate and mathworks profiles for any new updates.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
I carried out few experiments by combining 7 CNN based feature extractors and 7 machine learning classifier. In total there were 49 pairs.
For CNNs i used VGG16, Mobile net v2, Densenet121, Inception V3, ResNet 101, ResNet 152 and XceptionNet as feature extractors and passed the generated feature vector to ML classifiers to perform binary classification.
The ML classifiers i have used are support vector machine, k nearest neighbour, gaussian naive Bayes, decision tree, random forest, extra tress and Multi layer perceptron.
For all the evaluation metrics like accuracy, precision, recall and F1 score i achieved best results with the combination of ResNet 101 and Multi layer perceptron.
I'm not able to understand that why it is performing the best. Resnet152 has a deeper network and support vector machine generally perform well. In my case Resnet101 and multi layer perceptron is giving the best results.
Please help me to understand the reason behind it.
Relevant answer
Answer
In my opinion, if you will use pre-trained CNN based on ImageNet then better results can give Inception V3 or Xception.
I agree with Akhil Kumar that it will depend on a dataset.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
I have come across many research articles stating the values of rmse of deep learning models such as LSTM. Some results are between 0 and 1, while some are between 0 and 50. I want to know what an ideal rmse value is for such neural network models.
I have been trying to reduce the rmse values to less than 1 from values between 3 and 7.
Note: The error values are computed after unscalling back the dataset to its original form. The errors are, however, between 0 and 1 when computed with scaled data. But, I think computing with unscaled data makes more sense.
Relevant answer
Answer
You're welcome Hayatullahi Adeyemo and thanks for recommendation!
  • asked a question related to Neural Networks (Computer)
Question
3 answers
Generalized Regression Neural Network Model is available in nntool in Matlab? if not, then how can I use it?
Relevant answer
  • asked a question related to Neural Networks (Computer)
Question
5 answers
While training a neural network model for multiclass datasets using Keras, earth-analytics, and TensorFlow, variables "accuracy" and “loss” per epoch were carried out.
Now the questions are:
  1. How to interpret these variables?
  2. How do they affect the behavior of the trained model?
  3. How to define the appropriate epochs number for a better modeling performance?
  • asked a question related to Neural Networks (Computer)
Question
18 answers
Please elaborate with example
Relevant answer
Answer
Transfer learning models focus on storing knowledge gained while solving one problem and applying it to a different but related problem. Instead of training a neural network from scratch, many pre-trained models can serve as the starting point for training.
Also, you can check this link:
  • asked a question related to Neural Networks (Computer)
Question
4 answers
Following is the error faced while training the deep learning CNN model on the dataset. The dataset is of MRI images of prostate cancer in men and have dcm extension. I tried training the model with the dcm extension as well as by converting it to jpg but the error remains the same. Could anyone help me solving this error?
Code for reference:
import pydicom as dicom
PNG = False
folder_path = "/tmp/Prostate_dataset/Prostate_dataset-1"
jpg_folder_path = "/tmp/prostate_jpg"
images_path = os.listdir(folder_path)
for n, image in enumerate(images_path):
ds = dicom.dcmread(os.path.join(folder_path, image))
pixel_array_numpy = ds.pixel_array
if PNG == False:
image = image.replace('.dcm', '.jpg')
else:
image = image.replace('.dcm', '.png')
cv2.imwrite(os.path.join(jpg_folder_path, image), pixel_array_numpy)
if n % 50 == 0:
print('{} image converted'.format(n))
def get_data(path):
data = []
for img in os.listdir(path):
img_arr=os.path.join(path,img)
data_img=cv2.imread(img_arr)
data.append(data_img)
return np.array(data)
train = get_data(jpg_folder_path)
val = get_data(jpg_folder_path)
# Normalize the data
train = np.asarray(train)
train = train.reshape(-1, 256,256,1)
train = train.astype('float64')
train /= 255.0
val = np.asarray(val)
val = val.reshape(-1, 256,256,1)
val = val.astype('float64')
val /= 255.0
model = tf.keras.models.Sequential([
tf.keras.layers.Conv2D(128, (5,5), activation='relu', input_shape=(256,256,1)),
tf.keras.layers.BatchNormalization(axis=-1),
tf.keras.layers.Dropout(.5, noise_shape=None, seed=None),
tf.keras.layers.MaxPooling2D(2, 2),
tf.keras.layers.Conv2D(64, (5,5), activation='relu'),
tf.keras.layers.BatchNormalization(axis=1),
tf.keras.layers.Dropout(.5, noise_shape=None, seed=None),
tf.keras.layers.MaxPooling2D(2,2),
tf.keras.layers.Conv2D(16, (5,5), activation='relu'),
tf.keras.layers.BatchNormalization(axis=1),
tf.keras.layers.Dropout(.5, noise_shape=None, seed=None),
tf.keras.layers.MaxPooling2D(2,2),
tf.keras.layers.Conv2D(8, (5,5), activation='relu'),
tf.keras.layers.BatchNormalization(axis=1),
tf.keras.layers.Dropout(.5, noise_shape=None, seed=None),
tf.keras.layers.MaxPooling2D(2,2),
tf.keras.layers.Flatten(),
tf.keras.layers.Dense(128, activation='relu'),
tf.keras.layers.Dense(256, activation='softmax'),
])
model.summary()
model.compile(optimizer='adam',loss='categorical_crossentropy',metrics=['accuracy'])
model.fit(X_train, y_train, epochs = 10, verbose = 1, validation_data = (X_test, y_test))
test_loss = model.evaluate(X_test, y_test)
Error:
ValueError: in user code: /usr/local/lib/python3.7/dist-packages/keras/engine/training.py:853 train_function * return step_function(self, iterator) /usr/local/lib/python3.7/dist-packages/keras/engine/training.py:842 step_function ** outputs = model.distribute_strategy.run(run_step, args=(data,)) /usr/local/lib/python3.7/dist-packages/tensorflow/python/distribute/distribute_lib.py:1286 run return self._extended.call_for_each_replica(fn, args=args, kwargs=kwargs) /usr/local/lib/python3.7/dist-packages/tensorflow/python/distribute/distribute_lib.py:2849 call_for_each_replica return self._call_for_each_replica(fn, args, kwargs) /usr/local/lib/python3.7/dist-packages/tensorflow/python/distribute/distribute_lib.py:3632 _call_for_each_replica return fn(*args, **kwargs) /usr/local/lib/python3.7/dist-packages/keras/engine/training.py:835 run_step ** outputs = model.train_step(data) /usr/local/lib/python3.7/dist-packages/keras/engine/training.py:789 train_step y, y_pred, sample_weight, regularization_losses=self.losses) /usr/local/lib/python3.7/dist-packages/keras/engine/compile_utils.py:201 __call__ loss_value = loss_obj(y_t, y_p, sample_weight=sw) /usr/local/lib/python3.7/dist-packages/keras/losses.py:141 __call__ losses = call_fn(y_true, y_pred) /usr/local/lib/python3.7/dist-packages/keras/losses.py:245 call ** return ag_fn(y_true, y_pred, **self._fn_kwargs) /usr/local/lib/python3.7/dist-packages/tensorflow/python/util/dispatch.py:206 wrapper return target(*args, **kwargs) /usr/local/lib/python3.7/dist-packages/keras/losses.py:1666 categorical_crossentropy y_true, y_pred, from_logits=from_logits, axis=axis) /usr/local/lib/python3.7/dist-packages/tensorflow/python/util/dispatch.py:206 wrapper return target(*args, **kwargs) /usr/local/lib/python3.7/dist-packages/keras/backend.py:4839 categorical_crossentropy target.shape.assert_is_compatible_with(output.shape) /usr/local/lib/python3.7/dist-packages/tensorflow/python/framework/tensor_shape.py:1161 assert_is_compatible_with raise ValueError("Shapes %s and %s are incompatible" % (self, other))
ValueError: Shapes (None, 256, 256, 1) and (None, 256) are incompatible
Relevant answer
Answer
I think there might be a problem in python version you are using, try to use python 2 instead of python 3
In addition, have a look here also:
  • asked a question related to Neural Networks (Computer)
Question
4 answers
One of famous method in Machine Learning is Perceptron Rule. In following Book and it's R Perceptron function some of problems are not working correctly. I develop R code that worked with several examples and exercises.
# R Deep Learning Project, Y. Liu, P. Maldonado
One of simple problem is showing in this picture that separate points in two groups with correctly decision boundary line.
Relevant answer
Answer
I publish my Neural Network new approach codes in githup, as mirkhan64, in reviews. Thanks
  • asked a question related to Neural Networks (Computer)
Question
3 answers
While modeling Lynx series after applying log10 transformation, I get ARIMA(2,0,3) as the best fitted model using auto.arima in R. But I see Zhang 2003 (Time series forecasting using a hybrid ARIMA and neural network model, G. Peter Zhang) specifies, after applying log10 transformation, ARIMA(12,0,0) as the best fitted model using MATLAB. This has also been followed up by many researchers Khandelwal et al. 2015, Adhikari R and Agrawal R.K 2011, etc. Can some one clarify if R and MATLAB can provide different models as the best fitted model(s)?
Relevant answer
Answer
You can have different best models using different softwares. The estimation procedures used in the software could be different leading to this difference.
  • asked a question related to Neural Networks (Computer)
Question
15 answers
r2 for MLR = 10-20 % approx at various train:test ratio
but for ANN it is 1-4%.
why?
thanks in advance
Relevant answer
Answer
What is the size of your data? Is it big data or a small one? MLR is a parametric model which assumes certain criteria. Neural needs big data as it uses multiple and many parameters. That may be one reason for this difference. Pls read more about parametric vs non-parametric models
  • asked a question related to Neural Networks (Computer)
Question
3 answers
I am using nonlinear autoregressive neural network model (NAR-NN) to predict future values from historical record. I made the model using the ntstool in MATLAB but I couldn't apply & implement the model to predict future values. So, how to implement the model for future data prediction?
  • asked a question related to Neural Networks (Computer)
Question
4 answers
What should consider to compare performance of two neural network models? What are the parameters required to increase model accuracy?
Relevant answer
Answer
Different architectures have different footprints of computation and memory. The computation can be measured roughly by how long the achitecture takes to process and the memory is proportional to the number of parameters the model possesses. A fair comparison would fix these two.
To increase the accuracy, one might enhance the efficiency of the architecture so that with a given computation and memory budget, the model performs better.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
I want to use the General regression neural network(GRNN) model in r-software to simulate the reactions of aboveground biomasses of perennial trees to climate change factors (rainfall, minimum and maximum temperature). However, I felt compelled to utilize the program to install the GRNN package, preparing the dataset, and execute the model.
Is there anyone who can help me with how to install the GRNN package and executing the R model?
Relevant answer
  • asked a question related to Neural Networks (Computer)
Question
4 answers
Hi, Is there any way to include the tensorflow lite on the Teensy board? I have been searching online but it seems that CMSIS NN is the only way to run neural network model on Teensyduino. I have tried to include the Tensorflow lite library but it won't run. Thanks,
Relevant answer
Answer
What is tensorflow?
  • asked a question related to Neural Networks (Computer)
Question
6 answers
As I have observed, if we increase window size then it increases accuracy which in turn slowdowns the training. Contrarily, if we decrease input window size then accuracy decreases, and the network gets trained faster. Need suggestions for obtaining the best results from LSTM network with optimal window size.
Relevant answer
Answer
Thanks Shahboz!
You're right! Window size depends on multiple factors including model type and behavior of signal considered for experimentation. I got it resolved through considering sampling frequency of the signal and model complexity. Now, It works fine on my project.
  • asked a question related to Neural Networks (Computer)
Question
8 answers
Hi All,
I have recently started working on neural networks for hardware implementation where we normally work with fully connected networks for pattern classification. I am facing this issue as to how to report the pattern classification accuracy. There are 2 situations that mainly arises in this context:
1) How to report the training accuracy vs epoch data? whether this should be some sort of mean data over several runs for every epoch or the best performing data?
2) How to report the test accuracy for the trained model?
NB: There has been some discussions on stack exchange on this (https://stats.stackexchange.com/questions/205610/reporting-of-neural-network-accuracy-for-academic-publications) but its not very conclusive
Relevant answer
Dear, Kannan u.m!
I can develop a professional ANN (FANN) model for your task in python using KARAS and TensorFlow. Also, I can trained you this is skills. My contacts: shahboztjk@mail.ru or Whatsapp(+79322301628).
  • asked a question related to Neural Networks (Computer)
Question
4 answers
hi dears!
i have 4 inputs and one output neural network model. i want to extract the optimum input from the four inputs. some one who has MATLAB code, please share me.
thank you!
Relevant answer
Answer
I would suggest you to perform ANOVA test and check the effect of your main parameters on Output then decide about the considering them as input data of your network. for example, if the P-value of your first input was not less than 0.05, your input node will be reduced to three.
Good luck
  • asked a question related to Neural Networks (Computer)
Question
6 answers
Dear researchers
Objective:
We've applied machine learning methods such as artificial neural networks, random forest, and support vector machines to predict stroke patient's recovery.
Materials and methods:
We have stroke patients' clinical data from EMRs(electronic medical records) and their kinematic data obtained by the exoskeleton robot's sensor system(from gait training).
The clinical data are ordinal and categorical, and the kinematic data are time-series data.
Clinical data and kinematic data have been integrated into tabular data by applying moving windows to time-series data (obtained mean, std, median, max, and min).
Limitations:
In our experience, it was not easy to use all the data for training at once because the types and characteristics of clinical data and kinematic data were different.
Thus, we are applying the ensembling method to various neural network models.
(We've tried conventional bagging or stacking algorithms to the outputs of the neural networks.)
Question:
At this point, we would like to know some reasonable, preferred, recommended methods for ensembling the neural network models with different data learned separately. (i.e., how to combine a neural network model trained by clinical data and another model trained by kinematic data)
Relevant answer
Answer
  • asked a question related to Neural Networks (Computer)
Question
4 answers
The WuDao 2.0 model of China Neural Network (NN) with 1,75 trillion parameters topped the 1.6 trillion that Google unveiled in a similar model in January 2021. Is it the start of Race to Quadrillion parameters NN? Do you have additional information about the structure and design of such ultra big NN?
Relevant answer
Answer
This reminds me of the Japanese Fifth Generation program. Working bigger and harder with the same methods will probably not lead to major leaps.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
In order to make a comparison between experimental results and results obtained by machine learning models or regressions . I'd like to know how is it possible to predict results of (just ONE output ) according to the FOUR inputs illustrated in the attached picture using NEURAL NETWORKS using PYTHON where all data are numbers? and which is the best NEURAL NETWORK MODEL that gives the best prediction in this case ?
Relevant answer
Answer
As you are dealing with an apparently simple problem using real numbers,
artificial neural networks (ANN) may help you to achieve good performances (it depends all on your training data size).
You may want to start with a two-layer ANN (input, hidden, output):
- as for the input layer, the number of neurons is related to the number of features (let's say N) you have, leading to N + 1 input neurons (N features + 1 bias term)
- as for the hidden layer/layers, you can start with a single hidden layer, tuning the number of neurons. Then, add another hidden layer, tuning the number of neurons once again
- as for the output layer, in your case it includes only 1 neuron.
I recommend normalizing the data before feeding the network (you may use z-score transform - subtract the mean value and divide by the standard deviation)
I suggest you take a look at this link, as the problem is well stated and explained in the details, including python code.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
The best simulator that can aid the design of a Bayesian Neural Network model.
Relevant answer
Answer
hi, you can find Bayesian machine learning models in python library scikit-learn https://scikit-learn.org/stable/modules/naive_bayes.html
  • asked a question related to Neural Networks (Computer)
Question
5 answers
According to the nature of temporal data, strategies like k-fold cross validation is not an appropriate idea since we cannot remove the dimension of time. In this discussion we want to explore ideas about testing models for temporal data.
Relevant answer
Answer
Pooia Lalbakhsh you may want to employ deep learning models such as LSTM and GRU.
Good luck
  • asked a question related to Neural Networks (Computer)
Question
4 answers
I'm learning how to use the Recurrent Neural Network model (RNN). I'm not entirely sure about the feed-forward procedure in RNN. It includes, for example, input, hidden state, and output. As far as I know, the hidden state is a type of multi-layer perception (MLP). However, in this case, a hidden state is derived from both current input and a previously hidden state. Unfortunately, "we can note that everyone reported the total number of memory cells, but no one specified the number of neurons inside each memory cell," which confuse me.
Second, I am confused by the RNN backpropagation procedure. I searched Google, and everyone only mentioned the generic steps (calculations of gradients) but no one conducted step-by-step backpropagation on example. I'm desperate for the entire RNN training process (two iterations is sufficient), even on a single layer with three to four memory cells.
Can anyone available concrete examples of the RNN model training process?
Relevant answer
Answer
Backpropagation Through Time, or BPTT, is the application of the Backpropagationtraining algorithm to recurrent neural network applied to sequence data like a time series. A recurrent neural network is shown one input each timestep and predicts one output. Conceptually, BPTT works by unrolling all input timesteps.
Regards,
Shafagat
  • asked a question related to Neural Networks (Computer)
Question
12 answers
I am working on a novel image classification problem with only 128 images of one class and 25 images of another class. From CNN perspective, I have too little data to learn features well and on top of that, the dataset is highly imbalanced. What strategies should I adopt to improve the model in this scenario? Could other techniques could be tried (SVM ?)
  • asked a question related to Neural Networks (Computer)
Question
4 answers
Why in simulink neural network model reference control example the inputs to the controller consist of two delayed reference inputs, two delayed plant outputs, and one delayed controller output? Also why the plant has 2 inputs?
Relevant answer
Answer
It is not mentioned there but per my experience we give error and delta(error) to the controller. I assume that in this example but still I don’t see the point of giving two delayed reference inputs.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
In RNN, Yt is the output which is given by the equation;
Yt = Why.ht
where Why is the weight associated with the output layer and ht is the current state.
What is the need of having this weight at the output layer of an RNN? Is it anything related to the memory of the network?
  • asked a question related to Neural Networks (Computer)
Question
38 answers
Why most researchers are shifting from tensorFlow to Pytorch?
Relevant answer
Answer
Tensorflow creates static graphs, PyTorch creates dynamic graphs.
In Tensorflow, you have to define the entire computational graph of the model and then run your ML model. In PyTorch, you can define/manipulate/adapt your graph as you work. This is particularly helpful while using variable length inputs in RNNs.
Tensorflow has a steep learning curve. Building ML models in PyTorch feels more intuitive. PyTorch is a relatively new framework as compared to Tensorflow. So, in terms of resources, you will find much more content about Tensorflow than PyTorch. This I think will change soon.
Tensorflow is currently better for production models and scalability. It was built to be production ready. PyTorch is easier to learn and work with and, is better for some projects and building rapid prototypes.
  • asked a question related to Neural Networks (Computer)
Question
6 answers
Hi, I am working on an Industrial Process "Tennessee Eastman Process" and I need to train Neural Network model for its reactor part using BAT optimisation algorithm in MATLAB/Simulink.
MATLAB neural network app doesn't provide the option of selecting BAT learning algorithm while training the model.
Can anyone guide me or help me in finding the code that can show how to train ANN model using BAT algorithm?
I have attached the relevant paper which uses BAT algorithm to train ANN model.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
Performance forecasting...
I want to start the training of my ANN with 13 input data, three layers. my problem is, does my actual output has to be 13 input data as well? or can it be any number too.
How do i bring down the different unit of measurements? i have data with units of measurements in; hours, days, percentage, tons, TEU, metres, and monetary units. to make it simple, data representation.?
Relevant answer
Answer
I agree with Dr Radha Mohan Pattanayak. The output of your model depends on your regression task/problem. So, in case, you want to apply the Multi-Output Regression Model, I suggest to read this tutorial "How to Develop Multi-Output Regression Models with Python". https://machinelearningmastery.com/multi-output-regression-models-with-python/.
For the data representation, you can apply "How to Grid Search Data Preparation Techniques" https://machinelearningmastery.com/category/data-preparation/
  • asked a question related to Neural Networks (Computer)
Question
3 answers
I'm a hydrologist and pretty new to machine learning, but would like to use sensor data that I have (rainfall time series for example) and combine it with GIS data (grids/rasters of topography for example) as the input to a neural network to then produce a variable of interest (streamflow for example). I can easily take a 1d-array of daily rainfall and match it up with a 1d-array of stream flow and set up and train a regression perceptron (vanilla) network in Keras, but I am having a hard time wrapping my head around how it would be best to combine the single timeseries values with static but spatially distributed topography data to create a single input array of input/training data to then ultimately try to predict the 1d-array of streamflow data.
Any assistance regarding how to format these data sets would be appreciated. Thanks!
  • asked a question related to Neural Networks (Computer)
Question
7 answers
I'm working on a Project to improve the vehicle Position accuracy based on vehicle data. Basically I'm trying to make an Algorithm that can estimate the vehicle position with neural networks where my input data are the vehicle data like velocity, acceleration, braking pressure etc.. and my outputs are latitude and longitude. I''m doing a research about this but I didn't find any good papers about this topic. there is some good related papers but they are about using Kalman Filters or INS data but my Goal is to achieve this with using vehicle data from sensors like I said, for example velocity, acceleration, Braking, steer wheel, current Gear etc.. is there some good refences about this? I hope someone here have Experience about this and can help me with some advices
Relevant answer
Answer
Thanks to all of you. Please Notice that I'm using neural network and not Kalman Filter or such other Techniques
  • asked a question related to Neural Networks (Computer)
Question
4 answers
If a deep neural network model (Net A) produces higher accuracy than another deep net model (Net B) for a given dataset, will Net A always produce better accuracy than Net B for any dataset given that all other parameters are constant?
Relevant answer
Answer
In some cases yes but not always necessarily true. Neural network will try to learn latent (or hidden) features in the data. The latent space is different between datasets. Usually bigger network has greater capacity for learning but be aware of overfitting, if a very big network is used in a small dataset it might overfit by memorizing the data instead of learning the hidden features.
There is a body of work related to network capacity for learning
  • asked a question related to Neural Networks (Computer)
Question
4 answers
I have a multivariate time-series forecasting problem which I am trying to solve using LSTM models. I have transformed my data such that it looks back last 20 timestamps to reconstruct the records at current timestamp, using deep LSTM model. Following is the model architecture :
model3 = Sequential()
model3.add(LSTM(1000,input_shape=(train_X.shape[1], train_X.shape[2]), return_sequences = True))
model3.add(Dropout(0.2))
model3.add(LSTM(500, activation = 'relu', return_sequences = True))
model3.add(Dropout(0.2))
model3.add(LSTM(250, activation = 'relu', return_sequences = False))
model3.add(Dropout(0.2))
model3.add(Dense(train_y.shape[1]))
model3.compile(loss='mse', optimizer='adam', metrics = ['accuracy'])
model3.summary()
history3 = model3.fit(train_X, train_y, epochs=100, batch_size=36, validation_data=(test_X, test_y), verbose = 2, shuffle= False)
Attached is the graph of the neural network output. The 'validation loss' metrics from the test data has been oscillating a lot after epochs but not really decreasing.
Can anyone explain how to interpret the graph ? Also, what could be the potential ways to ensure that the model does get improved with every new epoch ?
Relevant answer
Answer
Actually the graph doesn't tell us the entire story, It looks like the validation loss is oscillating a lot! But in truth it appears that way b/c you y-axis is scaled from 0 to 0.12, which is a very small margin. Make this scale bigger and then you will see the validation loss is stuck at somewhere at 0.05. Of course these mild oscillations will naturally occur (that's a different discussion point).
In your case, it seems the model is overfitting a little, but just the loss function is not enough. Check the difference in training and validation accuracy and then it will be clearer.
To overcome overfitting, there are many things you can do
(i) Add more regularization
(ii) Data augmentation
(iii) Increase more data samples naturally
(iv) Reduce hidden layers
etc.
  • asked a question related to Neural Networks (Computer)
Question
5 answers
Cross-validation, it’s a model validation techniques for assessing how the results of a statistical analysis (model) will generalize to an independent data set
Relevant answer
Answer
Before of course you split entire data set in 75% for training 25% for evaluation both randomly , train the net on the 75% and evaluate on the rest of 25% . IMPORANT AFTER DID THIS DELETE NEWORK MODEL. Repeat all 10 times
  • asked a question related to Neural Networks (Computer)
Question
5 answers
If any such a related research article or text books are available kindly list it..
Relevant answer
Answer
Firstly, it is inefficient to update the hidden states of nodes iteratively for the fixed point. If relaxing the assumption of the fixed point, we can design a multi-layer GNN to get a stable representation of node and its neighborhood. Secondly, GNN uses the same parameters in the iteration while most popular neural networks use different parameters in different layers, which serve as a hierarchical feature extraction method. Moreover, the update of node hidden states is a sequential process which can benefit from the RNN kernel like GRU and LSTM. Thirdly, there are also some informative features on the edges which cannot be effectively modeled in the original GNN. For example, the edges in the knowledge graph have the type of relations and the message propagation through different edges should be different according to their types. Besides, how to learn the hidden states of edges is also an important problem. Lastly, it is unsuitable to use the fixed points if we focus on the representation of nodes instead of graphs because the distribution of representation in the fixed point will be much smooth in value and less informative for distinguishing each node
  • asked a question related to Neural Networks (Computer)
Question
4 answers
I am training a neural network model in R but the solutions are not converging during training. so as a result i am not getting the desired output during prediction (r2 coming below 1). Can anyone tell me how can I get a better training model? and if anyone can provide me a modified code of neural net in R?
Relevant answer
Answer
Ankush Rai yeah normalization of the data was done.
  • asked a question related to Neural Networks (Computer)
Question
13 answers
Hai,
I have dataset with 100s of independent features (column wise) and one column of dependent (categorical like good or bad). I would like to develop a neural network model which train on this dataset.
in this below mentioned matlab code, net = feedforwardnet(10); net = train(net,x,t); view(net) y = net(x); perf = perform(net,y,t)
x = (6500x100)double
t = (6500x1)categorical (good/bad)
Neural network training is not working. why? and how can I fix this problem?
Relevant answer
Answer
It seems that you have too much independent variables. The suggested relation the number of datasets (6500) to the number of connections between neurons should be around 10. Having 100 inputs and 10 neurons in one hidden layer you have 1001 connections. Probably your network is more complicated (and the relation datasets/connections less favourable).
You can:
1. Calculate mutual linear correlations between inputs and exclude from inputs these who are sronly correlated (one from tha pair).
or
2. Calculate linear correlations of a single input and output for each input, then choose to input the features strongly correlated with output. Finally you can add the features one by obe to input and observe if it improve result.
or
3. You can make principal component analysis to find which of inputs are the most influential on output and creating the input only from these features.
It is to remember that if you have categorical input with 4 possible categories, you should crate 4 inputs e.g if the variable belogns to the category 3 (out of 4), four inputs should have following values 0, 0, 1, 0..
Of course thare are more problems you can face in obtaining good results for such high number of inputs, but I hope you will manage them sucessfully.
Best regards
Hubert
  • asked a question related to Neural Networks (Computer)
Question
4 answers
is it okay to get data from ansys or abaqus and build a model of neural network Or I must get the data from a realistic field experiment ?
Relevant answer
Answer
One can trust the results from simulation software as long as the inputs given to them were validated. The model setup in and results from ansys/abaqus needs to be validated against either experimental or analytical results. It could be an older study for which experimental results already exist or it could be one of the many cases you want to explore. Without proper validation, you MAY find a lot of numerical inaccuracies.
  • asked a question related to Neural Networks (Computer)
Question
6 answers
hi, I would like to calculate the distance of a step with an accelerometer 6 axis 3 accelero 3 gyro attach to the belt of the human . When I integrate, I'm losing too much accuracy.
I see some research about a way to determine the distance of steeps with neural network models. Does anyone have an idea if it is really possible and is this method precise ?
If yes, do someone know the type of the model and the input parameters ?
If someone has another idea of who could I have the distance of one step with an accelerometer ?
Thanks
Relevant answer
Answer
Usage of any inertial sensors requires external correction (GPS in our case), otherwise the integration errors will be unconstrained due to the sensor's biases, which are immanent to this kind of sensors. Generally speaking, we really
"improve the signal of the acceleration" by GPS. From my point of view, solution of your problem doesn't exist without external correction.
Regards, Anatol Tunik.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
Hello everyone.
I am dealing with construction of extrapolation model with deep learning.
When the training dataset is a shape of (1,0,0,0), (0,1,0,0), (0,0,1,0), (0,0,0,1)
I want to build up a model that is able to reconstruct data that is shape of (1,1,0,0 ), (1,1,1,0).
Since the training data and test (or actually-using) data is different, it does not make sense for conventional facts related to deep learning.
The task now I am doing is, using the test data as a validation set of learning and finding the learning that minimize validation loss.
However, now I realize that the task costs too much resource and time. It can be same as doing some lottery.
Therefore, there should be some improvement for the methods.
If someone can give me some advice for this part, it would be a really big help.
Thank you so much.
Relevant answer
Answer
Neural Arithmetic Logic Units could be interesting in this area: https://arxiv.org/pdf/1808.00508.pdf
" 5 Conclusions
Current approaches to modeling numeracy in neural networks fall short because numerical representations fail to generalize outside of the range observed during training. We have shown how the NAC and NALU can be applied to rectify these two shortcomings across a wide variety of domains ..."
  • asked a question related to Neural Networks (Computer)
Question
4 answers
Hi
I am currently trying to interpret the predictions of a fully connected Neural Network on a regression task. The inputs are adjacency matrices for a brain neural networks (from diffusion MRI scans) to predict cognitive abilities. My model currently does a good job at predicting (mean error of 5%) - I would like to understand why it is making the decisions it is making (e.g. which regions of the brain are more important).
I have come across a lot of literature covering this problem, but only regarding constitutional neural networks, not fully connected ones.
Thanks!
Relevant answer
Answer
I can recommend the shap library: https://github.com/slundberg/shap
  • asked a question related to Neural Networks (Computer)
Question
10 answers
The achievement of high-quality results of deep learning is possible by appropriately setting of its parameters. The number of hidden layers and the number of neurons in each layer of a deep machine learning have main influence on the performance of the algorithm. Some manual parameter setting and grid search approaches somewhat ease the users' tasks in setting these important parameters. But, these two techniques can be very time-consuming.
Relevant answer
Answer
This paper will explain in details your question.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
I want to test MLP Neural network using leave one out cross validation, is there some approach to obtaining cross-validated R2 using in Matlab?
Best regards
Relevant answer
Answer
Dear Akhmad Faqih
yes I need a code CV if it is possible
thank you so much
  • asked a question related to Neural Networks (Computer)
Question
3 answers
I am doing thesis on baby cry prediction.
i had dataset of baby cries and non- baby cries of two classes.
i want use Mfcc feature extraction technique to identify important components of audio signal and train a model using this feature.
my question is how mfcc knows to select important components of signal.
i got a mfcc plot of baby cry.
can any one explain this plot?
Relevant answer
Answer
MFCCs are the Mel Frequency Cepstral Coefficients. MFCC takes into account human perception for sensitivity at appropriate frequencies by converting the conventional frequency to Mel Scale, and are thus suitable for speech recognition tasks quite well (as they are suitable for understanding humans and the frequency at which humans speak/utter). I would suggest you to try to extract MFCCs from your audio signal. We generally take 12-13 Mel Frequency coefficients into consideration as features when training models. Other than that, I believe it would be useful if you consider other features as well. For your task on baby cry prediction, I would suggest you to use Volume, Energy, Pitch, Zero Crossing Rate, Spectral Centroid etc. as some additional features along with MFCC. I also suggest you to use some feature selection techniques like Principal Component Analysis (PCA) or t-sne (t-distributed stochastic neighbor embedding) for identifying optimal features. This would help you get better results.
  • asked a question related to Neural Networks (Computer)
Question
9 answers
I want to know if I can consider an ELM NN as a Deep Learning approach even if it is a single layer neural network, and in Deep Learning we work with Multi Layer perceptrons.
Thanks in advance
Relevant answer
Answer
The word 'deep' in deep learning refers to at least one hidden layer. A model which cannot have more than one hidden layer should not be considered 'deep', in my opinion. However, considering rapidly developing fields like deep learning and machine learning, there is much room for debate in terms of nomenclature.
  • asked a question related to Neural Networks (Computer)
Question
4 answers
Hope you help me for the algorithm for the bayesian regularization Thank you
Relevant answer
Answer
Mohammed Abdullah Al-Hagery there is no link please help thank you
  • asked a question related to Neural Networks (Computer)
Question
10 answers
i want to detect baby cry in real-time and i am using windows system. I had extracted the mfcc features of baby cry and trained a neural network model.
i got model accuracy 99%. This model i export to raspberry pi for real-time testing and got good results. But the problem is lot of false positives.
the model is detecting every loud sounds it shows lot of false positives.
my aim is to detect only baby cry.
i also used pitch features but no changes in reducing false positives.
please help me to overcome this problem.
Relevant answer
Answer
Adding to Jiří Kroc , also look at your feature extraction process. Make sure you extract features from both time and frequency domain. In frequency domain, use windowing to extract features at different frequency bands. This might help in further reducing false positives.
Perform PCA to your data and look at the scatter for PC1 and PC2. You should see some groups in the scatter. If you see one big scatter then, you might have to consider collecting more data.
  • asked a question related to Neural Networks (Computer)
Question
14 answers
I am working on computer vision problem, we have a very images to train. I wanted to know is there any possible and practical way to generate synthetic images using deep learning.
Any suggestions are appreciated.
Relevant answer
Answer
Hi,
That is exactly what GAN does. Read about synthetic data augmentation using GAN (generative adversarial network). One of the MSc students I had used it to generate synthetic data to augment the MNIST data set.
  • asked a question related to Neural Networks (Computer)
Question
3 answers
I am having problem with prediction with my model trained on resnet50. I have 10 classes of Nepali numbers from (0 ...9). I have trained the model for 100 epochs with around 40,000 data . What is the issue with my model? I am having overfit? Or, the model I have used to train is just too complex for 10 class. I have also tried this model to predict on the training set but the prediction accuracy is very very poor around (10%).
Relevant answer
Answer
I think you are using few samples to learning a ResNet network. In particular, this network requires a lot of samples during the training, due to its high capacity. I had similar problems in CIFAR-10 using ResNet 50, and when I used data argumentation, the network works fine. Finally, I recommend you to decrease the learning rate, iteratively, after some epochs, for instance, after 80, 120, 160, 180 epochs.