Birds Images Prediction with Watson Visual Recognition Services from IBM-Cloud and Conventional Neural Network
Volume 7, Issue 6, Page No 181–188, 2022
Adv. Sci. Technol. Eng. Syst. J. 7(6), 181–188 (2022);
DOI: 10.25046/aj070619
Keywords: Bird classification, Gradient descent, Deep learning, Android app, Convolution
Bird watchers and people obsessed with raising and taming birds make a kind of motivation about our subject. It consists of the creation of an Android application called ”Birds Images Predictor” which helps users to recognize nearly 210 endemic bird species in the world. The proposed solution compares the performance of the python script, which realizes a convolutional neural network (CNN), and the performance of the cloud-bound mobile application using IBM’s visual recognition service to choose the platform one. android form. In the first solution we presented an architecture of a CNN model to predict bird class. While the other solution, which shows its effectiveness, is based on IBM’s visual recognition service, we connect the IBM project that contains the training images with our Android Studio project using an API key , and the IBM process classifies the image captured or downloaded from the application and returns the prediction result which indicates the type of bird. Our study highlights three major advantages of the solution using IBM’s visual recognition service compared to that of CNN, the first appears in the number of images used in training which is higher compared to the other and the strong distinction between bird types where images and bird positions come together in color. The second advantage is to create a trained model saved on the cloud in order to use it with each prediction the most difficult thing to do locally due to the low processing capacity of smartphones. The last advantage is reflected in the correct prediction with a certainty of 99% unlike the other solution due to the instability of the CNN model.
1 Introduction
This article is an extension of the work originally ”Bird Image Recognition and Classification Using IBMCloud’s Watson Visual Recognition Services and Conventional Neural Network (CNN)” presented in the conference ”International Conference on Electrical Engineering, Communication and Computing (ICECCE) 2021”[1].
Birds are among the living organisms that play an important role in the life cycle of the planet. They have become a center of interest for many people such as ornithologists who are interested in researching different types of birds, curious and passionate people who wish to live closer to them, without forgetting the sanctuaries, associations or large bird stores that are interested in breeding or protecting birds from extinction. These people often need to know certain characteristics of birds such as their diet, family (or gender) and housing in order to provide them with a suitable and comfort- able lifestyle. To make it easier for these people, we had the idea to
develop an intelligent bird recognition system.
Science has progressively evolved through several chronologi- cal stages. The first phase was based on an experimental paradigm, from the 17th century, which took into account the experiment, the observation of phenomena and the use of this experimentation [2]. The second phase was theoretical, it uses an idea or a hypothesis to explain an observed phenomenon, it is a way of synthesizing the acquired knowledge and inventing new experiments. In fact, there was a dynamic relationship between experimental and theoretical science since one completes the other [3]. Around the year 1995 another paradigm appeared, it is essentially interested in the pro- duction, processing and use of information after storage, it is the science of information. This latter one has been developed in an exponential way stimulating the birth of the 4th paradigm of the data oriented science which leads to significant advances in the statistical techniques and machine learning [4].
Machine learning is one of the disciplines of artificial intel-181 ligence [5], this notion of intelligence defines two fundamental approaches. The first is the connectionist approach, considers in- telligence as a result of all minimalist functions, which are called neurons, by connecting several neurons, which can solve more complex cases and have good results. When we talk about the con- nectionist approach, we say that it is inductive, we have an input and an output and from the observation we try to understand what it is in order to define one or more rules. On the other hand, there is the symbolic movement of things which affirms that intelligence is only symbols interpreted at a high level. For example when we say ”Every Cauchy sequence is a convergent sequence” it means that ”If U is a Cauchy sequence, then U is a convergent sequence”. This approach is deductive,it deduces the result by applying one or more inference rules on the input, which leads to NP-complete problems due to the exponential complexity of the rule application algorithm.
To achieve our solution, a comparable supervised machine learn- ing model using a convolutional neural network (CNN) [6] has been proposed. The first step is to perform a convolution [7] of raster or tensor image representation of our bird database in order to extract the main features. The second step is to perform the [7] flattening operation on the result of the convolutional layer to have a data format that can be injected into an artificial neural network (ANN). The third step is to train the trained model by back-propagating each ANN layer that are fully connected [8]. In the end, we get a well-trained model that is able to predict the type of a given bird from its image. All the steps mentioned above will be detailed in the third section.
The problem with this method is that the script takes a long time to train the model, updating the network weights each time with the gradient back-propagating algorithm that minimizes the error function, and carrying with small speed CPUs the problem becomes heavier. For this reason, we propose another method for prediction that uses IBM Cloud [9]. In fact, the visual recognition service of IBM Cloud forms the prediction classes once and saves them to exploit them when we need a prediction. This applied technique minimizes the prediction time: we don’t need to train the model each time and also it gives the possibility to link the IBM Cloud project with an Android studio project in order to generate an APK installable on an Android OS smart phone, which makes the solution portable and usable.
There are other bird class prediction methods in the literature such as The ”Application of Two-level Attention Models in Deep Convolutional Neural Network for Fine-grained Image Classifica- tion” [10]. This paper focuses on fine-grained classification by ap- plying visual attention using a deep neural network, It incorporates three types of attention: bottom-up, top-down at the object level and top-down at the part level. All of these attentions are combined to form a domain-specific deep network and then used to enhance aspects of objects and their discriminative features. Bottom-up at- tention proposes candidate patches, top-down object-level attention selects patches relevant to a certain object, and top-down part-level attention locates discriminative parts. This pipeline provides signifi- cant improve ments and achieves the best accuracy under the weak- est supervision conditions compared to subsets of the ILSVRC2012 dataset and the CUB200 2011 dataset.
The article ”Handcrafted features and late fusion with deep learn- ing for bird sound classification” [11], makes a study on acoustic features, visual features and deep learning for bird sound classifica- tion. In order to classify bird caller species a unified model built by combining convolution neural network layers ,convolution layers derives the important features and reduces the dimensionality, then fully connected conventional layers for classification. Based on the experimental results on 14 bird species, this method indicates that the proposed deep learning can achieve an F1 score of 94.36 %, superior to the use of acoustic feature approach which gives a score of 88.97% and the use of visual feature approach which gives a score of 88.87%. To further improve the classification performance, they merged the three approaches: acoustic feature approach, visual feature approach and deep learning approach, to expect a final better score of 95.95%.
The authors of the paper titled ”Image based Bird Species Identi- fication using Convolutional Neural Network” [12] have developed a deep learning model to recognize 60 species of birds in order to give the possibility to birders to easily identify the type of birds. The method implements a model to extract information from bird images using the Convolutional Neural Network (CNN) algorithm. The data is gathered from Microsoft’s Bing v7 image search API. The classification accuracy rate of CNN on the training set included 80% of data reached 93.19% and the accuracy rate on the test set that included on 20% of data reached just 84.91%. The experimen- tal study was conducted on Windows 10 operating system in Atom Editor with TensorFlow library.
There are other applications on the Play Store such as ”Merlin Bird ID” [13] whose purpose is to identify birds using the Photo ID pane which offers a short list of matches between photos taken or imported from photo gallery thanks to patterns trained on thousands of photos of each species. This application covers over 8000 species, has a code size of 25 MB and is trained on images downloaded from the eBird checklists and archived in the Macaulay library. This application can identify species that are currently included in the regional kits that contain the reindeer species in that region.
This article is organized as follows: in the second section we describe the approach adopted for the realization of the solution, in the third section we give a description of the method adopted during recognition, in the fourth section we will give the results obtained then we will devote the fifth section to the discussion and evaluation of the results and the last one to conclude the work and open another research track.
2 Materials and Mrthod
2.1 IBM Watson
Watson [14] is a cognitive artificial intelligence computer program developed by IBM. It has several sub-services, including Watson Visual Recognition Ref used primarily for visual content recogni- tion and analysis using machine learning. The IBM platform offers choices for managing application interfaces, among which the An- droid Studio is chosen. After authentication on the IBM Ref Cloud, we create a project as follows:

Figure 1: IBM Cloud interface to create new project
The project name and description (optional) must be defined.

Figure 2: Project information
The next step is to choose the Visual Recognition subsystem of Watson :

Figure 3: Creation of Visual Recognition Service
The visual recognition service helps to create a registered pattern of bird classes, from a database of images where each class must contain at least 10 different images to facilitate prediction.
On the Kaggle website[15] , a dataset of training images has been uploaded which contains approximately 210 bird types, where each type contains more than 10 images.
2.2 Firebase
To store the different characteristics of each bird, including name, family, diet and housing in addition to other similar images; the external Google ”Firebase” database [16] was used.

Figure 4: Realtime database of Firebase Storage
3 Recognition Method
Several innovations in various fields of science derive their oper- ating principle from the behavior of the human body, for example fundamental computer science is inspired by the nervous system to develop models of artificial intelligence. The human brain consists of more than 80 billion living cells [17] called ”neurons”, intercon- nected with each other to transmit or process chemical or electrical signals. A biological neuron consists of the dendrites (inputs) which are branches that receive information from other neurons, a Soma or cell nucleus that processes the information received from the dendrites, an axon as a cable used by neurons to send information to the output and synapses (outputs) that connects the axon and the other neural dendrites.
By analogy, a formal model has been proposed to be integrated into so-called intelligent systems, which can react almost like hu- mans; named artificial neuron or perceptrons [18]: it is a mathe- matical function that takes values into account to make a simple calculation and return one or more values as output. The calcula- tion performed by a perceptron is broken down into two phases: a calculation by a linear function(1):
f (x1, x2…, xn) = w1 x1 + w2 x2 + + wn xn (1)
where the wi are the weights and the xi are the input values with (1 i n), followed by an activation function [19] (or transfer function)φ which will decide whether the information will be trans- mitted to another neuron or not.
There are several activation functions that are chosen accord- ing to the type of problem to be solved, such as the Relu function defined by f (x) = max(0, x) and Softmax generally used to solve multi-classification whith single-label problems. Note that we can add an additional weight b j called bias [20] or error which measures the difference between the expected output value and the estimated output value y j and the estimated value to have an ”affine percep- tron”.
The juxtaposition of several neurons, grouped by layers, gives the notion of the artificial neural network [6]. Previously, the con- nectionist approach consists in finding a relation between the input and the output, more formally looking for a function F which con- nects the input to the output. In real life many types of relationships result in linear or quadratic functions, however there are more com- plex problems such as predicting the class of a bird which results
in a non-linear function. According to the universal approximation theorem [21], any continuous and bounded function can be approxi- mated by a neural network with n inputs and yi outputs, possessing a finite number of neurons in a single layer hidden with an activation function and a output whose activation function is linear.
To find the best function F (2) which connects the input to the output
F(Xi) ≈ Ywhith(Y = (y1, y2, …., ym)) (2)
It suffices to configure the parameters (the weights P = (w1, w2, …wn) and the bias b j) of the network, so that the mide error is minimal. To precisely measure the performance of the ap- proximation, we define a cost function [22]in our case Cross-entropy for categories(3) [23]:
E = −y j log(y˜j) + (1 − y j) log(1 − y˜j) (3)
It sum of Ei the local errors, generally convex and we are looking for its minimum, so we are faced with an optimization problem that can be solved by the method of descent of the gradient [24], the latter assimilates the notion of gradient [24] which is a generalization of the derivation in the case of functions with several variables. The derivation studies the variations of a function and tries to find its extrema; similarly the gradient is a vector which gathers the partial derivations of a function with several variables and which indicates the direction of the greatest slope of the function from a point. The calculation of the gradient vector of the cost function requires the calculation of the partial derivatives of E from those of F by the formula (4):
∂ f (x , …, x )
we say that there is a gradient descent in batches (or mini-batches). It is an intermediate method between the descent of the classical gradient (traversing all the data at each iteration) and the descent of the stochastic gradient (which uses only one data at each iteration), at the end of N iterations, we have traversed the entire dataset: this is called an epoch.
However, these variations do not always lead to the minimum point, which requires another variation in the learning rate, where δ must be chosen neither too large nor too small: if δ is too large, the Pk points will oscillate around the minimum, but if δ is too small, the Pk points will approach the minimum only after a very long time. For example, it is possible to choose, during training, a fairly large δk, then smaller and smaller over the iterations; either by a linear, quadratic or exponential decrease.
The proposed model applies gradient descent Adam [25] stands for Adaptive Momentum Estimation , which is a stochastic gradi- ent descent that combines between two methods Momentum and RMSprop. Momentum [25] is a method that helps to speed up stochastic gradient descent in the appropriate direction and dampen oscillations by adding fractions of previous gradients to the cur- rent gradient. RMSprop [25] is based on a decreasing variation of aggressive learning rate.

Back-propagation [8] is the application of gradient descent on an artificial neural network. Gradient descent is an optimization algorithm used to find the minimum of a function, based on the opposite vector [24] to the gradient vector. To find the minimum error, we start from the initial weights W0 = (w0, w1, …), the direction of the number of iterations fixed in advance or when the error tends to zero; according to the following recurrence formula (5)[8]:
Pk+1 = Pk − δ × grad(Eglobal(Pk)) (5)
There are three variants of the gradient descent, which is differen- tiated by the amount of data used in the calculate the gradient of the cost function. If we use the global error Eglobal, we say that there is a classical gradient descent. If we consider at each iter- ation a single gradient Ei instead of Eglobal, we say that there is a stochastic gradient descent, this technique reduces the amount of calculations, because the calculation of the gradient of Eglobal, involves the calculation of the gradient of each of the Ei, that is to say the partial derivation with respect to each of the weights w j. If we divide the data into packets of size K. For each packet (called a ”batch”), a gradient is calculated and an iteration is performed,
Our study is focused on a convolutional neural network (CNN) , inspired by the operating principle of the visual lobe of vertebrates. It is a type of artificial neural network suitable for image recognition which is composed of two parts: a convolution part and a classifi- cation part corresponds to a classic multilayer model. Convolution
- is a mathematical method that applies a convolutional product between an array (matrix representation of an image) and a pat- tern (filter or kernel), frequently used for image processing in order to bring a transformation or derive its main characteristics. This convolutional product [7] can be realized by a neural network with convolution neurons, in which also we apply the gradient descent algorithm to adjust their weights. The result of this convolution forms a feature map. The convolution part also consists of a pool- ing phase[7], the purpose of which is to reduce the dimensionality which aims to reduce the computational complexity while keeping
the important characteristics of the input image. There are two types of Pooling: max-Pooling and average-Pooling. The first takes the max of each of the sub-matrices (windows) while the second takes the average. The last phase of the convolution part is the flattening phase which converts the feature maps obtained after the Pooling operation into a column vector.
The result column vector will be the input of our artificial neural network which constitutes the second phase of the multilayer model, it is trained using the gradient back-propagation algorithm with two image databases of birds, one for training and another for testing.
A good learning model offers an optimal balance between bias and variance, it simultaneously minimizes these two error quantities, this problem is known as the bias–variance trade-off [20]. On the one hand, a high bias and a low variance cause underlearning[26]: lack of relevant relationships between input/output data; On the other hand, excessive data variance and low bias can lead to overfitting[26], where the model remains too specific to the training data and thus is unable to adapt to new data. To remedy this prob- lem, we use the Dropout technique. Dropout [8] is a technique that temporarily deactivates certain neurons of a layer during training with a given probability, therefore it sets the weights to zero for both evaluation and backpropagation.
Here is a summary diagram of all the steps mentioned above of our adapted version of the RNC (go to the appendix section to see the source code in python):

Figure 5: Descriptive schema of the applied convolutional neural network
The problem mentioned in this research is included in the family of image classification problems, so to clearly indicate to which class an image belongs, we need to minimize the gap (2) between each of the images of the input classes (training base) with the im- ages of the same classes of the output (test base), this division of the images into two: test and training, is done to remedy the problem of ias–variance 3. First of all, each RGB image is considered as a matrix of pixels to which convolution 3 is applied to reduce its size and keep only its relevant characteristics in order to transform it into a column vector that we introduce in a Multilayer Perceptron. This operation decreases the computing power required to process the data. The column vector enters the artificial nouron network in which the backpropagation 3 is applied based on the ADAM 3 algo- rithm, at each iteration of the process and over a series of epochs, the model is able distinguish between dominant features and some low-level features in images and classify them using the Softmax 3 activation function. Here are the essential steps for creating the CNN model:
- Create an image generator for the training data;
- Create an image generator for the training data;
- Upload the training images;
- Create two convolutional layers;
- Create two Pooling layers;
- Create the Flattening Layer;
- Treated the over-learning problem;
- Added a Multilayer Perceptron with one hidden layer and an output layer with 128 neurons which allows the detection of 210 types of

Figure 6: Model architecture summary
4 Experimental Results and Analysis
This study is a comparison between the results of two methods of predicting the class of a bird, such as the python script that imple- ments a convolutional neural network (CNN)[6] and the method of prediction by the IBM Cloud visual recognition service[9].
The first method of the CNN script was trained on 210 bird classes, while the second one offers the possibility to test up to 1010 classes. The latter minimises the error and prediction time, where the prediction reaches a percentage of servitude of 99% and a wait- ing time of less than 10 seconds. In addition, it offers the possibility to store the trained model in the cloud for later reuse or integration with a solution such as Java, Android Studio or Python. Thus, the process of learning bird images using the CNN method takes enough time for training despite being tested on a small dataset and gives a low prediction accuracy compared to the second method.
The proposed solution integrates the prediction method by IBM Cloud’s visual recognition service with an Android Studio project to generate an APK of an Android mobile application.
Here is a table of comparison between CNN model and visual recognition service from IBM Cloud :
Table 1: comparative table between CNN and VRS-IBM
| Comparison criterion | CNN | VR-IBM |
| Number of class | 210 | 1010 |
| Response time | 2s | 5s |
| Accuracy | 22% | no access |
| Loss | 64% | no access |
| Model storage | locally | on the cloud |
| Integration with mobile app | TensorFlow Lite | API |
Images show the instability of the CNN model:

Figure 7: accuracy and loss per epoch graphs
the two graphs above indicate the variation of accuracy and loss over epochs. The first accuracy graph illustrates that the accuracy of the validation increased linearly with the accuracy, but from epoch 18 it fluctuates between values very close and away from the ac- curacy, this means that the model learns more information at each time; similarly for the loss graph, the validation loss shows that it decreased linearly, but after 18-19 epoch it started to increase. This means that the model tried to store the data. We also notice that the error rate is 64% and the accuracy rate is 22%, which means that the model requires a good adjustment of the hyperparameters to have good results.
The application represents in the first interface ”figure 6” a list of bird images, when you press the button of the ”eye” icon a detail containing the characteristics (name, family, housing, food) of this race, will be displayed in the form of a table ”figure 7” as well as a small gallery of the same race to have more about this race.

Figure 8: Home interface
Predicting the breed of a bird from its image is done by two options, the first allows you to choose an already captured image saved in the smartphone gallery as indicated by the plus ”+” icon in ”Figure 6”, or by the second method which allows you to capture a new image instantly as illustrated by the camera icon in ”Figure 6”.

Figure 9: Detail interface
To detail the scenario, once the user uploads or captures an image of a bird, the image will be transmitted to the IBM Cloud Platform through the ”ClassifyOptions” object, thereafter the vi- sual recognition service will return the prediction result through the ”ClassifiedImages” object in the form of the name of the class to which the bird belongs and the percentage of similarity. After retrieving the name of this bird a query will be sent to the FireBase database with the bird name as parameter in order to receive their characteristics and display them in a table as shown in ”Figure 7”.
5 Discussion
Our article presents a demonstration of three main results concern- ing the recognition problem: the first result is the prediction of the class of bird given in the image captured or chosen by the user, this result reflects the correct prediction with certainty 99% unlike the other solution due to CNN model instability. The second result is the presentation of major characteristics of the predicted bird such as its family, its habitat and its food with the display of a small gallery of bird images from the database used, This result illustrates the model’s reliability in distinguishing between similar types in the face of the high number of images used in the training phase. The third result is the usability and easy portability of the solution on the phone in the form of an Android mobile application, the solution generates a trained model saved on the cloud in order to use it for each prediction the most difficult thing to be done locally due to the low processing capacity of smartphones. However, our study focuses on 210 classes of endemic bird species to make the exact classification in a targeted and efficient way. The proposed CNN model has been tested on training images, it returns the result with an accuracy of 22%. But due to the visual similarity between bird species, when using a new image, the model sometimes fails to distinguish between two types and ultimately leads to misclassifi- cation. On average, testing the data set gives a sensitivity of 64%. Therefore, the visual recognition service of the IBM Cloud method has been adopted to make the prediction of the class of bird species in order to have a simple, fast and efficient result. The latter applies a deep learning algorithm and gives a good numerical precision, which we bet on the success of this project, despite that it does not give access to visualize the model architecture and compare their metrics with that CNN model. have a simple, fast and effective result. The latter applies a deep learning algorithm and gives a good numerical precision, which we bet on the success of this project, de-
spite that it does not give access to visualize the model architecture and compare their metrics with that CNN model. have a simple, fast and effective result. The latter applies a deep learning algorithm and gives a good numerical precision, which we bet on the success of this project, despite that it does not give access to visualize the model architecture and compare their metrics with that CNN model.
In December 2021 the visual recognition service of Watson was obsolete [27], it is replaced by others like the machine learning service and to keep the usability of the application we proposed another solution consists in saving the trained model of the python script, after having improved the setting of the hyper-parameters, in the form of a file with extension ”h5” then with this file we will generate another TansorFlow file[28] with extension ”tflite” [29]in order to integrate it with the Android studio project by replacing the part of prediction with the services by an image injection on the recorded model.
The “Merlin Bird ID” app [13] is one of the powerful apps in this field of bird species identification, it is released on play store the first time in 2014; it is powered by Visipedia, relies on artifi- cial vision and deep learning, its last update was published in May 2022. It offers other functionality more than the features offered in our image recognition solution. “Merlin Bird ID” contains several pane among them, image recognition photo id pane to identify birds from photos by calling sets of millions of photos contributed by bird watchers in eBird.org at l using other user-provided informa- tion such as date and location; it brings together regional kits for example from the United States, Mexico, Central America, from Europe, North Africa and Australia. . . The Sound Id component is dedicated to audio recognition, it identifies the species from the sound and contains an audio recording library by region. The start pane id who Merlin starts by asking you a few simple questions and then magically gives you a list of birds that best match your description After choosing one you can see more photos, listen to recordings sound and read tips for identification. APK size rating our application is small with 23.64 MB than Merlin of 227 MB, so our study is limited to 210 types compared to the other which has more than 40,000 photos of birds (males, females and juveniles). Finally, Merlin is developed for different languages such as English, Spanish, Portuguese, Hebrew, German, Simplified Chinese and Tra- ditional Chinese unlike our solution which is limited to English language only. The remark that we have noticed as users of this application that always the download and installation of regional kits either in the Photo ID or Sound ID pane fails but is it because of weak connection or technical problem in the application.
The paper ”A new mobile application for agricultural pest recog- nition using deep learning in a cloud computing system” [30] is one of the research topics that exists in the literature in the sense of classification using a mobile application, this paper makes a comparison between the performance of two models, including the convolutional neural network (CNN) with connection hopping and the convolutional neural network without connection hopping. This research focuses on the development of an application that recognises weeds in agricultural areas, such as fields, greenhouses and large farms. The mobile application was developed by Apache Cordova, using Flask Framework to handle the HTTP POST re- quests sent, the python code executed and hosted in the Cloud on the Python Anywhere environment and Mysql as the database to
store all the pest information. So the application is based on Faster R-CNN, it was the result of this study, it gave good results which reached a score of 98.9%, regarding automatic control of pests, about five popular agricultural pests, crops compared to machine learning and deep learning classifiers in previous studies. This study also shows a major advantage in being able to detect pests in similar and complex backgrounds to the images tested. Furthermore, the Faster R-CNN is suitable for real-time identification of agricultural pests in the field, without prior knowledge of the number of objects in the acquired images. Finally, the prediction of the categories and positions of the different agricultural pests are accurately identified. There is also another achievement in the literature which is based on the visual recognition service of IBM Cloud: ”Personal robotic assistants: a proposal based on the intelligent services of the IBM cloud and additive manufacturing”[31], it proposes a robotic assistant as a personalized version of the Otto robot, which relies on additive manufacturing and an intelligent chatbot using IBM Cloud services, which concerns speech and image recognition, in noisy environments. This project uses OpenCV and IBM Cloud’s own service for visual recognition, in order to obtain better per- formance in image recognition both in virtual assistants and in the robotic assistant. So in this discussion we are interested in the image recognition part. The authors use techniques like Gaussian, Image Segmentation, Blur effects, and salt and pepper to add noise to the original images to determine if the machine learning model implemented in the robot (via IBM Watson Cloud) can identify the images. correctly based on 3 levels of acceptance that have only been applied to virtual assistants (telegram and messenger). The result of this article is a combination of precision obtained at the same time in all the assistants.
6 Conclusion
Several projects related to various fields of artificial intelligence exploit image recognition, which makes their success depends on finding the best solution to the recognition problem; our project is the study, realization and development of an Android mobile ap- plication named ”Birds images predictor” for the classification and recognition of bird images using IBM’s cloud visual recognition ser- vice which applies a machine learning algorithm more specifically a deep learning algorithm.
The application manages to identify about 210 types of birds from their images captured or chosen by the user, it can be improved by increasing the number of classes of birds, adding other additional options such as the geographical distribution on the map of each of these races. The architecture of the CNN model can also be im- proved by readjusting its hyperparameters to increase the prediction accuracy and to have a low error rate in order to compare it again with another solution in the same direction.
- F. Z. ELBouni, T. Elhariri, C. Zouitni, I. BenBahya, H. Elaboui, A. ELOuaazizi, “Bird image recognition and classification using Watson visual recognition services from IBMCloud and Conventional Neural Network,” the 3rd International Conference on Electrical, communication and Computer Engineering, (6), 978–1–6654–3897–1, 2021.
- P. Beck, A. Courdent, A. Feriati, P. Hyvon, “Sciences expe´rimentales et technologie,” 79, 2218945959, 9782218945953, 2012.
- M. Michel, A quoi sert l’histoire des sciences?, 2006, ISBN: 978-2-7592-0082- 5.
- Y. Li, S. Abdallah, “On hyperparameter optimization of machine learning algorithms:Theory and practice,” ELSEVIER, 22, 295–316, 2020.
- B. Gilles, F. Jym, “Langage naturel et intelligence artificielle,” Hal-Archive, 10, 1638–1580, 1988.
- A. Saad, A. M. Tareq, S. AL-ZAWI, “Understanding of a Convolutional Neural Network,” IEEE Access, 6, 978–1–5386–1949–0, 2017.
- S. Dominik, M. Andreas, B. Sven, “Evaluation of Pooling Operations in Con- volutional Architectures for Object Recognition,” conference, 10, 164, 53117.
- A. Bodin, F. Recher, “Alge`bre–Convolution,” in Deepmath mathe´matique simple des re´seaux de neurones (pas trop complique´s), Janvier-2021.
- IBM, “IBM Cloud,” http://www.ibm.com/cloud, 2022.
- X. Tianjun, X. Yichong, Y. Kuiyuan, Z. Jiaxing, P. Yuxin, Z. Zheng, “The Application of Two-level Attention Models in Deep Convolutional Neural Network for Fine-grained Image Classification,” IEEE Access, 9, 842–850, 2015.
- X. Jie, Z. Mingying, “Handcrafted features and late fusion with deep learning for bird sound classification,” IEEE Access, 8, 75–81, 2019.
- R. Satyam, G. Saiaditya, K. Sanu, S. Shidnal, “Image based Bird Species Identification using Convolutional Neural Network,” 2278–0181, 2020.
- C. University, “Merlin Bird ID,” https://merlin.allaboutbirds.org, 2022.
- IBM, “Watson Visual Recognition,” https://www.ibm.com/no-en/cloud/ watson-visual-recognition/pricing, 2022.
- Machine, D. Learning, “Kaggle,” https://www.kaggle.com/, 2022.
- Google, “Firebase,” https://firebase.google.com.
- S. S. Mader, Biologie humaine, 22 janvier 2010, ISBN: 978-2-8041-2117-4.
- C. Touzet, Les re´seaux de neurones artificiels, introduction au connexionnisme,Ph.D. thesis, 2016.
- E. Chigozie, W. Nwankpa, G. Ijomah Anthony, S. Marshall, “Activation Functions Comparison of Trends in Practice and Research for Deep Learning,” 19, 1811–03378, 2018.
- j. Dong, J. Jaeseung, W. Sang, “Prefrontal solution to the bias-variance tradeoff during reinforcement learning,” Elsevier, 37, 37, 110185, 2021.
- D. Andrea, S. Fabian, H. Johannes, N. Patrick, “Cybenko’s Theorem and the capability of a neural network as function approximator,” 32, 2019.
- S. Ve´zina, “fonction d’erreur,” in Le re´seau de neurones.
- S. Anna, A. Bosman, Engelbrecht, M. Helbig, “Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions,” Elsevier, 24, 113–136, 2020.
- A. Frouvelle, “Me´thodes de descente de gradient,” in Me´thodes nume´riques optimisation, 19 avril 2015”.
- P. K. Diederik, L. B. Jimmy, “ADAM:A method for stochastic optimization,” conference paper, 15, 2017.
- D. Yehuda, M. Vidya, G. B. Richard, “A Farewell to the Bias-Variance Trade-off? An Overview of the Theory of Overparameterized Machine Learning,” Elsevier, 48, 2021.
- IBM, “IBM Documentation,” https://www.ibm.com/docs/en/ app-connect/containers_cd?topic=SSTTDS_contcd/com.ibm.ace. icp.doc/localconn_ibmwatsonvr.html, 2022.
- “TensorFlow,” https://www.tensorflow.org/, 2022.
- “TensorFlow,” https://www.tensorflow.org/lite/guide?hl=fr, 2022.
- K. Mohamed Esmail, A. Fahad, S. Albusaymi, A. Sultan, “A new mobile application of agricultural pests recognition using deep learning in cloud computing system,” IEEE Access, 10, 66980–66989, 9 March 2021.
- S. Amendan o Murriol, E. Duta n Go, Christian mez Lema-Condo, V. Robles- Bykbaev, “Personal robotic assistants: a proposal based on the intelligent services of the IBM cloud and additive manufacturing,” IEEE, (6), 978–1–7281–9365–6, 2020.
- Kohinur Parvin, Eshat Ahmad Shuvo, Wali Ashraf Khan, Sakibul Alam Adib, Tahmina Akter Eiti, Mohammad Shovon, Shoeb Akter Nafiz, "Computationally Efficient Explainable AI Framework for Skin Cancer Detection", Advances in Science, Technology and Engineering Systems Journal, vol. 11, no. 1, pp. 11–24, 2026. doi: 10.25046/aj110102
- Jenna Snead, Nisa Soltani, Mia Wang, Joe Carson, Bailey Williamson, Kevin Gainey, Stanley McAfee, Qian Zhang, "3D Facial Feature Tracking with Multimodal Depth Fusion", Advances in Science, Technology and Engineering Systems Journal, vol. 10, no. 5, pp. 11–19, 2025. doi: 10.25046/aj100502
- François Dieudonné Mengue, Verlaine Rostand Nwokam, Alain Soup Tewa Kammogne, René Yamapi, Moskolai Ngossaha Justin, Bowong Tsakou Samuel, Bernard Kamsu Fogue, "Explainable AI and Active Learning for Photovoltaic System Fault Detection: A Bibliometric Study and Future Directions", Advances in Science, Technology and Engineering Systems Journal, vol. 10, no. 3, pp. 29–44, 2025. doi: 10.25046/aj100305
- Surapol Vorapatratorn, Nontawat Thongsibsong, "AI-Based Photography Assessment System using Convolutional Neural Networks", Advances in Science, Technology and Engineering Systems Journal, vol. 10, no. 2, pp. 28–34, 2025. doi: 10.25046/aj100203
- Win Pa Pa San, Myo Khaing, "Advanced Fall Analysis for Elderly Monitoring Using Feature Fusion and CNN-LSTM: A Multi-Camera Approach", Advances in Science, Technology and Engineering Systems Journal, vol. 9, no. 6, pp. 12–20, 2024. doi: 10.25046/aj090602
- Joshua Carberry, Haiping Xu, "GPT-Enhanced Hierarchical Deep Learning Model for Automated ICD Coding", Advances in Science, Technology and Engineering Systems Journal, vol. 9, no. 4, pp. 21–34, 2024. doi: 10.25046/aj090404
- Nguyen Viet Hung, Tran Thanh Lam, Tran Thanh Binh, Alan Marshal, Truong Thu Huong, "Efficient Deep Learning-Based Viewport Estimation for 360-Degree Video Streaming", Advances in Science, Technology and Engineering Systems Journal, vol. 9, no. 3, pp. 49–61, 2024. doi: 10.25046/aj090305
- Sami Florent Palm, Sianou Ezéckie Houénafa, Zourkalaïni Boubakar, Sebastian Waita, Thomas Nyachoti Nyangonda, Ahmed Chebak, "Solar Photovoltaic Power Output Forecasting using Deep Learning Models: A Case Study of Zagtouli PV Power Plant", Advances in Science, Technology and Engineering Systems Journal, vol. 9, no. 3, pp. 41–48, 2024. doi: 10.25046/aj090304
- Henry Toal, Michelle Wilber, Getu Hailu, Arghya Kusum Das, "Evaluation of Various Deep Learning Models for Short-Term Solar Forecasting in the Arctic using a Distributed Sensor Network", Advances in Science, Technology and Engineering Systems Journal, vol. 9, no. 3, pp. 12–28, 2024. doi: 10.25046/aj090302
- Toya Acharya, Annamalai Annamalai, Mohamed F Chouikha, "Enhancing the Network Anomaly Detection using CNN-Bidirectional LSTM Hybrid Model and Sampling Strategies for Imbalanced Network Traffic Data", Advances in Science, Technology and Engineering Systems Journal, vol. 9, no. 1, pp. 67–78, 2024. doi: 10.25046/aj090107
- Toya Acharya, Annamalai Annamalai, Mohamed F Chouikha, "Optimizing the Performance of Network Anomaly Detection Using Bidirectional Long Short-Term Memory (Bi-LSTM) and Over-sampling for Imbalance Network Traffic Data", Advances in Science, Technology and Engineering Systems Journal, vol. 8, no. 6, pp. 144–154, 2023. doi: 10.25046/aj080614
- Nizar Sakli, Chokri Baccouch, Hedia Bellali, Ahmed Zouinkhi, Mustapha Najjari, "IoT System and Deep Learning Model to Predict Cardiovascular Disease Based on ECG Signal", Advances in Science, Technology and Engineering Systems Journal, vol. 8, no. 6, pp. 08–18, 2023. doi: 10.25046/aj080602
- Zobeda Hatif Naji Al-azzwi, Alexey N. Nazarov, "MRI Semantic Segmentation based on Optimize V-net with 2D Attention", Advances in Science, Technology and Engineering Systems Journal, vol. 8, no. 4, pp. 73–80, 2023. doi: 10.25046/aj080409
- Kohei Okawa, Felix Jimenez, Shuichi Akizuki, Tomohiro Yoshikawa, "Investigating the Impression Effects of a Teacher-Type Robot Equipped a Perplexion Estimation Method on College Students", Advances in Science, Technology and Engineering Systems Journal, vol. 8, no. 4, pp. 28–35, 2023. doi: 10.25046/aj080404
- Yu-Jin An, Ha-Young Oh, Hyun-Jong Kim, "Forecasting Bitcoin Prices: An LSTM Deep-Learning Approach Using On-Chain Data", Advances in Science, Technology and Engineering Systems Journal, vol. 8, no. 3, pp. 186–192, 2023. doi: 10.25046/aj080321
- Mario Cuomo, Federica Massimi, Francesco Benedetto, "Detecting CTC Attack in IoMT Communications using Deep Learning Approach", Advances in Science, Technology and Engineering Systems Journal, vol. 8, no. 2, pp. 130–138, 2023. doi: 10.25046/aj080215
- Ivana Marin, Sven Gotovac, Vladan Papić, "Development and Analysis of Models for Detection of Olive Trees", Advances in Science, Technology and Engineering Systems Journal, vol. 8, no. 2, pp. 87–96, 2023. doi: 10.25046/aj080210
- Víctor Manuel Bátiz Beltrán, Ramón Zatarain Cabada, María Lucía Barrón Estrada, Héctor Manuel Cárdenas López, Hugo Jair Escalante, "A Multiplatform Application for Automatic Recognition of Personality Traits in Learning Environments", Advances in Science, Technology and Engineering Systems Journal, vol. 8, no. 2, pp. 30–37, 2023. doi: 10.25046/aj080204
- Anh-Thu Mai, Duc-Huy Nguyen, Thanh-Tin Dang, "Transfer and Ensemble Learning in Real-time Accurate Age and Age-group Estimation", Advances in Science, Technology and Engineering Systems Journal, vol. 7, no. 6, pp. 262–268, 2022. doi: 10.25046/aj070630
- Bahram Ismailov Israfil, "Deep Learning in Monitoring the Behavior of Complex Technical Systems", Advances in Science, Technology and Engineering Systems Journal, vol. 7, no. 5, pp. 10–16, 2022. doi: 10.25046/aj070502
- Nosiri Onyebuchi Chikezie, Umanah Cyril Femi, Okechukwu Olivia Ozioma, Ajayi Emmanuel Oluwatomisin, Akwiwu-Uzoma Chukwuebuka, Njoku Elvis Onyekachi, Gbenga Christopher Kalejaiye, "BER Performance Evaluation Using Deep Learning Algorithm for Joint Source Channel Coding in Wireless Networks", Advances in Science, Technology and Engineering Systems Journal, vol. 7, no. 4, pp. 127–139, 2022. doi: 10.25046/aj070417
- Tiny du Toit, Hennie Kruger, Lynette Drevin, Nicolaas Maree, "Deep Learning Affective Computing to Elicit Sentiment Towards Information Security Policies", Advances in Science, Technology and Engineering Systems Journal, vol. 7, no. 3, pp. 152–160, 2022. doi: 10.25046/aj070317
- Kaito Echizenya, Kazuhiro Kondo, "Indoor Position and Movement Direction Estimation System Using DNN on BLE Beacon RSSI Fingerprints", Advances in Science, Technology and Engineering Systems Journal, vol. 7, no. 3, pp. 129–138, 2022. doi: 10.25046/aj070315
- Jayan Kant Duggal, Mohamed El-Sharkawy, "High Performance SqueezeNext: Real time deployment on Bluebox 2.0 by NXP", Advances in Science, Technology and Engineering Systems Journal, vol. 7, no. 3, pp. 70–81, 2022. doi: 10.25046/aj070308
- Sreela Sreekumaran Pillai Remadevi Amma, Sumam Mary Idicula, "A Unified Visual Saliency Model for Automatic Image Description Generation for General and Medical Images", Advances in Science, Technology and Engineering Systems Journal, vol. 7, no. 2, pp. 119–126, 2022. doi: 10.25046/aj070211
- Idir Boulfrifi, Mohamed Lahraichi, Khalid Housni, "Video Risk Detection and Localization using Bidirectional LSTM Autoencoder and Faster R-CNN", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 6, pp. 145–150, 2021. doi: 10.25046/aj060619
- Seok-Jun Bu, Hae-Jung Kim, "Ensemble Learning of Deep URL Features based on Convolutional Neural Network for Phishing Attack Detection", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 5, pp. 291–296, 2021. doi: 10.25046/aj060532
- Osaretin Eboya, Julia Binti Juremi, "iDRP Framework: An Intelligent Malware Exploration Framework for Big Data and Internet of Things (IoT) Ecosystem", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 5, pp. 185–202, 2021. doi: 10.25046/aj060521
- Baida Ouafae, Louzar Oumaima, Ramdi Mariam, Lyhyaoui Abdelouahid, "Survey on Novelty Detection using Machine Learning Techniques", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 5, pp. 73–82, 2021. doi: 10.25046/aj060510
- Fatima-Ezzahra Lagrari, Youssfi Elkettani, "Traditional and Deep Learning Approaches for Sentiment Analysis: A Survey", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 5, pp. 01–07, 2021. doi: 10.25046/aj060501
- Anjali Banga, Pradeep Kumar Bhatia, "Optimized Component based Selection using LSTM Model by Integrating Hybrid MVO-PSO Soft Computing Technique", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 4, pp. 62–71, 2021. doi: 10.25046/aj060408
- Nadia Jmour, Slim Masmoudi, Afef Abdelkrim, "A New Video Based Emotions Analysis System (VEMOS): An Efficient Solution Compared to iMotions Affectiva Analysis Software", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 990–1001, 2021. doi: 10.25046/aj0602114
- Bakhtyar Ahmed Mohammed, Muzhir Shaban Al-Ani, "Follow-up and Diagnose COVID-19 Using Deep Learning Technique", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 971–976, 2021. doi: 10.25046/aj0602111
- Showkat Ahmad Dar, S Palanivel, "Performance Evaluation of Convolutional Neural Networks (CNNs) And VGG on Real Time Face Recognition System", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 956–964, 2021. doi: 10.25046/aj0602109
- Dominik Štursa, Daniel Honc, Petr Doležel, "Efficient 2D Detection and Positioning of Complex Objects for Robotic Manipulation Using Fully Convolutional Neural Network", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 915–920, 2021. doi: 10.25046/aj0602104
- Kenza Aitelkadi, Hicham Outmghoust, Salahddine laarab, Kaltoum Moumayiz, Imane Sebari, "Detection and Counting of Fruit Trees from RGB UAV Images by Convolutional Neural Networks Approach", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 887–893, 2021. doi: 10.25046/aj0602101
- Binghan Li, Yindong Hua, Yifeng Liu, Mi Lu, "Dilated Fully Convolutional Neural Network for Depth Estimation from a Single Image", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 801–807, 2021. doi: 10.25046/aj060292
- Binghan Li, Yindong Hua, Mi Lu, "Advanced Multiple Linear Regression Based Dark Channel Prior Applied on Dehazing Image and Generating Synthetic Haze", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 790–800, 2021. doi: 10.25046/aj060291
- Susanto Kumar Ghosh, Mohammad Rafiqul Islam, "Convolutional Neural Network Based on HOG Feature for Bird Species Detection and Classification", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 733–745, 2021. doi: 10.25046/aj060285
- Abraham Adiputra Wijaya, Inten Yasmina, Amalia Zahra, "Indonesian Music Emotion Recognition Based on Audio with Deep Learning Approach", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 716–721, 2021. doi: 10.25046/aj060283
- Shahnaj Parvin, Liton Jude Rozario, Md. Ezharul Islam, "Vehicle Number Plate Detection and Recognition Techniques: A Review", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 2, pp. 423–438, 2021. doi: 10.25046/aj060249
- Prasham Shah, Mohamed El-Sharkawy, "A-MnasNet and Image Classification on NXP Bluebox 2.0", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 1, pp. 1378–1383, 2021. doi: 10.25046/aj0601157
- Byeongwoo Kim, Jongkyu Lee, "Fault Diagnosis and Noise Robustness Comparison of Rotating Machinery using CWT and CNN", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 1, pp. 1279–1285, 2021. doi: 10.25046/aj0601146
- Alisson Steffens Henrique, Anita Maria da Rocha Fernandes, Rodrigo Lyra, Valderi Reis Quietinho Leithardt, Sérgio D. Correia, Paul Crocker, Rudimar Luis Scaranto Dazzi, "Classifying Garments from Fashion-MNIST Dataset Through CNNs", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 1, pp. 989–994, 2021. doi: 10.25046/aj0601109
- Reem Bayari, Ameur Bensefia, "Text Mining Techniques for Cyberbullying Detection: State of the Art", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 1, pp. 783–790, 2021. doi: 10.25046/aj060187
- Theerayod Wiangtong, Sethakarn Prongnuch, Suchada Sitjongsataporn, "Stochastic Behaviour Analysis of Adaptive Averaging Step-size Sign Normalised Hammerstein Spline Adaptive Filtering", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 1, pp. 577–586, 2021. doi: 10.25046/aj060162
- Imane Jebli, Fatima-Zahra Belouadha, Mohammed Issam Kabbaj, Amine Tilioua, "Deep Learning based Models for Solar Energy Prediction", Advances in Science, Technology and Engineering Systems Journal, vol. 6, no. 1, pp. 349–355, 2021. doi: 10.25046/aj060140
- Revanna Sidamma Kavitha, Uppara Eranna, Mahendra Nanjappa Giriprasad, "A Computational Modelling and Algorithmic Design Approach of Digital Watermarking in Deep Neural Networks", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 1560–1568, 2020. doi: 10.25046/aj0506187
- Xianxian Luo, Songya Xu, Hong Yan, "Application of Deep Belief Network in Forest Type Identification using Hyperspectral Data", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 1554–1559, 2020. doi: 10.25046/aj0506186
- Majdouline Meddad, Chouaib Moujahdi, Mounia Mikram, Mohammed Rziza, "Optimization of Multi-user Face Identification Systems in Big Data Environments", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 762–767, 2020. doi: 10.25046/aj050691
- Kin Yun Lum, Yeh Huann Goh, Yi Bin Lee, "American Sign Language Recognition Based on MobileNetV2", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 481–488, 2020. doi: 10.25046/aj050657
- Gede Putra Kusuma, Jonathan, Andreas Pangestu Lim, "Emotion Recognition on FER-2013 Face Images Using Fine-Tuned VGG-16", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 315–322, 2020. doi: 10.25046/aj050638
- Lubna Abdelkareim Gabralla, "Dense Deep Neural Network Architecture for Keystroke Dynamics Authentication in Mobile Phone", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 307–314, 2020. doi: 10.25046/aj050637
- Helen Kottarathil Joy, Manjunath Ramachandra Kounte, "A Comprehensive Review of Traditional Video Processing", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 274–279, 2020. doi: 10.25046/aj050633
- Andrea Generosi, Silvia Ceccacci, Samuele Faggiano, Luca Giraldi, Maura Mengoni, "A Toolkit for the Automatic Analysis of Human Behavior in HCI Applications in the Wild", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 185–192, 2020. doi: 10.25046/aj050622
- Fei Gao, Jiangjiang Liu, "Effective Segmented Face Recognition (SFR) for IoT", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 36–44, 2020. doi: 10.25046/aj050605
- Pamela Zontone, Antonio Affanni, Riccardo Bernardini, Leonida Del Linz, Alessandro Piras, Roberto Rinaldo, "Supervised Learning Techniques for Stress Detection in Car Drivers", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 6, pp. 22–29, 2020. doi: 10.25046/aj050603
- Redha Touati, Max Mignotte, Mohamed Dahmane, "A Circular Invariant Convolution Model-Based Mapping for Multimodal Change Detection", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 1288–1298, 2020. doi: 10.25046/aj0505155
- Sherif H. ElGohary, Aya Lithy, Shefaa Khamis, Aya Ali, Aya Alaa el-din, Hager Abd El-Azim, "Interactive Virtual Rehabilitation for Aphasic Arabic-Speaking Patients", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 1225–1232, 2020. doi: 10.25046/aj0505148
- Daniyar Nurseitov, Kairat Bostanbekov, Maksat Kanatov, Anel Alimova, Abdelrahman Abdallah, Galymzhan Abdimanap, "Classification of Handwritten Names of Cities and Handwritten Text Recognition using Various Deep Learning Models", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 934–943, 2020. doi: 10.25046/aj0505114
- Khalid A. AlAfandy, Hicham Omara, Mohamed Lazaar, Mohammed Al Achhab, "Using Classic Networks for Classifying Remote Sensing Images: Comparative Study", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 770–780, 2020. doi: 10.25046/aj050594
- Khalid A. AlAfandy, Hicham, Mohamed Lazaar, Mohammed Al Achhab, "Investment of Classic Deep CNNs and SVM for Classifying Remote Sensing Images", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 652–659, 2020. doi: 10.25046/aj050580
- Lana Abdulrazaq Abdullah, Muzhir Shaban Al-Ani, "CNN-LSTM Based Model for ECG Arrhythmias and Myocardial Infarction Classification", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 601–606, 2020. doi: 10.25046/aj050573
- Chigozie Enyinna Nwankpa, "Advances in Optimisation Algorithms and Techniques for Deep Learning", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 563–577, 2020. doi: 10.25046/aj050570
- Sathyabama Kaliyapillai, Saruladha Krishnamurthy, "Differential Evolution based Hyperparameters Tuned Deep Learning Models for Disease Diagnosis and Classification", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 253–261, 2020. doi: 10.25046/aj050531
- Hicham Moujahid, Bouchaib Cherradi, Oussama El Gannour, Lhoussain Bahatti, Oumaima Terrada, Soufiane Hamida, "Convolutional Neural Network Based Classification of Patients with Pneumonia using X-ray Lung Images", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 167–175, 2020. doi: 10.25046/aj050522
- Mohsine Elkhayati, Youssfi Elkettani, "Towards Directing Convolutional Neural Networks Using Computational Geometry Algorithms: Application to Handwritten Arabic Character Recognition", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 5, pp. 137–147, 2020. doi: 10.25046/aj050519
- Nghia Duong-Trung, Luyl-Da Quach, Chi-Ngon Nguyen, "Towards Classification of Shrimp Diseases Using Transferred Convolutional Neural Networks", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 4, pp. 724–732, 2020. doi: 10.25046/aj050486
- Tran Thanh Dien, Nguyen Thanh-Hai, Nguyen Thai-Nghe, "Deep Learning Approach for Automatic Topic Classification in an Online Submission System", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 4, pp. 700–709, 2020. doi: 10.25046/aj050483
- Adonis Santos, Patricia Angela Abu, Carlos Oppus, Rosula Reyes, "Real-Time Traffic Sign Detection and Recognition System for Assistive Driving", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 4, pp. 600–611, 2020. doi: 10.25046/aj050471
- Maroua Abdellaoui, Dounia Daghouj, Mohammed Fattah, Younes Balboul, Said Mazer, Moulhime El Bekkali, "Artificial Intelligence Approach for Target Classification: A State of the Art", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 4, pp. 445–456, 2020. doi: 10.25046/aj050453
- Roberta Avanzato, Francesco Beritelli, "A CNN-based Differential Image Processing Approach for Rainfall Classification", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 4, pp. 438–444, 2020. doi: 10.25046/aj050452
- Samir Allach, Mohamed Ben Ahmed, Anouar Abdelhakim Boudhir, "Deep Learning Model for A Driver Assistance System to Increase Visibility on A Foggy Road", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 4, pp. 314–322, 2020. doi: 10.25046/aj050437
- Rizki Jaka Maulana, Gede Putra Kusuma, "Malware Classification Based on System Call Sequences Using Deep Learning", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 4, pp. 207–216, 2020. doi: 10.25046/aj050426
- Van-Hung Le, Hung-Cuong Nguyen, "A Survey on 3D Hand Skeleton and Pose Estimation by Convolutional Neural Network", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 4, pp. 144–159, 2020. doi: 10.25046/aj050418
- Hai Thanh Nguyen, Nhi Yen Kim Phan, Huong Hoang Luong, Trung Phuoc Le, Nghi Cong Tran, "Efficient Discretization Approaches for Machine Learning Techniques to Improve Disease Classification on Gut Microbiome Composition Data", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 3, pp. 547–556, 2020. doi: 10.25046/aj050368
- Neptali Montañez, Jomari Joseph Barrera, "Automated Abaca Fiber Grade Classification Using Convolution Neural Network (CNN)", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 3, pp. 207–213, 2020. doi: 10.25046/aj050327
- Aditi Haresh Vyas, Mayuri A. Mehta, "A Comprehensive Survey on Image Modality Based Computerized Dry Eye Disease Detection Techniques", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 2, pp. 748–756, 2020. doi: 10.25046/aj050293
- Yeji Shin, Youngone Cho, Hyun Wook Kang, Jin-Gu Kang, Jin-Woo Jung, "Neural Network-based Efficient Measurement Method on Upside Down Orientation of a Digital Document", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 2, pp. 697–702, 2020. doi: 10.25046/aj050286
- Rajesh Kannan Megalingam, Santosh Tantravahi, Hemanth Sai Surya Kumar Tammana, Nagasai Thokala, Hari Sudarshan Rahul Puram, Naveen Samudrala, "ROS Based Multimode Control of Wheeled Robot", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 2, pp. 688–696, 2020. doi: 10.25046/aj050285
- Bokyoon Na, Geoffrey C Fox, "Object Classifications by Image Super-Resolution Preprocessing for Convolutional Neural Networks", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 2, pp. 476–483, 2020. doi: 10.25046/aj050261
- Bayan O Al-Amri, Mohammed A. AlZain, Jehad Al-Amri, Mohammed Baz, Mehedi Masud, "A Comprehensive Study of Privacy Preserving Techniques in Cloud Computing Environment", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 2, pp. 419–424, 2020. doi: 10.25046/aj050254
- Lenin G. Falconi, Maria Perez, Wilbert G. Aguilar, Aura Conci, "Transfer Learning and Fine Tuning in Breast Mammogram Abnormalities Classification on CBIS-DDSM Database", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 2, pp. 154–165, 2020. doi: 10.25046/aj050220
- Daihui Li, Chengxu Ma, Shangyou Zeng, "Design of Efficient Convolutional Neural Module Based on An Improved Module", Advances in Science, Technology and Engineering Systems Journal, vol. 5, no. 1, pp. 340–345, 2020. doi: 10.25046/aj050143
- Andary Dadang Yuliyono, Abba Suganda Girsang, "Artificial Bee Colony-Optimized LSTM for Bitcoin Price Prediction", Advances in Science, Technology and Engineering Systems Journal, vol. 4, no. 5, pp. 375–383, 2019. doi: 10.25046/aj040549
- Shun Yamamoto, Momoyo Ito, Shin-ichi Ito, Minoru Fukumi, "Verification of the Usefulness of Personal Authentication with Aerial Input Numerals Using Leap Motion", Advances in Science, Technology and Engineering Systems Journal, vol. 4, no. 5, pp. 369–374, 2019. doi: 10.25046/aj040548
- Kevin Yudi, Suharjito, "Sentiment Analysis of Transjakarta Based on Twitter using Convolutional Neural Network", Advances in Science, Technology and Engineering Systems Journal, vol. 4, no. 5, pp. 281–286, 2021. doi: 10.25046/aj040535
- Bui Thanh Hung, "Integrating Diacritics Restoration and Question Classification into Vietnamese Question Answering System", Advances in Science, Technology and Engineering Systems Journal, vol. 4, no. 5, pp. 207–212, 2019. doi: 10.25046/aj040526
- Alvina Aulia, Yogi Udjaja, Irma Kartika Wairooy, Adithia Panji Hutama, Dinar Khansa Shabira, Sabilal Muhtadin, "Android Application to Detect Cat Disease Using an Expert System", Advances in Science, Technology and Engineering Systems Journal, vol. 4, no. 5, pp. 158–162, 2019. doi: 10.25046/aj040521
- Priyamvada Chandel, Tripta Thakur, "Smart Meter Data Analysis for Electricity Theft Detection using Neural Networks", Advances in Science, Technology and Engineering Systems Journal, vol. 4, no. 4, pp. 161–168, 2019. doi: 10.25046/aj040420
- Nelly Elsayed, Anthony Maida, Magdy Bayoumi, "Effects of Different Activation Functions for Unsupervised Convolutional LSTM Spatiotemporal Learning", Advances in Science, Technology and Engineering Systems Journal, vol. 4, no. 2, pp. 260–269, 2019. doi: 10.25046/aj040234
- Ajees Arimbassery Pareed, Sumam Mary Idicula, "A Relation Extraction System for Indian Languages", Advances in Science, Technology and Engineering Systems Journal, vol. 4, no. 2, pp. 65–69, 2019. doi: 10.25046/aj040208
- Bok Gyu Han, Hyeon Seok Yang, Ho Gyeong Lee, Young Shik Moon, "Low Contrast Image Enhancement Using Convolutional Neural Network with Simple Reflection Model", Advances in Science, Technology and Engineering Systems Journal, vol. 4, no. 1, pp. 159–164, 2019. doi: 10.25046/aj040115
- An-Ting Cheng, Chun-Yen Chen, Bo-Cheng Lai, Che-Huai Lin, "Software and Hardware Enhancement of Convolutional Neural Networks on GPGPUs", Advances in Science, Technology and Engineering Systems Journal, vol. 3, no. 2, pp. 28–39, 2018. doi: 10.25046/aj030204
- Marwa Farouk Ibrahim Ibrahim, Adel Ali Al-Jumaily, "Auto-Encoder based Deep Learning for Surface Electromyography Signal Processing", Advances in Science, Technology and Engineering Systems Journal, vol. 3, no. 1, pp. 94–102, 2018. doi: 10.25046/aj030111
- Myoung-Wan Koo, Guanghao Xu, Hyunjung Lee, Jungyun Seo, "Retrieving Dialogue History in Deep Neural Networks for Spoken Language Understanding", Advances in Science, Technology and Engineering Systems Journal, vol. 2, no. 3, pp. 1741–1747, 2017. doi: 10.25046/aj0203213
- Marwa Farouk Ibrahim Ibrahim, Adel Ali Al-Jumaily, "Self-Organizing Map based Feature Learning in Bio-Signal Processing", Advances in Science, Technology and Engineering Systems Journal, vol. 2, no. 3, pp. 505–512, 2017. doi: 10.25046/aj020365
- R. Manju Parkavi, M. Shanthi, M.C. Bhuvaneshwari, "Recent Trends in ELM and MLELM: A review", Advances in Science, Technology and Engineering Systems Journal, vol. 2, no. 1, pp. 69–75, 2017. doi: 10.25046/aj020108