ensure that the model operates effectively in reality
and improves accuracy.
Table 2 shows the results of predictions based on
a single photo. Examining a photo of a rose, as shown
in Figure 4, it indicates the likelihood of it being a
dandelion. It may be due to the resolution of the
photos being compressed to 64 × 64 for training,
which leads to insufficient feature extraction during
training, resulting in inaccuracies during
classification, and ultimately failing to accurately
identify which category the image belongs to. The
possibility of dandelions in Table 2 should
theoretically be very low, but the identified results are
close to 50%.
Table 1 Results of various indicators for model training
Metric Epoch Loss Accurac
y
Precisio
n
Recall
Value 40 0.174
8
0.9657 0.2636 0.2727
Table 2 Results of Predictions for a Single Photo
Metric The possibility of
dandelions
Value 0.451324
Figure 4: Roses used for prediction (Mamaev, 2021).
4 CONCLUSIONS
This experiment used a convolutional neural network
(CNN) model, employing ReLu as the activation
function during training. It also incorporates early
stopping and model checkpoints to save the trained
model. Through the recognition of a single photo, it
was found that although the accuracy reached 0.96,
the precision was only 0.27. Therefore, there are some
potential issues during the model training process.
For example, issues such as excessive false positives,
over-prediction of positive cases, and data imbalance
need to be addressed. To achieve high accuracy in
identifying flower species, it is necessary to further
optimize the model, such as adjusting the learning
rate to minimize the risk of overfitting.By adjusting
the model, it may be possible to effectively identify
the features of images and accurately classify them,
thereby reducing the interference of the natural
environment on classification recognition in real-
world applications.
REFERENCES
Abadi, M., Barham, P., Chen, J., (2016). TensorFlow: A
system for large-scale machine learning. In 12th
USENIX Symposium on Operating Systems Design and
Implementation (OSDI 16) (pp. 265–283).
Agarap, A. F. (2018). Deep learning using rectified linear
units (relu). arXiv preprint arXiv:1803.08375.
Banerjee, K., Gupta, R. R., Vyas, K., et al. (2020).
Exploring alternatives to softmax function. arXiv
preprint arXiv:2011.11538.
Demir, F., Abdullah, D. A., & Sengur, A. (2020). A new
deep CNN model for environmental sound
classification. IEEE Access, 8, 66529 – 66537.
https://doi.org/10.1109/ACCESS.2020.2984903
Haloi, M., & Shekhar, S. (2021). Datum: A system for
tfrecord dataset management. GitHub. Retrieved from
https://github.com/openAGI/datum
Lee, J., Wang, Y., & Cho, S. (2022). Angular margin-
mining softmax loss for face recognition. IEEE Access,
10, 43071 – 43080.
https://doi.org/10.1109/ACCESS.2022.3168310
Liu, F., Zhang, Z., & Zhou, R. (2022). Automatic
modulation recognition based on CNN and GRU.
Tsinghua Science and Technology, 27(2), 422–431.
https://doi.org/10.26599/TST.2020.9010057
Saba, T., Rehman, A., Jamail, N. S. M., Marie-Sainte, S. L.,
Raza, M., & Sharif, M. (2021). Categorizing the
students’ activities for automated exam proctoring
using proposed Deep L2-GraftNet CNN network and
ASO based feature selection approach. IEEE Access, 9,
47639 – 47656.
https://doi.org/10.1109/ACCESS.2021.3068223
Wang, S. Y., Wang, O., Zhang, R., et al. (2020). CNN-
generated images are surprisingly easy to spot... for
now. In Proceedings of the IEEE/CVF Conference on
Computer Vision and Pattern Recognition (pp. 8695–
8704).
Yan, Y., Zhang, X., & Wu, B. (2019). Simplified early
stopping criterion for belief-propagation polar code
decoder based on frozen bits. IEEE Access, 7, 134691–
134696.
https://doi.org/10.1109/ACCESS.2019.2940135
Yan, Z., Guo, S., Xiao, G., & Zhang, H. (2020). On
combining CNN with non-local self-similarity based
image denoising methods. IEEE Access, 8, 14789–