Informacja

Drogi użytkowniku, aplikacja do prawidłowego działania wymaga obsługi JavaScript. Proszę włącz obsługę JavaScript w Twojej przeglądarce.

Wyszukujesz frazę "RNN" wg kryterium: Temat


Wyświetlanie 1-7 z 7
Tytuł:
Attention-based deep learning model for Arabic handwritten text recognition
Autorzy:
Aïcha Gader, Takwa Ben
Echi, Afef Kacem
Powiązania:
https://bibliotekanauki.pl/articles/2201264.pdf
Data publikacji:
2022
Wydawca:
Szkoła Główna Gospodarstwa Wiejskiego w Warszawie. Instytut Informatyki Technicznej
Tematy:
Arabic handwriting recognition
attention mechanism
BLSTM
CNN
CTC
RNN
Opis:
This work proposes a segmentation-free approach to Arabic Handwritten Text Recog-nition (AHTR): an attention-based Convolutional Neural Network - Recurrent Neural Network - Con-nectionist Temporal Classification (CNN-RNN-CTC) deep learning architecture. The model receives asinput an image and provides, through a CNN, a sequence of essential features, which are transferred toan Attention-based Bidirectional Long Short-Term Memory Network (BLSTM). The BLSTM gives features sequence in order, and the attention mechanism allows the selection of relevant information from the features sequences. The selected information is then fed to the CTC, enabling the loss calculation and the transcription prediction. The contribution lies in extending the CNN by dropout layers, batch normalization, and dropout regularization parameters to prevent over-fitting. The output of the RNN block is passed through an attention mechanism to utilize the most relevant parts of the input sequence in a flexible manner. This solution enhances previous methods by improving the CNN speed and performance and controlling over model over-fitting. The proposed system achieves the best accuracy of97.1% for the IFN-ENIT Arabic script database, which competes with the current state-of-the-art. It was also tested for the modern English handwriting of the IAM database, and the Character Error Rate of 2.9% is attained, which confirms the model’s script independence.
Źródło:
Machine Graphics & Vision; 2022, 31, 1/4; 49--73
1230-0535
2720-250X
Pojawia się w:
Machine Graphics & Vision
Dostawca treści:
Biblioteka Nauki
Artykuł
Tytuł:
An overview of deep learning techniques for short-term electricity load forecasting
Autorzy:
Adewuyi, Saheed
Aina, Segun
Uzunuigbe, Moses
Lawal, Aderonke
Oluwaranti, Adeniran
Powiązania:
https://bibliotekanauki.pl/articles/117932.pdf
Data publikacji:
2019
Wydawca:
Polskie Towarzystwo Promocji Wiedzy
Tematy:
Short-term Load Forecasting
Deep Learning Architectures
RNN
LSTM
CNN
SAE
prognozowanie obciążenia krótkoterminowego
architektura głębokiego uczenia
Opis:
This paper presents an overview of some Deep Learning (DL) techniques applicable to forecasting electricity consumptions, especially in the short-term horizon. The paper introduced key parts of four DL architectures including the RNN, LSTM, CNN and SAE, which are recently adopted in implementing Short-term (electricity) Load Forecasting problems. It further presented a model approach for solving such problems. The eventual implication of the study is to present an insightful direction about concepts of the DL methods for forecasting electricity loads in the short-term period, especially to a potential researcher in quest of solving similar problems.
Źródło:
Applied Computer Science; 2019, 15, 4; 75-92
1895-3735
Pojawia się w:
Applied Computer Science
Dostawca treści:
Biblioteka Nauki
Artykuł
Tytuł:
Forecasting future values of time series using the lstm network on the example of currencies and WIG20 companies
Prognozowanie przyszłych wartości szeregów czasowych z wykorzystaniem sieci lstm na przykładzie kursów walut i spółek WIG20
Autorzy:
Mróz, Bartosz
Nowicki, Filip
Powiązania:
https://bibliotekanauki.pl/articles/2016294.pdf
Data publikacji:
2020
Wydawca:
Politechnika Bydgoska im. Jana i Jędrzeja Śniadeckich. Wydawnictwo PB
Tematy:
recurrent neural network
RNN
gated recurrent unit
GRU
long short-term memory
LSTM
rekurencyjna sieć neuronowa
blok rekurencyjny
pamięć długookresowa
Opis:
The article presents a comparison of the RNN, GRU and LSTM networks in predicting future values of time series on the example of currencies and listed companies. The stages of creating an application which is a implementation of the analyzed issue were also shown – the selection of networks, technologies, selection of optimal network parameters. Additionally, two conducted experiments were discussed. The first was to predict the next values of WIG20 companies, exchange rates and cryptocurrencies. The second was based on investments in cryptocurrencies guided solely by the predictions of artificial intelligence. This was to check whether the investments guided by the predictions of such a program have a chance of effective earnings. The discussion of the results of the experiment includes an analysis of various interesting phenomena that occurred during its duration and a comprehensive presentation of the relatively high efficiency of the proposed solution, along with all kinds of graphs and comparisons with real data. The difficulties that occurred during the experiments, such as coronavirus or socio-economic events, such as riots in the USA, were also analyzed. Finally, elements were proposed that should be improved or included in future versions of the solution – taking into account world events, market anomalies and the use of supervised learning.
W artykule przedstawiono porównanie sieci RNN, GRU i LSTM w przewidywaniu przyszłych wartości szeregów czasowych na przykładzie walut i spółek giełdowych. Przedstawiono również etapy tworzenia aplikacji będącej realizacją analizowanego zagadnienia – dobór sieci, technologii, dobór optymalnych parametrów sieci. Dodatkowo omówiono dwa przeprowadzone eksperymenty. Pierwszym było przewidywanie kolejnych wartości spółek z WIG20, kursów walut i kryptowalut. Drugi opierał się na inwestycjach w kryptowaluty, kierując się wyłącznie przewidywaniami sztucznej inteligencji. Miało to na celu sprawdzenie, czy inwestowanie na podstawie przewidywania takiego programu pozwala na efektywne zarobki. Omówienie wyników eksperymentu obejmuje analizę różnych ciekawych zjawisk, które wystąpiły w czasie jego trwania oraz kompleksowe przedstawienie relatywnie wysokiej skuteczności proponowanego rozwiązania wraz z wszelkiego rodzaju wykresami i porównaniami z rzeczywistymi danymi. Analizowano również trudności, które wystąpiły podczas eksperymentów, takie jak koronawirus, wydarzenia społeczno-gospodarcze czy zamieszki w USA. Na koniec zaproponowano elementy, które należałoby ulepszyć lub uwzględnić w przyszłych wersjach rozwiązania, uwzględniając wydarzenia na świecie, anomalie rynkowe oraz wykorzystanie uczenia się nadzorowanego.
Źródło:
Zeszyty Naukowe. Telekomunikacja i Elektronika / Uniwersytet Technologiczno-Przyrodniczy w Bydgoszczy; 2020, 24; 13-30
1899-0088
Pojawia się w:
Zeszyty Naukowe. Telekomunikacja i Elektronika / Uniwersytet Technologiczno-Przyrodniczy w Bydgoszczy
Dostawca treści:
Biblioteka Nauki
Artykuł
Tytuł:
An optimized parallel implementation of non-iteratively trained recurrent neural networks
Autorzy:
El Zini, Julia
Rizk, Yara
Awad, Mariette
Powiązania:
https://bibliotekanauki.pl/articles/2031147.pdf
Data publikacji:
2021
Wydawca:
Społeczna Akademia Nauk w Łodzi. Polskie Towarzystwo Sieci Neuronowych
Tematy:
GPU implementation
parallelization
Recurrent Neural Network
RNN
Long-short Term Memory
LSTM
Gated Recurrent Unit
GRU
Extreme Learning Machines
ELM
non-iterative training
Opis:
Recurrent neural networks (RNN) have been successfully applied to various sequential decision-making tasks, natural language processing applications, and time-series predictions. Such networks are usually trained through back-propagation through time (BPTT) which is prohibitively expensive, especially when the length of the time dependencies and the number of hidden neurons increase. To reduce the training time, extreme learning machines (ELMs) have been recently applied to RNN training, reaching a 99% speedup on some applications. Due to its non-iterative nature, ELM training, when parallelized, has the potential to reach higher speedups than BPTT. In this work, we present Opt-PR-ELM, an optimized parallel RNN training algorithm based on ELM that takes advantage of the GPU shared memory and of parallel QR factorization algorithms to efficiently reach optimal solutions. The theoretical analysis of the proposed algorithm is presented on six RNN architectures, including LSTM and GRU, and its performance is empirically tested on ten time-series prediction applications. Opt- PR-ELM is shown to reach up to 461 times speedup over its sequential counterpart and to require up to 20x less time to train than parallel BPTT. Such high speedups over new generation CPUs are extremely crucial in real-time applications and IoT environments.
Źródło:
Journal of Artificial Intelligence and Soft Computing Research; 2021, 11, 1; 33-50
2083-2567
2449-6499
Pojawia się w:
Journal of Artificial Intelligence and Soft Computing Research
Dostawca treści:
Biblioteka Nauki
Artykuł
Tytuł:
Robust zeroing neural networks with two novel power-versatile activation functions for solving dynamic Sylvester equation
Autorzy:
Zhou, Peng
Tan, Mingtao
Powiązania:
https://bibliotekanauki.pl/articles/2173674.pdf
Data publikacji:
2022
Wydawca:
Polska Akademia Nauk. Czytelnia Czasopism PAN
Tematy:
recurrent neural network
RNN
zeroing neural network
ZNN
robust zeroing neural network
RZNN
fixed-time convergence
rekurencyjna sieć neuronowa
zerowanie sieci neuronowej
konwergencja w ustalonym czasie
Opis:
In this work, two robust zeroing neural network (RZNN) models are presented for online fast solving of the dynamic Sylvester equation (DSE), by introducing two novel power-versatile activation functions (PVAF), respectively. Differing from most of the zeroing neural network (ZNN) models activated by recently reported activation functions (AF), both of the presented PVAF-based RZNN models can achieve predefined time convergence in noise and disturbance polluted environment. Compared with the exponential and finite-time convergent ZNN models, the most important improvement of the proposed RZNN models is their fixed-time convergence. Their effectiveness and stability are analyzed in theory and demonstrated through numerical and experimental examples.
Źródło:
Bulletin of the Polish Academy of Sciences. Technical Sciences; 2022, 70, 3; art. no. e141307
0239-7528
Pojawia się w:
Bulletin of the Polish Academy of Sciences. Technical Sciences
Dostawca treści:
Biblioteka Nauki
Artykuł
Tytuł:
Influence of modelling phase transformations with the use of LSTM network on the accuracy of computations of residual stresses for the hardening process
Autorzy:
Wróbel, Joanna
Kulawik, Adam
Powiązania:
https://bibliotekanauki.pl/articles/27311451.pdf
Data publikacji:
2023
Wydawca:
Polska Akademia Nauk. Czasopisma i Monografie PAN
Tematy:
hardening process
temperature
phase transformations in the solid state
effective stresses
numerical modelling
RNN
recurrent neural network
proces hartowania
temperatura
przemiany fazowe w stanie stałym
modelowanie numeryczne
rekurencyjna sieć neuronowa
naprężenie efektywne
Opis:
Replacing mathematical models with artificial intelligence tools can play an important role in numerical models. This paper analyses the modeling of the hardening process in terms of temperature, phase transformations in the solid state and stresses in the elastic-plastic range. Currently, the use of artificial intelligence tools is increasing, both to make greater generalizations and to reduce possible errors in the numerical simulation process. It is possible to replace the mathematical model of phase transformations in the solid state with an artificial neural network (ANN). Such a substitution requires an ANN network that converts time series (temperature curves) into shares of phase transformations with a small training error. With an insufficient training level of the network, significant differences in stress values will occur due to the existing couplings. Long-Short-Term Memory (LSTM) networks were chosen for the analysis. The paper compares the differences in stress levels with two coupled models using a macroscopic model based on CCT diagram analysis and using the Johnson-Mehl-Avrami-Kolmogorov (JMAK) and Koistinen-Marburger (KM) equations, against the model memorized by the LSTM network. In addition, two levels of network training accuracy were also compared. Considering the results obtained from the model based on LSTM networks, it can be concluded that it is possible to effectively replace the classical model in modeling the phenomena of the heat treatment process.
Źródło:
Bulletin of the Polish Academy of Sciences. Technical Sciences; 2023, 71, 4; art. no. e145681
0239-7528
Pojawia się w:
Bulletin of the Polish Academy of Sciences. Technical Sciences
Dostawca treści:
Biblioteka Nauki
Artykuł
Tytuł:
A class of neuro-computational methods for assamese fricative classification
Autorzy:
Patgiri, C.
Sarma, M.
Sarma, K. K.
Powiązania:
https://bibliotekanauki.pl/articles/91763.pdf
Data publikacji:
2015
Wydawca:
Społeczna Akademia Nauk w Łodzi. Polskie Towarzystwo Sieci Neuronowych
Tematy:
neuro-computational classifier
fricative phonemes
Assamese language
Recurrent Neural Network
RNN
neuro fuzzy classifier
linear prediction cepstral coefficients
LPCC
self-organizing map
SOM
adaptive neuro-fuzzy inference system
ANFIS
klasyfikator neuronowy
klasyfikator neuronowo rozmyty
sieć Kohonena
Opis:
In this work, a class of neuro-computational classifiers are used for classification of fricative phonemes of Assamese language. Initially, a Recurrent Neural Network (RNN) based classifier is used for classification. Later, another neuro fuzzy classifier is used for classification. We have used two different feature sets for the work, one using the specific acoustic-phonetic characteristics and another temporal attributes using linear prediction cepstral coefficients (LPCC) and a Self Organizing Map (SOM). Here, we present the experimental details and performance difference obtained by replacing the RNN based classifier with an adaptive neuro fuzzy inference system (ANFIS) based block for both the feature sets to recognize Assamese fricative sounds.
Źródło:
Journal of Artificial Intelligence and Soft Computing Research; 2015, 5, 1; 59-70
2083-2567
2449-6499
Pojawia się w:
Journal of Artificial Intelligence and Soft Computing Research
Dostawca treści:
Biblioteka Nauki
Artykuł
    Wyświetlanie 1-7 z 7

    Ta witryna wykorzystuje pliki cookies do przechowywania informacji na Twoim komputerze. Pliki cookies stosujemy w celu świadczenia usług na najwyższym poziomie, w tym w sposób dostosowany do indywidualnych potrzeb. Korzystanie z witryny bez zmiany ustawień dotyczących cookies oznacza, że będą one zamieszczane w Twoim komputerze. W każdym momencie możesz dokonać zmiany ustawień dotyczących cookies