Academic literature on the topic 'Encoder and decoder feature'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Encoder and decoder feature.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Encoder and decoder feature"

1

Shim, Jae-hun, Hyunwoo Yu, Kyeongbo Kong, and Suk-Ju Kang. "FeedFormer: Revisiting Transformer Decoder for Efficient Semantic Segmentation." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 2 (2023): 2263–71. http://dx.doi.org/10.1609/aaai.v37i2.25321.

Full text
Abstract:
With the success of Vision Transformer (ViT) in image classification, its variants have yielded great success in many downstream vision tasks. Among those, the semantic segmentation task has also benefited greatly from the advance of ViT variants. However, most studies of the transformer for semantic segmentation only focus on designing efficient transformer encoders, rarely giving attention to designing the decoder. Several studies make attempts in using the transformer decoder as the segmentation decoder with class-wise learnable query. Instead, we aim to directly use the encoder features as the queries. This paper proposes the Feature Enhancing Decoder transFormer (FeedFormer) that enhances structural information using the transformer decoder. Our goal is to decode the high-level encoder features using the lowest-level encoder feature. We do this by formulating high-level features as queries, and the lowest-level feature as the key and value. This enhances the high-level features by collecting the structural information from the lowest-level feature. Additionally, we use a simple reformation trick of pushing the encoder blocks to take the place of the existing self-attention module of the decoder to improve efficiency. We show the superiority of our decoder with various light-weight transformer-based decoders on popular semantic segmentation datasets. Despite the minute computation, our model has achieved state-of-the-art performance in the performance computation trade-off. Our model FeedFormer-B0 surpasses SegFormer-B0 with 1.8% higher mIoU and 7.1% less computation on ADE20K, and 1.7% higher mIoU and 14.4% less computation on Cityscapes, respectively. Code will be released at: https://github.com/jhshim1995/FeedFormer.
APA, Harvard, Vancouver, ISO, and other styles
2

Wen, Ying, Kai Xie, and Lianghua He. "Segmenting Medical MRI via Recurrent Decoding Cell." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 07 (2020): 12452–59. http://dx.doi.org/10.1609/aaai.v34i07.6932.

Full text
Abstract:
The encoder-decoder networks are commonly used in medical image segmentation due to their remarkable performance in hierarchical feature fusion. However, the expanding path for feature decoding and spatial recovery does not consider the long-term dependency when fusing feature maps from different layers, and the universal encoder-decoder network does not make full use of the multi-modality information to improve the network robustness especially for segmenting medical MRI. In this paper, we propose a novel feature fusion unit called Recurrent Decoding Cell (RDC) which leverages convolutional RNNs to memorize the long-term context information from the previous layers in the decoding phase. An encoder-decoder network, named Convolutional Recurrent Decoding Network (CRDN), is also proposed based on RDC for segmenting multi-modality medical MRI. CRDN adopts CNN backbone to encode image features and decode them hierarchically through a chain of RDCs to obtain the final high-resolution score map. The evaluation experiments on BrainWeb, MRBrainS and HVSMR datasets demonstrate that the introduction of RDC effectively improves the segmentation accuracy as well as reduces the model size, and the proposed CRDN owns its robustness to image noise and intensity non-uniformity in medical MRI.
APA, Harvard, Vancouver, ISO, and other styles
3

Sun, Jun, Junbo Zhang, Xuesong Gao, et al. "Fusing Spatial Attention with Spectral-Channel Attention Mechanism for Hyperspectral Image Classification via Encoder–Decoder Networks." Remote Sensing 14, no. 9 (2022): 1968. http://dx.doi.org/10.3390/rs14091968.

Full text
Abstract:
In recent years, convolutional neural networks (CNNs) have been widely used in hyperspectral image (HSI) classification. However, feature extraction on hyperspectral data still faces numerous challenges. Existing methods cannot extract spatial and spectral-channel contextual information in a targeted manner. In this paper, we propose an encoder–decoder network that fuses spatial attention and spectral-channel attention for HSI classification from three public HSI datasets to tackle these issues. In terms of feature information fusion, a multi-source attention mechanism including spatial and spectral-channel attention is proposed to encode the spatial and spectral multi-channels contextual information. Moreover, three fusion strategies are proposed to effectively utilize spatial and spectral-channel attention. They are direct aggregation, aggregation on feature space, and Hadamard product. In terms of network development, an encoder–decoder framework is employed for hyperspectral image classification. The encoder is a hierarchical transformer pipeline that can extract long-range context information. Both shallow local features and rich global semantic information are encoded through hierarchical feature expressions. The decoder consists of suitable upsampling, skip connection, and convolution blocks, which fuse multi-scale features efficiently. Compared with other state-of-the-art methods, our approach has greater performance in hyperspectral image classification.
APA, Harvard, Vancouver, ISO, and other styles
4

Alharbi, Majed, Ahmed Stohy, Mohammed Elhenawy, Mahmoud Masoud, and Hamiden El-Wahed Khalifa. "Solving Traveling Salesman Problem with Time Windows Using Hybrid Pointer Networks with Time Features." Sustainability 13, no. 22 (2021): 12906. http://dx.doi.org/10.3390/su132212906.

Full text
Abstract:
This paper introduces a time efficient deep learning-based solution to the traveling salesman problem with time window (TSPTW). Our goal is to reduce the total tour length traveled by -*the agent without violating any time limitations. This will aid in decreasing the time required to supply any type of service, as well as lowering the emissions produced by automobiles, allowing our planet to recover from air pollution emissions. The proposed model is a variation of the pointer networks that has a better ability to encode the TSPTW problems. The model proposed in this paper is inspired from our previous work that introduces a hybrid context encoder and a multi attention decoder. The hybrid encoder primarily comprises the transformer encoder and the graph encoder; these encoders encode the feature vector before passing it to the attention decoder layer. The decoder consists of transformer context and graph context as well. The output attentions from the two decoders are aggregated and used to select the following step in the trip. To the best of our knowledge, our network is the first neural model that will be able to solve medium-size TSPTW problems. Moreover, we conducted sensitivity analysis to explore how the model performance changes as the time window width in the training and testing data changes. The experimental work shows that our proposed model outperforms the state-of-the-art model for TSPTW of sizes 20, 50 and 100 nodes/cities. We expect that our model will become state-of-the-art methodology for solving the TSPTW problems.
APA, Harvard, Vancouver, ISO, and other styles
5

Ai, Xinbo, Yunhao Xie, Yinan He, and Yi Zhou. "Improve SegNet with feature pyramid for road scene parsing." E3S Web of Conferences 260 (2021): 03012. http://dx.doi.org/10.1051/e3sconf/202126003012.

Full text
Abstract:
Road scene parsing is a common task in semantic segmentation. Its images have characteristics of containing complex scene context and differing greatly among targets of the same category from different scales. To address these problems, we propose a semantic segmentation model combined with edge detection. We extend the segmentation network with an encoder-decoder structure by adding an edge feature pyramid module, namely Edge Feature Pyramid Network (EFPNet, for short). This module uses edge detection operators to get boundary information and then combines the multiscale features to improve the ability to recognize small targets. EFPNet can make up the shortcomings of convolutional neural network features, and it helps to produce smooth segmentation. After extracting features of the encoder and decoder, EFPNet uses Euclidean distance to compare the similarity between the presentation of the encoder and the decoder, which can increase the decoder’s ability to restore from the encoder. We evaluated the proposed method on Cityscapes datasets. The experiment on Cityscapes datasets demonstrates that the accuracies are improved by 7.5% and 6.2% over the popular SegNet and ENet. And the ablation experiment validates the effectiveness of our method.
APA, Harvard, Vancouver, ISO, and other styles
6

Jiang, S. L., G. Li, W. Yao, Z. H. Hong, and T. Y. Kuc. "DUAL PYRAMIDS ENCODER-DECODER NETWORK FOR SEMANTIC SEGMENTATION IN GROUND AND AERIAL VIEW IMAGES." ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences XLIII-B2-2020 (August 12, 2020): 605–10. http://dx.doi.org/10.5194/isprs-archives-xliii-b2-2020-605-2020.

Full text
Abstract:
Abstract. Semantic segmentation is a fundamental research task in computer vision, which intends to assign a certain category to every pixel. Currently, most existing methods only utilize the deepest feature map for decoding, while high-level features get inevitably lost during the procedure of down-sampling. In the decoder section, transposed convolution or bilinear interpolation was widely used to restore the size of the encoded feature map; however, few optimizations are applied during up-sampling process which is detrimental to the performance for grouping and classification. In this work, we proposed a dual pyramids encoder-decoder deep neural network (DPEDNet) to tackle the above issues. The first pyramid integrated and encoded multi-resolution features through sequentially stacked merging, and the second pyramid decoded the features through dense atrous convolution with chained up-sampling. Without post-processing and multi-scale testing, the proposed network has achieved state-of-the-art performances on two challenging benchmark image datasets for both ground and aerial view scenes.
APA, Harvard, Vancouver, ISO, and other styles
7

Abdulaziz AlArfaj, Abeer, and Hanan Ahmed Hosni Mahmoud. "A Moving Object Tracking Technique Using Few Frames with Feature Map Extraction and Feature Fusion." ISPRS International Journal of Geo-Information 11, no. 7 (2022): 379. http://dx.doi.org/10.3390/ijgi11070379.

Full text
Abstract:
Moving object tracking techniques using machine and deep learning require large datasets for neural model training. New strategies need to be invented that utilize smaller data training sizes to realize the impact of large-sized datasets. However, current research does not balance the training data size and neural parameters, which creates the problem of inadequacy of the information provided by the low visual data content for parameter optimization. To enhance the performance of moving object tracking that appears in only a few frames, this research proposes a deep learning model using an abundant encoder–decoder (a high-resolution transformer (HRT) encoder–decoder). An HRT encoder–decoder employs feature map extraction that focuses on high resolution feature maps that are more representative of the moving object. In addition, we employ the proposed HRT encoder–decoder for feature map extraction and fusion to reimburse the few frames that have the visual information. Our extensive experiments on the Pascal DOC19 and MS-DS17 datasets have implied that the HRT encoder–decoder abundant model outperforms those of previous studies involving few frames that include moving objects.
APA, Harvard, Vancouver, ISO, and other styles
8

Wang, Hongquan, Xinshan Zhu, Chao Ren, Lan Zhang, and Shugen Ma. "A Frequency Attention-Based Dual-Stream Network for Image Inpainting Forensics." Mathematics 11, no. 12 (2023): 2593. http://dx.doi.org/10.3390/math11122593.

Full text
Abstract:
The rapid development of digital image inpainting technology is causing serious hidden danger to the security of multimedia information. In this paper, a deep network called frequency attention-based dual-stream network (FADS-Net) is proposed for locating the inpainting region. FADS-Net is established by a dual-stream encoder and an attention-based blue-associative decoder. The dual-stream encoder includes two feature extraction streams, the raw input stream (RIS) and the frequency recalibration stream (FRS). RIS directly captures feature maps from the raw input, while FRS performs feature extraction after recalibrating the input via learning in the frequency domain. In addition, a module based on dense connection is designed to ensure efficient extraction and full fusion of dual-stream features. The attention-based associative decoder consists of a main decoder and two branch decoders. The main decoder performs up-sampling and fine-tuning of fused features by using attention mechanisms and skip connections, and ultimately generates the predicted mask for the inpainted image. Then, two branch decoders are utilized to further supervise the training of two feature streams, ensuring that they both work effectively. A joint loss function is designed to supervise the training of the entire network and two feature extraction streams for ensuring optimal forensic performance. Extensive experimental results demonstrate that the proposed FADS-Net achieves superior localization accuracy and robustness on multiple datasets compared to the state-of-the-art inpainting forensics methods.
APA, Harvard, Vancouver, ISO, and other styles
9

Li, Xin, Feng Xu, Runliang Xia, et al. "Encoding Contextual Information by Interlacing Transformer and Convolution for Remote Sensing Imagery Semantic Segmentation." Remote Sensing 14, no. 16 (2022): 4065. http://dx.doi.org/10.3390/rs14164065.

Full text
Abstract:
Contextual information plays a pivotal role in the semantic segmentation of remote sensing imagery (RSI) due to the imbalanced distributions and ubiquitous intra-class variants. The emergence of the transformer intrigues the revolution of vision tasks with its impressive scalability in establishing long-range dependencies. However, the local patterns, such as inherent structures and spatial details, are broken with the tokenization of the transformer. Therefore, the ICTNet is devised to confront the deficiencies mentioned above. Principally, ICTNet inherits the encoder–decoder architecture. First of all, Swin Transformer blocks (STBs) and convolution blocks (CBs) are deployed and interlaced, accompanied by encoded feature aggregation modules (EFAs) in the encoder stage. This design allows the network to learn the local patterns and distant dependencies and their interactions simultaneously. Moreover, multiple DUpsamplings (DUPs) followed by decoded feature aggregation modules (DFAs) form the decoder of ICTNet. Specifically, the transformation and upsampling loss are shrunken while recovering features. Together with the devised encoder and decoder, the well-rounded context is captured and contributes to the inference most. Extensive experiments are conducted on the ISPRS Vaihingen, Potsdam and DeepGlobe benchmarks. Quantitative and qualitative evaluations exhibit the competitive performance of ICTNet compared to mainstream and state-of-the-art methods. Additionally, the ablation study of DFA and DUP is implemented to validate the effects.
APA, Harvard, Vancouver, ISO, and other styles
10

Geng, Yaogang, Hongyan Mei, Xiaorong Xue, and Xing Zhang. "Image-Caption Model Based on Fusion Feature." Applied Sciences 12, no. 19 (2022): 9861. http://dx.doi.org/10.3390/app12199861.

Full text
Abstract:
The encoder–decoder framework is the main frame of image captioning. The convolutional neural network (CNN) is usually used to extract grid-level features of the image, and the graph convolutional neural network (GCN) is used to extract the image’s region-level features. Grid-level features are poor in semantic information, such as the relationship and location of objects, while regional features lack fine-grained information about images. To address this problem, this paper proposes a fusion-features-based image-captioning model, which includes the fusion feature encoder and LSTM decoder. The fusion-feature encoder is divided into grid-level feature encoder and region-level feature encoder. The grid-level feature encoder is a convoluted neural network embedded in squeeze and excitation operations so that the model can focus on features that are highly correlated to the title. The region-level encoder employs node-embedding matrices to enable models to understand different node types and gain richer semantics. Then the features are weighted together by an attention mechanism to guide the decoder LSTM to generate an image caption. Our model was trained and tested in the MS COCO2014 dataset with the experimental evaluation standard Bleu-4 score and CIDEr score of 0.399 and 1.311, respectively. The experimental results indicate that the model can describe the image in detail.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Encoder and decoder feature"

1

Kalchbrenner, Nal. "Encoder-decoder neural networks." Thesis, University of Oxford, 2017. http://ora.ox.ac.uk/objects/uuid:d56e48db-008b-4814-bd82-a5d612000de9.

Full text
Abstract:
This thesis introduces the concept of an encoder-decoder neural network and develops architectures for the construction of such networks. Encoder-decoder neural networks are probabilistic conditional generative models of high-dimensional structured items such as natural language utterances and natural images. Encoder-decoder neural networks estimate a probability distribution over structured items belonging to a target set conditioned on structured items belonging to a source set. The distribution over structured items is factorized into a product of tractable conditional distributions over individual elements that compose the items. The networks estimate these conditional factors explicitly. We develop encoder-decoder neural networks for core tasks in natural language processing and natural image and video modelling. In Part I, we tackle the problem of sentence modelling and develop deep convolutional encoders to classify sentences; we extend these encoders to models of discourse. In Part II, we go beyond encoders to study the longstanding problem of translating from one human language to another. We lay the foundations of neural machine translation, a novel approach that views the entire translation process as a single encoder-decoder neural network. We propose a beam search procedure to search over the outputs of the decoder to produce a likely translation in the target language. Besides known recurrent decoders, we also propose a decoder architecture based solely on convolutional layers. Since the publication of these new foundations for machine translation in 2013, encoder-decoder translation models have been richly developed and have displaced traditional translation systems both in academic research and in large-scale industrial deployment. In services such as Google Translate these models process in the order of a billion translation queries a day. In Part III, we shift from the linguistic domain to the visual one to study distributions over natural images and videos. We describe two- and three- dimensional recurrent and convolutional decoder architectures and address the longstanding problem of learning a tractable distribution over high-dimensional natural images and videos, where the likely samples from the distribution are visually coherent. The empirical validation of encoder-decoder neural networks as state-of- the-art models of tasks ranging from machine translation to video prediction has a two-fold significance. On the one hand, it validates the notions of assigning probabilities to sentences or images and of learning a distribution over a natural language or a domain of natural images; it shows that a probabilistic principle of compositionality, whereby a high- dimensional item is composed from individual elements at the encoder side and whereby a corresponding item is decomposed into conditional factors over individual elements at the decoder side, is a general method for modelling cognition involving high-dimensional items; and it suggests that the relations between the elements are best learnt in an end-to-end fashion as non-linear functions in distributed space. On the other hand, the empirical success of the networks on the tasks characterizes the underlying cognitive processes themselves: a cognitive process as complex as translating from one language to another that takes a human a few seconds to perform correctly can be accurately modelled via a learnt non-linear deterministic function of distributed vectors in high-dimensional space.
APA, Harvard, Vancouver, ISO, and other styles
2

Padinjare, Sainath. "VLSI implementation of a turbo encoder/decoder /." Internet access available to MUN users only, 2003. http://collections.mun.ca/u?/theses,162832.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Weitzman, Jonathan M. "SELECTABLE PERMUTATION ENCODER/DECODER FOR A QPSK MODEM." International Foundation for Telemetering, 2003. http://hdl.handle.net/10150/605817.

Full text
Abstract:
International Telemetering Conference Proceedings / October 20-23, 2003 / Riviera Hotel and Convention Center, Las Vegas, Nevada<br>An artifact of QPSK modems is ambiguity of the recovered data. There are four variations of the output data for a given input data stream. All are equally probable. To resolve this ambiguity, the QPSK data streams can be differentially encoded before modulation and differentially decoded after demodulation. The encoder maps each input data pair to a phase angle change of the QPSK carrier. In the demodulator, the inverse is performed - each phase change of the input QPSK carrier is mapped to an output data pair. This paper discusses a very simple and unique differential encoder/decoder that handles all possible data pair/phase change permutations.
APA, Harvard, Vancouver, ISO, and other styles
4

Mejdi, Sami. "Encoder-Decoder Networks for Cloud Resource Consumption Forecasting." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-291546.

Full text
Abstract:
Excessive resource allocation in telecommunications networks can be prevented by forecasting the resource demand when dimensioning the networks and the allocation the necessary resources accordingly, which is an ongoing effort to achieve a more sustainable development. In this work, traffic data from cloud environments that host deployed virtualized network functions (VNFs) of an IP Multimedia Subsystem (IMS) has been collected along with the computational resource consumption of the VNFs. A supervised learning approach was adopted to address the forecasting problem by considering encoder-decoder networks. These networks were applied to forecast future resource consumption of the VNFs by regarding the problem as a time series forecasting problem, and recasting it as a sequence-to-sequence (seq2seq) problem. Different encoder-decoder network architectures were then utilized to forecast the resource consumption. The encoder-decoder networks were compared against a widely deployed classical time series forecasting model that served as a baseline model. The results show that while the considered encoder-decoder models failed to outperform the baseline model in overall Root Mean Squared Error (RMSE) and Mean Absolute Error (MAE), the forecasting capabilities were more resilient to degradation over time. This suggests that the encoder-decoder networks are more appropriate for long-term forecasting, which is an agreement with related literature. Furthermore, the encoder-decoder models achieved competitive performance when compared to the baseline, despite being treated with limited hyperparameter-tuning and the absence of more sophisticated functionality such as attention. This work has shown that there is indeed potential for deep learning applications in forecasting of cloud resource consumption.<br>Överflödig allokering av resurser I telekommunikationsnätverk kan förhindras genom att prognosera resursbehoven vid dimensionering av dessa nätverk. Detta görs i syfte att bidra till en mer hållbar utveckling. Inför detta prjekt har trafikdata från molnmiljön som hyser aktiva virtuella komponenter (VNFs) till ett IÅ Multimedia Subsystem (IMS) samlats in tillsammans med resursförbrukningen av dessa komponenter. Detta examensarbete avhandlar hur effektivt övervakad maskininlärning i form av encoder-decoder nätverk kan användas för att prognosera resursbehovet hos ovan nämnda VNFs. Encoder-decoder nätverken appliceras genom att betrakta den samlade datan som en tidsserie. Problemet med att förutspå utvecklingen av tidsserien formuleras sedan som ett sequence-2-sequence (seq2seq) problem. I detta arbete användes en samling encoder-decoder nätverk med olika arkitekturer för att prognosera resursförbrukningen och dessa jämfördes med en populär modell hämtad från klassisk tidsserieanalys. Resultaten visar att encoder-decoder nätverken misslyckades med att överträffa den klassiska tidsseriemodellen med avseende på Root Mean Squeared Error (RMSE) och Mean Absolut Error (MAE). Dock visar encoder-decoder nätverken en betydlig motståndskraft mot prestandaförfall över tid i jämförelse med den klassiska tidsseriemodellen. Detta indikerar att encoder-decoder nätverk är lämpliga för prognosering över en längre tidshorisont. Utöver detta visade encoder-decoder nätverken en konkurrenskraftig förmåga att förutspå det korrekta resursbehovet, trots en begränsad justering av disponeringsparametrarna och utan mer sofistikerad funktionalitet implementerad som exempelvis attention.
APA, Harvard, Vancouver, ISO, and other styles
5

Mejdi, Sami. "Encoder-Decoder Networks for Cloud Resource Consumption Forecasting." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-294066.

Full text
Abstract:
Excessive resource allocation in telecommunications networks can be prevented by forecasting the resource demand when dimensioning the networks and then allocating the necessary resources accordingly, which is an ongoing effort to achieve a more sustainable development. In this work, traffic data from cloud environments that host deployed virtualized network functions (VNFs) of an IP Multimedia Subsystem (IMS) has been collected along with the computational resource consumption of the VNFs. A supervised learning approach was adopted to address the forecasting problem by considering encoder-decoder networks. These networks were applied to forecast future resource consumption of the VNFs by regarding the problem as a time series forecasting problem, and recasting it as a sequence-to-sequence (seq2seq) problem. Different encoder-decoder network architectures were then utilized to forecast the resource consumption. The encoder-decoder networks were compared against a widely deployed classical time series forecasting model that served as a baseline model. The results show that while the considered encoder-decoder models failed to outperform the baseline model in overall Root Mean Squared Error (RMSE) and Mean Absolute Error (MAE), the forecasting capabilities were more resilient to degradation over time. This suggests that the encoder-decoder networks are more appropriate for long-term forecasting, which is in agreement with related literature. Furthermore, the encoder-decoder models achieved competitive performance when compared to the baseline, despite being treated with limited hyperparameter-tuning and the absence of more sophisticated functionality such as attention. This work has shown that there is indeed potential for deep learning applications in forecasting of cloud resource consumption.<br>Överflödig allokering av resurser i telekommunikationsnätverk kan förhindras genom att prognosera resursbehoven vid dimensionering av dessa nätverk. Detta görs i syfte att bidra till en mer hållbar utveckling. Infor  detta  projekt har  trafikdata från molnmiljon som hyser aktiva virtuella komponenter (VNFs) till ett  IP Multimedia Subsystem (IMS) samlats in tillsammans med resursförbrukningen  av dessa komponenter. Detta examensarbete avhandlar hur effektivt övervakad maskininlärning i form av encoder-decoder natverk kan användas för att prognosera resursbehovet hos ovan nämnda VNFs. Encoder-decoder nätverken appliceras genom att betrakta den samlade datan som en tidsserie. Problemet med att förutspå utvecklingen av tidsserien formuleras sedan som ett sequence-to-sequence (seq2seq) problem. I detta arbete användes en samling encoder-decoder nätverk med olika arkitekturer for att prognosera resursförbrukningen och dessa jämfördes med en populär modell hämtad från klassisk tidsserieanalys. Resultaten visar att encoder- decoder nätverken misslyckades med att överträffa den klassiska tidsseriemodellen med avseende på Root Mean Squared Error (RMSE) och Mean Absolute Error (MAE). Dock visade encoder-decoder nätverken en betydlig motståndskraft mot prestandaförfall över tid i jämförelse med den klassiska tidsseriemodellen. Detta indikerar att encoder-decoder nätverk är lämpliga för prognosering över en längre tidshorisont. Utöver detta visade encoder-decoder nätverken en konkurrenskraftig förmåga att förutspå det korrekta resursbehovet, trots en begränsad justering av disponeringsparametrarna och utan mer sofistikerad funktionalitet implementerad som exempelvis attention.
APA, Harvard, Vancouver, ISO, and other styles
6

Correia, Tiago Miguel Pina. "FPGA implementation of Alamouti encoder/decoder for LTE." Master's thesis, Universidade de Aveiro, 2013. http://hdl.handle.net/10773/12679.

Full text
Abstract:
Mestrado em Engenharia Electrónica e Telecomunicações<br>Motivados por transmissões mais rápidas e mais fiáveis num canal sem fios, os sistemas da 4G devem proporcionar processamento de dados mais rápido a baixa complexidade, elevadas taxas de dados, assim como robustez na performance reduzindo também, a latência e os custos de operação. LTE apresenta, na sua camada física, tecnologias como OFDM e MIMO que prometem alcançar elevadas taxas de dados e aumentar a eficiência espectral. Especificamente a camada física do LTE emprega OFDMA para downlink e SC-FDMA para uplink. A tecnologia MIMO permite também melhorar significativamente o desempenho dos sistemas OFDM com as vantagens de multiplexação e diversidade espacial diminuindo o efeito de desvanecimento de multi-percurso no canal. Nesta dissertação são implementados um codificador e um descodificador com base no algoritimo de Alamouti num sistema MISO nomeadamente para serem incluídos num OFDM transceiver que segue as especificações da camada física do LTE. A codificação/descodificação de Alamouti realiza-se no espaço e frequência e os blocos foram projetados e simulados em Matlab através do ambiente Simulink com o auxílio dos blocos da Xilinx inseridos no seu software System Generator para DSP. Pode-se concluir que os blocos baseados no algoritmo de Alamouti foram implementados em hardware com sucesso.<br>Motivated by faster transmissions and more reliable wireless channel, future 4G systems should provide faster data processing at low complexity, high data rates, as well as robustness in performance while also reducing the latency and operating costs. LTE presents in its physical layer technologies such as OFDM and MIMO that promise to achieve high data rates and increase spectral efficiency. Specifically the physical layer of LTE employs OFDMA on the downlink and SC-FDMA for uplink. MIMO technology also allows to significantly improve the performance of OFDM systems with the advantages of multiplexing and spatial diversity by decreasing the effect of multipath fading in the channel. In this thesis we implemented an encoder and a decoder based on an Alamouti algorithm in a MISO system namely to be added to an OFDM transceiver that follows closely the LTE physical layer specifications. Alamouti coding/decoding is performed in frequency and space and the blocks were projected and simulated in Matlab using Simulink environment through the Xilink's blocks in the System Generator for DSP. One can conclude that the blocks based on Alamouti algorithm were well-implemented.
APA, Harvard, Vancouver, ISO, and other styles
7

Leivas, Oliveira Gabriel [Verfasser], Thomas [Akademischer Betreuer] Brox, and Wolfram [Akademischer Betreuer] Burgard. "Encoder-decoder methods for semantic segmentation: efficiency and robustness aspects." Freiburg : Universität, 2019. http://d-nb.info/1191689476/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Kopparthi, Sunitha. "Flexible encoder and decoder designs for low-density parity-check codes." Diss., Manhattan, Kan. : Kansas State University, 2010. http://hdl.handle.net/2097/4190.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Pisacane, Claudia. "Skopos Theory La figura del traduttore come decoder e re-encoder." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2015. http://amslaurea.unibo.it/8926/.

Full text
Abstract:
La Skopos Theory è una teoria introdotta nel mondo della traduzione dal linguista tedesco Hans Joseph Vermeer. Skopos è una parola di derivazione greca che significa “fine” o “scopo”. La teoria elaborata da Vermeer si basa sull’idea che ogni testo abbia uno skopos che determina i metodi e le strategie secondo le quali esso debba essere tradotto. Oltre alla Skopos Theory, che sarà la base della tesi, i testi a seguire verranno analizzati seguendo altri autori, quali Mona Baker e Laurence Venuti, che si rifanno all’idea di skopos e analizzano molto dettagliatamente la figura del traduttore come de-coder e re-encoder del testo.
APA, Harvard, Vancouver, ISO, and other styles
10

Nina, Oliver A. Nina. "A Multitask Learning Encoder-N-Decoder Framework for Movie and Video Description." The Ohio State University, 2018. http://rave.ohiolink.edu/etdc/view?acc_num=osu1531996548147165.

Full text
APA, Harvard, Vancouver, ISO, and other styles
More sources

Books on the topic "Encoder and decoder feature"

1

Jet Propulsion Laboratory (U.S.), ed. A software simulation study of a (255,223) Reed-Solomon encoder/decoder. National Aeronautics and Space Administration, Jet Propulsion Laboratory, California Institute of Technology, 1985.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
2

Sari, Mehmet. Designing fast Golay encoder/decoder in Xilinx XACT with Mentor Graphics CAD interface. Naval Postgraduate School, 1997.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
3

Madadi, H. Assessment of a wide area digital paging system (microprocessor based decoder/encoder) and VHF Channel characterization for urban and suburban areas. University of Birmingham, 1985.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
4

McKenzie, Stephen Scott. A systolic array implementation of a Reed-Solomon encoder and decoder. 1985.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
5

Bluedorn, Harvey. Handy English Encoder Decoder: All the Spelling and Phonics Rules You Could Ever Want to Know. Trivium Pursuit, 2004.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
6

Wireless Radio Frequency Module Using PIC Microcontroller.: The Basics of Wireless Radio Frequency Communications With Using Latest & An Advanced MCU Named PIC. Public Domain, 2012.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
7

Lamel, Lori, and Jean-Luc Gauvain. Speech Recognition. Edited by Ruslan Mitkov. Oxford University Press, 2012. http://dx.doi.org/10.1093/oxfordhb/9780199276349.013.0016.

Full text
Abstract:
Speech recognition is concerned with converting the speech waveform, an acoustic signal, into a sequence of words. Today's approaches are based on a statistical modellization of the speech signal. This article provides an overview of the main topics addressed in speech recognition, which are, acoustic-phonetic modelling, lexical representation, language modelling, decoding, and model adaptation. Language models are used in speech recognition to estimate the probability of word sequences. The main components of a generic speech recognition system are, main knowledge sources, feature analysis, and acoustic and language models, which are estimated in a training phase, and the decoder. The focus of this article is on methods used in state-of-the-art speaker-independent, large-vocabulary continuous speech recognition (LVCSR). Primary application areas for such technology are dictation, spoken language dialogue, and transcription for information archival and retrieval systems. Finally, this article discusses issues and directions of future research.
APA, Harvard, Vancouver, ISO, and other styles

Book chapters on the topic "Encoder and decoder feature"

1

Gong, Yansheng, and Wenfeng Jing. "A Fully-Nested Encoder-Decoder Framework for Anomaly Detection." In Proceeding of 2021 International Conference on Wireless Communications, Networking and Applications. Springer Nature Singapore, 2022. http://dx.doi.org/10.1007/978-981-19-2456-9_75.

Full text
Abstract:
AbstractAnomaly detection is an important branch of computer vision. At present, a variety of deep learning models are applied to anomaly detection. However, the lack of abnormal samples makes supervised learning difficult to implement. In this paper, we mainly study abnormal detection tasks based on unsupervised learning and propose a Fully-Nested Encoder-decoder Framework. The main part of the proposed generating model consists of a generator and a discriminator, which are adversarially trained based on normal data samples. In order to improve the image reconstruction capability of the generator, we design a Fully-Nested Residual Encoder-decoder Network, which is used to encode and decode the images. In addition, we add residual structure into both encoder and decoder, which reduces the risk of overfitting and enhances the feature expression ability. In the test phase, a distance measurement model is used to determine whether the test sample is abnormal. The experimental results on the CIFAR-10 dataset demonstrate the excellent performance of our method. Compared with the existing models, our method achieves the state-of-the-art result.
APA, Harvard, Vancouver, ISO, and other styles
2

Liu, Hongyu, Bin Jiang, Yibing Song, Wei Huang, and Chao Yang. "Rethinking Image Inpainting via a Mutual Encoder-Decoder with Feature Equalizations." In Computer Vision – ECCV 2020. Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-58536-5_43.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Li, Mingxiao. "A High-Efficiency Knowledge Distillation Image Caption Technology." In Proceeding of 2021 International Conference on Wireless Communications, Networking and Applications. Springer Nature Singapore, 2022. http://dx.doi.org/10.1007/978-981-19-2456-9_92.

Full text
Abstract:
AbstractImage caption is wildly considered in the application of machine learning. Its purpose is describing one given picture into text accurately. Currently, it uses the Encoder-Decoder architecture from deep learning. To further increase the semantic transmitted after distillation by feature representation, this paper proposes a knowledge distillation framework to increase the results of the teacher section, extracting features by different semantic levels from different fields of view, and the loss function adopts the method of label normalization. Handle unmatched image-sentence pairs. In order to achieve the purpose of a more efficient process. Experimental results prove that this knowledge distillation architecture can strengthen the semantic information transmitted after distillation in the feature representation, achieve a more efficient training model on less data, and obtain a higher accuracy rate.
APA, Harvard, Vancouver, ISO, and other styles
4

Liu, Hongyu, Bin Jiang, Yibing Song, Wei Huang, and Chao Yang. "Correction to: Rethinking Image Inpainting via a Mutual Encoder-Decoder with Feature Equalizations." In Computer Vision – ECCV 2020. Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-58536-5_47.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Duan, Lijuan, Xuan Feng, Jie Chen, and Fan Xu. "An Automated Method with Feature Pyramid Encoder and Dual-Path Decoder for Nuclei Segmentation." In Pattern Recognition and Computer Vision. Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-60633-6_28.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Qiu, Yi, Long Cheng, Man Xu, Jing Chen, and Hongjie Wu. "Feature Extraction Approach for Predicting Protein-DNA Binding Residues Using Transformer Encoder-Decoder Architecture." In Advanced Intelligent Computing in Bioinformatics. Springer Nature Singapore, 2024. http://dx.doi.org/10.1007/978-981-97-5689-6_21.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Li, Zihan, Wei Ding, Inal Mashukov, Scott Crouter, and Ping Chen. "A Multi-view Feature Construction and Multi-Encoder-Decoder Transformer Architecture for Time Series Classification." In Advances in Knowledge Discovery and Data Mining. Springer Nature Singapore, 2024. http://dx.doi.org/10.1007/978-981-97-2266-2_19.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Lu, Xuesong, and Yuchuan Qiao. "Multi-channel Image Registration of Cardiac MR Using Supervised Feature Learning with Convolutional Encoder-Decoder Network." In Biomedical Image Registration. Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-50120-4_10.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Sellami, Akrem, and Salvatore Tabbone. "EDNets: Deep Feature Learning for Document Image Classification Based on Multi-view Encoder-Decoder Neural Networks." In Document Analysis and Recognition – ICDAR 2021. Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-86337-1_22.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Yang, Liu, Boyu Wang, Jack C. P. Cheng, Peipei Liu, and Hoon Sohn. "Real-Time Geometry Assessment Using Laser Line Scanner During Laser Powder Directed Energy Deposition Additive Manufacturing of SS316L Component with Sharp Feature." In CONVR 2023 - Proceedings of the 23rd International Conference on Construction Applications of Virtual Reality. Firenze University Press, 2023. http://dx.doi.org/10.36253/10.36253/979-12-215-0289-3.97.

Full text
Abstract:
Directed energy deposition (DED) is a major metal additive manufacturing (AM) technology that is increasingly used in many industries due to its ability to manufacture complex components of arbitrary shapes and sizes. However, a lack of timely geometry assessment and the consequent geometry control hinders the development of DED towards zero defect manufacturing. In this study, a real-time geometry assessment methodology is developed for laser pow-der directed energy deposition (LP-DED). A geometry assessment system is developed using a laser line scanner capable of inspecting the melt pool area, the just solidified area, as well as layer-wise inspection. An image processing method with an encoder-decoder based profile completion network was developed to obtain accurate track profile in images from real-time inspection. Experiments have been conducted to validate the proposed methodology by depositing multi-layer X-shape objects
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Encoder and decoder feature"

1

Ramon, Raz, Hadar Cohen-Duwek, and Elishai Ezra Tsur. "ED-DCFNet: an unsupervised encoder-decoder neural model for event-driven feature extraction and object tracking." In 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE, 2024. http://dx.doi.org/10.1109/cvprw63382.2024.00224.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Liu, Boyu. "Comparative Analysis of Encoder-Only, Decoder-Only, and Encoder- Decoder Language Models." In International Conference on Data Science and Engineering. SCITEPRESS - Science and Technology Publications, 2024. http://dx.doi.org/10.5220/0012829800004547.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Zhang, Jinchao, Qun Liu, and Jie Zhou. "ME-MD: An Effective Framework for Neural Machine Translation with Multiple Encoders and Decoders." In Twenty-Sixth International Joint Conference on Artificial Intelligence. International Joint Conferences on Artificial Intelligence Organization, 2017. http://dx.doi.org/10.24963/ijcai.2017/474.

Full text
Abstract:
The encoder-decoder neural framework is widely employed for Neural Machine Translation (NMT) with a single encoder to represent the source sentence and a single decoder to generate target words. The translation performance heavily relies on the representation ability of the encoder and the generation ability of the decoder. To further enhance NMT, we propose to extend the original encoder-decoder framework to a novel one, which has multiple encoders and decoders (ME-MD). Through this way, multiple encoders extract more diverse features to represent the source sequence and multiple decoders capture more complicated translation knowledge. Our proposed ME-MD framework is convenient to integrate heterogeneous encoders and decoders with multiple depths and multiple types. Experiment on Chinese-English translation task shows that our ME-MD system surpasses the state-of-the-art NMT system by 2.1 BLEU points and surpasses the phrase-based Moses by 7.38 BLEU points. Our framework is general and can be applied to other sequence to sequence tasks.
APA, Harvard, Vancouver, ISO, and other styles
4

SharifiPour, Sasan, Hossein Fayyazi, and Mohammad Sabokro. "Unsupervised Feature Selection using Encoder-Decoder Networks." In 2020 6th Iranian Conference on Signal Processing and Intelligent Systems (ICSPIS). IEEE, 2020. http://dx.doi.org/10.1109/icspis51611.2020.9349608.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Wang, Qixiang, Shanfeng Wang, Maoguo Gong, and Yue Wu. "Feature Hashing for Network Representation Learning." In Twenty-Seventh International Joint Conference on Artificial Intelligence {IJCAI-18}. International Joint Conferences on Artificial Intelligence Organization, 2018. http://dx.doi.org/10.24963/ijcai.2018/390.

Full text
Abstract:
The goal of network representation learning is to embed nodes so as to encode the proximity structures of a graph into a continuous low-dimensional feature space. In this paper, we propose a novel algorithm called node2hash based on feature hashing for generating node embeddings. This approach follows the encoder-decoder framework. There are two main mapping functions in this framework. The first is an encoder to map each node into high-dimensional vectors. The second is a decoder to hash these vectors into a lower dimensional feature space. More specifically, we firstly derive a proximity measurement called expected distance as target which combines position distribution and co-occurrence statistics of nodes over random walks so as to build a proximity matrix, then introduce a set of T different hash functions into feature hashing to generate uniformly distributed vector representations of nodes from the proximity matrix. Compared with the existing state-of-the-art network representation learning approaches, node2hash shows a competitive performance on multi-class node classification and link prediction tasks on three real-world networks from various domains.
APA, Harvard, Vancouver, ISO, and other styles
6

Yu, Yunlong, Dingyi Zhang, and Zhong Ji. "Masked Feature Generation Network for Few-Shot Learning." In Thirty-First International Joint Conference on Artificial Intelligence {IJCAI-22}. International Joint Conferences on Artificial Intelligence Organization, 2022. http://dx.doi.org/10.24963/ijcai.2022/513.

Full text
Abstract:
In this paper, we present a feature-augmentation approach called Masked Feature Generation Network (MFGN) for Few-Shot Learning (FSL), a challenging task that attempts to recognize the novel classes with a few visual instances for each class. Most of the feature-augmentation approaches tackle FSL tasks via modeling the intra-class distributions. We extend this idea further to explicitly capture the intra-class variations in a one-to-many manner. Specifically, MFGN consists of an encoder-decoder architecture, with an encoder that performs as a feature extractor and extracts the feature embeddings of the available visual instances (the unavailable instances are seen to be masked), along with a decoder that performs as a feature generator and reconstructs the feature embeddings of the unavailable visual instances from both the available feature embeddings and the masked tokens. Equipped with this generative architecture, MFGN produces nontrivial visual features for the novel classes with limited visual instances. In extensive experiments on four FSL benchmarks, MFGN performs competitively and outperforms the state-of-the-art competitors on most of the few-shot classification tasks.
APA, Harvard, Vancouver, ISO, and other styles
7

Chen, Xueying, Rong Zhang, and Pingkun Yan. "Feature Fusion Encoder Decoder Network for Automatic Liver Lesion Segmentation." In 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI). IEEE, 2019. http://dx.doi.org/10.1109/isbi.2019.8759555.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Tao, Zhiqiang, Hongfu Liu, Jun Li, Zhaowen Wang, and Yun Fu. "Adversarial Graph Embedding for Ensemble Clustering." In Twenty-Eighth International Joint Conference on Artificial Intelligence {IJCAI-19}. International Joint Conferences on Artificial Intelligence Organization, 2019. http://dx.doi.org/10.24963/ijcai.2019/494.

Full text
Abstract:
Ensemble clustering generally integrates basic partitions into a consensus one through a graph partitioning method, which, however, has two limitations: 1) it neglects to reuse original features; 2) obtaining consensus partition with learnable graph representations is still under-explored. In this paper, we propose a novel Adversarial Graph Auto-Encoders (AGAE) model to incorporate ensemble clustering into a deep graph embedding process. Specifically, graph convolutional network is adopted as probabilistic encoder to jointly integrate the information from feature content and consensus graph, and a simple inner product layer is used as decoder to reconstruct graph with the encoded latent variables (i.e., embedding representations). Moreover, we develop an adversarial regularizer to guide the network training with an adaptive partition-dependent prior. Experiments on eight real-world datasets are presented to show the effectiveness of AGAE over several state-of-the-art deep embedding and ensemble clustering methods.
APA, Harvard, Vancouver, ISO, and other styles
9

Liu, Yang, Deyan Xie, Quanxue Gao, Jungong Han, Shujian Wang, and Xinbo Gao. "Graph and Autoencoder Based Feature Extraction for Zero-shot Learning." In Twenty-Eighth International Joint Conference on Artificial Intelligence {IJCAI-19}. International Joint Conferences on Artificial Intelligence Organization, 2019. http://dx.doi.org/10.24963/ijcai.2019/421.

Full text
Abstract:
Zero-shot learning (ZSL) aims to build models to recognize novel visual categories that have no associated labelled training samples. The basic framework is to transfer knowledge from seen classes to unseen classes by learning the visual-semantic embedding. However, most of approaches do not preserve the underlying sub-manifold of samples in the embedding space. In addition, whether the mapping can precisely reconstruct the original visual feature is not investigated in-depth. In order to solve these problems, we formulate a novel framework named Graph and Autoencoder Based Feature Extraction (GAFE) to seek a low-rank mapping to preserve the sub-manifold of samples. Taking the encoder-decoder paradigm, the encoder part learns a mapping from the visual feature to the semantic space, while decoder part reconstructs the original features with the learned mapping. In addition, a graph is constructed to guarantee the learned mapping can preserve the local intrinsic structure of the data. To this end, an L21 norm sparsity constraint is imposed on the mapping to identify features relevant to the target domain. Extensive experiments on five attribute datasets demonstrate the effectiveness of the proposed model.
APA, Harvard, Vancouver, ISO, and other styles
10

Lin, Zhihui, Chun Yuan, and Maomao Li. "HAF-SVG: Hierarchical Stochastic Video Generation with Aligned Features." In Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20}. International Joint Conferences on Artificial Intelligence Organization, 2020. http://dx.doi.org/10.24963/ijcai.2020/138.

Full text
Abstract:
Stochastic video generation methods predict diverse videos based on observed frames, where the main challenge lies in modeling the complex future uncertainty and generating realistic frames. Numerous of Recurrent-VAE-based methods have achieved state-of-the-art results. However, on the one hand, the independence assumption of the variables of approximate posterior limits the inference performance. On the other hand, although these methods adopt skip connections between encoder and decoder to utilize multi-level features, they still produce blurry generation due to the spatial misalignment between encoder and decoder features at different time steps. In this paper, we propose a hierarchical recurrent VAE with a feature aligner, which can not only relax the independence assumption in typical VAE but also use a feature aligner to enable the decoder to obtain the aligned spatial information from the last observed frames. The proposed model is named Hierarchical Stochastic Video Generation network with Aligned Features, referred to as HAF-SVG. Experiments on Moving-MNIST, BAIR, and KTH datasets demonstrate that hierarchical structure is helpful for modeling more accurate future uncertainty, and the feature aligner is beneficial to generate realistic frames. Besides, the HAF-SVG exceeds SVG on both prediction accuracy and the quality of generated frames.
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Encoder and decoder feature"

1

Johnson, Zachary. Heatmap Router: An encoder-decoder approach to PCB routing. Iowa State University, 2022. http://dx.doi.org/10.31274/cc-20240624-819.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Ramakrishnan, Aravind, Fangyu Liu, Angeli Jayme, and Imad Al-Qadi. Prediction of Pavement Damage under Truck Platoons Utilizing a Combined Finite Element and Artificial Intelligence Model. Illinois Center for Transportation, 2024. https://doi.org/10.36501/0197-9191/24-030.

Full text
Abstract:
For robust pavement design, accurate damage computation is essential, especially for loading scenarios such as truck platoons. Studies have developed a framework to compute pavement distresses as function of lateral position, spacing, and market-penetration level of truck platoons. The established framework uses a robust 3D pavement model, along with the AASHTOWare Mechanistic–Empirical Pavement Design Guidelines (MEPDG) transfer functions to compute pavement distresses. However, transfer functions include high variability and lack physical significance. Therefore, as an improvement to effectively predict permanent deformation, this study utilized a conventional Burger’s model, incorporating a nonlinear power-law dashpot, in lieu of a transfer function. Key components, including stress increments and the Jacobian, were derived for implementation in ABAQUS as a user subroutine. Model parameters were determined through asphalt concrete (AC) flow number and dynamic modulus tests. Using a nonlinear power-law dashpot, the model accurately characterized rutting under varying conditions. The Burger’s model was both verified and validated to check the accuracy of implementation and representative of the actual behavior, respectively. Initially developed in 1D domain, the validated Burger’s model was integrated into the robust 3D finite element (FE) pavement model to predict permanent deformation. A new load-pass approach (LPA) enabled reduction in computational domain and cost, along with implementing transient loads more efficiently. The combined integration of the LPA and the Burger’s model into the pavement model effectively captured the rutting progression per loading cycle. Moreover, a graph neural network (GNN) was established to extend the prediction power of the framework, while strategically limiting the FE numerical matrix. The FE model data was transformed into a graph structure, converting FE model components into corresponding graph nodes and edges. The GNN-based pavement simulator (GPS) was developed to model 3D pavement responses, integrating three key components: encoder, processor, and decoder. The GPS model employed two-layer multilayer perceptrons (MLP) for the encoder and decoder, while utilizing graph network (GN) technology for the processor. Validation occurred through two case studies—OneStep and Rollout—with results compared against FE model data as ground truth. Results demonstrated that the GPS model provides an accurate and computationally efficient alternative to traditional 3D pavement FE simulations.
APA, Harvard, Vancouver, ISO, and other styles
3

Patwa, B., P. L. St-Charles, G. Bellefleur, and B. Rousseau. Predictive models for first arrivals on seismic reflection data, Manitoba, New Brunswick, and Ontario. Natural Resources Canada/CMSS/Information Management, 2022. http://dx.doi.org/10.4095/329758.

Full text
Abstract:
First arrivals are the primary waves picked and analyzed by seismologists to infer properties of the subsurface. Here we try to solve a problem in a small subsection of the seismic processing workflow: first break picking of seismic reflection data. We formulate this problem as an image segmentation task. Data is preprocessed, cleaned from outliers and extrapolated to make the training of deep learning models feasible. We use Fully Convolutional Networks (specifically UNets) to train initial models and explore their performance with losses, layer depths, and the number of classes. We propose to use residual connections to improve each UNet block and residual paths to solve the semantic gap between UNet encoder and decoder which improves the performance of the model. Adding spatial information as an extra channel helped increase the RMSE performance of the first break predictions. Other techniques like data augmentation, multitask loss, and normalization methods, were further explored to evaluate model improvement.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!