Academic literature on the topic 'Pytorch geometric'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Pytorch geometric.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Pytorch geometric"

1

Zaidi, Adnan Haider. "Federated Multi-Layer Energy Optimization for Earth-to-Orbit Smart Grid Systems GNN-RL Architecture." International Journal of Engineering Research and Applications 15, no. 6 (2025): 145–50. https://doi.org/10.9790/9622-1506145150.

Full text
Abstract:
This paper presents a Python-only framework using federated Graph Neural Networks (GNNs) and Reinforcement Learning (RL) for smart grid optimization from Earth to Low Earth Orbit (LEO). Built entirely in Google Colab using PyTorch, PyTorch-Geometric, and Flower, this modular solution enables learning across ground grids, UAVs, and satellite nodes using synthetic data. Six Jupyter notebooks simulate a real-time, multilayer smart energy network, achieving over 93% forecasting accuracy, fast recovery, and scalable control—all without proprietary software.
APA, Harvard, Vancouver, ISO, and other styles
2

Sønstebø, Jakob L., Herman Brunborg, and Mikkel Elle Lepperød. "Spikeometric: Linear Non-Linear Cascade Spiking Neural Networks with Pytorch Geometric." Journal of Open Source Software 8, no. 89 (2023): 5451. http://dx.doi.org/10.21105/joss.05451.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Sineglazov, Victor, and Kyrylo Bylym. "Twitter Fake News Detection Using Graph Neural Networks." Electronics and Control Systems 4, no. 78 (2023): 26–33. http://dx.doi.org/10.18372/1990-5548.78.18259.

Full text
Abstract:
This article is devoted to the intellectual processing of text information for the purpose of detecting rail news. To solve the given task, the use of deep graph neural networks is proposed. Fake news detection based on user preferences is augmented with deeper graph neural network topologies, including Hierarchical Graph Pooling with Structure Learning, to improve the graph convolution operation and capture richer contextual relationships in news graphs. The paper presents the possibilities of extending the framework of fake news detection based on user preferences using deep graph neural networks to improve fake news recognition. Evaluation on the FakeNewsNet dataset (a subset of Gossipcop) using the PyTorch Geometric and PyTorch Lightning frameworks demonstrates that the developed deep graph neural network model achieves 94% accuracy in fake news classification. The results show that deeper graph neural networks with integrated text and graph features offer promising options for reliable and accurate fake news detection, paving the way for improved information quality in social networks and beyond.
APA, Harvard, Vancouver, ISO, and other styles
4

Betkier, Igor, Mateusz Oszczypała, Janusz Pobożniak, Sergiusz Sobieski, and Przemysław Betkier. "PocketFinderGNN: A manufacturing feature recognition software based on Graph Neural Networks (GNNs) using PyTorch Geometric and NetworkX." SoftwareX 23 (July 2023): 101466. http://dx.doi.org/10.1016/j.softx.2023.101466.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Villanueva-Domingo, Pablo, Francisco Villaescusa-Navarro, Daniel Anglés-Alcázar, et al. "Inferring Halo Masses with Graph Neural Networks." Astrophysical Journal 935, no. 1 (2022): 30. http://dx.doi.org/10.3847/1538-4357/ac7aa3.

Full text
Abstract:
Abstract Understanding the halo–galaxy connection is fundamental in order to improve our knowledge on the nature and properties of dark matter. In this work, we build a model that infers the mass of a halo given the positions, velocities, stellar masses, and radii of the galaxies it hosts. In order to capture information from correlations among galaxy properties and their phase space, we use Graph Neural Networks (GNNs), which are designed to work with irregular and sparse data. We train our models on galaxies from more than 2000 state-of-the-art simulations from the Cosmology and Astrophysics with MachinE Learning Simulations project. Our model, which accounts for cosmological and astrophysical uncertainties, is able to constrain the masses of the halos with a ∼0.2 dex accuracy. Furthermore, a GNN trained on a suite of simulations is able to preserve part of its accuracy when tested on simulations run with a different code that utilizes a distinct subgrid physics model, showing the robustness of our method. The PyTorch Geometric implementation of the GNN is publicly available on GitHub (https://github.com/PabloVD/HaloGraphNet).
APA, Harvard, Vancouver, ISO, and other styles
6

Tang, T., T. Chen, B. Zhu, and Y. Ye. "MU-NET: A MULTISCALE UNSUPERVISED NETWORK FOR REMOTE SENSING IMAGE REGISTRATION." International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences XLIII-B3-2022 (May 30, 2022): 537–44. http://dx.doi.org/10.5194/isprs-archives-xliii-b3-2022-537-2022.

Full text
Abstract:
Abstract. Registration for multi-sensor or multi-modal image pairs with a large degree of distortions is a fundamental task for many remote sensing applications. To achieve accurate and low-cost remote sensing image registration, we propose a multiscale unsupervised network (MU-Net). Without costly ground truth labels, MU-Net directly learns the end-to-end mapping from the image pairs to their transformation parameters. MU-Net performs a coarse-to-fine registration pipeline by stacking several deep neural network models on multiple scales, which prevents the backpropagation being falling into a local extremum and resists significant image distortions. In addition, a novel loss function paradigm is designed based on structural similarity, which makes MU-Net suitable for various types of multi-modal images. MU-Net is compared with traditional feature-based and area-based methods, as well as supervised and other unsupervised learning methods on the Optical-Optical, Optical-Infrared, Optical-SAR and Optical-Map datasets. Experimental results show that MU-Net achieves more robust and accurate registration performance between these image pairs with geometric and radiometric distortions.We share the datasets and the code implemented by Pytorch at https://github.com/yeyuanxin110/MU-Net.
APA, Harvard, Vancouver, ISO, and other styles
7

Yendrapati Ravindra Babu. "Enhancing Distributed Workflow Optimization with Graph Neural Networks and Deep Learning Techniques." Journal of Information Systems Engineering and Management 10, no. 3 (2025): 1255–70. https://doi.org/10.52783/jisem.v10i3.8862.

Full text
Abstract:
This research presents a distributed processing framework using Graph Neural Networks (GNNs) for workflow scheduling and data routing in large-scale systems. Our adaptive GNN architecture dynamically models computing workflows, where nodes represent tasks and edges capture dependencies and communication patterns. Evaluated on Microsoft Azure Datacenter Traces (25 days, 11,000 machines) and Amazon AWS CloudWatch Metrics (10 days, 5,000 machines), our framework achieves 43% lower processing latency and 39% reduced memory footprint compared to DAG-based schedulers, maintaining 99.7% accuracy. The GNN-based topology optimization predicts optimal data routing paths with 91% accuracy, reducing storage overhead by 45% versus shortest-path algorithms (62% accuracy). Using PyTorch Geometric on a 180-node cluster, the system reduces network congestion by 35% and improves space utilization by 42% over baseline methods. Our multi-layer graph attention mechanism with dynamic edge weight updates accelerates workflow optimization by 46% while using 33% less memory. Under sudden workload variations, the framework sustains 92% performance stability and 98.5% data accuracy, surpassing traditional systems (68% stability). It achieves a time-space optimization ratio of 0.85 (vs. 0.62 in conventional systems), processing 1,100 tasks/hr with 95% resource efficiency. Additionally, it improves memory utilization by 41%, maintaining a ±0.3% accuracy deviation across workloads, setting new benchmarks in distributed processing.
APA, Harvard, Vancouver, ISO, and other styles
8

Lin, Yi-Hsin, Yu-Hung Ting, Yi-Cyun Huang, Kai-Lun Cheng, and Wen-Ren Jong. "Integration of Deep Learning for Automatic Recognition of 2D Engineering Drawings." Machines 11, no. 8 (2023): 802. http://dx.doi.org/10.3390/machines11080802.

Full text
Abstract:
In an environment where manufacturing precision requirements are increasing, complete project plans can consist of hundreds of engineering drawings. The presentation of these drawings often varies based on personal preferences, leading to inconsistencies in format and symbols. The lack of standardization in these aspects can result in inconsistent interpretations during subsequent analysis. Therefore, proper annotation of engineering drawings is crucial as it determines product quality, subsequent inspections, and processing costs. To reduce the time and cost associated with interpreting and analyzing drawings, as well as to minimize human errors in judgment, we developed an engineering drawing recognition system. This study employs geometric dimensioning and tolerancing (GD&T) in accordance with the ASME (American Society of Mechanical Engineers) Y14.5 2018 specification to describe the language of engineering drawings. Additionally, PyTorch, OpenCV, and You Only Look Once (YOLO) are utilized for training. Existing 2D engineering drawings serve as the training data, and image segmentation is performed to identify objects such as dimensions, tolerances, functional frames, and geometric symbols in the drawings using the network model. By reading the coordinates corresponding to each object, the correct values are displayed. Real-world cases are utilized to train the model with multiple engineering drawings containing mixed features, resulting in recognition capabilities surpassing those of single-feature identification. This approach improves the recognition accuracy of deep learning models and makes engineering drawing and image recognition more practical. The recognition results are directly stored in a database, reducing product verification time and preventing errors that may occur due to manual data entry, thereby avoiding subsequent quality control issues. The accuracy rates achieved are as follows: 85% accuracy in detecting views in 2D engineering drawings, 70% accuracy in detecting annotation groups and annotations, and 80% accuracy in text and symbol recognition.
APA, Harvard, Vancouver, ISO, and other styles
9

Frank, Franz, and Fernando Bacao. "Advanced Genetic Programming vs. State-of-the-Art AutoML in Imbalanced Binary Classification." Emerging Science Journal 7, no. 4 (2023): 1349–63. http://dx.doi.org/10.28991/esj-2023-07-04-021.

Full text
Abstract:
The objective of this article is to provide a comparative analysis of two novel genetic programming (GP) techniques, differentiable Cartesian genetic programming for artificial neural networks (DCGPANN) and geometric semantic genetic programming (GSGP), with state-of-the-art automated machine learning (AutoML) tools, namely Auto-Keras, Auto-PyTorch and Auto-Sklearn. While all these techniques are compared to several baseline algorithms upon their introduction, research still lacks direct comparisons between them, especially of the GP approaches with state-of-the-art AutoML. This study intends to fill this gap in order to analyze the true potential of GP for AutoML. The performances of the different tools are assessed by applying them to 20 benchmark datasets of the imbalanced binary classification field, thus an area that is a frequent and challenging problem. The tools are compared across the four categories average performance, maximum performance, standard deviation within performance, and generalization ability, whereby the metrics F1-score, G-mean, and AUC are used for evaluation. The analysis finds that the GP techniques, while unable to completely outperform state-of-the-art AutoML, are indeed already a very competitive alternative. Therefore, these advanced GP tools prove that they are able to provide a new and promising approach for practitioners developing machine learning (ML) models. Doi: 10.28991/ESJ-2023-07-04-021 Full Text: PDF
APA, Harvard, Vancouver, ISO, and other styles
10

ЗАЦ, Олександр, Сергій ШМАТКОВ, Вікторія СТРІЛЕЦЬ та Григорій ЛІТВІНОВ. "МОДЕЛЬ МАРШРУТИЗАЦІЇ В КОМП’ЮТЕРНИХ МЕРЕЖАХ НА ОСНОВІ ГРАФОВИХ НЕЙРОННИХ МЕРЕЖ". Information Technology: Computer Science, Software Engineering and Cyber Security, № 1 (30 квітня 2025): 73–81. https://doi.org/10.32782/it/2025-1-11.

Full text
Abstract:
У комп’ютерних мережах ефективна маршрутизація є ключовим елементом для забезпечення надійної та швидкої передачі даних. Останнім часом набуває популярності використання графових нейронних мереж для вирішення задач маршрутизації. Графові нейронні мережі дозволяють моделювати складні взаємозв’язки в мережах та адаптуватися до змінних умов, що робить їх перспективним інструментом для оптимізації процесів передачі даних. Мета роботи полягає у розробці ефективної моделі маршрутизації в комп’ютерних мережах для забезпечення зниження середньої затримки, покращення пропускної здатності та рівномірного розподілу навантаження між вузлами, з урахуванням динамічних умов мережі та складної топології. Методологія. У дослідженні проаналізовано застосування графових нейронних мереж для розв’язання задачі маршрутизації. Задача маршрутизації представлена як задача машинного навчання, а саме класифікації ребер графа мережі, які формують оптимальний маршрут. Для побудови моделі оптимізації використані сучасні інструменти і технології, такі як Mininet, PyTorch Geometric та Ryu SDN Controller. Наукова новизна. Запропонована модель маршрутизації в комп’ютерній мережі створена на основі згорткової графової нейронної мережі з двома згортковими шарами, що дозволяє врахувати складні топології мереж. Додавання шару регуляризації дозволило запобігти перенавчанню моделі. Висновки. Розроблено модель маршрутизації в комп’ютерних мережах з використанням графових нейронних мереж. Результати тестування та інтеграції моделі довели, що GNN здатні ефективно вирішувати складні задачі маршрутизації, враховуючи динамічні зміни в мережі та складну топологію. Модель може бути використана для моніторингу та управління трафіком у реальних комп’ютерних мережах, особливо за умов великих навантажень або потенційних атак.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Pytorch geometric"

1

Liberatore, Lorenzo. "Introduction to geometric deep learning and graph neural networks." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2022. http://amslaurea.unibo.it/25339/.

Full text
Abstract:
This thesis proposes an introduction to the fundamental concepts of supervised deep learning. Starting from Rosemblatt's Perceptron we will discuss the architectures that, in recent years, have revolutioned the world of deep learning: graph neural networks, which led to the formulation of geometric deep learning. We will then give a simple example of graph neural network, discussing the code that composes it and then test our architecture on the MNISTSuperpixels dataset, which is a variation of the benchmark dataset MNIST.
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Pytorch geometric"

1

Rozemberczki, Benedek, Paul Scherer, Yixuan He, et al. "PyTorch Geometric Temporal: Spatiotemporal Signal Processing with Neural Machine Learning Models." In CIKM '21: The 30th ACM International Conference on Information and Knowledge Management. ACM, 2021. http://dx.doi.org/10.1145/3459637.3482014.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Klipfel, Astrid, Yaël Frégier, Adlane Sayede, and Zied Bouraoui. "Optimized Crystallographic Graph Generation for Material Science." In Thirty-Second International Joint Conference on Artificial Intelligence {IJCAI-23}. International Joint Conferences on Artificial Intelligence Organization, 2023. http://dx.doi.org/10.24963/ijcai.2023/836.

Full text
Abstract:
Graph neural networks are widely used in machine learning applied to chemistry, and in particular for material science discovery. For crystalline materials, however, generating graph-based representation from geometrical information for neural networks is not a trivial task. The periodicity of crystalline needs efficient implementations to be processed in real-time under a massively parallel environment. With the aim of training graph-based generative models of new material discovery, we propose an efficient tool to generate cutoff graphs and k-nearest-neighbours graphs of periodic structures within GPU optimization. We provide pyMatGraph a Pytorch-compatible framework to generate graphs in real-time during the training of neural network architecture. Our tool can update a graph of a structure, making generative models able to update the geometry and process the updated graph during the forward propagation on the GPU side. Our code is publicly available at https://github.com/aklipf/mat-graph.
APA, Harvard, Vancouver, ISO, and other styles
3

Bhattacharya, Chandrachur, Joshua Christopher, David Thierry, Munidhar Biruduganti, Sarang Supekar, and Debolina Dasgupta. "Data-Driven Surrogate Modeling of Microturbine Combustors Burning Hydrogen Blends." In ASME Turbo Expo 2023: Turbomachinery Technical Conference and Exposition. American Society of Mechanical Engineers, 2023. http://dx.doi.org/10.1115/gt2023-103229.

Full text
Abstract:
Abstract Hydrogen is a good candidate as an alternative fuel for power generation, due to the absence of carbon. Retrofitting available combustion systems to run hydrogen or its blends with natural gas is an effective option to decarbonize this sector. This demands a reassessment of the system response for the entire range of operation. Experimentation can be expensive and prohibitive from safety standpoints, and though computational models can provide data to fill the gaps, they come with their own sets of challenges. Data-driven surrogates use machine learning methods to learn from the sparse available data and can fill this knowledge gap by learning statistical correlations that describe the system. In this paper, a data-driven surrogate model is developed for a Capstone C65 microturbine combustor that is modified to burn pure natural gas and blends with up to 60% hydrogen for various power demands. Gaussian process (GP) regression modeling is used to learn from this dataset to emulate the system characteristics. Active learning is also invoked to learn a good model using as few data points as possible. Additional data is, then, generated using Reynolds-averaged Navier-Stokes (RANS) simulations of the same combustor geometry across a wider range of operating conditions i.e., 0–100% hydrogen for 0–65kW power loads. This provides low-fidelity information that can be included in a multi-fidelity learning setting which shows distinct improvements over the single-fidelity model. The novel multi-input-multi-output multi-fidelity GP surrogate modeling framework used in this work, is developed in-house using GPyTorch, which runs on a PyTorch backend, and is capable of superior scaling compared to traditional GP approaches.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!