gradient accumulation,model training
Accumulate gradients over multiple mini-batches before updating weights.
3,145 technical terms and definitions
Accumulate gradients over multiple mini-batches before updating weights.
Group gradients for efficient communication.
Gradient clipping bounds gradient norms preventing privacy leakage and training instability.
Cap gradient magnitude to prevent exploding gradients.
Reduce gradient communication.
Maintain gradient flow in sparse networks.
Make gradients uninformative.
Regularize gradient magnitude (GANs).
Quantize gradients for transmission.
Reverse gradients for adversarial training.
Aggregate gradients across devices.
Optimize architecture with gradients.
Gradient-based pruning estimates weight importance using gradient information.
Use gradients to determine importance.
Interfaces between crystallites.
Analyze grain boundary structure and energy.
Energy per area of boundary.
Impurities collect at boundaries.
Increase grain size to reduce resistance.
Grammar-based decoding generates text following formal grammar specifications.
Grammar-based generation uses formal grammars to ensure syntactic validity of generated graphs.
Graph Recurrent Attention Networks generate graphs through sequential block-wise generation with recurrent state tracking for scalability.
Granger causality tests whether past values of one time series provide statistically significant information for predicting another series.
Granger non-causality tests null hypothesis that past values of one series don't help predict another.
Use attention in GNNs.
Graph completion predicts missing nodes or edges in incomplete graphs for knowledge graph construction.
Graph convolution generalizes convolutional operations to irregular graph structures by aggregating features from neighboring nodes with learnable weights.
Convolutional operations on graphs.
Generate new graphs.
Most expressive message-passing GNN.
Operator encoding graph structure.
GNNs operate on graph-structured data. Message passing between nodes. Social networks, molecules.
Apply Neural ODEs to graph-structured data.
Operators on graph-structured data.
Graph optimization transforms computation graphs improving efficiency through fusion reordering and elimination.
Downsample graphs in GNNs.
Graph recurrence applies RNNs to sequences of graph snapshots learning temporal graph dynamics.
Graph serialization stores model structure and parameters in portable format for deployment.
Graph U-Net applies encoder-decoder architecture with skip connections to graphs for node classification.
Upsample graphs.
Variational autoencoder for graphs.
Wavelet transforms on graphs.
Use graph networks for relational tasks.
Graph Autoregressive Flow generates graphs by sequentially adding nodes and edges with normalizing flows.
Sequential graph generation.
GraphNVP applies normalizing flows to graph generation enabling exact likelihood computation and efficient sampling of molecular structures.
GraphRNN generates graphs sequentially by modeling node and edge formation as a sequence generation problem using recurrent neural networks.
RNN-based graph generation.
GraphSAGE generates node embeddings by sampling and aggregating features from local neighborhoods enabling inductive learning on unseen graphs.
Inductive GNN with sampling.