Pages that link to "Category:Research"
The following pages link to Category:Research:
Displayed 50 items.
- Deep & Cross Network for Ad Click Predictions/paper/zh (← links)
- Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts/paper/zh (← links)
- Searching for Activation Functions/paper/zh (← links)
- Decoupled Weight Decay Regularization/es (← links)
- Deep & Cross Network for Ad Click Predictions/es (← links)
- Deep & Cross Network for Ad Click Predictions/paper/es (← links)
- Deep & Cross Network for Ad Click Predictions/zh (← links)
- Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer/es (← links)
- Language Modeling with Gated Convolutional Networks/paper/es (← links)
- Dropout: A Simple Way to Prevent Neural Networks from Overfitting/zh (← links)
- Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer/zh (← links)
- Language Modeling with Gated Convolutional Networks/zh (← links)
- Searching for Activation Functions/es (← links)
- Language Modeling with Gated Convolutional Networks/paper/en (← links)
- Searching for Activation Functions/zh (← links)
- Language Modeling with Gated Convolutional Networks/es (← links)
- Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts/zh (← links)
- Deep & Cross Network for Ad Click Predictions/paper/en (← links)
- Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts/es (← links)
- MTGR: Industrial-Scale Generative Recommendation Framework in Meituan/en (← links)
- Language Modeling with Gated Convolutional Networks/en (← links)
- Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts/en (← links)
- Wide & Deep Learning for Recommender Systems/en (← links)
- Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer/paper/en (← links)
- Decoupled Weight Decay Regularization/paper/en (← links)
- MTGR: Industrial-Scale Generative Recommendation Framework in Meituan/paper/en (← links)
- Searching for Activation Functions/paper/en (← links)
- Searching for Activation Functions/en (← links)
- Incorporating Nesterov Momentum into Adam/en (← links)
- Decoupled Weight Decay Regularization/en (← links)
- Dropout: A Simple Way to Prevent Neural Networks from Overfitting/en (← links)
- Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts/paper/en (← links)
- Deep & Cross Network for Ad Click Predictions/en (← links)
- Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer/en (← links)
- Translations:Adam A Method for Stochastic Optimization/1/en (← links)
- Translations:BERT Pre-training of Deep Bidirectional Transformers/1/en (← links)
- Translations:Batch Normalization Accelerating Deep Network Training/1/en (← links)
- Translations:Language Models are Few-Shot Learners/1/en (← links)
- Translations:Efficient Estimation of Word Representations/1/en (← links)
- Translations:Dropout A Simple Way to Prevent Overfitting/1/en (← links)
- Translations:Deep Residual Learning for Image Recognition/1/en (← links)
- Translations:Attention Is All You Need/1/en (← links)
- Translations:ImageNet Classification with Deep CNNs/1/en (← links)
- Translations:Generative Adversarial Nets/1/en (← links)
- Translations:Dropout A Simple Way to Prevent Overfitting/1/es (← links)
- Translations:Dropout A Simple Way to Prevent Overfitting/1/zh (← links)
- Translations:Adam A Method for Stochastic Optimization/1/es (← links)
- Translations:Adam A Method for Stochastic Optimization/1/zh (← links)
- Translations:Attention Is All You Need/1/es (← links)
- Translations:Attention Is All You Need/1/zh (← links)