Peer Learning: Learning Complex Policies in Groups from Scratch via Action Recommendations.
|
AAAI |
2024 |
0 |
Parameterized Projected Bellman Operator.
|
AAAI |
2024 |
0 |
Structure-Aware E(3)-Invariant Molecular Conformer Aggregation Networks.
|
ICML |
2024 |
0 |
Open X-Embodiment: Robotic Learning Datasets and RT-X Models : Open X-Embodiment Collaboration.
|
ICRA |
2024 |
0 |
What Matters for Active Texture Recognition With Vision-Based Tactile Sensors.
|
ICRA |
2024 |
0 |
Time-Efficient Reinforcement Learning with Stochastic Stateful Policies.
|
ICLR |
2024 |
0 |
Multi-Task Reinforcement Learning with Mixture of Orthogonal Experts.
|
ICLR |
2024 |
0 |
Robust Adversarial Reinforcement Learning via Bounded Rationality Curricula.
|
ICLR |
2024 |
0 |
CrossQ: Batch Normalization in Deep Reinforcement Learning for Greater Sample Efficiency and Simplicity.
|
ICLR |
2024 |
0 |
Domain Randomization via Entropy Maximization.
|
ICLR |
2024 |
0 |
MoVEInt: Mixture of Variational Experts for Learning Human-Robot Interactions From Demonstrations.
|
IEEE Robotics and Automation Letters |
2024 |
0 |
A Comparison of Imitation Learning Algorithms for Bimanual Manipulation.
|
IEEE Robotics and Automation Letters |
2024 |
0 |
Variational Hierarchical Mixtures for Probabilistic Learning of Inverse Dynamics.
|
TPAMI |
2024 |
0 |
Learning-Based Design and Control for Quadrupedal Robots With Parallel-Elastic Actuators.
|
IEEE Robotics and Automation Letters |
2023 |
0 |
Hierarchical Policy Blending as Inference for Reactive Robot Control.
|
ICRA |
2023 |
0 |
Safe Reinforcement Learning of Dynamic High-Dimensional Robotic Tasks: Navigation, Manipulation, Interaction.
|
ICRA |
2023 |
0 |
Start State Selection for Control Policy Learning from Optimal Trajectories.
|
ICRA |
2023 |
0 |
SE(3)-DiffusionFields: Learning smooth cost functions for joint grasp and motion optimization through diffusion.
|
ICRA |
2023 |
0 |
Clustering of Motion Trajectories by a Distance Measure Based on Semantic Features.
|
Humanoids |
2023 |
0 |
Model-Based Uncertainty in Value Functions.
|
AISTATS |
2023 |
0 |
Disentangling Interaction Using Maximum Entropy Reinforcement Learning in Multi-Agent Systems.
|
ECAI |
2023 |
0 |
Improved Algorithms for Stochastic Linear Bandits Using Tail Bounds for Martingale Mixtures.
|
NIPS/NeurIPS |
2023 |
0 |
Pseudo-Likelihood Inference.
|
NIPS/NeurIPS |
2023 |
0 |
Accelerating Motion Planning via Optimal Transport.
|
NIPS/NeurIPS |
2023 |
0 |
Diminishing Return of Value Expansion Methods in Model-Based Reinforcement Learning.
|
ICLR |
2023 |
0 |
LS-IQ: Implicit Reward Regularization for Inverse Reinforcement Learning.
|
ICLR |
2023 |
0 |
Placing by Touching: An Empirical Study on the Importance of Tactile Sensing for Precise Object Placing.
|
IROS |
2023 |
0 |
Motion Planning Diffusion: Learning and Planning of Robot Motions with Diffusion Models.
|
IROS |
2023 |
0 |
Digital Twin of a Driver-in-the-Loop Race Car Simulation With Contextual Reinforcement Learning.
|
IEEE Robotics and Automation Letters |
2023 |
0 |
PAC-Bayes Bounds for Bandit Problems: A Survey and Experimental Comparison.
|
TPAMI |
2023 |
0 |
Continuous-Time Fitted Value Iteration for Robust Policies.
|
TPAMI |
2023 |
0 |
A Deterministic Approximation to Neural SDEs.
|
TPAMI |
2023 |
0 |
Composable energy policies for reactive motion generation and reinforcement learning.
|
IJRR |
2023 |
0 |
Combining physics and deep learning to learn continuous-time dynamics models.
|
IJRR |
2023 |
0 |
Learning to Control Highly Accelerated Ballistic Movements on Muscular Robots.
|
Robotics and Autonomous Systems |
2023 |
0 |
MILD: Multimodal Interactive Latent Dynamics for Learning Human-Robot Interaction.
|
Humanoids |
2022 |
0 |
HEBO: An Empirical Study of Assumptions in Bayesian Optimisation.
|
JAIR |
2022 |
34 |
Active Exploration for Robotic Manipulation.
|
IROS |
2022 |
0 |
Learning Implicit Priors for Motion Optimization.
|
IROS |
2022 |
7 |
Curriculum Reinforcement Learning via Constrained Optimal Transport.
|
ICML |
2022 |
5 |
Dimensionality Reduction and Prioritized Exploration for Policy Search.
|
AISTATS |
2022 |
1 |
Boosted Curriculum Reinforcement Learning.
|
ICLR |
2022 |
2 |
Graph-based Reinforcement Learning meets Mixed Integer Programs: An application to 3D robot assembly discovery.
|
IROS |
2022 |
0 |
Monte-Carlo Robot Path Planning.
|
IEEE Robotics and Automation Letters |
2022 |
1 |
Controlling the Cascade: Kinematic Planning for N-ball Toss Juggling.
|
IROS |
2022 |
1 |
Regularized Deep Signed Distance Fields for Reactive Motion Generation.
|
IROS |
2022 |
7 |
Improving Sample Efficiency of Example-Guided Deep Reinforcement Learning for Bipedal Walking.
|
Humanoids |
2022 |
0 |
PAC-Bayesian lifelong learning for multi-armed bandits.
|
DMKD |
2022 |
3 |
Robot Learning of Mobile Manipulation With Reachability Behavior Priors.
|
IEEE Robotics and Automation Letters |
2022 |
4 |
Adapting Object-Centric Probabilistic Movement Primitives with Residual Reinforcement Learning.
|
Humanoids |
2022 |
0 |
Integrated Bi-Manual Motion Generation and Control shaped for Probabilistic Movement Primitives.
|
Humanoids |
2022 |
0 |
Inferring Smooth Control: Monte Carlo Posterior Policy Iteration with Gaussian Processes.
|
CoRL |
2022 |
0 |
Information-Theoretic Safe Exploration with Gaussian Processes.
|
NIPS/NeurIPS |
2022 |
0 |
Learning Stable Vector Fields on Lie Groups.
|
IEEE Robotics and Automation Letters |
2022 |
0 |
Benchmarking Structured Policies and Policy Optimization for Real-World Dexterous Object Manipulation.
|
IEEE Robotics and Automation Letters |
2022 |
0 |
Continuous Action Reinforcement Learning From a Mixture of Interpretable Experts.
|
TPAMI |
2022 |
0 |
Batch Reinforcement Learning With a Nonparametric Off-Policy Policy Gradient.
|
TPAMI |
2022 |
0 |
Composable Energy Policies for Reactive Motion Generation and Reinforcement Learning.
|
RSS |
2021 |
14 |
Value Iteration in Continuous Actions, States and Time.
|
ICML |
2021 |
8 |
Directed Acyclic Graph Neural Network for Human Motion Prediction.
|
ICRA |
2021 |
3 |
Gaussian Approximation for Bias Reduction in Q-Learning.
|
JMLR |
2021 |
4 |
SKID RAW: Skill Discovery From Raw Trajectories.
|
IEEE Robotics and Automation Letters |
2021 |
11 |
Learn2Assemble with Structured Representations and Search for Robotic Architectural Construction.
|
CoRL |
2021 |
14 |
Neural Posterior Domain Randomization.
|
CoRL |
2021 |
11 |
A Probabilistic Interpretation of Self-Paced Learning with Applications to Reinforcement Learning.
|
JMLR |
2021 |
8 |
Latent Derivative Bayesian Last Layer Networks.
|
AISTATS |
2021 |
5 |
An Empirical Analysis of Measure-Valued Derivatives for Policy Gradients.
|
IJCNN |
2021 |
5 |
Efficient and Reactive Planning for High Speed Robot Air Hockey.
|
IROS |
2021 |
6 |
Robot Reinforcement Learning on the Constraint Manifold.
|
CoRL |
2021 |
16 |
Model Predictive Actor-Critic: Accelerating Robot Skill Acquisition with Deep Reinforcement Learning.
|
ICRA |
2021 |
22 |
Convex optimization with an interpolation-based projection and its application to deep learning.
|
MLJ |
2021 |
0 |
Robust Value Iteration for Continuous Control Tasks.
|
RSS |
2021 |
11 |
Learning Human-like Hand Reaching for Human-Robot Handshaking.
|
ICRA |
2021 |
5 |
Convex Regularization in Monte-Carlo Tree Search.
|
ICML |
2021 |
0 |
A Variational Infinite Mixture for Probabilistic Inverse Dynamics Learning.
|
ICRA |
2021 |
0 |
Contextual Latent-Movements Off-Policy Optimization for Robotic Manipulation Skills.
|
ICRA |
2021 |
0 |
Differentiable Physics Models for Real-world Offline Model-based Reinforcement Learning.
|
ICRA |
2021 |
0 |
Data-Efficient Domain Randomization With Bayesian Optimization.
|
IEEE Robotics and Automation Letters |
2021 |
0 |
Assessing Transferability From Simulation to Reality for Reinforcement Learning.
|
TPAMI |
2021 |
0 |
MushroomRL: Simplifying Reinforcement Learning Research.
|
JMLR |
2021 |
0 |
A Probabilistic Framework for Imitating Human Race Driver Behavior.
|
IEEE Robotics and Automation Letters |
2020 |
6 |
Learning attribute grammars for movement primitive sequencing.
|
IJRR |
2020 |
7 |
Model-Based Quality-Diversity Search for Efficient Robot Learning.
|
IROS |
2020 |
15 |
Multi-Sensor Next-Best-View Planning as Matroid-Constrained Submodular Maximization.
|
IEEE Robotics and Automation Letters |
2020 |
8 |
ImitationFlow: Learning Deep Stable Stochastic Dynamic Systems by Normalizing Flows.
|
IROS |
2020 |
24 |
Self-Paced Deep Reinforcement Learning.
|
NIPS/NeurIPS |
2020 |
22 |
Deep Adversarial Reinforcement Learning for Object Disentangling.
|
IROS |
2020 |
1 |
Probabilistic Approach to Physical Object Disentangling.
|
IEEE Robotics and Automation Letters |
2020 |
3 |
Multi-agent active information gathering in discrete and continuous-state decentralized POMDPs by policy graph improvement.
|
Autonomous Agents and Multi-Agent Systems |
2020 |
9 |
High Acceleration Reinforcement Learning for Real-World Juggling with Binary Rewards.
|
CoRL |
2020 |
12 |
Redundancy resolution under hard joint constraints: a generalized approach to rank updates.
|
IROS |
2020 |
1 |
Learning Hierarchical Acquisition Functions for Bayesian Optimization.
|
IROS |
2020 |
0 |
Learning Control Policies from Optimal Trajectories.
|
ICRA |
2020 |
1 |
Numerical Quadrature for Probabilistic Policy Search.
|
TPAMI |
2020 |
9 |
Underactuated Waypoint Trajectory Optimization for Light Painting Photography.
|
ICRA |
2020 |
1 |
Plucking Motions for Tea Harvesting Robots Using Probabilistic Movement Primitives.
|
IEEE Robotics and Automation Letters |
2020 |
12 |
A Nonparametric Off-Policy Policy Gradient.
|
AISTATS |
2020 |
11 |
Sharing Knowledge in Multi-Task Deep Reinforcement Learning.
|
ICLR |
2020 |
66 |
Bayesian Online Prediction of Change Points.
|
UAI |
2020 |
0 |
Generalized Mean Estimation in Monte-Carlo Tree Search.
|
IJCAI |
2020 |
0 |
Real Time Trajectory Prediction Using Deep Conditional Generative Models.
|
IEEE Robotics and Automation Letters |
2020 |
0 |
Multimodal Uncertainty Reduction for Intention Recognition in Human-Robot Interaction.
|
IROS |
2019 |
18 |
Reinforcement Learning of Trajectory Distributions: Applications in Assisted Teleoperation and Motion Planning.
|
IROS |
2019 |
7 |
TD-regularized actor-critic methods.
|
MLJ |
2019 |
0 |
Experience Reuse with Probabilistic Movement Primitives.
|
IROS |
2019 |
9 |
Deep Lagrangian Networks for end-to-end learning of energy-based control for under-actuated systems.
|
IROS |
2019 |
42 |
Switching Linear Dynamics for Variational Bayes Filtering.
|
ICML |
2019 |
37 |
Building a Library of Tactile Skills Based on FingerVision.
|
Humanoids |
2019 |
6 |
Generalized Multiple Correlation Coefficient as a Similarity Measurement between Trajectories.
|
IROS |
2019 |
4 |
Self-Paced Contextual Reinforcement Learning.
|
CoRL |
2019 |
30 |
Reinforcement learning of motor skills using Policy Search and human corrective advice.
|
IJRR |
2019 |
10 |
Chance-Constrained Trajectory Optimization for Non-linear Systems with Unknown Stochastic Dynamics.
|
IROS |
2019 |
5 |
Receding Horizon Curiosity.
|
CoRL |
2019 |
8 |
Compatible natural gradient policy search.
|
MLJ |
2019 |
0 |
Learning Replanning Policies With Direct Policy Search.
|
IEEE Robotics and Automation Letters |
2019 |
4 |
Projections for Approximate Policy Iteration Algorithms.
|
ICML |
2019 |
8 |
Learning Intention Aware Online Adaptation of Movement Primitives.
|
IEEE Robotics and Automation Letters |
2019 |
12 |
Information Gathering in Decentralized POMDPs by Policy Graph Improvement.
|
AAMAS |
2019 |
6 |
Entropic Risk Measure in Policy Search.
|
IROS |
2019 |
18 |
Stochastic Optimal Control as Approximate Input Inference.
|
CoRL |
2019 |
16 |
HJB Optimal Feedback Control with Deep Differential Value Functions and Action Constraints.
|
CoRL |
2019 |
12 |
Exploration Driven by an Optimistic Bellman Equation.
|
IJCNN |
2019 |
1 |
Local Online Motor Babbling: Learning Motor Abundance of a Musculoskeletal Robot Arm
|
IROS |
2019 |
0 |
Learning to Serve: An Experimental Study for a New Learning From Demonstrations Framework.
|
IEEE Robotics and Automation Letters |
2019 |
0 |
Online Learning of an Open-Ended Skill Library for Collaborative Tasks.
|
Humanoids |
2018 |
16 |
Inverse Reinforcement Learning via Nonparametric Spatio-Temporal Subgoal Modeling.
|
JMLR |
2018 |
7 |
Inducing Probabilistic Context-Free Grammars for the Sequencing of Movement Primitives.
|
ICRA |
2018 |
4 |
Using probabilistic movement primitives in robotics.
|
Autonomous Robots |
2018 |
138 |
Control of Musculoskeletal Systems Using Learned Dynamics Models.
|
IEEE Robotics and Automation Letters |
2018 |
13 |
Mixture of Attractors: A Novel Movement Primitive Representation for Learning Motor Skills From Demonstrations.
|
IEEE Robotics and Automation Letters |
2018 |
14 |
Domain Randomization for Simulation-Based Policy Optimization with Transferability Assessment.
|
CoRL |
2018 |
38 |
Regularizing Reinforcement Learning with State Abstraction.
|
IROS |
2018 |
11 |
Sample and Feedback Efficient Hierarchical Reinforcement Learning from Human Preferences.
|
ICRA |
2018 |
19 |
Learning Coupled Forward-Inverse Models with Combined Prediction Errors.
|
ICRA |
2018 |
2 |
Approximate Value Iteration Based on Numerical Quadrature.
|
IEEE Robotics and Automation Letters |
2018 |
2 |
Online optimal trajectory generation for robot table tennis.
|
Robotics and Autonomous Systems |
2018 |
45 |
Utilizing Human Feedback in POMDP Execution and Specification.
|
Humanoids |
2018 |
6 |
A kernel-based approach to learning contact distributions for robot manipulation tasks.
|
Autonomous Robots |
2018 |
12 |
PIPPS: Flexible Model-Based Policy Search Robust to the Curse of Chaos.
|
ICML |
2018 |
0 |
Model-Free Trajectory-based Policy Optimization with Monotonic Improvement.
|
JMLR |
2018 |
0 |
Anticipatory action selection for human-robot table tennis.
|
Artificial Intelligence |
2017 |
31 |
Learning inverse dynamics models in O(n) time with LSTM networks.
|
Humanoids |
2017 |
47 |
Whole-body multi-contact motion in humans and humanoids: Advances of the CoDyCo European project.
|
Robotics and Autonomous Systems |
2017 |
19 |
Layered direct policy search for learning hierarchical skills.
|
ICRA |
2017 |
17 |
Stability of Controllers for Gaussian Process Dynamics.
|
JMLR |
2017 |
28 |
Probabilistic Prioritization of Movement Primitives.
|
IEEE Robotics and Automation Letters |
2017 |
21 |
Efficient online adaptation with stochastic recurrent neural networks.
|
Humanoids |
2017 |
3 |
A comparison of distance measures for learning nonparametric motor skill libraries.
|
Humanoids |
2017 |
3 |
Online Learning with Stochastic Recurrent Neural Networks using Intrinsic Motivation Signals.
|
CoRL |
2017 |
4 |
Hybrid control trajectory optimization under uncertainty.
|
IROS |
2017 |
13 |
Context-driven movement primitive adaptation.
|
ICRA |
2017 |
5 |
Non-parametric Policy Search with Limited Information Loss.
|
JMLR |
2017 |
26 |
Learning movement primitive libraries through probabilistic segmentation.
|
IJRR |
2017 |
49 |
Model-based contextual policy search for data-efficient generalization of robot skills.
|
Artificial Intelligence |
2017 |
72 |
A learning-based shared control architecture for interactive task execution.
|
ICRA |
2017 |
40 |
Local Bayesian Optimization of Motor Skills.
|
ICML |
2017 |
16 |
Generalized exploration in policy search.
|
MLJ |
2017 |
0 |
Empowered skills.
|
ICRA |
2017 |
4 |
A Comparison of Autoregressive Hidden Markov Models for Multimodal Manipulations With Variable Masses.
|
IEEE Robotics and Automation Letters |
2017 |
1 |
Guiding Trajectory Optimization by Demonstrated Distributions.
|
IEEE Robotics and Automation Letters |
2017 |
39 |
Goal-driven dimensionality reduction for reinforcement learning.
|
IROS |
2017 |
12 |
State-Regularized Policy Search for Linearized Dynamical Systems.
|
ICAPS |
2017 |
6 |
Active Incremental Learning of Robot Movement Primitives.
|
CoRL |
2017 |
42 |
Policy Search with High-Dimensional Context Variables.
|
AAAI |
2017 |
0 |
Phase estimation for fast action recognition and trajectory generation in human-robot collaboration.
|
IJRR |
2017 |
0 |
Probabilistic movement primitives for coordination of multiple human-robot collaborative tasks.
|
Autonomous Robots |
2017 |
0 |
Active tactile object exploration with Gaussian processes.
|
IROS |
2016 |
66 |
Demonstration based trajectory optimization for generalizable robot motions.
|
Humanoids |
2016 |
38 |
Deep spiking networks for model-based planning in humanoids.
|
Humanoids |
2016 |
10 |
Incremental imitation learning of context-dependent motor skills.
|
Humanoids |
2016 |
17 |
Probabilistic inference for determining options in reinforcement learning.
|
MLJ |
2016 |
0 |
Stable reinforcement learning with autoencoders for tactile and visual data.
|
IROS |
2016 |
135 |
Movement primitives with multiple phase parameters.
|
ICRA |
2016 |
6 |
Stability of Controllers for Gaussian Process Forward Models.
|
ICML |
2016 |
34 |
Jointly learning trajectory generation and hitting point prediction in robot table tennis.
|
Humanoids |
2016 |
29 |
Using probabilistic movement primitives for striking movements.
|
Humanoids |
2016 |
19 |
Learning soft task priorities for control of redundant robots.
|
ICRA |
2016 |
35 |
A new trajectory generation framework in robotic table tennis.
|
IROS |
2016 |
5 |
Probabilistic decomposition of sequential force interaction tasks into Movement Primitives.
|
IROS |
2016 |
10 |
Hierarchical Relative Entropy Policy Search.
|
JMLR |
2016 |
0 |
Acquiring and Generalizing the Embodiment Mapping From Human Observations to Robot Skills.
|
IEEE Robotics and Automation Letters |
2016 |
21 |
A lightweight robotic arm with pneumatic muscles for robot learning.
|
ICRA |
2016 |
32 |
Catching heuristics are optimal control policies.
|
NIPS/NeurIPS |
2016 |
29 |
Manifold Gaussian Processes for regression.
|
IJCNN |
2016 |
0 |
Probabilistic progress prediction and sequencing of concurrent movement primitives.
|
IROS |
2015 |
9 |
Learning optimal striking points for a ping-pong playing robot.
|
IROS |
2015 |
22 |
Learning multiple collaborative tasks with a mixture of Interaction Primitives.
|
ICRA |
2015 |
99 |
Learning inverse dynamics models with contacts.
|
ICRA |
2015 |
56 |
Learning of Non-Parametric Control Policies with High-Dimensional State Features.
|
AISTATS |
2015 |
34 |
Evaluation of tactile feature extraction for interactive object recognition.
|
Humanoids |
2015 |
58 |
Reinforcement learning vs human programming in tetherball robot games.
|
IROS |
2015 |
15 |
Learning movement primitive attractor goals and sequential skills from kinesthetic demonstrations.
|
Robotics and Autonomous Systems |
2015 |
55 |
Optimizing robot striking movement primitives with Iterative Learning Control.
|
Humanoids |
2015 |
4 |
Active reward learning with a novel acquisition function.
|
Autonomous Robots |
2015 |
51 |
Towards learning hierarchical skills for multi-phase manipulation tasks.
|
ICRA |
2015 |
109 |
Learning robot in-hand manipulation with tactile features.
|
Humanoids |
2015 |
137 |
Stabilizing novel objects by learning to predict tactile slip.
|
IROS |
2015 |
86 |
A Probabilistic Framework for Semi-autonomous Robots Based on Interaction Primitives with Phase Estimation.
|
ISRR |
2015 |
15 |
Extracting low-dimensional control variables for movement primitives.
|
ICRA |
2015 |
44 |
First-person tele-operation of a humanoid robot.
|
Humanoids |
2015 |
27 |
Combined pose-wrench and state machine representation for modeling Robotic Assembly Skills.
|
IROS |
2015 |
15 |
A comparison of contact distribution representations for learning to predict object interactions.
|
Humanoids |
2015 |
5 |
Learning torque control in presence of contacts using tactile sensing from robot skin.
|
Humanoids |
2015 |
13 |
Model-free Probabilistic Movement Primitives for physical interaction.
|
IROS |
2015 |
17 |
Probabilistic segmentation applied to an assembly task.
|
Humanoids |
2015 |
38 |
Learning motor skills from partially observed movements executed at different speeds.
|
IROS |
2015 |
25 |
Policy Evaluation with Temporal Differences: A Survey and Comparison (Extended Abstract).
|
ICAPS |
2015 |
0 |
Model-Based Relative Entropy Stochastic Search.
|
NIPS/NeurIPS |
2015 |
0 |
Policy search for learning robot control using sparse data.
|
ICRA |
2014 |
18 |
Learning to sequence movement primitives from demonstrations.
|
IROS |
2014 |
45 |
Tools for simulating humanoid robot dynamics: A survey based on user feedback.
|
Humanoids |
2014 |
64 |
Learning robot tactile sensing for object manipulation.
|
IROS |
2014 |
67 |
Special issue on autonomous grasping and manipulation.
|
Autonomous Robots |
2014 |
12 |
Interaction primitives for human-robot cooperation tasks.
|
ICRA |
2014 |
199 |
Learning interaction for collaborative tasks with probabilistic movement primitives.
|
Humanoids |
2014 |
88 |
An experimental comparison of Bayesian optimization for bipedal locomotion.
|
ICRA |
2014 |
72 |
Latent space policy search for robotics.
|
IROS |
2014 |
25 |
Dimensionality reduction for probabilistic movement primitives.
|
Humanoids |
2014 |
33 |
Predicting object interactions from contact distributions.
|
IROS |
2014 |
20 |
Multi-task policy search for robotics.
|
ICRA |
2014 |
111 |
Robust policy updates for stochastic optimal control.
|
Humanoids |
2014 |
8 |
Active Reward Learning.
|
RSS |
2014 |
98 |
Natural evolution strategies.
|
JMLR |
2014 |
9 |
Policy Search for Path Integral Control.
|
ECML/PKDD |
2014 |
41 |
Generalizing pouring actions between objects using warped parameters.
|
Humanoids |
2014 |
33 |
Learning to predict phases of manipulation tasks as hidden states.
|
ICRA |
2014 |
45 |
Sample-based informationl-theoretic stochastic optimal control.
|
ICRA |
2014 |
23 |
Policy evaluation with temporal differences: a survey and comparison.
|
JMLR |
2014 |
0 |
Towards Robot Skill Learning: From Simple Skills to Table Tennis.
|
ECML/PKDD |
2013 |
39 |
Probabilistic interactive segmentation for anthropomorphic robots in cluttered environments.
|
Humanoids |
2013 |
10 |
Probabilistic Movement Primitives.
|
NIPS/NeurIPS |
2013 |
444 |
Reinforcement learning in robotics: A survey.
|
IJRR |
2013 |
2459 |
Learning sequential motor tasks.
|
ICRA |
2013 |
49 |
A probabilistic approach to robot trajectory generation.
|
Humanoids |
2013 |
20 |
Feedback error learning for rhythmic motor primitives.
|
ICRA |
2013 |
7 |
Autonomous reinforcement learning with hierarchical REPS.
|
IJCNN |
2013 |
5 |
Learning responsive robot behavior by imitation.
|
IROS |
2013 |
39 |
Model-based imitation learning by probabilistic trajectory matching.
|
ICRA |
2013 |
46 |
Data-Efficient Generalization of Robot Skills with Contextual Policy Search.
|
AAAI |
2013 |
124 |
Alignment-based transfer learning for robot models.
|
IJCNN |
2013 |
65 |
Probabilistic movement modeling for intention inference in human-robot interaction.
|
IJRR |
2013 |
0 |
Learning to select and generalize striking movements in robot table tennis.
|
IJRR |
2013 |
0 |
A kernel-based approach to direct action perception.
|
ICRA |
2012 |
85 |
Algorithms for Learning Markov Field Policies.
|
NIPS/NeurIPS |
2012 |
0 |
Generalization of human grasping for multi-fingered robot hands.
|
IROS |
2012 |
98 |
Learning concurrent motor skills in versatile solution spaces.
|
IROS |
2012 |
45 |
Learning throwing and catching skills.
|
IROS |
2012 |
10 |
Maximally informative interaction learning for scene exploration.
|
IROS |
2012 |
43 |
Probabilistic Modeling of Human Movements for Intention Inference.
|
RSS |
2012 |
52 |
Robot Skill Learning.
|
ECAI |
2012 |
9 |
Reinforcement learning to adjust parametrized motor primitives to new situations.
|
Autonomous Robots |
2012 |
0 |
Structured Apprenticeship Learning.
|
ECML/PKDD |
2012 |
26 |
Learning tracking control with forward models.
|
ICRA |
2012 |
10 |
Toward fast policy search for learning legged locomotion.
|
IROS |
2012 |
33 |
Point cloud completion using extrusions.
|
Humanoids |
2012 |
28 |
A brain-robot interface for studying motor learning after stroke.
|
IROS |
2012 |
20 |
Hierarchical Relative Entropy Policy Search.
|
AISTATS |
2012 |
0 |
Learning visual representations for perception-action systems.
|
IJRR |
2011 |
41 |
Trajectory planning for optimal robot catching in real-time.
|
ICRA |
2011 |
112 |
Modeling Opponent Actions for Table-Tennis Playing Robot.
|
AAAI |
2011 |
1 |
Model learning for robot control: a survey.
|
Cognitive Processing |
2011 |
455 |
A flexible hybrid framework for modeling complex manipulation tasks.
|
ICRA |
2011 |
8 |
Learning elementary movements jointly with a higher level task.
|
IROS |
2011 |
16 |
A Non-Parametric Approach to Dynamic Programming.
|
NIPS/NeurIPS |
2011 |
19 |
Learning robot grasping from 3-D images with Markov Random Fields.
|
IROS |
2011 |
45 |
Relative Entropy Inverse Reinforcement Learning.
|
AISTATS |
2011 |
315 |
Balancing Safety and Exploitability in Opponent Modeling.
|
AAAI |
2011 |
19 |
Learning inverse kinematics with structured prediction.
|
IROS |
2011 |
49 |
Learning task-space tracking control with kernels.
|
IROS |
2011 |
6 |
Learning anticipation policies for robot table tennis.
|
IROS |
2011 |
19 |
Reinforcement Learning to Adjust Robot Movements to New Situations.
|
IJCAI |
2011 |
0 |
Policy search for motor primitives in robotics.
|
MLJ |
2011 |
0 |
Movement extraction by detecting dynamics switches and repetitions.
|
NIPS/NeurIPS |
2010 |
60 |
Learning table tennis with a Mixture of Motor Primitives.
|
Humanoids |
2010 |
97 |
Combining active learning and reactive control for robot grasping.
|
Robotics and Autonomous Systems |
2010 |
164 |
Using model knowledge for learning inverse dynamics.
|
ICRA |
2010 |
171 |
Reinforcement Learning to adjust Robot Movements to New Situations.
|
RSS |
2010 |
203 |
A biomimetic approach to robot table tennis.
|
IROS |
2010 |
113 |
Learning probabilistic discriminative models of grasp affordances under limited supervision.
|
IROS |
2010 |
21 |
Relative Entropy Policy Search.
|
AAAI |
2010 |
605 |
Switched Latent Force Models for Movement Segmentation.
|
NIPS/NeurIPS |
2010 |
37 |
Incremental Sparsification for Real-time Online Model Learning.
|
AISTATS |
2010 |
25 |
Movement templates for learning of hitting and batting.
|
ICRA |
2010 |
169 |
Learning Visual Representations for Interactive Systems.
|
ISRR |
2009 |
11 |
An Expectation Maximization Algorithm for Continuous Markov Decision Processes with Arbitrary Reward.
|
AISTATS |
2009 |
33 |
Learning complex motions by sequencing simpler motion templates.
|
ICML |
2009 |
47 |
Efficient Sample Reuse in EM-Based Policy Search.
|
ECML/PKDD |
2009 |
17 |
Active learning using mean shift optimization for robot grasping.
|
IROS |
2009 |
35 |
Sparse online model learning for robot control with support vector regression.
|
IROS |
2009 |
31 |
Guest editorial: Special issue on robot learning, Part A.
|
Autonomous Robots |
2009 |
2 |
Learning motor primitives for robotics.
|
ICRA |
2009 |
236 |
Towards Motor Skill Learning for Robotics.
|
ISRR |
2009 |
0 |
Guest editorial: Special issue on robot learning, Part B.
|
Autonomous Robots |
2009 |
0 |
Real-time learning of resolved velocity control on a Mitsubishi PA-10.
|
ICRA |
2008 |
7 |
Local Gaussian Process Regression for Real Time Online Model Learning.
|
NIPS/NeurIPS |
2008 |
242 |
Local Gaussian process regression for real-time model-based robot control.
|
IROS |
2008 |
138 |
Learning to Control in Operational Space.
|
IJRR |
2008 |
183 |
Operational Space Control: A Theoretical and Empirical Comparison.
|
IJRR |
2008 |
353 |
Policy Gradients with Parameter-Based Exploration for Control.
|
ICANN |
2008 |
70 |
Fitted Q-iteration by Advantage Weighted Regression.
|
NIPS/NeurIPS |
2008 |
58 |
Model-Based Reinforcement Learning with Continuous States and Actions.
|
ESANN |
2008 |
23 |
Policy Search for Motor Primitives in Robotics.
|
NIPS/NeurIPS |
2008 |
712 |
Learning perceptual coupling for motor primitives.
|
IROS |
2008 |
90 |
Episodic Reinforcement Learning by Logistic Reward-Weighted Regression.
|
ICANN |
2008 |
10 |
Using Bayesian Dynamical Systems for Motion Template Libraries.
|
NIPS/NeurIPS |
2008 |
26 |
Learning Inverse Dynamics: a Comparison.
|
ESANN |
2008 |
89 |
Adaptive Importance Sampling with Automatic Model Selection in Value Function Approximation.
|
AAAI |
2008 |
0 |
A unifying framework for robot control with redundant DOFs.
|
Autonomous Robots |
2008 |
0 |
Reinforcement Learning for Operational Space Control.
|
ICRA |
2007 |
28 |
Applying the Episodic Natural Actor-Critic Architecture to Motor Primitive Learning.
|
ESANN |
2007 |
28 |
Towards compliant humanoids-an experimental assessment of suitable task space position/orientation controllers.
|
IROS |
2007 |
5 |
Reinforcement learning by reward-weighted regression for operational space control.
|
ICML |
2007 |
197 |
Solving Deep Memory POMDPs with Recurrent Policy Gradients.
|
ICANN |
2007 |
159 |
Policy Gradient Methods for Robotics.
|
IROS |
2006 |
559 |
Reinforcement Learning for Parameterized Motor Primitives.
|
IJCNN |
2006 |
30 |
Learning Operational Space Control.
|
RSS |
2006 |
22 |
A Bayesian Approach to Nonlinear Parameter Identification for Rigid Body Dynamics.
|
RSS |
2006 |
63 |
A unifying methodology for the control of robotic systems.
|
IROS |
2005 |
42 |
Comparative experiments on task space control with redundancy resolution.
|
IROS |
2005 |
82 |
Natural Actor-Critic.
|
ECML/PKDD |
2005 |
0 |
Learning Movement Primitives.
|
ISRR |
2003 |
0 |