Add on 9 April, 2019
The Rainbow Team at IRISA/Inria Rennes – France has an open position for a post-doc : https://urldefense.proofpoint.com/v2/url?u=https-3A__team.inria.fr_rainbow_adapt-2Dsimulator-2Den_&d=DwIFaQ&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=qRfbrKs6RoHWy5VsBWhOrZ4hESa0BXUOs1Pma3HWUCI&s=RPHfWlaa8P3c4Uc1cjnbl5ROH7U3XXYbGQdUW403bes&e= Keywords: Power wheelchair simulator, Virtual Reality, Physics engine, motion cues, haptics, user tests ------------------------------ Context The ADAPT project, co-financed by the European Regional Development Fund as part of the INTERREG VA France (Channel) England program, began in January 2017 and will end in January 2021. The aim of the project is to design and to develop assistive devices for empowering disabled people through robotics. INSA Rennes, an engineering school, is involved in the ADAPT project through 3 research units: IRISA-IETR-LGCGM. IRISA (Rainbow team) is specialized in computer vision and robotics. IETR (VAADER team) works on the design of embedded systems for signal processing. LGCGM (ECSR team) carries out work in mechanical processes and systems. As part of the ADAPT project and through past collaborations, in particular with the Pôle Saint Hélier , a rehabilitation center in Rennes, INSA teams have developed a smart powered wheelchair (PW) that compensates user disabilities by using driving assistance technologies. In parallel, for an immersive experience of the smart PW, a virtual reality (VR) driving simulator has been developed. This simulator includes immersive visual rendering capabilities (immersive room / VR headset), interfaces compatible with those of the PW and a haptic platform for replicating wheelchair motions. ------------------------------ Post-doctoral subject The goal of the post-doctoral work is directly linked to the wheelchair simulator, and will be in particular dedicated to motion cues. The idea is to combine software development for Virtual Reality (based on Robot Operating System, Unity 3D,…) add a 4 degrees of freedom motion platform (1 translation and 3 rotations). Within the project team working on the simulator, the researcher will have to propose solutions to increase the Sense of Presence felt by the user, and to prevent from cybersickness for different indoor/outdoor scenarios: • Design of a new physics engine to replicate motion dynamics of a powered wheelchair, such as collisions, bumps, floor irregularities, caster wheel kinematics… • Design of motion cues and haptics: o the 4 degrees of freedom should be properly used to match motion cues; o the behavior of the platform should be finely tuned and adapted to each user with disabilities. ------------------------------ Skill / Requirements The applicant should have a strong interest in working in a multidisciplinary team, including the medical staff of the rehabilitation center of Pôle Saint Hélier. Ideally, she/he has an experience in VR and/or robotics domains: • Simulation and/or motion platform control; • C/C++ coding; • Familiarity with ROS; • Familiarity with system dynamics and mechanics is considered as an asset. ------------------------------ Conditions The position is full-time for 1 year. Salary : around 2200€/month (net salary) Localisation : IRISA / INSA de Rennes – Campus de Beaulieu – Rainbow team ------------------------------ To apply Applications consist of a cover letter describing how your research experience is relevant to the position and a resume sent to ronan.le-breton at insa-rennes.fr, sylvain.guegan at insa-rennes.fr and marie.babel at irisa.fr.
Add on 9 April, 2019
Flight dynamics and control lab (https://urldefense.proofpoint.com/v2/url?u=http-3A__fdcl.seas.gwu.edu_&d=DwIFAg&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=2TG3BlKY-_ZmQgqvdzl4wtZswuuBHCtY-BCjUTqqRcw&s=9hPu6CHtgi-9JUwL7fJWvTQKl-qe8VIMjgA-qLpBcng&e=) at the George Washington University, Washington DC is looking for new doctoral students. There are two open positions for the following projects. 1. Aerial Mapping and Situation Awareness This project aims to develop aerial robotic systems to monitor complex urban environments, to create a real time map, and to share it with other autonomous vehicles. Backgrounds in the following topics are desired. a. SLAM b. ROS c. Machine learning d. Image processing This position will be funded by NSF. 2. Dynamics and Control of Flapping Wing Unmanned Aerial Vehicle This project aims to uncover the biomechanics of Monarch butterflies exhibiting the longest flight range among insects, and to utilize it to develop bio-inspired control schemes for long-range flapping-wing micro aerial vehicles. Backgrounds in the following topics are desired. a. Nonlinear dynamics b. Nonlinear control theory c. Aerodynamics d. Geometric mechanics For every position, it is expected that the candidates have expertise in structured programming. To apply, or to request more information, please contact Prof. Taeyoung Lee at tylee at gwu.edu. The application should include CV, a list of publications, and a list of reference, and it should be submitted as a single PDF file.
Add on 3 April, 2019
Title: Advanced and reliable autonomy in aerial robotics for high speed trajectory tracking using deep neural networks Research area and project description: Applications are invited for a fully funded PhD studentship position within the Department of Engineering at Aarhus University, Denmark starting in Fall 2019. At Aarhus University, you will have thrilling opportunity to apply the latest machine learning/artificial intelligence technologies to solve real-world problems, such as high speed, agile manoeuvres for drones. We are looking for qualified, talented, and enthusiastic PhD students who wish to investigate the embedded guidance, control and navigation problem of unmanned aerial systems using artificial intelligence/machine learning methods. Our aim is to leverage the current state-of-the-art autonomy level towards smarter robots, which will learn and interact with their environment, collaborate with people and other robots, plan their future actions, and execute the given task accurately. What you stand to gain: a fully funded PhD position for 3 years (starting Fall 2019) at the Department of Engineering, Aarhus University; a fun environment to drive your passion for robotics. The research will be carried out under the supervision of Associate Professor Erdal Kayacan (https://urldefense.proofpoint.com/v2/url?u=http-3A__www.erdal.info&d=DwICAg&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=ZfDBlGDIbei49bxzkRoCkvmNDrsukj3uJasSVv7VaMc&s=I7OdiKWZfb4vK8_p2hdIj_anQqe-AJjvPx1NLG-Dml8&e=). Qualifications and specific competences: Required: A Master's degree in mechanical engineering, electrical engineering, aerospace engineering, computer science/engineering, control theory, mechatronics, applied mathematics, or other related disciplines Excellent verbal and writing skills in English with very good communication skills Experience in Robot Operating System (ROS) Concrete knowledge in C/C++ Preferred: Hands on experience in UAVs and basic understanding of UAV models Demonstration of research activities (conference or journal papers) --- Here is the official link for the PhD applications: https://urldefense.proofpoint.com/v2/url?u=http-3A__phd.scitech.au.dk_for-2Dapplicants_apply-2Dhere_may-2D2019_advanced-2Dand-2Dreliable-2Dautonomy-2Din-2Daerial-2Drobotics-2Dfor-2Dhigh-2Dspeed-2Dtrajectory-2Dtracking-2Dusing-2Ddeep-2Dneural-2Dnetworks_&d=DwICAg&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=ZfDBlGDIbei49bxzkRoCkvmNDrsukj3uJasSVv7VaMc&s=TKRc_ghvHZkP6_wyRFy2C4FyzzUNyyNIDTOLN-OClbg&e= --- Get your graduate degree at a top 100 university! The Graduate School of Science and Technology at Aarhus University is looking for a bright, enthusiastic and ambitious candidate who is interested in pursuing a PhD education. The University is located in Aarhus, Denmark, which provides international students with a safe and stable environment, a high standard of living and a wealth of social opportunities. Besides having an excellent reputation that enables our PhD graduates to find outstanding employment prospects, Aarhus University offers attractive working conditions, research support and campus resources. Application deadline is 1 May 2019, with a starting date of 1 August 2019 or later.
Add on 10 April, 2019
The University of Birmingham Extreme Robotics Lab, is one of the leading university robotics labs in Europe dedicated to practical applications of robotics and AI to extreme environments. A PhD student is sought to create and contribute to the creation of knowledge by undertaking a specified range of activities within Recycling of Lithium Ion Battery (RELIB) project, using expertise in Computer Vision and/or vision-guided robotics, to work with a team to develop integrated robotic systems for practical applications, driven by major industrial and societal problems. The UoB robotics team are collectively working on robotic grasping and manipulation; planning and reasoning for fixed and/or mobile manipulators; robotic vehicles and navigation; robot dynamics and control and Computer Vision, machine learning and AI. In particular we are looking for PhD students with expertise in vision, but with a heavy emphasis on applying state-of-the-art vision methods and sensing for robotics in automotive environments: https://urldefense.proofpoint.com/v2/url?u=https-3A__relib.org.uk_&d=DwIFaQ&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=X06VOjNzAW9TW0La9aD7coy2dCY4qQvIWM1iDoSWDHo&s=ksGSXoCgNAqdkZysCBeO-urEB2HaT-LWSb9tLZAIhUA&e= The PhD candidate will focus on developing vision-guided robot control methods, understand the relevant recent literature, in order to plan experiments and papers. Recycling of Lithium Ion Battery (RELIB) project The ReLiB team at Extreme Robotics Lab will tackle the most demanding technical challenges in sensing and robotic sorting. The ReLiB project will have a significant impact on the safety, economics and efficiency of battery recycling whilst minimizing the environmental impact of these processes. The aim of this project is to implement the robotics techniques to remove the battery pack lid and to present individual modules to the sensor array and disassemble at cell level. Robots should connect packs, modules, cells to electrical testing machinery at various stages. The research activities of this PhD will be focused on developing the robotic manipulation techniques for disassembling the battery cells and implementing these techniques to the real industrial environments. The industrial infrastructure is prepared and the candidate will test his developed techniques to remove and disassemble the battery modules in this environment. The successful applicant will work closely with roboticists and vision researchers at Extreme Robotics Lab. The applicant should develop vision guided robotic manipulation techniques with advanced force-feedback control to unbolt, cut, grasp and manipulate the battery modules. Main Duties The responsibilities may include some but not all the responsibilities outlined below: • To use expertise in Computer Vision and/or vision-guided robotics, to develop integrated robotic systems for disassembly applications, driven by automotive industry. • Within the project scope of RELIB project, generate novel ideas that progress beyond the current state-of-the-art in computer vision, AI/machine learning, and vision-based robot control. • Develop new methods for combining these ideas with, and/or adding new directions to, the UoB robotics and vision and robotics research portfolio. • To understand the project scope, and appreciate the practical industrial applications which motivate our robotics, vision and AI research. • While having expertise in Computer Vision, be also willing and able to collaborate with robotic manipulation experts, to help develop integrated vision-guided robot control methods. • Independently understand the relevant recent literature, in order to plan experiments and papers. • Plan experiments and gather the data needed to publish ideas in leading international journals and conference proceedings. • Write up work as papers, of sufficient quality to be published in leading international journals and conference proceedings. • To take personal responsibility for ensuring that outlined parts of project deliverables are completed. Knowledge, Skills, Qualifications and Experience Required • 1st class degree in robotics, AI, machine learning, control engineering, mechanical engineering or a relevant discipline, at Masters level, or an equivalent overseas degree • English language requirement: https://urldefense.proofpoint.com/v2/url?u=https-3A__www.birmingham.ac.uk_postgraduate_pgt_requirements-2Dpgt_international_index.aspx&d=DwIFaQ&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=X06VOjNzAW9TW0La9aD7coy2dCY4qQvIWM1iDoSWDHo&s=NiArxRCVTkbTLQfMZs0DbYn3UhOcmagKewVNCPrU2U0&e= • Demonstrable knowledge of mathematics and algorithms for computer vision and relevant machine learning methods. • Demonstrated expertise and experience in 3D vision, i.e. working with point-cloud data and devices, is especially desirable. • Strong programming skills, including extensive knowledge and experience in C++. Matlab, Python, JAVA also desirable. • Ability to use OpenCV, VISP, PCL and other relevant open-source vision software libraries. ROS experience may also be useful, but ability to combine vision libraries with our lab’s bespoke C++ robot control software library is also important. To apply: Send you CV and cover letter summarising your research interests and suitability for the position to Dr Alireza Rastegarpanah: [a.rastegarpanah at bham.ac.uk] Start Date: Starting as early as possible ----- Dr. Alireza Rastegarpanah, Ph.D., MEng, MIEEE Faraday Institution Robotic Scientist Faraday Institution United Kingdom Mobile: +44 (0)7429224991 e-mail: a.rastegarpanah at bham.ac.uk W: www.relib.org.uk
Add on 22 March, 2019
At the BCAI at Bosch, we are building a new group for machine learning for robot manipulation. The group will be located at the University of Tübingen and led by Prof. Gerhard Neumann. The BCAI is a leading lab in machine learning and artificial intelligence which regularly publishes at top venues such as NIPS and ICML. Our aspiration for the new group is to become one of the leading robotics groups in world. We aim to make robot manipulation useable in the real world, such as cluttered unknown environments consisting of unknown, possibly deformable objects where fine and forceful manipulations need to be applied to the scene such as object assembly. Our research will be focused on the development of new learning algorithms that are useable in these scenarios. We are looking for new PhD students in the following areas: * Deep learning methods for robot vision and robot control * Model-based and model-free deep reinforcement learning * Imitation learning and inverse reinforcement learning * Hierarchical reinforcement learning * Uncertainty quantification in deep learning models * Learning from human feedback * Information-theoretic learning methods * Efficient robot movement representations * Skill libraries and hierarchical skill learning * Teleoperation and shared control for user guidance Your profile: * Personality: highly motivated to work on these challenging topics and develop a personal research agenda * Working Practice: team-oriented, open-minded and inquisitive * Experience and Knowledge: excellent skills in Python, ROS, C++ or Matlab and excellent math skills, knowledge in the areas of machine learning and robotics is preferential * Languages: very good in English (written and spoken) * Education: excellent degree (Master) in computer science, electrical engineering, mechanical engineering, math or related fields with excellent marks. For questions concerning the positions, please contact Gerhard Neumann (Gerhard.Neumann at de.bosch.com). Please apply at: https://urldefense.proofpoint.com/v2/url?u=http-3A__www.jobstairs.de_stellenangebot-2Dphd-2Dmachine-2Dlearning-2Dfor-2Drobot-2Dmanipulation-2Dbosch-2Dgroup-2D71272-2Drenningen-2D-2D1251342.html&d=DwIFAw&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=m7VC_U282klCL2E4wostOD9-APFOWrRorUjlxZwFHb0&s=s5xAtZYLuaBmKQRatm1GoatUhp5O0PBTRDkNX46hOkw&e= Mit freundlichen Grüßen / Best regards Gerhard Neumann Information Theoretic Reinforcement Learning (CR/PJ-AI-R33) Gerhard.Neumann at de.bosch.com<mailto:Gerhard.Neumann at de.bosch.com>
Add on 7 March, 2019
The Unmanned Systems Laboratory at the University of Tennessee, Chattanooga is searching for MS and PhD students to start in summer or fall of 2019. The successful candidates will work with department faculty, researchers, and graduate and undergraduate students on a variety of unmanned systems projects. In particular, we are looking for students who will contribute toward our research efforts in developing cooperative control, coordination, sensing, and communication technologies for multiple robotic systems. Required: - B.S. or M.S. degree in Engineering or a related field. - Must be a U.S. citizen or permanent resident. - Programming experience. - Strong background in cooperative control, cyber-physical systems, or related topics. Desired: - Familiarity with ROS, C, C++, Matlab/Simulink, and Python. - Experience with robotics hardware (i.e. data acquisition and signal processing, real-time feedback control, networking) - Experience flying fixed wing and/or multi-rotor unmanned aerial vehicles. (FAA Part-107 certification a plus) The compensation is competitive and negotiable. For more information, contact Dean Daniel Pack at daniel-pack at utc.edu or Zachary Ruble at zachary-ruble at utc.edu .
Add on 1 March, 2019
Technische Universität München (TUM) is one of Europe’s top universities. It is committed to excellence in research and teaching, interdisciplinary education, and the active promotion of promising young scientists. The university also forges strong links with companies and scientific institutions across the world. The Chair of Robotics Science and Systems Intelligence is a member of the newly founded TUM Munich School of Robotics and Machine Intelligence. The research focus of RSI is the development of control algorithms, mechatronics, intelligent robotics and prosthetics, robot learning algorithms, foundations of machine intelligence, as well as nonlinear control and systems theory. Furthermore, we offer unified teaching in terms of lectures, laboratories and student projects with a tight connection to cutting-edge research. Description Safety issues during physical human-robot interaction has attracted increasing attention from researchers in various fields of robotics. This is motivated by the fact that human and robot will work intensively and closely together. In order to ensure human safety for various human-robot contact scenarios, potential threats should be studied and safe human–robot interaction strategies should be developed. This can include: • Design and analysis of collision experiments and/or testing devices • Development and verification of collision simulations • Survey of biomechanics and forensics literature • Motion planning and/or control schemes for ensuring human safety • Prediction and control for optimal safety and performance trade-off • Adaption to the sensor field of view We are looking for individuals with experience in one or more of the following areas and strong interest to further develop their skills: • Masters-level degree in Robotics, Mechatronics, Computer Science or closely related • Strong background in robotics and control • Proficient programming experience (e.g. C++, MATLAB/Simulink, ROS) • Strong communication skills, including fluency in written and spoken English • Enthusiastic and highly motivated to complete a PhD • Prior experience in safety or human-robot interaction is a plus How to apply? Applications should include the following documents: • A motivation letter describing your research interests, your qualifications, and why you would be a suitable candidate • A detailed CV • Academic transcripts from your Bachelor’s and Master’s degrees • Email addresses of at least two references Interested applicants should send us the necessary documents via email to: saeed.abdolshah at tum.de quoting “PhD Application, Safe Human-Robot Interaction” in the e-mail subject. These positions will remain open until filled. Only shortlisted candidates will be notified. TUM has been pursuing the strategic goal of substantially increasing the diversity of its staff. As an equal opportunity and affirmative action employer, TUM explicitly encourages nominations of and applications from women as well as from all others who would bring additional diversity dimensions to the university’s research and teaching strategies. Preference will be given to disabled candidates with equal qualifications. International candidates are highly encouraged to apply. Sincerely, ------------------------------------- Saeed Abdolshah, PhD Postdoc researcher Munich School of Robotics and Machine Intelligence Technische Universität München Heßstraße 134 80797 München Email: saeed.abdolshah at tum.de Phone: +49 89 298 29413
Add on 5 October, 2018
The Human-Centered Assistive Robotics (HCR) Lab at Technical University of Munich (TUM) is looking for a talented Postdoctoral researcher in the field of robot learning. The expected research area includes cognitive robot motor skill learning and reinforcement learning with the focus on learning dynamics and context based skill manipulation learning.
The appointed researcher will be based at the Human-Centered Assistive Robotics group (link) at TUM under the supervision of Prof. Dongheui Lee. The new position offers the opportunity to engage in national and international collaboration, especially tight collaboration with the Institute of Robotics and Mechatronics at German Aerospace Center (DLR). The research group has access to a number of robotic platforms such as light weight robotic arms, humanoid robots, motion capturing systems, etc at TUM and DLR.
* PhD degree in engineering, computer science, or equivalent.
* Excellent mathematical and coding skills (C/C++, Matlab, ROS, Python).
* Excellent skills and experience in machine learning, robotics, and control
* Proficiency in writing and speaking English.
* Publications in recognized conferences/journals
* Interested to be engaged in teaching and education at TUM
Start: The position is available immediately.
Applications. Interested candidates are invited to send application to email@example.com including a CV, transcripts, a motivation letter, and a list of references.
Prof. Dr. Dongheui Lee
Human-centered Assistive Robotics
Technical University of Munich (TUM)
Tel. +49 89 289 25780
Institute of Robotics and Mechatronics
German Aerospace Center (DLR)
Add on 4 October, 2018
an exciting opportunity opens for a 3-years Postdoc position at ARQ (Advanced Robotics @ Queen Mary), the newly established Robotics centre of
the Queen Mary University of London, to work on a UK EPSRC-funded project on dexterous robotic manipulation.
Apply here: link
Deadline for applications: 01-Nov-2018 (interviews are expected to be held shortly thereafter).
Informal enquiries should be addressed to Dr Lorenzo Jamone (firstname.lastname@example.org).
Role: Postdoctoral Research Assistant - QMUL16208.
Research group: Advanced Robotics @ Queen Mary (ARQ).
Department: School of Electronic Engineering & Computer Science (EECS).
University: Queen Mary University of London (QMUL).
Salary: £33,615 - £37,411 (Grade 5).
Duration: 3 years.
The main task of the Postdoc will be to work on the EPSRC UK project “MAN^3: huMAN-inspired robotic MANipulation for advanced MANufacturing”.
The project is led by the Queen Mary University of London, in collaboration with three companies based in London: Shadow Robot, Google DeepMind and
Ocado. The research will be carried out at ARQ, the Advanced Robotics centre at Queen Mary, and within the school of EECS (Electronic Engineering and Computer Science), under the supervision of Dr Lorenzo Jamone.
The overall aim of the project is to improve dexterous robotic manipulation in terms of autonomy, adaptability, flexibility and robustness: in other
words, to make it more human-like. Data-driven learning and control, bootstrapped by human demonstrations, is the most appealing solution to
achieve this improvement. Also, human demonstrations will be more effective if the human has access to the same sensory feedback the robot has (nothing
more, nothing less) and if multimodal sensory feedback is available (vision, touch, proprioception); this is now possible due to the recent
advances in VR and sensing technologies. This leads to the three main objectives of MAN^3: developing a system for natural human-robot
demonstration of manipulation actions through immersive teleoperation (O1); delivering a solution for robust robotic grasping and dexterous in-hand
manipulation based on human demonstrations and robot learning (O2); applying the proposed solution to the human-demonstrated robotic picking,
handling and placing of delicate food items (O3).
Candidates must have a PhD degree in Robotics, Electronic or Mechanical Engineering, Computer Science, or related areas, with a solid publication
track record. Candidates should have good knowledge of computer programming (C++/Python) and software integration for robotics, robot control, and
basics of signal processing and electronics. Good knowledge of machine learning, estimation and optimization is highly desirable. Experience with real robotic hardware is essential, and experience with robotic manipulation, robot vision and tactile sensing would be highly beneficial. Solid knowledge of Linux and ROS environments is also highly desirable.
The post is a full time, fixed term appointment for the period of 36 months or until 31 December 2021 (whichever is the shorter), with tentative
starting date on 1 January 2019. Starting salary will be in the range of £33,615 - £37,411 per annum inclusive of London Allowance. Benefits include 30 days annual leave, pension scheme and interest-free season ticket loan.
Candidates must be able to demonstrate their eligibility to work in the UK in accordance with the Immigration, Asylum and Nationality Act 2006. Where
required this may include entry clearance or continued leave to remain under the Points Based Immigration Scheme.
QMUL (Queen Mary University of London) is a leading research oriented UK university, ranking top 20 in the UK and top 100 in the world (link). Located in the booming creative high-tech cluster of east London (link), the university hosts several multidisciplinary research groups in cognitive science, intelligent sensing, computer vision, multimedia, advanced materials, control, aerospace, and more. QMUL has recently opened a new centre for robotics (ARQ - Advanced Robotics @ Queen Mary, link), which is rapidly expanding and now counts more than 40 active brains including faculty members, postdocs, PhD students and a few non-completely-brainless robots. Join us!
Lorenzo Jamone - Lecturer in Robotics ARQ - Advanced Robotics at Queen Mary School of Electronic Engineering and Computer Science - link
Queen Mary University of London - link link
Add on 4 October, 2018
Flight dynamics and control lab of the George Washington University at Washington DC is looking for new doctoral students. There are three open positions for the following projects.
1. Aerial Mapping and Situation Awareness
This project aims to develop aerial robotic systems to monitor a complex urban environments, create a real-time map, and share it with other autonomous vehicles. Backgrounds in the following topics are desired.
c. Machine learning
d. Image processing
This position will be funded by NSF.
2. Dynamics and Control of Flapping Wing Unmanned Aerial Vehicle
This project aims to uncover the biomechanics of Monarch butterflies exhibiting the longest flight range among insects, and to utilize it to develop bio-inspired control schemes for long-range flapping-wing micro aerial vehicles. Backgrounds in the following topics are desired.
a. Nonlinear dynamics
b. Nonlinear control theory
d. Geometric mechanics
This position will be funded by NSF.
3. Uncertainty Propagation for Hybrid Systems
This project is to construct comprehensive computational techniques for stochastic analysis of hybrid systems evolving on a nonlinear configuration manifold, including uncertainty propagation, Bayesian estimation, and stochastic optimal control schemes.
In particular, backgrounds in the following topics are desired.
a. Stochastic analysis
b. Hybrid systems
d. Geometric mechanics
This position will be funded by AFOSR.
For every position, it is expected that the candidates have expertise in structured programming.
To apply, or to request more information, please contact Prof. Taeyoung Lee at email@example.com. The application should include CV, a list of publications, and a list of reference, and it should be submitted as a single PDF file.
Add on 1 October, 2018
[jobs] Postdoctoral Research Fellow in AI for Long-Term Autonomy
School of Computer Science, University of Lincoln, UK
Salary: From £33,199 per annum
Please note this post is full time and fixed term for up to 2 years.
Closing Date: Wednesday 31 October 2018
Interview Date: Wednesday 15 November 2018†
Enabling robotic systems to perform autonomously in complex, real-world scenarios over extended time periods (i.e. weeks, months, or years) poses many challenges. Some of these have been investigated by sub-disciplines of Artificial Intelligence (AI) including navigation and mapping, perception, knowledge representation and reasoning, planning, interaction, and learning.
The University of Lincoln is seeking to appoint a Postdoctoral Research Fellow to join the Lincoln Centre for Autonomous Systems (L-CAS). The successful candidate will be employed as part of the Horizon 2020 project ILIAD, involving fleets of robots that operate in intralogistics applications with a high demand on flexibility in environments shared with humans (see iliad-project.eu).
We are looking to recruit a postdoctoral Research Fellow with relevant experience and/or a keen interest in a number of research areas, including (not exclusively):
* AI methods for long-term robot autonomy;
* Persistent mapping and navigation;
* Long-term robotic perception;
* Human activity monitoring and prediction;
* Machine learning for autonomous robots;
* Systems integration, including ROS.
Applicants should have, or expect to soon obtain, a PhD in a relevant area. You must have excellent mathematical and coding skills (C++/Python, ROS). The new position offers the opportunity to engage in international collaboration within an ambitious team, to work with state-of-the-art robotic hardware and software, and to benefit from excellent support to produce and disseminate original research contributions in the leading conferences and journals.
This post will be fixed term for up to 24 months from 1st Jan 2019, with excellent opportunities for ongoing career development and progression. The successful candidate will contribute to the Universityís ambition to achieve international recognition as a research intensive institution and will be expected to design, conduct and manage original research in the above subject areas as well contribute to the wider activities of L-CAS.
Meet us at IROS 2018: Informal enquiries about the post can be made to Prof Tom Duckett (email: tduckett at lincoln.ac.uk), Prof Marc Hanheide (email: mhanheide at lincoln.ac.uk), or Dr Greg Cielniak (email: gcielniak at lincoln.ac.uk). We are all attending IROS 2018, so please feel free to contact us if you would like to meet for a chat during the conference.
To apply, please use the online application form at: link
Professor Tom Duckett FRSA
School of Computer Science
University of Lincoln
Add on 26 September, 2018
The Morelab.org has a number of fully funded PhD/Master positions. PhD and Master applicants with the background in Engineering and Computer Science (Robotics, Control Engineering, Robot Vision and AI) are preferable. Good knowledge in Math and practical experience in Computer Programming are required. Hands-on skills in ROS and Matlab are highly expected.
*Position 1*: High-level Perception for Human-robot Interaction and Collaboration
Good knowledge in computer vision/robot vision is required. Practical experience in sensor fusion and data association is plus.
*Position 2*: Socially Capable Mobile Robot Navigation
Good knowledge in robot perception, path planning, motion planning and control is required. A strong background in algorithms is plus. Hands-on skills of system integration are highly expected.
*Position 3*: Deep Reinforcement Learning for Socially aware Robot Navigation
Fundamental knowledge in machine learning for robotics is required. Practical experience of applied AI in robotics and autonomous systems is highly expected.
*Position 4*: Multi-robot Systems/Swarm Robotics
Solid background in Math and Control Theory is required. Algorithmic thinking and programming skills are required for this project.
You can find our demonstration of our research projects at: here
Please contact the lab director, Trung Dung Ngo (firstname.lastname@example.org) for more information. You should send your CV, degrees and transcripts, samples of your research work along with a research statement expressing your interests in a position.
*If you are in IROS2018, please send an email for a private meeting. You can also join me at the workshop "Robotic Co-worker 4.0".*
*Trung Dung Ngo, Ph.D, P.Eng, SMIEEE*
*Director of the More-Than-One Robotics Laboratory*
*Lead Researcher for Center for Excellence in Robotics and Industrial Automation*
*Faculty of Sustainable Design Engineering*
*University of Prince Edward Island*
*550 University Avenue, Charlottetown*
*Prince Edward Island**, CANADA C1A 4P3*
*Email: email@example.com , firstname.lastname@example.org *
*Faculty: www.upei.ca/engineering <https://urldefense.proofpoint.com/v2/url?u=http-3A__www.upei.ca_engineering&d=DwIBaQ&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=g5vaL8_NOiHQl4xSBWtfl3-gXGQ6a8RKZm4hzOtpWwM&s=HGMygzzf_HxUqVNBBgfOxQe7kDGiVQhpv5n080dWs1c&e=> *
Add on 25 September, 2018
the Agile Robotics and Perception Lab at the New York University is looking for 2 outstanding PhD students in the area of robotics, perception and machine learning
- *Position 1: Vision-based Agile Autonomous Navigation of Micro Aerial Vehicles*
- *Position 2 Human-aware perception and control of Micro Aerial Vehicles*
The lab performs fundamental and applied research in the area of robotics. The main mission of the lab is to create agile autonomous flying machines that can navigate all by themselves and collaborate with humans in different tasks using only onboard cameras, without relying on external infrastructure, such as GPS or motion capture systems. These machines need to be active and they should react to and navigate within their environment capturing its key properties.
Some related videos can be found at here
The lab has been founded by Prof. Giuseppe Loianno and is located in New York. The lab is fully equipped with a flying space and multiple aerial and ground robots.
The lab currently collaborates on numerous projects with worldwide partners institutions and companies (e.g. University of Pennsylvania, ARL, Qualcomm, TEPCO, SWRI, MIT, University of Zurich and ETH Zurich,..) You will get a very competitive salary and access to excellent research facilities (motion capture, 3D printing, a large flying arena, electronic and machine workshops).
New York is well known to be one of the most vibrant and exciting cities in the world. The lab and the school have a very enjoyable work atmosphere and organize many social events.
Candidates should send an email to loiannog at nyu.edu with subject line PhD application Position #. Candidates will need a one-page research interest
statement, CV, and 3 contact references when submitting the application. Candidates without the requested requirements will not be considered for
*Position 1: Vision-based Agile Autonomous Navigation of Micro Aerial Vehicles*
The candidate will create revolutionary algorithms combining perception action loop paradigms and machine learning schemes to enable and enhance agile navigation of multiple aerial platforms in dynamic and unstructured GPS-denied environments using vision as the main sensing modality. The position involves the development of novel algorithms to achieve robust and control of single and interconnected small scale Micro Aerial Vehicles (MAVs) and the development of high-performance MAV systems that can operate in extreme and challenging environments.
- Master degree or outstanding bachelor students in computer engineering, computer science, electrical engineering, mechanical engineering, robotics, physics, or related fields
- Excellent written and spoken English skills
- Be self-motivated and independent
- A strong interest in robotics, programming, computer vision, and mathematics
- Strong Experience with C/C++, ROS, OpenCV, TensorFlow, GPU optimization skills
*Nice to have*
- Background in filtering approaches to fuse multiple sensor data and real time control
- Previous experience in closed loop control, computer vision and state estimation
- Familiarity with LIDAR/Vision-based SLAM, Machine Learning for Semantic Classification
- Familiarity with tools such as Git is desirable
*Position 2 Human-aware perception and control of Micro Aerial Vehicles*
The candidate will develop new fundamental theories and algorithms to enable flying machines to collaborate with human in unstructured environments, at high speed and in a resilient way. The proposed approaches will help to obtain new ways of human-robot interaction helpful in tasks such as collaborative mapping of building, environment inspection, and transportation. The candidate will also study human behaviors and try to analyze reactive policies the drone has to take to accommodate human intention. This position involves to work using wearable different devices and communication protocols to interact with the robots.
- Master degree or outstanding bachelor student in computer engineering, computer science, electrical engineering, mechanical engineering, robotics,
physics, or related fields
- Excellent written and spoken English skills
- Experience with C/C++, Python, and ROS, Machine Learning tools, GPU optimization skills
- Be self-motivated and independent
- A strong interest in robotics, programming, and mathematics
*Nice to have*
• Familiarity with tools such as Git is desirable
• Previous works using wearable devices
Dr. Giuseppe Loianno
Director of the Agile Robotics and Perception Lab
NYU Tandon School of Engineering
Main organizer of the Vision-based Drones: What's Next? workshop at IROS
2018 in Madrid
Add on 24 September, 2018
-- OPEN POSITION --
A postdoctoral fellow position on multirobot systems and active sensing.
-- MENTOR --
Prof. Alberto Quattrini Li
Robotics Lab, Computer Science Department, Dartmouth College
-- DESCRIPTION --
The successful postdoc candidate will work on multirobot systems for active sensing -- with a main focus on the marine domain -- designing algorithms and assessing them through simulations and field trials with surface and underwater robots.
The postdoc is expected to carry out research; write papers for conferences and journals; participate in preparing proposals for research funding and in writing project reports; assist in organizing research meetings and workshops; and mentor graduate students.
The expected duration of the position is one year with renewal subject to funding availability and performance. Note that benefits will be provided.
Positions are available immediately with a possible option of starting in Winter, 2019. Candidates applying before October 31, 2018 will be given full consideration; applications will continue to be accepted until the positions are filled.
We recognize that strength comes through diversity and actively seek and welcome people with diverse backgrounds, experiences, and identities. We encourage women and minority candidates to apply.
-- QUALIFICATIONS --
- Ph.D. in Robotics, AI, Computer Science, Engineering
- A strong theoretical and practical background in multirobot coordination and decision-making under uncertainty
- Programming skills including experiences using C++ and/or Python
- Great presentation skills, both written and oral
- Experiences with ROS, networking, computer vision, machine learning techniques (deep learning, Gaussian Processes, …), and real robots will be a plus.
-- APPLICATION PACKAGE --
To apply, send a single file containing:
2. One page summary of research interests
3. One page thesis abstract
3. Names and contact information of at least three references
4. Dates of availability
5. A personal website, if available, where further details can be found
with a subject line:
POSTDOC APPLICATION 2018: LastName, Firstname
Applicants should also indicate if they will be in Madrid during IROS2019 for a possible meeting.
-- ROBOTICS @ DARTMOUTH --
The robotics laboratories at Dartmouth have a variety of robotics platforms, currently including 19 Turtlebot 3, 5 Parrot AR.Drone 2.0, an Adept SCARA arm, an ABB 6-DOF robot arm. More ground and marine platforms are being ordered, including a BlueROV2.
The lab also contains a small shop for construction of robot parts and prototypes. For rapid prototyping, we have an FDM 2000 rapid prototyping machine, band saw, drill press, and hand tools, as well as access to Dartmouth Thayer Engineering School machine shop and rapid prototyping facilities.
-- DARTMOUTH AND SURROUNDING AREA HIGHLIGHTS --
Dartmouth College, a member of the Ivy League, is located in Hanover, New Hampshire (on the Vermont border). Home to Dartmouth College, the Upper Connecticut Valley is a vibrant, academic and professional community offering excellent schools, lively arts, and an unmatched quality of life in a beautiful setting. Amenities associated with urban areas in Boston MA, Burlington VT, and Montreal QC are all within a few hours drive.
Alberto Quattrini Li
Department of Computer Science
Add on 18 September, 2018
Open Postdoctoral Positions in Robotics and Machine learning
The Interactive Robotics Laboratory (Ben Amor Lab) at Arizona State University invites applications for multiple postdoctoral positions with expertise in robotics, machine learning, and human-robot interaction. Positions are for 1 year each with the potential to be extended based on project performance. Postdoctoral fellows will collaborate on two projects funded by NASA and the National Science Foundation.
The focus of our research is on the development of novel reinforcement and imitation learning techniques for robot motor skill acquisition, e.g., grasping, walking, or human-robot collaboration. We are particularly interested in developing sample-efficient learning methods that incorporate multimodal input data. We are looking for talented postdoctoral fellows with a strong interest in one or multiple of the following topics:
- Imitation Learning
- Reinforcement Learning
- Machine Learning for Computer Vision
- Human-Robot Interaction and Collaboration
- (Bi-Manual) Grasping and Manipulation
- Robot Control and Learning for Control
Applications to: email@example.com
Initial Review Date: October 29, 2018
- Excellent written and verbal communication skills,
- Strong background in machine learning, mathematics, statistics, programming,
- A Ph.D. degree in Computer Science, Mathematics, Mechanical Engineering, or related topics,
- Excellent academic track record, with publications at top conferences, e.g., IROS, ICRA, RSS, AAAI, ICML, NIPS and Journals
- Experience with Linux, ROS, Git, OpenCV, PyTorch or any Deep Learning libary, and other robotics-related software platforms
- Ability to work in a collaborative team
About the ASU - Interactive Robotics Lab
Postdoctoral fellows will be embedded in a vibrant research group lead by Heni Ben Amor which is centered in Tempe, USA. The research group has strong collaborations with leading robotics and machine learning research groups in the US, Europe and Japan. Successful applicants will be encouraged to engage in these collaborations by visiting our collaborators abroad. Besides academic labs, we also closely collaborate with industrial partners including Honda Research Institute, Intel, Elementary Robotics and others.
Please submit as one PDF document a detailed letter of application stating qualifications and experience; curriculum vitae; and ideally two writing samples to Heni
Ben Amor at: firstname.lastname@example.org
Initial review of applications will begin on October 29, 2018 and continue every week thereafter until the search is closed.
About the Work Environment
Arizona State University (www.asu.edu) is a public research university with ~83,000 students spread over five campuses across the Phoenix metropolitan area, Arizona. ASU is classified as a research university by the Carnegie Classification of Institutions of Higher Education and is ranked among top research universities in the U.S., both private and public. Arizona State University has developed a new model for the American Research University, creating an institution that is committed to excellence, access and impact. ASU measures itself by those it includes, not by those it excludes. ASU pursues research that contributes to the public good, and ASU assumes major
responsibility for the economic, social and cultural vitality of the communities that surround it.
Add on 17 September, 2018
PhD / Research Assistant / Post-doc Positions in Perception, Manipulation, and Tactile Sensing for Bipedal Humanoid Robotics
Mechatronics and Haptic Interfaces Lab; Rice University; Houston, TX USA jointly with NASA Johnson Space Center's Robotic Systems Technology Branch; Houston TX
Rice University's Mechatronics and Haptic Interfaces Lab, directed by Professor Marcia O'Malley, is seeking candidates to fulfill three research and development positions in the areas of robotic perception, manipulation and tactile sensing development. Individuals recruited to fill these positions will be collaborating closely with NASA Johnson Space Center, and will be focused on bipedal humanoid application development.
Position types: We are recruiting prospective post-docs, full-time research assistants (with at least an MS degree), and/or PhD students to fulfill these roles. Salary/stipend will be commensurate with prior experience. Prospective PhD students are expected to join Rice University's PhD program in Mechanical Engineering effective January 2019.
Perception: A successful candidate will have a background in perception development for robotic systems, with a focus on fusion of LIDAR and image data to detect both objects and obstacles in a robot's environment. The candidate will be developing perception algorithms and pipelines for the Valkyrie bipedal humanoid, which uses a Carnegie Robotics MultiSense SL head for its primary vision based sensing. The ideal candidate will have experience developing robotic software using ROS on Linux Ubuntu operating systems, Python and C++ programming experience, and experience using the Git version control system. It is highly desired that the candidate has
experience in porting perception algorithms to physical robots and testing those algorithms in real-world environments.
Manipulation: A successful candidate will have a background in manipulation development for robotic systems, with a focus on intelligent grasping of objects. The candidate will be developing intelligent grasping capabilities for the Valkyrie bipedal humanoid, integrating perception data with haptic/force data to increase the current autonomous manipulation capabilities of Valkyrie. The ideal candidate will have experience developing robotic software using ROS on Linux Ubuntu operating systems,
Python and C++ programming experience, and experience using the Git version control system. It is highly desired that the candidate has experience in dual-armed manipulation and/or humanoid manipulation. It is also desired that the candidate has experience with physical robots and testing manipulation algorithms in real-world environments.
Tactile Sensing: A successful candidate will have a background in tactile sensing (systems integration and data processing) for robotic systems. The candidate will analyze existing tactile sensing in NASA's Valkyrie robot hand, and will explore the potential to increase sensing capabilities with existing modular fingers. The candidate will formulate and plan a technology development path to improve finger and sensor design and integration to support improved manipulation capabilities to handle fine motions without haptic feedback to an operator. The ideal candidate will have experience developing robotic software using ROS on Linux Ubuntu operating systems, Python, and C++ programming experience. Experience in both hardware integration and algorithm development and experimental validation is desired.
Each position is expected to be funded for twelve months. The candidates must be U.S. citizens or permanent residents, as the majority of his/her duties will be performed at NASA Johnson Space Center in Houston, TX. Candidates are expected to produce at least one scientific publication while funded and will be expected to provide documentation of all development and a final report upon conclusion of the position.
Application Process: Interested applicants should send, in a single email with 'RICE-NASA' in the subject line, to email@example.com, the following
1. In the body of the email:
a. A summary of research background and interests (<300 words)
b. Position sought (Post-doc, full time research assistant (not degree seeking), or PhD student (degree seeking))
c. If applicable, MS and/or Doctoral dissertation topic
c. Desired / available start date
2. As attachments (pdf)
a. Curriculum vitae including list of publications
b. Contact information for at least 3 references
Review of applications will begin immediately, and applicants are strongly encouraged to apply by October 1, 2018 for strongest consideration. Start dates are intended to be between October 1 and December 31, 2018. Applications for these positions received after December 1, 2018, will not be considered.
Rice University, an equal opportunity employer and a Tier 1 Research University, is located in the vibrant urban setting of Houston, TX, the fourth largest city in the U.S. The MAHI Lab, directed by Prof Marcia O'Malley, focuses on the design, manufacture, and evaluation of mechatronic or robotic systems to model, rehabilitate, enhance or augment the human sensorimotor control system. To this end, we employ analytical and experimental approaches from the field of dynamic systems and controls, with
inspiration from human motor control and neuroscience. link
The Lyndon B. Johnson Space Center (JSC) is the National Aeronautics and Space Administration's Manned Spacecraft Center, where human spaceflight training, research, and flight control are conducted. The Robotic Systems Technology Branch is responsible for the research, engineering, development, integration, and application of robotic hardware and software technologies for specific flight and ground robotic system applications in support of human spaceflight. The vision is to have humans and robots working side-by-side in the pursuit of exploration. Advanced robotic systems technology efforts include both remotely controlled and autonomous robots for space and terrestrial application, as well as intelligent robotics for high value functionality. Technology focal areas within the Branch include: Autonomy, Dexterous Manipulation, Human-Robot Interaction, Embedded Control Systems, Simulation, Tele-presence, Power Dense Actuation, and Sensing. link
Marcia K. O'Malley (Marcie), Ph.D.
Stanley C. Moore Professor of Mechanical Engineering
Director of Graduate Studies in Mechanical Engineering
Special Advisor to the Provost on Education and Research Initiatives for Collaborative Health (ENRICH)