This thesis proposes a method that uses deep learning to implement a conversational robot. The ability of this robot includes: 1. turn on/turn off Wi-Fi. 2. turn on/turn off Bluetooth. 3. darken/brighten screen brightness. 4. turn up/trun down volume. 5. turn on camera. 6. basic self-introduction. Our robot also supports multi-task command such as "turn on Bluetooth and Wi-Fi, and turn down the screen brightness", which is out of many virtual assistant’s ability such as SIRI (Speech Interpretation and Recognition Interface). In this thesis, we will introduce how we create our training data, build our models, experimental result, and future work. We will also introduce development of this domain and explain why we use deep learning and seq2seq-based model to implement this conversational robot.