7. Japan: Frontlines of “Robo-Conversion”

For waiters and waitresses in Japan, jobs are plentiful. For every applicant there are 3.8 wait jobs to choose from, which means 2.8 wait jobs go wanting. For drivers, there are 2.7 jobs per applicant, while for builders the ratio is 3.9 to 1.

That, of course, means that things don’t look as plentiful if you own a restaurant, a trucking company or want to build a shopping mall. Finding enough workers is just about impossible. A recent Financial Times headline says it all: Japan worker shortage has only one winner so far: robots.

6. Plagiarism and Myopia in AI

I have criticized the ImageNet Contest in an open letter to Fei-Fei Li:


I quote: “The AIML Contest series was meant to overcome major limitations of ImageNet Contests and other contests. It is meant for AMD, where a single learning engine must autonomously develop for any practical sensing modality and motor modality, including vision, audition, and natural language acquisition (text). It is meant for any practical tasks that a human teacher in the environment has on mind; however, the task to be learned is not known to the programmer of the learning engine. I invite you to take a look at the Rules of AIML Contest, which was successfully conducted during 2016 without getting a penny of government funding. AIML Contest will be conducted again in 2017.”

Please give your inputs for the BMI 2017 program:


Best regards,

Prof. John Weng, Michigan State University, USA

5. Some Thoughts on Humanoid Robotics Research

Without a human body, human intelligence cannot be developed on its own. This fact equally applies to robots. Most importantly, it explains why many research teams working in the field of robotics do build physical prototypes.

In recent decades, it has been an ambition of human beings to develop human-like machines, which are called robots in general and humanoid robots in particular. So far, tremendous progresses have been made in science and engineering for the understanding of design, analysis and control of grasping, biped walking and manipulating mechanisms. In parallel, a great deal of wonderful results has been generated from the investigation on various principles behind learning, and comprehension of language and images, etc.

Fueled by the excitements and impressive shows of HONDA’s and SONY’s humanoid robots, should we believe that a new era of humanoid robots has just started, and that more excitements are waiting ahead?

In fact, humanoid robots are human-made inventions and creatures. Therefore, the development of humanoid robots is basically invention-centric research, which should not be shadowed by our limited understanding about the nature. Instead, we have the full freedom in exercising our creativity for the invention of humanoid robots.

On the other hand, humanoid robotics research will gain a lot of benefits from the results of the discovery on how human body and mind work. As the outcome of invention will stimulate discovery-centric research in various ways, humanoid robots which are embodiment of mind and body, are undoubtedly ideal platforms for us to validate, or apply, theories from the study in neuroscience, psychology, learning, and cognition.

We all know that the embodiment of mechanics, electronics, control, communication, perception, decision-making, artificial psychology, and machine intelligence has greatly enlarged the scope of scientific investigation into the engineering approaches and principles underlying the development of humanoid robots. Because both the discovery-centric research and the invention-centric research in humanoid robotics could make progress hand-in-hand, this opens a new horizon in which fruitful results are expected to emerge in various forms of new theories, new technologies and new products. In comparison with industrial robotics research, humanoid robotics research will certainly offer much more opportunities and inspirations for new inventions and new discoveries.

Therefore, our goal toward this research direction is to adopt an integrative approach which aims at developing human-like Artificial Self-Intelligence (or artificial life) which could autonomously learn and develop its physical, intellectual, and emotional abilities through the interaction with human beings and environments. Some of our research focuses include: cognitive vision, cognitive speech, machine perception, machine learning, intelligent & real-time OS, image understanding, natural language understanding, conversational dialogue, machine translation, etc.

Ming Xie



4. Some Thoughts on Artificial Intelligence Research

With the increase in performance and capabilities of today’s computers, researchers and scientists are excited about the possibility of computerizing human intelligence in the form of computer programs so as to make computers possess a certain degree of human-like intelligence. In the middle of last century, this human endeavor has produced a new technical term that was named as: Artificial Intelligence (AI).

However, until today, there is no common consensus on the definition of Artificial Intelligence. For example, it is still not clear about whether Artificial Intelligence literally refers to computerized human-intelligence or machine’s self-intelligence.

As we know, the study of Artificial Intelligence historically focuses on problem-solving (i.e. analysis) and learning. The difficult issue of synthesis (i.e. creativity) has not yet been received much attention. Although artificial intelligence may literally mean man-made intelligence inside machines or robots, the contents discussed under the traditional paradigm of artificial intelligence suggests that it implicitly deals with computerized human-intelligence or computational intelligence (i.e. rationality).

Then, we can raise this question: What do we mean by intelligence from an engineering point of view? Refer to my book on Fundamentals of Robotics published in 2003.  One possible definition of intelligence is as follows: “Intelligence is the self-ability which links perception to actions so as to achieve intended outcomes. Intelligence is a measurable attribute, and is inversely proportional to the effort spent in achieving an intended goal.”

In a conference held in Italy in 2004, I have further made a concise definition of machine (or robot) intelligence as follows: “Robot intelligence is an attribute engendered by a robot’s brain, under the governance of causality and rationality. Causality is the study of intelligence without considering motivation (i.e. value and belief), while rationality is the study of intelligence in relation to motivation.”

In view of above definitions, it is clear to us that the achievement made so far in the field of AI is still very limited. Many questions remain un-answered, for example: How to make machines to acquire and learn knowledge by themselves? How to make machines to communicate knowledge or meanings by themselves? How to make machines to understand knowledge and meanings by themselves? How to make machines to synthesize (i.e. create) knowledge or meanings by themselves?

Therefore, it is still a tremendous challenge for us to develop innate algorithms or engineering principles underlying Artificial Self-Intelligence (AsI). And, our goal toward this research direction is to develop practical engineering solutions to the problem of how to make future machines and robots to autonomously learn, understand, synthesize, and communicate meanings.

Ming Xie


1. Can Robots Learn Languages the Way Children Do?

Share your responses to the question of “Can Robots Learn Languages the Way Children Do?”

You can find my original response here.

Below is my another version of response:

Can Robots Learn Language the Way Children do?

 Ming Xie

Nanyang Technological University

Singapore 639798

The question here is how to design the mind which enables robots to learn languages by experiencing the real world in the same way a child does. In particular, what does it meant by “experiencing the real world”?  Could the real world be modelled? How does the modelling of the real world help robots to autonomously learn, analyze and synthesize human languages through the interaction with the physical world (i.e. environment)? What is the definition of meanings?  What is the relationship between physical meanings and languages? What is the principle behind the process of learning human languages?  These are the issues that should be addressed with regard to the design of the mind for robots to learn and to understand human languages through interaction with the real world ([1],[2]).

Let’s first examine the issue of “what is the definition of meanings?”.

The understanding of the definition of meanings will help us to find the appropriate principle behind the design of a robot’s mind. In our opinions, the real world should be divided into both the physical world (i.e. environment) and the conceptual worlds (i.e. texts in various languages). Therefore, the meanings of a word in a human language will consist of two parts: a) the physical meanings of the entity, referenced by the word, in the physical world, and b) the conceptual meanings of the word itself, in various human languages. Here, we consider that an entity’s properties (i.e. geometrical, mechanical, chemical, electrical, etc) as well as its constraints (i.e. kinematic constraint, dynamic constraint, etc) are the physical meanings of the entity. Due to the nature of constraints, when multiple entities co-exist in a common space of the physical, interactions among these entities will occur. And, these interactions will create the concepts such as actions, behaviours, events, episodes and stories, etc. Therefore, along the history of mankind, the process of encoding the meanings in the physical world gives rise to the invention of human languages. On the basis of the inventive nature of human languages, we advocate that the use of human languages creates the so-called conceptual worlds. That is to say that a conceptual world is the set of texts in one human language, which describes the meanings of the physical world. Hence, multiple human languages produce multiple conceptual worlds. Most importantly, the properties and constraints of a word in a particular human language are simply the conceptual meanings of the word itself. For example, nouns, verbs, adjectives, proverbs are properties of words, while noun-phases and verb-phases are constraints of words. In summary, properties and constraints in the physical world as well as the conceptual worlds define what we call the meanings.

Then, let’s examine the issue of “what is the relationship between physical meanings and languages?”.

We all know that languages are the inventions of human beings for the purpose of encoding the physical meanings of entities in the physical world. Interestingly enough, the relationship between physical meanings and languages is similar to the relationship between scenes and cameras. For example, we can say that cameras are the inventions of human beings for the purpose of projecting the appearances of scenes into images. In a similar way, we can say that languages are the inventions of human beings for the purpose of projecting the physical meanings of entities into texts. In robot vision, one of the tasks is to do reconstruction or photo interpretation, which aims at reconstructing the scenes from given videos or images. Similarly, in robot hearing, one of the tasks is to do reconstruction or text understanding, which aims at reconstructing the physical meanings from given texts or sounds of texts.

Now, we come to this important question of “Can robots learn language the way children do?”

As mentioned above, properties and constraints are the contents of knowledge or meanings. And, the best way of representing these knowledge or meanings is the use of human languages. Therefore, the mastery of human languages is crucial to the development of robots of tomorrow which are capable of interact and communicate with human beings. Human children have the innate capability of mastering or learning any human language. This capability depends on two important factors. The first factor is the built-in blueprint of the mind which is the foundation of learning human languages. The second factor is the chance of interaction in the physical world during the process of learning human languages. As a result, if we could make robots of tomorrow to also gain these two conditions (i.e. the built-in blueprint of the mind similar to human beings’ one, and the ability of interacting in the physical world) , then robots will be able to learn language the way children do.

Actually, we have initiated the project with the aim of developing a robotics mind under the name of KnowNet, which is a software with the functionality such as teacher-assisted learning of human languages, vision-guided learning of human languages, visualization of physical meanings in 3D virtual space, text understanding, text synthesis, speech recognition, speech synthesis, conversational dialogue and multiple language translation, etc.


  1. Xie, Jayakumar. S. Kandhasamy and H.F. Chia. Meaning-centric Framework for Natural Text/Scene Understanding by Robots,International Journal of Humanoid Robotics,  1(2), June 2004.
  2. Jayakumar S. Kandhasamy, Organized Memory for Natural Text Understanding and Its Meaning Visualization by Machine,PhD thesis (Under Review), School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore (2005).