Skip to content Skip to sidebar Skip to footer

Human-computer interaction

Start developing a more efficient AI assistant by first understanding and replicating the unpredictable actions of humans.

Artificial Intelligence (AI) researchers at MIT and the University of Washington have created a model that can predict a human's decision-making behaviour by learning from their past actions. The model incorporates the understanding that humans can behave sub-optimally due to computational constraints — essentially the idea that humans can't spend indefinitely long periods considering the…

Read More

An improved, more efficient method to inhibit AI chatbots from delivering harmful responses.

Read More

The automated platform instructs users on the appropriate instances to engage with an AI assistant.

Researchers at MIT and the MIT-IBM Watson AI Lab have developed an AI system designed to educate users on when to trust an AI's decision-making process - for instance, a radiologist determining if a patient's X-ray shows signs of pneumonia. The training system identifies scenarios where the human should not trust the AI model, automatically…

Read More

The system instructs users on determining the appropriate times to work in conjunction with an AI assistant.

Researchers from MIT and the MIT-IBM Watson AI Lab have developed a system to teach users of artificial intelligence (AI) technology when they should or shouldn't trust its outcomes. This could be particularly beneficial in the medical field, where errors could have serious repercussions. The team created an automated system to teach a radiologist how…

Read More

The automated platform instructs users on the appropriate instances to engage with an AI assistant.

Researchers at MIT and the MIT-IBM Watson AI Lab have developed a method of teaching users when to collaborate with an artificial intelligence (AI) assistant. The model creates a customised onboarding process, educating users on when to trust or ignore an AI model’s advice. The training process can detect situations where the AI model is…

Read More

The automated system guides users on when to partner with an artificial intelligence assistant.

Researchers at the Massachusetts Institute of Technology and the MIT-IBM Watson AI Lab have developed an onboarding system that trains humans when and how to collaborate with Artificial Intelligence (AI). The fully automated system learns to customize the onboarding process according to the tasks performed, making it usable across a variety of scenarios where AI…

Read More

The automated platform instructs users about the appropriate time to partner with an AI assistant.

Researchers at MIT and the MIT-IBM Watson AI Lab have developed a system that teaches users when to trust AI and when to ignore it, and it has already led to an approximately 5% increase in accuracy during image prediction tasks. The researchers designed a customised onboarding process, which is when the user is familiarized…

Read More

The automated system educates users on the appropriate times to partner with an AI assistant.

Researchers from MIT and the MIT-IBM Watson AI Lab have developed a system that instructs users when to trust an AI system’s decision-making. In medicine, there might be instances like a radiologist using an AI model for reading X-rays where human intervention can make a difference. However, clinicians are uncertain whether to lean on the…

Read More

The automated system instructs users on the appropriate time to work together with an AI assistant.

Researchers from MIT and the MIT-IBM Watson AI Lab have developed an onboarding process which teaches users how to effectively collaborate with artificial intelligence (AI) assistants. The system was designed to provide guidance to users and to improve collaboration between humans and AI. The automated system learns how to create the onboarding process by gathering…

Read More

The automated system provides guidance on when to engage with an AI assistant for collaboration.

Researchers from MIT and the MIT-IBM Watson AI Lab have developed an automated system that trains users on when to collaborate with an AI assistant. In medical fields such as radiology, this system could guide a practitioner on when to trust an AI model’s diagnostic advice. The researchers claim that their onboarding procedure led to…

Read More

The automated system instructs users on the appropriate times to engage with an AI assistant.

Researchers at MIT and the MIT-IBM Watson AI Lab have developed a system that trains users on when to trust an AI model's advice. This automated system essentially creates an onboarding process based on a specific task performed by a human and an AI model. It then uses this data to develop training exercises, helping…

Read More

An automated platform instructs users on the appropriate timing for partnership with an AI assistant.

Researchers at MIT and the MIT-IBM Watson AI Lab have outlined an onboarding process, which includes training for users of artificial intelligence (AI) tools to better comprehend and utilise them. With a 5% accuracy improvement, the system setup enables a user to discern when to collaborate with AI by providing a personalised training programme. The AI…

Read More