Pub Date : 2024-01-02DOI: 10.1017/s089006042300015x
Farnaz Tehranchi, Amirreza Bagherzadeh, Frank E. Ritter
User models that can directly use and learn how to do tasks with unmodified interfaces would be helpful in system design to compare task knowledge and times between interfaces. Including user errors can be helpful because users will always make mistakes and generate errors. We compare three user models: an existing validated model that simulates users’ behavior in the Dismal spreadsheet in Emacs, a newly developed model that interacts with an Excel spreadsheet, and a new model that generates and fixes user errors. These models are implemented using a set of simulated eyes and hands extensions. All the models completed a 14-step task without modifying the system that participants used. These models predict that the task in Excel is approximately 20% faster than in Dismal, including suggesting why, where, and how much Excel is a better design. The Excel model predictions were compared to newly collected human data (N = 23). The model’s predictions of subtask times correlate well with the human data (r2 = .71). We also present a preliminary model of human error and correction based on user keypress errors, including 25 slips. The predictions to data comparison suggest that this interactive model that includes errors moves us closer to having a complete user model that can directly test interface design by predicting human behavior and performing the task on the same interface as users. The errors from the model’s hands also allow further exploration of error detection, error correction, and different knowledge types in user models.
{"title":"A user model to directly compare two unmodified interfaces: a study of including errors and error corrections in a cognitive user model","authors":"Farnaz Tehranchi, Amirreza Bagherzadeh, Frank E. Ritter","doi":"10.1017/s089006042300015x","DOIUrl":"https://doi.org/10.1017/s089006042300015x","url":null,"abstract":"<p>User models that can directly use and learn how to do tasks with unmodified interfaces would be helpful in system design to compare task knowledge and times between interfaces. Including user errors can be helpful because users will always make mistakes and generate errors. We compare three user models: an existing validated model that simulates users’ behavior in the Dismal spreadsheet in Emacs, a newly developed model that interacts with an Excel spreadsheet, and a new model that generates and fixes user errors. These models are implemented using a set of simulated eyes and hands extensions. All the models completed a 14-step task without modifying the system that participants used. These models predict that the task in Excel is approximately 20% faster than in Dismal, including suggesting why, where, and how much Excel is a better design. The Excel model predictions were compared to newly collected human data (N = 23). The model’s predictions of subtask times correlate well with the human data (r2 = .71). We also present a preliminary model of human error and correction based on user keypress errors, including 25 slips. The predictions to data comparison suggest that this interactive model that includes errors moves us closer to having a complete user model that can directly test interface design by predicting human behavior and performing the task on the same interface as users. The errors from the model’s hands also allow further exploration of error detection, error correction, and different knowledge types in user models.</p>","PeriodicalId":501676,"journal":{"name":"AI EDAM","volume":"49 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139077383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Engineering design has proven to be a rich context for applying artificial intelligence (AI) methods, but a categorization of such methods applied in AI-based design research works seems to be lacking. This paper presents a focused literature review of AI-based methods mapped to the different stages of the engineering design process and describes how these methods assist the design process. We surveyed 108 AI-based engineering design papers from peer-reviewed journals and conference proceedings and mapped their contribution to five stages of the engineering design process. We categorized seven AI-based methods in our dataset. Our literature study indicated that most AI-based design research works are targeted at the conceptual and preliminary design stages. Given the open-ended, ambiguous nature of these early stages, these results are unexpected. We conjecture that this is likely a result of several factors, including the iterative nature of design tasks in these stages, the availability of open design data repositories, and the inclination to use AI for processing computationally intensive tasks, like those in these stages. Our study also indicated that these methods support designers by synthesizing and/or analyzing design data, concepts, and models in the design stages. This literature review aims to provide readers with an informative mapping of different AI tools to engineering design stages and to potentially motivate engineers, design researchers, and students to understand the current state-of-the-art and identify opportunities for applying AI applications in engineering design.
{"title":"Mapping artificial intelligence-based methods to engineering design stages: a focused literature review","authors":"Pranav Milind Khanolkar, Ademir Vrolijk, Alison Olechowski","doi":"10.1017/s0890060423000203","DOIUrl":"https://doi.org/10.1017/s0890060423000203","url":null,"abstract":"<p>Engineering design has proven to be a rich context for applying artificial intelligence (AI) methods, but a categorization of such methods applied in AI-based design research works seems to be lacking. This paper presents a focused literature review of AI-based methods mapped to the different stages of the engineering design process and describes how these methods assist the design process. We surveyed 108 AI-based engineering design papers from peer-reviewed journals and conference proceedings and mapped their contribution to five stages of the engineering design process. We categorized seven AI-based methods in our dataset. Our literature study indicated that most AI-based design research works are targeted at the conceptual and preliminary design stages. Given the open-ended, ambiguous nature of these early stages, these results are unexpected. We conjecture that this is likely a result of several factors, including the iterative nature of design tasks in these stages, the availability of open design data repositories, and the inclination to use AI for processing computationally intensive tasks, like those in these stages. Our study also indicated that these methods support designers by synthesizing and/or analyzing design data, concepts, and models in the design stages. This literature review aims to provide readers with an informative mapping of different AI tools to engineering design stages and to potentially motivate engineers, design researchers, and students to understand the current state-of-the-art and identify opportunities for applying AI applications in engineering design.</p>","PeriodicalId":501676,"journal":{"name":"AI EDAM","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138577413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}