Understanding Artifact Evaluation in Design Science Research

Slide Note
Embed
Share

Explore the intricate process of evaluating artifacts in design science research with insights from Gondy Leroy, Ph.D., a seasoned expert in Management Information Systems. Discover the fundamental concepts, diverse study types, experiment design essentials, basic statistics, and contextual frameworks to enhance your understanding of this critical research area.


Uploaded on Sep 15, 2024 | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. Download presentation by click this link. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

E N D

Presentation Transcript


  1. TUTORIAL 2019 - EVALUATION OF ARTIFACTS IN DESIGN SCIENCE - PART 1 GONDY LEROY, PH.D. MANAGEMENT INFORMATION SYSTEMS UNIVERSITY OF ARIZONA GONDYLEROY@EMAIL.ARIZONA.EDU

  2. PRESENTER BACKGROUND Education B.S. and M.S. in Cognitive, Experimental Psychology, University of Leuven, Leuven, Belgium M.S. and Ph.D. in Management Information Systems, University of Arizona, Tucson, AZ Relevant Experience Principal Investigator $2.3M of funded research, by NIH, NSF, AHRQ, Microsoft Research Book Designing User Studies in Informatics , Springer (August 2011) Current Position and Contact Information Gondy Leroy, PhD Professor, Management Information Systems Director - Tomorrow's Leaders Equipped for Diversity Eller College of Management, University of Arizona http://nlp.lab.arizona.edu/

  3. ADDITIONAL RESOURCES Content based on Designing User Studies in Informatics Gondy Leroy, Ph.D. Springer, 2011 Freely available in most academic libraries (chapters can be downloaded electronically) Book information bridges gap between informatics, design science, and behavioral sciences. explains what an experimenter should pay attention to and why Practical, to-the-point, hands-on Contains a cookbook with step-by-step instructions for different types of artifact evaluations

  4. OVERVIEW 1. Design science and study types 2. Experiment design Independent, dependent, nuisance, confounding variables Short exercises 3. Basic statistics t-test, ANOVA 4. Exercise (Extra Materials not covered but available)

  5. DESIGN SCIENCE CONTEXT PART 1

  6. DESIGN SCIENCE Two paradigms in IS: behavioral and design science (Hevner et al 2004) Behavioral paradigm Develop theories, test theories Theories to understand, predict organizational and human phenomena relevant to information systems Design science Roots in engineering and AI (Simon 1996) Problem solving paradigm: create artifact to solve a problem Evaluation Can use many approaches to evaluate artifacts Simulation, mathematical description, impact studies Focus on correctness, effectiveness, efficiency Studies to evaluate impact Work with users in some cases Work with experts in some cases Work with gold standards in some cases

  7. DEVELOPMENT LIFE CYCLE Cyclical nature of development in informatics Digital libraries, online communities, mobile apps, Education, business, medical informatics, biology, commerce, Good fit of cycle (regardless of which one is adopted) with different types of user studies Importance of testing EARLY and FREQUENTLY: Requirements Analysis - most errors originate here, 60% undiscovered until user acceptance testing (Gartner Group, 2009) Example: Paper prototyping of interfaces: easy to accept changes Remodeling, renewing, improving: test new features against previous version Algorithm development and evaluation: 1) possibility to pinpoint location of strengths and weaknesses 2) Possibility of batch process evaluations System evaluation: test additional features, test synergy of all components together

  8. FOCUSING THE STUDY Increase your chances of getting published and funded by focusing the study according to these three principles: Define the goal of the system Will help choose comparison points (=define independent variable) and pinpoint measures (=dependent variables) Keep stakeholders in mind Why is the study conducted (to improve?), who is interested in it (what are they interested in?) Will help relate to people evaluating your artifact and study Timeline and Development Cycle What is available for testing, what can be tested in design phase? Will help design appropriate (series of) studies Example: reducing the number of no-shows at a clinic Goal: have more patients show up for appointment, don t increase workload of staff Stakeholders: patients (will need to be easy, no effort), staff at clinic (low training, easy to manage), purchaser at clinic (demonstrated effect on no- shows) Timeline and Development Cycle: paper prototyping for interface, comparison after implementation with existing system

  9. DIFFERENT STUDY TYPES (1/3) Naturalistic observation Individuals in their natural setting, no intrusion, ideally, people not aware Passive form of research, observation: Observe in person Use technology to observe (tracking, video, alerts, ) Case Studies, Field Studies and Descriptive Studies Several types of each exist Help explain and answer difficult questions, e.g., Why was the system not accepted? Can consider characteristics of work environment, culture, lifestyle and personal preferences when searching for explanations systematically controlled in experiments) can be combined with action research Action Research Case studies + direct involvement of the researcher Goal is to solve a problem or improve an existing situation Less role of observer but iterative and error-correcting approach

  10. DIFFERENT STUDY TYPES (2/3) Surveys Surveys are useful to measure opinions, intentions, feelings and beliefs, Dangers: 1) often hastily constructed, misconception about easiness of constructing a survey and validity, 2) Not a measure of behaviors or actions 3) Few people in IS properly trained to design surveys Correlation Studies About changes in variables, to find where change in one variable coincides with change in another. Involve many variables and many data points (surveys and large population samples) Do not attempt to discover what causes a change

  11. DIFFERENT STUDY TYPES (3/3) Quasi-experiment No randomization (main difference with experiments) Often useful when groups pre-exist because Geographical constraints, e.g., population across a country, in different cities social constraints, e.g., siblings in families time constraints, e.g., comparable group already studied in past Experiments (focus of tutorial) Goal is to evaluate hypotheses about causal relations between variables In informatics: evaluate the impact, benefit, advantages, disadvantages or other effects of information systems, algorithms, or interfaces, ... New/improved system is compared to other systems or under different conditions and evaluated for its impact

  12. TWO DIFFERENT KINDS OF EXPERIMENTS RELATED TO STAGES OF DEVELOPMENT Early Stages: focusing on algorithm/system development Indirect involvement of users Batch-process approach to experiments Example Advantages: large scale, efficient, quick turn around Example Dangers: confusing development and evaluation, forgetting design considerations (randomization, double blind evaluation, ) Later Stages: focusing on human-computer interaction, longitudinal evaluations of impact Direct involvement of users Example Advantages: rich data, useful information for product improvement Example Dangers: users not representative, IRB slows down process

  13. EXPERIMENTAL DESIGN IN A NUTSHELL PART 2

  14. THREE STEPS TO DESIGN A STUDY STEP 1: What is goal The answer helps define independent variables Independent variables: what will be manipulated STEP 2: How will we know that goal is reached The answer helps define dependent variables Dependent variables: what will be measured STEP 3: What can affect the system, users, use, actions, opinions, The answer helps define confounded and nuisance variables Confounded variables: 2 variables that change from one treatment to another Nuisance variables: variables that add errors and variance but are of no interest to the researcher (they should be controlled)

  15. STEP 1: CHOOSE THE INDEPENDENT VARIABLES Define the goal of the study Examples: Evaluate a new system where there was no system before? Evaluate whether a system is better than another system Evaluate whether different types of users get better use with system Independent Variable (IV) Other terms: treatment, intervention Manipulated by researcher Goal of a user study is to compare the results for different treatments Types of Independent Variable (IV) Qualitative Independent Variables Describe different kinds of treatments Quantitative Independent Variables Describes different amounts of a given treatment.

  16. STEP 1: CHOOSE THE INDEPENDENT VARIABLES There can be multiple independent variables In informatics, often only 1 or 2. Seldom more. Critical to make this a TRUE experiment Assignment to a condition/treatment has to be done randomly

  17. STEP 1: FIND THE INDEPENDENT VARIABLES EXERCISE Y. Gu, G. Leroy, D. Kauchak, "When synonyms are not enough: Optimal parenthetical insertion for text simplification," Accepted for the AMIA Fall Symposium, November 2017, Washington DC. Abstract As more patients use the Internet to answer health-related queries, simplifying medical information is becoming increasingly important. To simplify medical terms when synonyms are unavailable, we must add multi-word explanations. Following a data-driven approach, we conducted two user studies to determine the best formulation for adding explanatory content as parenthetical expressions. Study 1 focused on text with a single difficult term (N=260). We examined the effects of different types of text, types of content in parentheses, difficulty of the explanatory content, and position of the term in the sentence on actual difficulty, perceived difficulty, and reading time. We found significant support that enclosing the difficult term in parentheses is best for difficult text and enclosing the explanation in parentheses is best for simple text. Study 2 (N=116) focused on lists with multiple difficult terms. The same interaction is present although statistically insignificant, but parenthetical insertion can still significantly simplify text.

  18. STEP 1: FIND THE INDEPENDENT VARIABLES EXERCISE Y. Gu, G. Leroy, D. Kauchak, "When synonyms are not enough: Optimal parenthetical insertion for text simplification," Accepted for the AMIA Fall Symposium, November 2017, Washington DC. Abstract As more patients use the Internet to answer health-related queries, simplifying medical information is becoming increasingly important. To simplify medical terms when synonyms are unavailable, we must add multi-word explanations. Following a data-driven approach, we conducted two user studies to determine the best formulation for adding explanatory content as parenthetical expressions. Study 1 focused on text with a single difficult term (N=260). We examined the effects of different types of text, types of content in parentheses, difficulty of the explanatory content, and position of the term in the sentence on actual difficulty, perceived difficulty, and reading time. We found significant support that enclosing the difficult term in parentheses is best for difficult text and enclosing the explanation in parentheses is best for simple text. Study 2 (N=116) focused on lists with multiple difficult terms. The same interaction is present although statistically insignificant, but parenthetical insertion can still significantly simplify text.

  19. STEP 1: FIND THE INDEPENDENT VARIABLES EXERCISE G. Leroy, "Persuading Consumers to Form Precise Search Engine Queries", American Medical Informatics (AMIA) Fall Symposium, San Francisco, November 14-18, 2009. Abstract Today s search engines provide a single textbox for searching. This input method has not changed in decades and, as a result, consumer search behaviour has not changed either: few and imprecise keywords are used. Especially with health information, where incorrect information may lead to unwise decisions, it would be beneficial if consumers could search more precisely. We evaluated a new user interface that supports more precise searching by using query diagrams. In a controlled user study, using paper based prototypes, we compared searching with a Google interface with drawing new or modifying template diagrams. We evaluated consumer willingness and ability to use diagrams and the impact on query formulation. Users had no trouble understanding the new search method. Moreover, they used more keywords and relationships between keywords with search diagrams. In comparison to drawing their own diagrams, modifying existing templates led to more searches being conducted and higher creativity in searching.

  20. STEP 1: FIND THE INDEPENDENT VARIABLES EXERCISE G. Leroy, "Persuading Consumers to Form Precise Search Engine Queries", American Medical Informatics (AMIA) Fall Symposium, San Francisco, November 14-18, 2009. Abstract Today s search engines provide a single textbox for searching. This input method has not changed in decades and, as a result, consumer search behaviour has not changed either: few and imprecise keywords are used. Especially with health information, where incorrect information may lead to unwise decisions, it would be beneficial if consumers could search more precisely. We evaluated a new user interface that supports more precise searching by using query diagrams. In a controlled user study, using paper based prototypes, we compared searching with a Google interface with drawing new or modifying template diagrams. We evaluated consumer willingness and ability to use diagrams and the impact on query formulation. Users had no trouble understanding the new search method. Moreover, they used more keywords and relationships between keywords with search diagrams. In comparison to drawing their own diagrams, modifying existing templates led to more searches being conducted and higher creativity in searching.

  21. STEP 1: FIND THE INDEPENDENT VARIABLES EXERCISE C. H. Ku, A. Iriberri, and G. Leroy, Crime Information Extraction from Police and Witness Narrative Reports, 2008 IEEE International Conference on Technologies for Homeland Security, May 12-13, 2008 Abstract To solve crimes, investigators often rely on interviews with witnesses, victims, or criminals themselves. The interviews are transcribed and the pertinent data is contained in narrative form. To solve one crime, investigators may need to interview multiple people and then analyze the narrative reports. There are several difficulties with this process: interviewing people is time consuming, the interviews sometimes conducted by multiple officers need to be combined, and the resulting information may still be incomplete. For example, victims or witnesses are often too scared or embarrassed to report or prefer to remain anonymous. We are developing an online reporting system that combines natural language processing with insights from the cognitive interview approach to obtain more information from witnesses and victims. We report here on information extraction from police and witness narratives. We achieved high precision, 94% and 96%, and recall, 85% and 90%, for both narrative types.

  22. STEP 1: FIND THE INDEPENDENT VARIABLES EXERCISE C. H. Ku, A. Iriberri, and G. Leroy, Crime Information Extraction from Police and Witness Narrative Reports, 2008 IEEE International Conference on Technologies for Homeland Security, May 12-13, 2008 Abstract To solve crimes, investigators often rely on interviews with witnesses, victims, or criminals themselves. The interviews are transcribed and the pertinent data is contained in narrative form. To solve one crime, investigators may need to interview multiple people and then analyze the narrative reports. There are several difficulties with this process: interviewing people is time consuming, the interviews sometimes conducted by multiple officers need to be combined, and the resulting information may still be incomplete. For example, victims or witnesses are often too scared or embarrassed to report or prefer to remain anonymous. We are developing an online reporting system that combines natural language processing with insights from the cognitive interview approach to obtain more information from witnesses and victims. We report here on information extraction from police and witness narratives. We achieved high precision, 94% and 96%, and recall, 85% and 90%, for both narrative types.

  23. STEP 1: FIND THE INDEPENDENT VARIABLES EXERCISE G. Leroy, A. Lally, and H. Chen. The Use of Dynamic Contexts to Improve Casual Internet Searching, ACM Transactions on Information Systems (ACM - TOIS), vol. 21 (3), pp 229-253, July 2003. Abstract Research has shown that most users online information searches are suboptimal. Query optimization based on a relevance feedback or genetic algorithm using dynamic query contexts can help casual users search the Internet. These algorithms can draw on implicit user feedback based on the surrounding links and text in a search engine result set to expand user queries with a variable number of keywords in two manners. Positive expansion adds terms to a user s keywords with a Boolean and, negative expansion adds terms to the user s keywords with a Boolean not. Each algorithm was examined for three user groups, high, middle, and low achievers, who were classified according to their overall performance. The interactions of users with different levels of expertise with different expansion types or algorithms were evaluated. The genetic algorithm with negative expansion tripled recall and doubled precision for low achievers, but high achievers displayed an opposed trend and seemed to be hindered in this condition. The effect of other conditions was less substantial.

  24. STEP 1: FIND THE INDEPENDENT VARIABLES EXERCISE G. Leroy, A. Lally, and H. Chen. The Use of Dynamic Contexts to Improve Casual Internet Searching, ACM Transactions on Information Systems (ACM - TOIS), vol. 21 (3), pp 229-253, July 2003. Abstract Research has shown that most users online information searches are suboptimal. Query optimization based on a relevance feedback or genetic algorithm using dynamic query contexts can help casual users search the Internet. These algorithms can draw on implicit user feedback based on the surrounding links and text in a search engine result set to expand user queries with a variable number of keywords in two manners. Positive expansion adds terms to a user s keywords with a Boolean and, negative expansion adds terms to the user s keywords with a Boolean not. Each algorithm was examined for three user groups, high, middle, and low achievers, who were classified according to their overall performance. The interactions of users with different levels of expertise with different expansion types or algorithms were evaluated. The genetic algorithm with negative expansion tripled recall and doubled precision for low achievers, but high achievers displayed an opposed trend and seemed to be hindered in this condition. The effect of other conditions was less substantial.

  25. STEP 2: CHOOSE THE DEPENDENT VARIABLES What needs to be measured to know if goal was reached Examples: Were enough relevant articles found? Did people like using the system (or did they get frustrated)? Dependent Variable (DV) Other terms: outcome or response variable Metrics chosen by researcher Goal is to use metrics to compare different treatments, preferably use complementary measures to assess the impact of a treatment Choose a relevant dependent variable Keep the stakeholders in mind, what do they care about? Development phase affects choice Early on: do usability in formative phases of development Then: relevance, completeness of results, Later: cost savings, risk taking, improved decision making, Historically used metrics are good to include Because they are probably already well understood Because it will be expected by stakeholders Good starting point: effectiveness, efficiency and satisfaction (aka: outcome measures, performance measures, satisfaction measures)

  26. STEP 2: CHOOSE THE DEPENDENT VARIABLES Commonly used metrics for outcome measures Precision, recall, F-measure Accuracy, true positive, true negative, false positive, false negative, specificity, sensitivity Counts Commonly used metrics for performance measures Time (to completion) Errors Usability when measured in an objective manner: counting events, errors or measuring task completion. interesting measures compare different users on their training or task completion times. E.g., comparison between novice and expert users. Satisfaction and Acceptance Usually measured with one survey. NOTE: users need to be satisfied in short term before they will accept in the long term

  27. STEP 1: FIND THE DEPENDENT VARIABLES EXERCISE D. Kauchak, G. Leroy and A. Hogue, "Measuring Text Difficulty Using Parse-Tree Frequency", Journal of the Association for Information Science and Technology (JASIST), 68, 9, 2088-2100, 2017. Abstract - Text simplification often relies on dated, unproven readability formulas. As an alternative and motivated by the success of term familiarity, we test a complementary measure: grammar familiarity. Grammar familiarity is measured as the frequency of the 3rd level sentence parse tree and is useful for evaluating individual sentences. We created a database of 140K unique 3rd level parse structures by parsing and binning all 5.4M sentences in English Wikipedia. We then calculated the grammar frequencies across the corpus and created 11 frequency bins. We evaluate the measure with a user study and corpus analysis. For the user study, we selected 20 sentences randomly from each bin, controlling for sentence length and term frequency, and recruited 30 readers per sentence (N=6,600) on Amazon Mechanical Turk. We measured actual difficulty (comprehension) using a Cloze test, perceived difficulty using a 5-point Likert scale, and time taken. Sentences with more frequent grammatical structures, even with very different surface presentations, were easier to understand, perceived as easier and took less time to read. Outcomes from readability formulas correlated with perceived but not with actual difficulty. Our corpus analysis shows how the metric can be used to understand grammar regularity in a broad range of corpora.

  28. STEP 1: FIND THE DEPENDENT VARIABLES EXERCISE D. Kauchak, G. Leroy and A. Hogue, "Measuring Text Difficulty Using Parse-Tree Frequency", Journal of the Association for Information Science and Technology (JASIST), 68, 9, 2088-2100, 2017. Abstract - Text simplification often relies on dated, unproven readability formulas. As an alternative and motivated by the success of term familiarity, we test a complementary measure: grammar familiarity. Grammar familiarity is measured as the frequency of the 3rd level sentence parse tree and is useful for evaluating individual sentences. We created a database of 140K unique 3rd level parse structures by parsing and binning all 5.4M sentences in English Wikipedia. We then calculated the grammar frequencies across the corpus and created 11 frequency bins. We evaluate the measure with a user study and corpus analysis. For the user study, we selected 20 sentences randomly from each bin, controlling for sentence length and term frequency, and recruited 30 readers per sentence (N=6,600) on Amazon Mechanical Turk. We measured actual difficulty (comprehension) using a Cloze test, perceived difficulty using a 5-point Likert scale, and time taken. Sentences with more frequent grammatical structures, even with very different surface presentations, were easier to understand, perceived as easier and took less time to read. Outcomes from readability formulas correlated with perceived but not with actual difficulty. Our corpus analysis shows how the metric can be used to understand grammar regularity in a broad range of corpora.

  29. STEP 1: FIND THE DEPENDENT VARIABLES EXERCISE G. Leroy and T.C. Rindflesch, "Effects of Information and Machine Learning Algorithms on Word Sense Disambiguation with Small Datasets," International Journal of Medical Informatics, 74, 7-8, 573-585, 2005. Abstract - Current approaches to word sense disambiguation use and combine various machine-learning techniques. Most refer to characteristics of the ambiguous word and surrounding words and are based on hundreds of examples. Unfortunately, developing large training sets is time-consuming. We investigate the use of symbolic knowledge to augment machine learning techniques for small datasets. UMLS semantic types assigned to concepts found in the sentence and relationships between these semantic types form the knowledge base. A na ve Bayes classifier was trained for 15 words with 100 examples for each. The most frequent sense of a word served as the baseline. The effect of increasingly accurate symbolic knowledge was evaluated in eight experimental conditions. Performance was measured by accuracy based on 10-fold cross-validation. The best condition used only the semantic types of the words in the sentence. Accuracy was then on average 10% higher than the baseline; however, it varied from 8% deterioration to 29% improvement. In a follow- up evaluation, we noted a trend that the best disambiguation was found for words that were the least troublesome to the human evaluators.

  30. STEP 1: FIND THE DEPENDENT VARIABLES EXERCISE G. Leroy and T.C. Rindflesch, "Effects of Information and Machine Learning Algorithms on Word Sense Disambiguation with Small Datasets," International Journal of Medical Informatics, 74, 7-8, 573-585, 2005. Abstract - Current approaches to word sense disambiguation use and combine various machine-learning techniques. Most refer to characteristics of the ambiguous word and surrounding words and are based on hundreds of examples. Unfortunately, developing large training sets is time-consuming. We investigate the use of symbolic knowledge to augment machine learning techniques for small datasets. UMLS semantic types assigned to concepts found in the sentence and relationships between these semantic types form the knowledge base. A na ve Bayes classifier was trained for 15 words with 100 examples for each. The most frequent sense of a word served as the baseline. The effect of increasingly accurate symbolic knowledge was evaluated in eight experimental conditions. Performance was measured by accuracy based on 10-fold cross-validation. The best condition used only the semantic types of the words in the sentence. Accuracy was then on average 10% higher than the baseline; however, it varied from 8% deterioration to 29% improvement. In a follow- up evaluation, we noted a trend that the best disambiguation was found for words that were the least troublesome to the human evaluators.

  31. STEP 1: FIND THE DEPENDENT VARIABLES EXERCISE G. Leroy, S. Helmreich, and J. Cowie, "The Influence of Text Characteristics on Perceived and Actual Difficulty of Health Information", International Journal of Medical Informatics, 79 (6), 438-449, 2010. Abstract - Purpose: Willingness and ability to learn from health information in text are crucial for people to be informed and make better medical decisions. These two user characteristics are influenced by the perceived and actual difficulty of text. Our goal is to find text features that are indicative of perceived and actual difficulty so that barriers to reading can be lowered and understanding of information increased. Methods: We systematically manipulated three text characteristics, overall sentence structure (active, passive, extraposed-subject, or sentential-subject), noun phrases complexity (simple or complex), and function word density (high or low), which are more fine-grained metrics to evaluate text than the commonly used readability formulas. We measured perceived difficulty with individual sentences by asking consumers to choose the easiest and most difficult version of a sentence. We measured actual difficulty with entire paragraphs by posing multiple-choice questions to measure understanding and retention of information in easy and difficult versions of the paragraphs. Results: Based on a study with 86 participants, we found that low noun phrase complexity and high function words density lead to sentences being perceived as simpler. In the sentences with passive, sentential-subject, or extraposed-subject sentences, both main and interaction effects were significant (all p < .05). In active sentences, only noun phrase complexity mattered (p < .001). For the same group of participants, simplification of entire paragraphs based on these three linguistic features had only a small effect on understanding (p = .99) and no effect on retention of information. Conclusions: Using grammatical text features, we could measure and improve the perceived difficulty of text. In contrast to expectations based on readability formulas, these grammatical manipulations had limited effects on actual difficulty and so were insufficient to simplify the text and improve understanding. Future work will include semantic measures and overall text composition and their effects on perceived and actual difficulty.

  32. STEP 1: FIND THE DEPENDENT VARIABLES EXERCISE G. Leroy, S. Helmreich, and J. Cowie, "The Influence of Text Characteristics on Perceived and Actual Difficulty of Health Information", International Journal of Medical Informatics, 79 (6), 438-449, 2010. Abstract - Purpose: Willingness and ability to learn from health information in text are crucial for people to be informed and make better medical decisions. These two user characteristics are influenced by the perceived and actual difficulty of text. Our goal is to find text features that are indicative of perceived and actual difficulty so that barriers to reading can be lowered and understanding of information increased. Methods: We systematically manipulated three text characteristics, overall sentence structure (active, passive, extraposed-subject, or sentential-subject), noun phrases complexity (simple or complex), and function word density (high or low), which are more fine-grained metrics to evaluate text than the commonly used readability formulas. We measured perceived difficulty with individual sentences by asking consumers to choose the easiest and most difficult version of a sentence. We measured actual difficulty with entire paragraphs by posing multiple-choice questions to measure understanding and retention of information in easy and difficult versions of the paragraphs. Results: Based on a study with 86 participants, we found that low noun phrase complexity and high function words density lead to sentences being perceived as simpler. In the sentences with passive, sentential-subject, or extraposed-subject sentences, both main and interaction effects were significant (all p < .05). In active sentences, only noun phrase complexity mattered (p < .001). For the same group of participants, simplification of entire paragraphs based on these three linguistic features had only a small effect on understanding (p = .99) and no effect on retention of information. Conclusions: Using grammatical text features, we could measure and improve the perceived difficulty of text. In contrast to expectations based on readability formulas, these grammatical manipulations had limited effects on actual difficulty and so were insufficient to simplify the text and improve understanding. Future work will include semantic measures and overall text composition and their effects on perceived and actual difficulty.

  33. CONFOUNDED VARIABLES Confounding = when the effects of 2 (or more) variables cannot be separated from each other A variable, other than the independent variable, may have caused the effect Reduces the internal validity: unsure whether the independent variable caused the effect Random assignment to experimental condition is essential in avoiding confounding (but not always sufficient) Example: An experiment with 2 conditions to test a DSS for managers: old vs. new system Old system tested with experienced managers (25+ yrs) New system tested with inexperienced managers (1 yr)

  34. NUISANCE VARIABLES Nuisance variables add variation to the study outcome not due to the independent variables of no interest to the experimenter. reduce the chance of detecting the systematic impact of the independent variable Noise = if the variation is unsystematic E.g., conduct experiment at different times of day. At some times, the environment was noisy (train passes by, class changes and noise in hallway, people partying) affects performance of subjects Bias = if the variation is systematic E.g., conduct experiment and for each level of IV have different graduate student conduct study. One student is constantly chatting with friends during the experiment affects performance of subjects (true story)

  35. BASIC STATISTICS PART 3

  36. BASIC DESIGNS: TESTING WITH PEOPLE Example study: app vs consultant for weight loss intervention Between-subjects designs: - Each subject participates in only one experimental condition - Participants assigned to the app or consultant Within-subjects designs: - Each subject participates in only all experimental conditions - Participants assigned to app and consultant (make sure to reverse order for half of them)

  37. BASIC DESIGNS: TESTING WITHOUT PEOPLE Example study: Google Translate vs New Program for automated email translation Between-subjects designs: - Each subject participates in only one experimental condition - Each email assigned to Google or New Program Within-subjects designs: - Each subject participates in only all experimental conditions - Each email assigned to Google and New Program

  38. STATS FOR BASIC DESIGNS

  39. BETWEEN-SUBJECTS DESIGN: STATS When every treatment has a group of different subjects Statistics for 1 variable with 2 treatments: Independent samples t-test Independent Variable 1 Level 1 Level 2 n1-n30 n31-n60 Statistics for 2+ variables or 3+ treatments: ANOVA Independent Variable 1 Level 1 Level 2 Independent Variable 2 Independent Variable 2 Level 1 Level 2 Level 1 Level 2 n1-n30 n31-n60 n61-n90 n91-n120

  40. WITHIN-SUBJECTS DESIGN: STATS When subjects participate in multiple treatments Statistics for 1 variable with 2 treatments: Paired-samples t-test Independent Variable 1 Level 1 Level 2 n1-n30 n1-n30 Statistics for 2+variables or 3+ treatments: Repeated-Measures ANOVA Independent Variable 1 Level 1 Level 2 Independent Variable 2 Independent Variable 2 Level 1 Level 2 Level 1 Level 2 n1-n30 n1-n30 n1-n30 n1-n30

  41. SUMMARY T-test Comparison between 2 treatments Useful for both between- and within-subjects designs: Independent samples vs. Paired samples Bonferroni adjustment needed when many test are conducted ANOVA Comparison between 3 or more treatments Useful for both between- and within-subjects designs: ANOVA vs. repeated measures ANOVA Omnibus test: Main and Interaction effects Test if there is any significant difference between treatment (main effect) Post-hoc analysis needed to pinpoint which pairs of treatment are different A note about Multivariate ANOVA (MANOVA) Multiple dependent variables may inspire to MANOVA MANOVA has underlying factor analysis MANOVA appropriate when there are many dependent variables for which a simpler structures (fewer factors) are of interest Example: 35 measures of intelligence MANOVA can indicate when there are a few factors that contribute to results, e.g., 7 measure may load on verbal component , 8 measures may load on abstract component ,

  42. EXERCISE PART 4

  43. CHOOSE A TOPIC DESIGN A STUDY You have developed an app to help people diet using principles from psychology (e.g., the new noom ) You have developed an improved dashboard to track tasks/completion/personnel in a business You have developed text mining algorithms that can predict outbreaks of asthma from Twitter You have developed automated translation algorithms to translate legal text into layperson text

  44. DECIDE - What is new - How can you show that it works - What can influence that decision? Independent Variables? (what do you manipulate) Dependent Variables? (what are the important measures) Other?

  45. ADDITIONAL INFORMATION

  46. ERRORS TO AVOID

  47. ERRORS TO AVOID Trade-offs! Know your target population and sample Randomization is crucial in avoiding bias Of subjects assigned to conditions Facilitators assigned to conditions Order of conditions Of output for evaluation

  48. AVOID SUBJECT-RELATED BIAS when subjects act in a certain way because they are participating in a study Subject related bias, examples: Good subject effect: different behaviors because observed (want to look good) Volunteer effect (selection bias): volunteers of experiment are found to have different traits. E.g., may be healthier in medical studies, may need money from participation (drug interactions possible in clinical studies) How to avoid Location: clinic? Lab? Hospital? At work? Boss s office? Next to train station? try to avoid influences Explain importance of being honest Limit interaction with facilitator. Be careful of using only computer-based instructions Provide anonymity or confidentiality Single-blind studies = subject does not know experimental condition Make user task realistic

  49. AVOID EXPERIMENTER-RELATED BIAS Effects that are the result of experimenter/facilitator behaviors Experimenter related bias, examples: Experimenter effects are related to experimenter behaviors. Most famous example: Clever Hans (a horse) How to avoid Behaviors professional and courteous interaction standardized (practiced) instructions Working with multiple evaluators/facilitators (but not one per condition!) Reduce interaction with study subjects Double-blind study designs = both subject and facilitator are unaware of experimental condition Easier in IS than expected E.g., when comparing different algorithms

  50. AVOID DESIGN-RELATED BIAS Several biases introduced by use of a particular experimental design Within-subjects design: subjects participate in multiple conditions Minimize order-effects Cross-over design: balance order of conditions for subjects, e.g., A-B for half of subjects and B-A for other half When many conditions, randomize order of conditions per subject Leave enough time between conditions Between-subjects design: subjects participate in only one condition Minimize effects of time/day Randomize assignment to conditions Avoid assignment of one facilitator per condition

Related