Lighthouse Studio

MaxDiff Tutorial and Example

Creating a Sawtooth Software Account

 

If you aren't already registered to use Lighthouse Studio with a Sawtooth Software Account, the first time you launch the software (Start | Programs | Sawtooth Software | Sawtooth Software Lighthouse Studio) you are shown a dialog wherein you specify your Email and a Password and click Create New Account.  After creating the account you will receive an email asking you to confirm account creation by following the link in the email. Once your account is confirmed, return to the software to Sign in.

 

When signed in to your Sawtooth Software Account you will see your account email address at the upper-right hand side of the application. You can click on that link to view your license information, go to your account online, or sign out.  

 


Introduction

 

This tutorial leads you step-by-step through a simple MaxDiff study, from thinking about a research problem and planning a questionnaire to entering a list of MaxDiff items in Lighthouse Studio, generating and testing the questionnaire, and analyzing the results.  

 

Lighthouse Studio is a powerful program, and this tutorial only touches on the basics for running a MaxDiff experiment.  You'll become aware of different or more sophisticated approaches by reading other sections in the on-line help and by experimenting with Lighthouse Studio.

 

(Note: this example is based on a presentation delivered at the 2003 Sawtooth Software Conference by Amanda Kraus, Diana Lien, and Bryan Orme.  Though many of the items we use in this example are identical to the ones used in the conference presentation, the data we have developed for this example are purely fictional.  The suggested steps represent one approach to this type of research problem, but may not necessarily reflect the single "best" way to conduct such a research study using Lighthouse Studio.)

 


The Research Problem

 

As the economy improved in 2003, the Navy faced declining reenlistment rates among its sailors.  The Navy is looking for ways to improve the reenlistment rate.  Previous research has shown that reenlistment terms, quality of life issues, and the work environment are all significant factors when sailors face the reenlistment decision.  Pure economic aspects matter as well, such as the amount of the reenlistment bonus.  It's 2003 and you have been brought on as a consultant to help the military design and execute a research study to prioritize a list of possible job and quality of life enhancements as well as reenlistment terms.  The Navy is keenly interested in how reenlistment bonuses may affect sailors' desire to enlist relative to other non-monetary issues.  Those improvements which are most likely to increase the reenlistment rates at the lowest expense to the military should be implemented first.  They would like to implement this study over the Web, and expect that between 1,000 and 2,000 sailors might complete the study.

 

At first, you consider whether a conjoint method such as CBC would be appropriate.  You ask your military client, Jane, if the resulting model needs to be able to project the likely reenlistment rate when combinations of job enhancements are taken together.  Jane assures you that although that type of analysis would be of interest to military planners, the primary consideration is the ability to prioritize the items from most to least effective in terms of improving reenlistment rates.  The planners need to be able to directly compare the effect of each job enhancement or monetary bonus.

 

After pondering the problem, you recognize that a strength of MaxDiff is its ability to place multiple items on the same scale using a trade-off methodology.  With conjoint analysis (such as CBC), the client would not be able to directly compare the preference for levels of one job-related item to another.  MaxDiff doesn't formally support adding the preference scores across items and predicting the overall benefit of the job package or likelihood of reenlistment given a specific job package (as would CBC), but that is not a requirement for the study.  Therefore, you decide to use MaxDiff.

 


Developing a List of Items

 

You meet with Jane to discuss the elements (items) to be studied in the MaxDiff exercise.  Jane explains that a series of focus groups had been conducted over the last four months wherein the reasons for reenlistment versus leaving the Navy were discussed.  Hundreds of ideas had been presented, but there seemed to be some common themes.  Furthermore, Jane had discussed the myriad ideas with military planners.  Many of the ideas were clearly off-the-table (such as "king-sized beds on submarines," "more parties," and "supermodels as supervisors").  However, there were a number of areas of overlap: job conditions that the military was in the position to change and those that also were on the minds of the sailors.

 

The list included the following:

 

Length of reenlistment obligation

Location or duty assignment guarantees for next assignment

Faster promotion

Living conditions when in port (2-person barracks vs. 4-person barracks)

Increase in sea pay

Percent of on-job time using your skills and training

Time off each week allowed for on-line college coursework

Online career counseling services

Personal counseling services

Reenlistment bonus

 

This looked like a good start, but it seemed obvious to both you and Jane that some further refinement and quantification of benefits was needed.  The items needed to carry specific meaning to the sailors and the results needed to be actionable for the military planners.  Most of the items seemed to require multiple levels, whereas some seemed to stand on their own.  After thinking about the reasonable items (and some back-and-forth with Jane's internal clients, the military planners) a final list was drawn up:

 

Factors Affecting Reenlistment Decision:

3-year reenlistment obligation

5-year reenlistment obligation

Location guarantee for next assignment

Duty guarantee for next assignment

Both location and duty guarantees for next assignment

Get promoted 6-months sooner than expected

Live in 2-person barracks when in port

Live in 4-person barracks when in port

Choice between living in 2-person barracks or 4-person barracks when in port

$50 per month pay increase when out at sea

$75 per month pay increase when out at sea

Half of on-job time using your skills and training

Three-quarters of on-job time using your skills and training

Time off each week allowed for on-line college coursework

Receive online career counseling services

Receive in-person personal counseling services

Receive phone-based personal counseling services

Current reenlistment bonus ($2,500)

Increased reenlistment bonus ($5,000)

Increased reenlistment bonus ($10,000)

 

Next, you worked with Jane to decide which items would need to be prohibited from displaying with each other versus which could freely combine with all other items within the same set.  For example, it would not make sense to ask sailors to choose between a $50 pay increase versus a $100 pay increase when out at sea.  Every rational sailor would choose higher pay over lower, so this would be a wasted comparison.  Other items such as living in a 2-person barracks versus a 4-person barracks didn't seem to have a rational preference order (though having a choice between 2-person barracks or 4-person barracks would seem to be preferred to no choice).  Perhaps some sailors would prefer to live with more people, whereas others would prefer just one roommate.

 

You organized the list into groupings of items (listed as groups A through F), where the items within groups A through E would require at least some prohibitions (restrictions that they could not be displayed together within the same choice set).

 

A) Prohibit these two items:

3-year reenlistment obligation

5-year reenlistment obligation

 

B) Prohibit third item with appearing with either item 1 or item 2:

Location guarantee for next assignment

Duty guarantee for next assignment

Both location and duty guarantees for next assignment

 

C) Prohibit third item with appearing with either item 1 or item 2:

Live in 2-person barracks when in port

Live in 4-person barracks when in port

Choice between living in 2-person barracks or 4-person barracks when in port

 

D) Prohibit these two items:

$50 per month pay increase when out at sea

$75 per month pay increase when out at sea

 

E) Prohibit these items from appearing together:

Current reenlistment bonus ($2,500)

Increased reenlistment bonus ($5,000)

Increased reenlistment bonus ($7,500)

 

F) No prohibitions with other items:

Half of on-job time using your skills and training

Three-quarters of on-job time using your skills and training

Get promoted 6-months sooner than expected

Time off each week allowed for on-line college coursework

Receive online career counseling services

Receive in-person personal counseling services

Receive phone-based personal counseling services

 


Scripting the Questionnaire

 

A real research project would involve other questions about each sailor, including gender, current rank or job specification, current location, inclination to reenlist, etc.  Standard survey questions (such as select questions with radio buttons or check-boxes) can be developed within Lighthouse Studio (of which MaxDiff is a subcomponent).  Rather than take time within this tutorial to add such classification questions, we'll focus on the MaxDiff section of the survey.  Interested users can look to the Lighthouse Studio Tutorial within this same online help and manual for further help with standard survey questions.

 

You decide to include the following introductory text leading into the MaxDiff questions:

 

Start:

 

Thank you for taking time to help the Navy consider ways to improve your working environment and career development.  Our goal is to make the Navy a satisfying career for you and to improve your desire to reenlist when the decision arises.

 

We would like you to imagine that you are facing the decision whether to reenlist or not--today.  Please think about the issues involved, including future career planning, family/relationship considerations, job location, and compensation.

 

Click the Next button below to continue...

 

IntroMaxDiff:

 

If you were making the decision to reenlist or not today, there would certainly be many factors that affect your decision to reenlist or not.  Real life decisions are not always easy.  

 

We want to gain some insight into how you make this decision by asking you to evaluate a series of tradeoff scenarios.  In each scenario, we'll show you four possible factors affecting your career in the Navy and the reenlistment decision.  

 

We'll ask which factor (among this set of four) would make you most want to reenlist, and which would make you least want to reenlist.  

 

We need to ask you repeated tradeoff scenarios (involving different factors each time) so that we can learn what is truly important to you when deciding whether to reenlist.

 

 

Layout for MaxDiff Questions

 

Next, you need to decide the layout of the MaxDiff questions.  Research has suggested that showing more than about five items at a time in MaxDiff questions may not be very beneficial.  Some of the items in your study are a bit on the wordy side, so you decide to show just four items at a time.

 

The MaxDiff documentation suggests showing each item between 3 to 5 times on average per respondent across the sets.  With 20 items in the study, the number of questions to display so that each item is shown on average 3 times per respondent is equal to:

 

 3(K/k)

 

 where K is the total number items in the exercise, and k is the number of items per set

         

Therefore, you would need to ask each respondent a minimum of 3(20/4) = 15 sets.  The "nice to have" number of sets would be 5(20/4) = 25.  After discussing with the client, you decide that respondents should be able to complete 20 sets with little problem (each item displayed on average 4 times to each respondent).  You estimate that each set will take about 12-15 seconds to complete, for a total average interviewing time of about 4-5 minutes for the MaxDiff section.  

 

To try to keep respondents engaged and working productively through the MaxDiff exercise, you plan to show respondents a progress bar and also provide a count-down message regarding how many tasks remain.

 

Here's an example of the MaxDiff question layout you plan to use:

 

maxdifftut9

 

After the fifth MaxDiff question, you plan to show some text that provides encouragement and a countdown, such as:

 

ProgressCounter:

 

Good work so far.  We know that some of these tradeoffs are challenging.  

 

We need to ask you to evaluate 15 more sets so that we can better determine how you make your reenlistment decision.

 

 

And at the end, you want to display a screen that thanks respondents, and then immediately sends them to a Navy site www.navy.mil.

 

Finish:

 

Thank you very much for your help.  Your data will be used to help Navy planners as they work to make the Navy an even better career choice.  

 

Click the Next button to finish submitting your data and to return to the Navy's official website.

 


Using Lighthouse Studio to Write the Questionnaire

 

Start by opening the Sawtooth Software Lighthouse Studio.  If you have a standard installation, you start it by clicking Start | Programs | Sawtooth Software | Sawtooth Software Lighthouse Studio.  MaxDiff is one component within Lighthouse Studio. (Although not emphasized here, Lighthouse Studio also may be used for conducting paper-based MaxDiff projects.)

 

After starting Lighthouse Studio, choose File | New Study….  The New Lighthouse Study dialog appears, with your cursor active in the Name field.

 

New Study

 

The Location field lists the folder (directory) that Lighthouse Studio currently is using to store studies.  You can use any folder you like for this tutorial project.  You can browse to and create new folders by clicking the Browse... button to the right of the Location field, and then (after browsing to the folder in which you want to create a new sub-folder to contain this study) by clicking the "new folder" icon.  Each studyname in Lighthouse Studio has a maximum number of twenty-four characters (either letters or numbers), and Lighthouse Studio requires that the study be located within its own folder.  The folder that Lighthouse Studio automatically creates for your study also carries the same name as the study.  For this tutorial, you might create a name such as navyproject.  From the New Lighthouse Study dialog, specify navyproject as the studyname.  

 

Click the OK button.  You are returned to the main menu, and a new "Study Navigator" window is displayed.  

 

As you work with this study, items appear in the Study Navigator window, listing the functional areas you have been using or that are now available to you.  This list of items provides a Shortcut link to access those parts of Lighthouse Studio.  Alternatively, you can access those same areas by clicking icon buttons on the toolbar or by using the pull-down menus.  

 


Adding the Introduction Screens to the Survey

 

(If you are already familiar with Text and Select questions within Lighthouse surveys, you may choose to skip forward to the section entitled "Adding MaxDiff Questions to the Survey").

 

Let's enter the introductory text using the Write Questionnaire dialog.  You can access the Write Questionnaire dialog by selecting Compose | Write Questionnaire… or by clicking the pencil icon on the toolbar.  

 

The introductory text and survey questions would most likely be initially developed within a word processing document.  Assuming you really had such a document, you might press Ctrl-C to copy, and Ctrl-V to paste the information into Lighthouse Studio.  We suggest you simply copy-and-paste the text within this document into your Lighthouse questions rather than re-type the text (if viewing this document with Acrobat Reader, you can use the "text select" icon from the Acrobat Reader toolbar.)

 

The first question to compose is the introductory screen:

 

Intro:

 

Thank you for taking time to help the Navy consider ways to improve your working environment and career development.  Our goal is to make the Navy a satisfying career for you and to improve your desire to reenlist when the decision arises.

 

We would like you to imagine that you are facing the decision whether to reenlist or not--today.  Please think about the issues involved, including future career planning, family/relationship considerations, job location, and compensation.

 

Click the Next button below to continue...

 

We'll insert this introductory screen into a new Text question as the first question of the survey.

 

Most question types in Lighthouse Studio have "Header 1", "Header 2," "Body" and "Footer" sections.  These are "text" areas in which you can insert any text (including HTML).  When the question is viewed with a web browser, the sections are organized roughly as follows:

 

acatut20

 

For this introductory screen (that includes three paragraphs of information), it seems to make sense to place the first paragraph in the "Header 1" area, the second paragraph in the "Header 2" area, and the third paragraph in the "Body" area.

 

Type (or cut and paste) the following text into the text areas in the Text question.  Put the first paragraph in Header 1, the second paragraph in Header 2, and the third paragraph in Body.  

 

Thank you for taking time to help the Navy consider ways to improve your working environment and career development.  Our goal is to make the Navy a satisfying career for you and to improve your desire to reenlist when the decision arises.

 

We would like you to imagine that you are facing the decision whether to reenlist or not--today.  Please think about the issues involved, including future career planning, family/relationship considerations, job location, and compensation.

 

Click the Next button below to continue...

 

The main dialog only shows a portion of the text, but you can click the pencil icon adjacent to each text area to expand the text editing window.

 

To preview how the question will look when it runs on the respondent's computer, click the Preview button.   The question is displayed in "Preview" mode (no data are saved, and no special "error checking" or "validation" is performed for the question).

 

maxdifftut1

 

Lighthouse Studio automatically places paragraph breaks (extra blank lines) between the Header 1, Header 2, and Body sections.  If you put all the text in a single section, you may see that when the web browser interprets the text, it all runs together without any blank lines between the paragraphs (unless you insert some HTML instructions to force blank lines between paragraphs).  We'll talk about using HTML within your text to take greater control over the layout, font, and style later in this tutorial.

 

Click the OK button on the Preview window to close that window and return to the previous Lighthouse Studio dialog.

 

After viewing the text in preview mode, you might decide that it is really too small, or that you want to make the text bold.  You can modify the font and sizes of the font using the paintbrush icon.  Or, you can select a different Style (selecting styles was described in the 45-minute tutorial for Lighthouse Studio).  After changing any setting, click Preview again.

 

After you are happy with the layout of the introductory screen, click OK to return to the Write Questionnaire dialog.

 

Remember that when you add a new question to the List of Questions, it is inserted directly below the question (or page break) that was highlighted when you clicked the Add... button.  If you need to move a question once it has been added to the List of Questions, simply highlight the question to be moved and click the up or down buttons to move the questions within the list (you can also highlight a question and click Ctrl-X or Ctrl-C to cut or copy and Ctrl-V to paste questions within the current questionnaire, or even to another Lighthouse Studio study.)

 

Now, add a the second introductory screen on a separate page.  The text for that screen is as follows:

 

If you were making the decision to reenlist or not today, there would certainly be many factors that affect your decision to reenlist or not.  Real life decisions are not always easy.  

 

We want to gain some insight into how you make this decision by asking you to evaluate a series of tradeoff scenarios.  In each scenario, we'll show you five possible factors affecting your career in the Navy and the reenlistment decision.  

 

We'll ask which factor (among this set of five) would make you most want to reenlist, and which would make you least want to reenlist.  

 

We need to ask you repeated tradeoff scenarios (involving different factors each time) so that we can learn what is truly important to you when deciding whether to reenlist.

 

Note that there already is a page break following the Start question.  Highlight the page break and click Add....  Select Text as the question type with IntroMaxDiff as the question name.  Click OK to add the question to your survey.

 

This introductory screen has four paragraphs and it also includes two words that need to be underlined (most and least).  Therefore, we'll show you how to how to format text in HTML.  While we're showing you how to add underlining to text, we will show you how to insert paragraph breaks as well.

 

Copy and paste the full text for the IntroMaxDiff question from this document into the Header 1 field.  Next, click the pencil icon adjacent to the Header 1 text area to expand the text editing window.  Highlight the first full paragraph with the mouse and click the "paragraph break" icon which inserts the HTML text <p> at the beginning of the paragraph and </p> at the end of the paragraph.  Repeat the same steps for all four paragraphs.

 

Next, we need to underline the words most and least.  Highlight "most" and click the "underline" icon.  Repeat the same for the word "least".  When you are finished, the dialog should look like:

 

maxdifftut2

 

Click OK to close the text editor dialog and Preview to see how it is displayed in the browser window.

 

At this point, we are ready to add the MaxDiff questions to the questionnaire.

 


Adding MaxDiff Questions to the Survey

 

In this section, we'll add the MaxDiff questions to the survey, directly following the two introductory screens we previously added.  From the Write Questionnaire dialog, insert a page break (click Add Page Break) beneath the IntroMaxDiff question.  While the page break beneath IntroMaxDiff is highlighted, click Add... and select MaxDiff as the Exercise type.  Click Add.  Specify a name for this exercise, such as Reenlist.  (Note: you can have more than one MaxDiff exercise within the same Lighthouse questionnaire.  In this tutorial, we'll only insert a single exercise.)   Click OK and the MaxDiff Exercise dialog is shown.

 

The MaxDiff Exercise dialog has six tabs:

 

Question Text  (here you type the text to display directly above and below the question)

Label Text  (here you specify the text to display directly next to the item text)

Items  (here you specify the list of items in the exercise)

Format  (you may format your questionnaire, including widths and colors from this tab)

Design  (the number of items to display per set and number of sets is specified here)

Skip Logic  (this dialog is inactive and only present for consistency with other dialogs in Lighthouse Studio)

 

We'll begin by specifying the 20 items used in this exercise within the Items tab.  Click the Items tab.  A list box is displayed which allows you to specify a new list of items to be used in this Lighthouse study.  Or, you can select from an existing list of items.  Since we have not yet specified a list within this study, we'll add a new one.  By default, the list name is provided as ReenlistList.  Let's go ahead and use this default name.

 

(Note that other options such as making items on the list "Other Specify" items or "None of the Above" options are available.  These are for consistency with other dialogs in Lighthouse Studio, but are of no use in MaxDiff exercises.)

 

From this dialog, we could click the Add... button and type the items individually, but it is easier to copy-and-paste the list of items from an existing document.  Copy-and-Paste the list of items by highlighting the items below...

 

3-year reenlistment obligation

5-year reenlistment obligation

Location guarantee for next assignment

Duty guarantee for next assignment

Both location and duty guarantees for next assignment

Get promoted 6-months sooner than expected

Live in 2-person barracks when in port

Live in 4-person barracks when in port

Choice between living in 2-person barracks or 4-person barracks when in port

$50 per month pay increase when out at sea

$75 per month pay increase when out at sea

Half of on-job time using your skills and training

Three-quarters of on-job time using your skills and training

Time off each week allowed for on-line college coursework

Receive online career counseling services

Receive in-person personal counseling services

Receive phone-based personal counseling services

Current reenlistment bonus ($2,500)

Increased reenlistment bonus ($5,000)

Increased reenlistment bonus ($10,000)

 

...and pressing Ctrl-C to copy them into the clipboard, and then switching to the Items tab and clicking the Paste list member(s) from the clipboard icon.  The 20 items are all automatically inserted within the list.

 

Next, we'll specify the text to display above the MaxDiff sets into the Question Text tab.  Click the Question Text tab.  Some default text is already available within this question.  Paste the following text into the Header 1 field, replacing the default text:

 

Imagine you were making the decision to reenlist in the Navy today.

 

Which of the following four factors would make you most and least likely to want to reenlist?

 

Insert paragraph breaks and underlining as we showed you earlier in this tutorial.  

 

Header 2 has some scripting commands that add a counter to each MaxDiff question: 1 of 20, 2 of 20, etc.  We don't need to modify those scripting commands, so let's just leave it as is.

 

We are almost finished formatting the basics for our MaxDiff exercise.  Click the Label Text tab, and type the labels to display over the radio buttons respondents click to indicate best and worst items.  The text to use is as follows:

 

Best label:  Makes me <u>Most</u> want to reenlist

Worst label:  Makes me <u>Least</u> want to reenlist

 

Next, we need to tell Lighthouse Studio how many MaxDiff questions (sets) to display in the questionnaire.  Click the Design tab.  Basic design settings for your MaxDiff exercise are displayed:

 

maxdifftut5

 

You may recall that we previously decided that each set would present 4 items at a time and that the respondent would receive 20 questions (so each item was shown four times on average).  Specify 4 in the Itemps per Question (Set) field.  Specify 20 in the Questions (Sets) per Respondent field.

 

At this point, we might preview the MaxDiff question to see how we are doing so far.  Click the Preview button, and you may see something like this:

 

maxdifftut4

 

Note that by default the preview displays the first four items in our list therefore some of the items that we wanted to prohibit are being displayed together.  We can add the prohibited combinations under this same Design tab.  

 

Inserting prohibitions is a capability provided within the Advanced Settings on this dialog.  (After retiring the Preview dialog) click Show Advanced Settings to reveal additional controls, including the Prohibitions... button.

 

Click the Prohibitions... button.  We need to prohibit the following pairs of items for our exercise:

 

Prohibit this item...

...from appearing with this item.

3-year reenlistment obligation

5-year reenlistment obligation

Location guarantee for next assignment

Both location and duty guarantees for next assignment

Duty guarantee for next assignment

Both location and duty guarantees for next assignment

Live in 2-person barracks when in port

Choice between living in 2-person barracks or 4-person barracks when in port

Live in 4-person barracks when in port

Choice between living in 2-person barracks or 4-person barracks when in port

$50 per month pay increase when out at sea

$75 per month pay increase when out at sea

Current reenlistment bonus ($2,500)

Increased reenlistment bonus ($5,000)

Current reenlistment bonus ($2,500)

Increased reenlistment bonus ($10,000)

Increased reenlistment bonus ($5,000)

Increased reenlistment bonus ($10,000)

 

Click an item to prohibit from the list on the left, then click the item to prohibit from the list on the right.  Continue picking items until all prohibitions have been specified.  Click the OK button when finished.

 

So that the designer runs quickly for the purposes of this tutorial, specify 30 iterations.

 

maxdifftut30iterations

 

(Note: specifying prohibitions can cause the design algorithm to process more slowly than if no prohibitions are specified.  With the 9 prohibitions in this example, it takes about 10 minutes to run 1000 iterations of the design.  Yet, the designs found with even 30 iterations are extremely good quality, so we recommend reducing the number of iterations attempted to 30 for the purposes of this tutorial.)

 

Now, we are ready to "generate the design."  The design is all the combinations of items that respondents will see across their 20 choice sets (questions).  Each respondent evaluates the same number of items, so the questionnaire design can be built prior to collecting the data.

 

The sets need to be chosen carefully, so each item appears about an equal number of times, and that each item is shown about an equal number of times with every other item (non-prohibited item, that is).  Click Generate Design, and the software tries 30 separate attempts (recall we reduced the attempts from 1000 to 30 to save time in this tutorial) to create a balanced experiment (using a different random number start each time).  The best attempt is selected, and the results are shown to you in a design report.

 

When the computation finishes, you are shown a report (which you have the opportunity to save to a text file if you wish).

 

We'll explain the report, one step at a time.  First, design specifications are reported:

 

Design Parameters:

Items: 20

Items per Question (Set): 4

Questions (Sets) per Respondent: 20

Versions: 300

Iterations: 30

Random Seed: 1

Prohibitions: 9

Chosen Iteration: 30

 

Above, we see that the software made 30 total attempts (iterations) to generate a design plan, and the 30th attempt was chosen as "best."  In total, the software created 300 versions of the questionnaire (where each version features different item combinations within the sets) to display to respondents.  We generate multiple versions of the questionnaire to improve the stability of the item scores and also to reduce psychological order and context effects that might occur if every respondent received the same combinations of items.  Although you plan to interview 1000 or more respondents, it would probably be overkill to generate even more unique questionnaire versions.  (Some might argue that 300 versions is already overkill!)  Additional versions would provide imperceptibly little measurable benefit, though the software lets you specify up to 999 versions.  Each respondent will receive one of these versions of the questionnaire, and many of the questionnaires will therefore be repeated.

 

Below, the one-way balance is reported.  This shows how many times each item is displayed across all 300 versions of the questionnaire plan, where each version has 20 sets displaying 4 items at a time.  If the design was perfectly balanced, each item would be displayed 1200 times.  As can be seen, the design is indeed perfect with respect to one-way frequencies:

 

One-Way Frequencies:

 

Item

Times Used

1

1200

2

1200

3

1200

4

1200

5

1200

6

1200

7

1200

8

1200

9

1200

10

1200

11

1200

12

1200

13

1200

14

1200

15

1200

16

1200

17

1200

18

1200

19

1200

20

1200

 

 

Mean

1,200.00

Standard Deviation

0.00

 

 

The quality of the design is summarized as the standard deviation across the 20 item frequencies, or 0 in this case (no deviation from perfection).  This number is used to classify the quality of the design for multiple attempts (iterations) the software makes.  Attempts that demonstrate a lower standard deviation across the frequency counts are deemed better.

 

Next, the report summarizes how many times each item appeared together with every other item within sets.  Ideally, each item would appear an equal number of times with every other item.  You may recall that we prohibited nine combinations from occurring, so there cannot be perfect balance.  These prohibited combinations are shown with 0 frequency in the table below (for example, item 1 with item 2; item 4 with item 5, etc.).

 

Item

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

1

1200

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

2

0

1200

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

3

202

201

1200

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

4

201

201

202

1200

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

5

213

213

0

0

1200

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

6

190

188

190

190

201

1200

 

 

 

 

 

 

 

 

 

 

 

 

 

 

7

201

200

201

202

214

190

1200

 

 

 

 

 

 

 

 

 

 

 

 

 

8

201

201

202

202

213

190

202

1200

 

 

 

 

 

 

 

 

 

 

 

 

9

212

213

214

214

226

203

0

0

1200

 

 

 

 

 

 

 

 

 

 

 

10

200

201

200

202

213

189

201

201

212

1200

 

 

 

 

 

 

 

 

 

 

11

200

201

202

201

212

189

202

201

213

0

1200

 

 

 

 

 

 

 

 

 

12

190

190

189

190

202

178

190

190

201

190

189

1200

 

 

 

 

 

 

 

 

13

189

190

190

189

203

178

191

190

202

189

189

178

1200

 

 

 

 

 

 

 

14

189

189

190

190

203

178

190

191

202

189

189

178

178

1200

 

 

 

 

 

 

15

190

189

190

190

202

179

189

190

202

190

189

178

179

178

1200

 

 

 

 

 

16

189

189

190

190

202

178

190

190

202

189

189

178

178

179

178

1200

 

 

 

 

17

189

189

191

190

202

178

190

189

202

189

190

178

178

178

178

177

1200

 

 

 

18

215

215

215

215

227

203

216

215

227

215

215

204

203

203

203

205

204

1200

 

 

19

215

215

215

216

227

205

216

216

228

215

214

203

203

203

202

203

204

0

1200

 

20

214

215

216

215

227

203

215

216

227

215

215

204

203

203

204

204

204

0

0

1200

 

Off Diagonal Elements (adjusted for prohibitions)

Mean

199

 

 

 

 

 

Standard Deviation

12.7

 

 

 

 

 

 

Although the balance across the non-prohibited two-way frequency cells is not nearly as good as the one-way counts (the nine prohibitions of course prohibited the MaxDiff designer from finding a balanced design), the results are still quite acceptable in practice. How would you know if you were in deep trouble due to prohibitions?  If you have too many prohibitions, you would receive an error message reporting that it is impossible for the designer to ensure connectivity across items in the study.  If you have to prohibit a single item from appearing with more than half of the other items in the study, then this may be excessive.  Again, we'd like to stress that it is much more important that the one-way frequencies be nearly balanced than to achieve approximate balance in the two-way frequency table.

 

(Note for advanced users:  We were interested in the potential loss in design efficiency resulting from these nine prohibitions, so we performed a simple test.  We assumed 250 respondents receiving this study as just designed versus the same study in which there were no prohibitions.  The design with no prohibitions had perfect one-way balance, and nearly perfect two-way balance (low of 189 and high of 190 counts per two way cells).  Upon estimating MNL parameters and studying the standard errors, we determined that the difference in overall design efficiency between these two designs was only 1%.  For those familiar with conjoint analysis, the important point is that prohibitions are much less detrimental to MaxDiff designs than conjoint designs.  That is because prohibitions in conjoint are within-concept whereas prohibitions in MaxDiff are between-concept.  And, the nine prohibitions in MaxDiff eliminate a much smaller proportion of the full-factorial combinations than would nine prohibitions in a typical conjoint plan where each attribute has three to five levels.)

 

Finally, the report shows how often each item appears in the four positions within sets (top, bottom, etc.).  This type of balance has no bearing on statistical design efficiency.  But, balance here helps reduce psychological order effects (e.g. we wouldn't want a given item to always appear in the top position each time).  The positional balance for our design is nearly perfect:

 

Positional Balance:

 

 

Item/Position

1

2

3

4

1

300

300

300

300

2

300

300

300

300

3

300

300

300

300

4

300

300

300

300

5

300

300

300

300

6

300

300

300

300

7

299

301

300

300

8

300

300

300

300

9

300

300

300

300

10

300

300

300

300

11

300

300

300

300

12

299

300

301

300

13

300

300

299

301

14

300

300

300

300

15

301

299

300

300

16

300

300

300

300

17

301

300

300

299

18

300

300

300

300

19

300

300

300

300

20

300

300

300

300

 

Mean

300

Standard Deviation

0.35

 

 

At this point, we have the basics programmed for this MaxDiff study.  You can click OK to return to the Write Questionnaire dialog, where you'll see the twenty MaxDiff questions automatically inserted within your project:

 

maxdifftut6

 

 


Inserting the "Progress" and "Finish" Questions

 

When respondents have completed five MaxDiff questions, we want to display a screen that says:

 

ProgressCounter:

 

Good work so far.  We know that some of these tradeoffs are challenging.  

 

We need to ask you to evaluate 15 more sets so that we can better determine how you make your reenlistment decision.

 

 

Insert a new Text question by highlighting the Page Break directly beneath the Reenlist_5 question, and then clicking Add....  Select Text as the question type and name the question ProgressCounter.  Copy-and-paste the text for this question from this document and format the text into paragraphs.  Preview the question to make sure it looks as expected.  Click OK to return to the Write Questionnaire dialog, and insert another page break directly following this ProgressCounter question.

 

Last, we need to insert the final screen in which we tell respondents that they have completed the survey and link them to the Navy's official site.  

 

Finish:

 

Thank you very much for your help.  Your data will be used to help Navy planners as they work to make the Navy an even better career choice.  

 

Click the Next button to finish submitting your data and to return to the Navy's official website.

 

To do this, scroll to the bottom of the list of questions on the Write Questionnaire dialog.  Highlight the final question in the list (Reenlist_20) and insert a page break.  Then, click the Add... button and specify a Text question named Finish.

 

Copy-and-paste the text for this question from this document, implement any formatting you desire.  Click OK to return to the Write Questionnaire dialog.

 

Next, insert a page break directly following the Finish question.  Then, insert a new Terminate question directly following that entitled FinishLink.  No text should be shown on this page.  This page simply redirects respondents immediately to the Navy's website.  While editing the FinishLink question, click the Settings tab.  Note that by default the Terminate Respondent box is checked indicating that respondents seeing this page should be terminated and marked as completed records.  Check the second Link to a Questionnaire or Website box and type within the provided field so that the respondent is sent to http://www.navy.mil.  When you are finished, that dialog should look like:

 

maxdifftut7

 


Page Layout and Study Settings

 

It may be helpful at this point to review how Lighthouse Studio breaks the survey into separate pages, and some basic global settings that affect the look and functionality of your questionnaire.  We also want to add a progress bar to our survey.  The Write Questionnaire dialog also displays the page breaks that are currently set for your questionnaire.

 

Return to the Write Questionnaire dialog.  This dialog shows how the various questions we've specified (or that Lighthouse Studio has automatically added to the list) are arranged across different pages.  Page breaks are indicated by "---------------<Page>".  

 

From the Write Questionnaire dialog, click the Survey Settings icon (alternatively, click Compose | Survey Settings...).  This dialog controls global aspects of the questionnaire (meaning that the settings here control the general behavior of the questionnaire rather than page- or question-specific aspects.)  Examples include the fonts and colors used in the questionnaire, the wording for the "Next" button, and the global header and footer that can appear at the top or bottom of each page of the survey.  The details of this dialog are beyond the scope of this unit, but you can learn more about the items on the dialog by pressing the F1 key.  However, we do wish to insert a progress bar.  This is easily done by clicking the Progress Bar tab, and checking the Automatically show the progress bar box with the position selected as Page Footer (you'll note it is already checked by default).

 

Save your changes and exit the Write Questionnaire dialog.

 


Test Running Your Survey Using "Local Test Server"

 

Although the Preview Survey function is nice, you generally will want to test your survey in runtime mode, with any skip patterns, response verification, and randomizations in place.  You could upload your files and the Perl scripts to the Web Server, but Lighthouse Studio provides a more convenient way to test run your survey locally on your PC or laptop.

 

Lighthouse Studio uses its own internal web server to run the questionnaire just like a remote server would.

 


Local Test Server

 

From the Write Questionnaire dialog, Click the Test Survey button .  

 

Your browser opens up the first survey page.  This is your survey, as it would appear to a respondent over the Internet (there are slight visual differences from browser to browser).

 

To close the survey, simply close the browser window (using the X in the upper-right hand corner).  To run the survey again, click the Test Survey button.

 

After you are convinced that the survey is functioning properly, you should also pretest your survey among your colleagues.  They can give you feedback regarding the usability of the survey, and you can examine the resulting Counts data or perhaps even compute individual-level scores using the Analysis | Analysis Manager dialog (assuming you have enough data) to make sure the data at least have face validity.

 

At this point, you would set up the study typically using the hosting service provided by Sawtooth Software.  You would invite the respondents to complete the survey (probably using an email message with a clickable link).

 


Data Management

 

Lighthouse Studio includes an Admin Module so that you can monitor or access your project from any computer connected to the Web.  The Admin Module is password protected, and your passwords for access are specified on the Field | Hosting Management dialog.  Lighthouse Studio generates random administrative access passwords whenever you create a new study, but you can modify them to suit your needs.

 

To download the data for our study, you would browse to the administrative module for your study on the web site (again, we are speaking hypothetically, as for this tutorial study there is no such site set up).  Once at the Admin Module, you would download your data (by clicking Download from the main menu), making sure to save the data file to the same folder on your hard drive in which you developed the Lighthouse project.  Alternatively, from the Lighthouse Studio interface, you can simply click Field | Data Management then Get Data to download your data without logging into the Admin Module.

 

Once you have downloaded the data, you are ready to analyze the data by Counting analysis (Analysis | Analysis Manager... and selecting Counts as the Analysis Type) or through more sophisticated methods such as latent class or hierarchical Bayes analysis also available on the Analysis Manager dialog..  

 


Analysis: Counts and Computed Scores

 

At this point, we would like you to open the Tutorial study which contains hypothetical MaxDiff data for you to analyze.  Click Help | Sample Studies | Sample MaxDiff.

 

We'll assume that you have collected data over the web and have downloaded the data to the folder on your hard drive containing the study files (as described in the previous section).

 

A quick summary method called "counts" is often used for summarizing respondents' preferences.  Counting analysis simply looks at how many times an item was available within sets and counts how many times it was chosen either best or worst.  The percent of times an item was chosen as best or worst (when it was available) is termed the "count proportion."

 

Run counting analysis for the tutorial data set by clicking Analysis | Analysis Manager... by choosing Counts as the Analysis Type and then by clicking Run.  The MaxDiff Counts report is displayed:

 

MaxDiff Counts Analysis                                                        

                                                 

MaxDiff Exercise               Reenlist

Sets Included                  All

Respondents Included           All respondents included

Total Number of Respondents    250

Total Number of Sets           5000

 

                                      Times    Times    Best       Times     Worst      

                              Item    Shown-   Selected Count      Selected  Count

                              Number  Best     Best     Proportion Worst     Proportion

 

3-year reenlistment obligation    1      999     310       0.310     177      0.177

5-year reenlistment obligation    2     1000      72       0.072     561      0.561

Location guarantee for next assi  3      999     191       0.191     286      0.286

Duty guarantee for next assignme  4     1000     202       0.202     283      0.283

Both location and duty guarantees 5     1000     361       0.361     135      0.135

Get promoted 6-months sooner than 6     1000     183       0.183     307      0.307

Live in 2-person barracks when in 7     1000     252       0.252     231      0.231

Live in 4-person barracks when in 8     1000     208       0.208     276      0.276

Choice between living in 2-person 9     1000     294       0.294     186      0.186

$50 per month pay increase when o 10    1000     230       0.230     271      0.271

$75 per month pay increase when o 11    1000     246       0.246     225      0.225

Half of on-job time using your sk 12    1001     185       0.185     311      0.311

Three-quarters of on-job time usi 13    1001     199       0.199     251      0.251

Time off each week allowed for on 14    1000     190       0.190     309      0.309

Receive online career counseling  15    1000     234       0.234     268      0.268

Receive in-person personal counse 16    1001     231       0.231     238      0.238

Receive phone-based personal coun 17    1000     195       0.195     279      0.279

Current reenlistment bonus ($2,50 18    1000     276       0.276     189      0.189

Increased reenlistment bonus ($5, 19     999     386       0.386     110      0.110

Increased reenlistment bonus ($10 20    1000     555       0.555     107      0.107

 

The report shows results for 250 respondents and 5000 sets (remember, each respondent received 20 choice sets; 250 x 20 = 5000).  You may recall that we had originally generated 300 questionnaire versions for respondents to potentially complete.  We only interviewed 250 respondents, so not all questionnaire versions were used (which isn't a concern, since each questionnaire version independently has excellent qualities).  However, we see from the "Number of Times Shown" column that each item was shown almost exactly the same number of times.  In the "Times Selected Best" column, we see how many times each item was selected best (across all sets and respondents).  Dividing the "Times Selected Best" by the "Times Shown Best" results in the "Best Count Proportion."  The higher the proportion, the more impact this item has on reenlistment.

 

Counts proportions are a good way to get a quick (top-line) look at the data.  This method of analysis is very intuitive.  But, we don't suggest you use counts results in final reporting.  Because of prohibitions in your design or slight imbalances, the data are not as accurate as when using logit, latent class, or HB estimation routines, all automatically provided under Analysis | Analysis Manager....  

 

The item selected best most often was "Increased reenlistment bonus ($10,000)."  Thus, we can see that these sailors (at least our hypothetical ones) pay most attention to monetary rewards.  Even the $5,000 reenlistment bonus tends to trump every other issue.  If you are interested, you can also examine the proportion of respondents choosing each item as worst.  This data set was generated with artificial symmetric properties, so there is no difference between bests and worsts other than random noise.  However, this isn't always the case with real data sets, where respondents may not use a consistent scoring approach when judging bests and worsts.

 

Let's turn our attention to a more powerful and accurate method of analysis: hierarchical Bayes (HB).  A typical data set may take about 10 to 15 minutes to run using HB (very large data sets may take hours).  HB estimates a set of scores for each respondent.  These scores may be saved to an Excel-compatible file.  How HB actually works is beyond the scope of this tutorial and we refer you to technical papers on HB on our website, www.sawtoothsoftware.com.

 

Click Analysis | Analysis Manager... then select HB as the Analysis Type, then click Run.  First, the software reads the data set and prepares it in a more compact way for more efficient processing.  As part of this step, the software reports how many respondents it has found and how many times items were chosen in each position of the question.

 

** Data Summary **

Total respondents = 250

Total 'Best' choices = 5000

Total 'Worst' choices = 5000

Average sets per respondent = 20.00

 

'Best' responses by position:

 (1) 24.38%

 (2) 26.34%

 (3) 25.08%

 (4) 24.20%

 

'Worst' responses by position:

 (1) 24.76%

 (2) 25.28%

 (3) 24.78%

 (4) 25.18%

 

Click Estimate Scores to continue.  HB uses an iterative process that takes 1000s of iterations to converge on a stable solution.  By default, the first 20,000 iterations are called "burn-in" iterations that are discarded as iterations prior to obtaining convergence.  The last 10,000 by default are called the "used" iterations and are averaged to produce individual-level scores for your MaxDiff items.

 

Partway through the estimation, the MaxDiff Score Estimation dialog may look something like this:

 

maxdifftut11

 

This display shows how well the estimated scores are stabilizing over the planned thousands of iterations.  The graphic plots the estimated scores at each iteration of the process.  The estimates first all start at 0, and then trend toward their final values (which may be either negative or positive).  Once the process has "converged" the estimates will tend to randomly "wobble" up and down, but there should be no noticeable trend remaining.  

 

The graphic is divided into two regions: a gray region (iterations prior to assumed convergence) and a white region (iterations after assumed convergence).  Once the iterations cross into the white region, you should look for no noticeable trend in the data.  The plotted scores should just continue to wobble randomly up and down, with no strong and sustained move in any specific direction.  If there is noticeable trend after the plotted values are in the white region of the graphic, then you should repeat the estimation using more iterations (under the Advanced button).

 

(Note that for sake of brevity we didn't include "constraints" in this estimation to force the scores for some items to be higher than others, such as $10,000 bonus higher than $5,000 bonus.  You do not need to enforce such constraints during estimation, though some researchers choose to do so.)

 

When the estimation finishes, the graphical plot of the estimation remains on the screen.  Click the button stating: The estimation is complete.  Click here to close.

 

When you click the "close" button, a report is given which resembles a spreadsheet, with multiple "sheets" that can be accessed by clicking the tabs at the bottom of the dialog labeled "Summary," "Settings Summary," "Probability Scale," "Zero-Anchored Interval Scale," and "Zero-Centered Raw Scores."  The first (Summary) sheet contains a summary of the final scores, 95% confidence intervals,and other information regarding the settings for your run:

 

scores_output

 

This is a summary report averaged across all respondents in the analysis.  Note that the output has multiple tabs that you can click to access.  The third (Probability Scale) and fourth (Zero-Anchored Interval Scale) sheets contain the scores for each respondent, including a fit statistic.  Most researchers will pay most attention to the Probability Scale scores as they are easiest to interpret and present to others.  These individual-level item scores are positive values ranging from 0 to 100 that sum to 100.  This approach has the valuable property of ratio-scaling.  That is to say, an item with a score of 20 is twice as important (or preferred) as an item with a score of 10.

 

Respondents who answer randomly will generally have low fit statistics.  We have provided guidelines (fit statistic thresholds) for identifying "bad" respondents elsewhere in this documentation.  You can use the suggested cutoff rules to discard respondents that are most likely answering randomly.

 

You may click the Save Report icon to save this report to an Excel format.

Created with Help & Manual 8 and styled with Premium Pack Version 4 © by EC Software