Yell webpage

Yell is a leader in the business directory with over 2.9 million listed businesses across the whole of the UK. It works closely with a number of leading global technology companies in order to deliver the best possible products and advertising results for its customers. For over 50 years they have supported the marketing needs of local business offering a number of solutions and tools to support every business and enabling them to quickly create an online presence. Using their relationships with the leading players in the industry - including Apple, Facebook, Google, Microsoft and Amazon, Yell ensures their solutions are cutting-edge, using automation and machine learning to deliver the best returns. Yell’s main purpose is to help the business to be found, trusted and chosen over the competition.

Website:
https://www.yell.com/

The challenge

The Problem:
Content is unique to Yell and cannot be mimicked by Google on it’s own search page via Google my business. Yell’s marcomms team work very hard to curate this content by targeting users through email and other marketing mediums. Yell’s SEO lead believes that valuable opportunities are being missed with the current simple format that the content is delivered with.

The initial offering from Yell to customers is to set up a business listing online and provide as much content as possible about it. Generally speaking, the more content the customer provides about their business the more useful that listing becomes to users and the more likely they are to contact that business. Furthermore, the more information added to a business profile page the more likely search engines are to find it. The way we can create value for our customers is by encouraging them to give good descriptions, good photos, providing a great user experience and helping them by presenting their reviews as best we can.

Currently an SEO exercise is taking place to ensure content parity between the mobile and desktop versions of Yell.com. Content should match between each version for optimal SEO scores.

My role:
As a senior UX & UI Designer, I’m regularly communicating with all key stakeholders while discussing the projects with the product lead and the dev team. Together with my lead UX manager we are using various usability testing methods to produce quality research by scheduling unmoderated usability tests, surveys, card sorting and planning user interviews. Once we have the insights we build wireframes and prototypes using the Yell Design System to improve the performance and the processes. Outside of this I’m also advising on other UX and UI issues for other Yell products, working closely with SEO team, marketing team, outsource dev team and partners to put together design ideas for features ideas.

Research: Improving the Review Summary

Understanding the problem

During the planning and research session, we identified a possible opportunity to improve the value of the ‘review summary‘ by adding additional filtering and sorting features. To address this Yell’s SEO and UX team proposed additional features for upcoming testing:

  1. Integrating a review summary box, highlighting top attributes.

  2. Adding ability to filter reviews by clicking on attribute tags.

  3. Adding ability to sort reviews.

Assumptions have been mapped out below which we intend to explore and gain more insight around to validate if the review summary will gain additional value with the ability to filter and sort reviews to the business listing:

The purpose of this test is to establish if a ‘reviews summary’, with sorting and filtering features provides value to the business listing for a user. As with all research it’s essential to define and establish the users and business goals. From meetings with the stakeholders and planning design sprints including the dev team as well, we were able to gain valuable insights through our Adobe Analytics and Mouseflow tools. Based on this data, we were able to map out our assumptions and considerations.

Screenshot 2021-05-03 at 21.09.13.png

Assumptions:

YES.

As a customer, I want to have a review summary that doesn’t misinterpret how a product/service has been rated.

  • As a customer, I rely heavily on the review summary to provide the overall picture of how other users experienced the product/service.

  • As a customer, I understand the meaning behind those summaries and filters.

  • As a customer, I want to see the ‘reviews summary’ and ‘filters’ for the business at first glance.

  • As a customer, I want to filter the given reviews based on specific criteria.

  • As a customer, I can easily navigate through the review summary and learn more about the value of the products/services.

  • As a customer, I want to rate the reviews from the other customers as ‘Helpful’.

NO.

  • As a customer, I don’t need a review distribution summary as I can judge the product/service by the given reviews.

  • As a customer, I don’t rely heavily on the review distribution summary.

  • As a customer, I don’t know what the ‘reviews summary’ box and ‘filters’ mean.

  • As a customer, I don’t want to filter the given reviews on specific criteria.

  • As a customer, I have difficulties finding the real value of the products/services based on the review distribution summary.

  • As a customer, I don’t want to rate the reviews from the other customers.

Considerations:

  • Building the reviews tag will require significant backend build, therefore knowing its value beforehand is critical.

  • A PGS’ AI team is likely to be involved therefore third party cost needs planning, and this research will help determine potential return.

Personas

Based on the information I gathered, I used the personas method to better understand my users. Before I created the personas I was able to learn more about our users through previous user interviews and recorded calls, so I could better understand our users needs, experiences, behaviours and goals. Due to sensitivity of the data I’m unable to show our findings and our validated user personas, however, I am able to share below some user personas I created before the validation phase.

User+Persona+-+Customer@2x.jpg
User+Persona+-+Merchant@2x.jpg

Usability test

Method:

We will create a mockup of several BIP (business profile page) pages which are very similar and expose a panel of users to them in Usability Hub. The panel demographics will be quite general as is the audience of Yell.com:

  • Users aged 20-65

  • Male and female

  • UK based

  • Mobile based

  • Employment status employed, self employed and retired

The test starts with an instruction screen and a short description. After that a mobile screen with a design screenshot is presented which includes a free text field, asking the user - ‘Which 3 bits of information on this profile page are most likely to influence your decision in selecting a product or service from the business’. After this several follow up questions will be asked to gain more insights about what a user was thinking or feeling about the experience.

Due to the sensitivity of the data I’m unable to share all the custom questions and actions that were required to be performed or answered by the users, but here is an example for one of the screenshots that was used in the test:.

Task & Metrics

To get a better understanding of the problem we are facing, we define the following tasks, metrics:

Task:

  • Presenting screen with questions.

  • Selecting a screen.

  • Follow up questions.

Metrics:

  • The number of users who select the card displaying ‘reviews summary’ and ‘filter’ vs those who don’t.

  • Which 3 bits of information is specified by users as most valuable.

  • Number of users who understand correctly what  ‘review summary’ and ‘filter’ icons are intended to mean.

  • The number of users using ‘Sort by:’ functionality and finding it valuable.

  • The number of users who said the ‘Helpful’ button would impact their decision on choosing the business.

Hypothesis

H1: Majority of users select a listing with ‘reviews summary’ with ‘filters’ and ‘sorting’ features over those without additional features - users understand and perceive them to add value.

H2: Majority of users do not select a listing with ‘reviews summary’ with ‘filters’ and ‘sorting’ features - users don’t perceive noticeable value in them.

H3: Majority of users select a listing with ‘reviews summary’ with ‘filters’ and ‘sorting’ features over those without, but fail to offer correct explanations of what they mean - the ‘review summary’ and ‘filters’ add little perceived value, but the additional facets aid perceived quality somehow.

H4: Majority of users find the ‘Helpful’ button as a positive impact on their decision on choosing the business.

User findings

The unmoderated usability test was run through 30 users and the main purpose was to identify the value of the ‘review summary’ including filtering and sorting features.

Before we write down the findings, we write all the user feedback and quotes on notes and sort them by topics (Affinity Diagramming) using the InVision Freehand. Once the notes are organised, we study them and turn them into findings with determined severity.

Findings research 1.png

For easy reference, the findings in the report are identified as High, Med(ium), Low, or Good and appear as follows:

  1. GOOD: The designs were simple and users understood it.

  2. HIGH: Users could not complete the task or experienced significant frustration.

  3. MED: Users had some difficulties completing the task, but could complete it.

  4. LOW: Users experienced minor frustration or issues when completing the task.

Recommendations are written in Blue. One recommendation may address several usability issues.

 

Example of finding:

GOOD: Users do find the additional review information features useful.

Users like the new proposed design screen with the new review information and they do find them useful. Some of the users feedback:

Yes, it’s broken down into key parts and is clear and easy to understand - 27 years old, male, UK

Yes. Very much. It makes reviews more specific and targeted. - 47, male, UK

Yes, it gives a nice summary without me having to read every review. - 33 female, UK

UX team recommendation: Consider redesigning the profile page and adding additional review information, which would give the users the context they need to move forward.

Example of hypothesis result:

To validate our hypothesis we defined the research as: ‘We will run the test with 30 users to get more information about the interest for reviews summary filters and sorting features, and ask follow-up questions to identify the user’s emotion associated with them. We will know the hypothesis is valid when more than 75% identify positive value using the features.

H1: Majority of users select a listing with ‘reviews summary’ with ‘filters’ and ‘sorting’ features over those without additional features - users understand and perceive them to add value.

Hypothesis H1 pass.

H4: Majority of users find the ‘Helpful’ button as a positive impact on their decision on choosing the business.

Hypothesis H4 failed

From here, we were able to get valuable insights to determined what to do next. Taking a moment to understand the questions and assumptions we had helped us to build a large database of insights about our users, which saved us time in spending months building a feature that users didn’t need.

Research: Identifying ‘Credentials’ Value

Understanding the problem

During the SEO parity exercise, a problem was identified around the value of ‘Credential’ icons, which are shown on the S-page (search page) for each business listings card. Currently, cards on the S-page display different content on desktop and mobile. This is a problem, particularly on mobile where there is limited space and therefore only the most informative and valuable content should be prioritised. One piece of content called into question is a feature called ‘Credentials’, which uses icons to convey a business’s facets such as associations, accreditations, qualifications and so on.

The ‘Credentials’ shown on S-page are represented with specific icons, whereas on the BIP page (business page) there is additional text to support the meaning of the specific icon. Currently there is no space for the credentials icons on the mobile S-page, so in order to gain parity to content on desktop they were removed from S-page for both platforms, which additionally gains SEO benefits.

Further conversations presented a legal challenge to this change. If the removal reduces the effectiveness of the business listing we’d be taking a risk. Therefore before making final recommendations, we are testing if the icons are understood or have any significant impact on the UX. If not they will stay removed from the S-page.

Each business card on the S-page desktop view can contain over ten elements, including for example; a profile picture, the credentials icons, name of the business etc. Being faced with all of this information, customers need to decide if they will choose the specific business or not. Assumptions have been mapped out below which we intend to explore and gain insight around to validate if ‘Credentials’ icons are recognisable enough and add value to the business listing card.

The purpose of this test is to establish if users are able to recognise the ‘Credentials’ icons on S-page business listing and to determine whether or not those icons provide value.

Based on the data, we were able to map out our assumptions and considerations.

Assumptions:

YES.

As a customer, I want to have a review summary so I don’t misinterpret how a product/service has been rated.

  • As a customer, I can notice the icons easily.

  • As a customer, I can understand the meaning behind those icons.

  • As a customer, I want to see the ‘Credentials’ for the business at first glance.

NO.

  • As a customer, I don’t know what those icons mean.

  • As a customer, those icons don't represent anything to me.

  • As a customer, I can barely see them amongst the rest of the information.

Considerations:

  • Presenting the ‘Credentials’ icon on the business listing card is important for the merchants, as it represents extra value for their business.

  • ‘Credentials’ are an extra feature and merchants want to see what they are paying for.

  • Presenting the ‘Credentials’ icon at first glance is important for the Sales Team to present and sell the feature.

Personas

Based on the information I gathered, I used the personas method to better understand my users. Before I created the personas I was able to learn more about our users through previous user interviews and recorded calls, so I could better understand our users needs, experiences, behaviours and goals. Due to sensitivity of the data I’m unable to show our findings and our validated user personas, however, I am able to share below a user persona I created before the validation phase.

User Persona - Customer@2x.jpg

Usability test

Method:

We will create a mockup of several S-page cards which are very similar and expose a panel of users to them in Usability Hub. The panel demographics will be quite general as is the audience of Yell.com:

  • Users aged 20-6o

  • Male and female

  • UK based

  • Desktop based

  • Employment status employed, self employed and retired

The test starts with an instruction screen and a short description. After that a desktop screen with a design screenshot is presented which includes a task to perform - ‘Look over the listings and click on the one you are most likely to select’ and a free text field into which they are asked to input  - ‘What were the main reasons you selected that listing over the other two?’. After this, several follow up questions will be asked to gain more insights about what a user was thinking or feeling about the experience.

The test will be run twice to remove any bias based on the business card position during the task presentation.

Due to the sensitivity of the data I’m unable to share all the custom questions and actions that were required to be performed or answered by the users, but here is an example for one of the screenshots that was used in the test:.

Credentials 2@2x.png

Task & Metrics

To get a better understanding of the problem we are facing, we define the following tasks, metrics:

Task:

  • Selecting a card

  • Follow up questions.

Metrics:

  • The number of respondents who select the card displaying ‘Credentials’ vs those who don’t

  • Which 3-5 bits of information is specified by users and most valuable.

  • Number of users who understand correctly what ‘Credentials’ icons are intended to mean.

Hypothesis

H1: Majority of users select a listing with ‘Credentials’ icons over those without icons - users understand the icons and perceive them to add value.

H2: Majority of users select a listing without ‘Credentials’ icons over those with icons - users do not understand the icons or perceive them to have little value.

H3: Majority of users select a listing with ‘Credentials’ icons over those without icons, but fail to offer correct explanations of what they mean - the icons add little perceived value, but the additional facets aid perceived quality somehow.

User findings

The unmoderated usability test was run through 25 users and the main purpose was to identify the value of ‘Credentials’ icons and if those icons are factored in user’s decision making.

credentials.png

For easy reference, the findings in the report are identified as High, Med(ium), Low, or Good and appear as follows:

  1. GOOD: The designs were simple and users understood it.

  2. HIGH: Users could not complete the task or experienced significant frustration.

  3. MED: Users had some difficulties completing the task, but could complete it.

  4. LOW: Users experienced minor frustration or issues when completing the task.

Recommendations are written in Blue. One recommendation may address several usability issues.

 

Example of finding:

HIGH: Users do not find the icons useful to them and the majority of them do not mentioned them as the main factor in their decision making.

icon useful.png

Some of the users feedback:

No, I'd not really know what they are or use them to make a final decision. - 35 years old, female, UK

No. Unclear to me what they mean. If a key was provided, perhaps - 41, male, UK

no, I don't know what they stand for. - 24 male, UK

UX team recommendation: The vast majority of users did not find the credentials icons on the S-page a factor in their decision making or useful when they were directed to them specifically. The recommendation is the credential icons to be removed from the S-page on both web and mobile.