By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: A Human-centred Framework for Designing Complex AI Solutions for Strategic Decision-making | HackerNoon
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > Computing > A Human-centred Framework for Designing Complex AI Solutions for Strategic Decision-making | HackerNoon
Computing

A Human-centred Framework for Designing Complex AI Solutions for Strategic Decision-making | HackerNoon

News Room
Last updated: 2025/07/14 at 2:32 PM
News Room Published 14 July 2025
Share
SHARE

At Tesco Technology, designing a complex internal simulation platform for high-stakes decisions made one thing clear: even the most advanced AI tools can be misleading if they are not intuitive, user-friendly, and not designed to work for the humans using them. This raised a fundamental question: how can we make advanced data-driven tools truly user-centric? To answer this, we needed to design a framework for complex AI systems that bridges the gap between human-centred design and AI development.

This practical framework helps designers, data scientists, and engineers find common ground. It aligns the voice and needs of users with the design of AI systems, ensuring that human insights influence not only the interfaces but also the underlying logic and behaviour of AI tools, so that users trust and actually enjoy using them. This framework guides you through every phase of product development, from early research to post-launch feedback, highlighting where and how humans should be ‘in’ and ‘on’ the loop. This enables complex AI systems not just functional, but transparent, learnable, and trustworthy, systems that are designed to support decision-making, not replace human decision-makers, especially in strategic contexts and without a data science background.

Introduction

In the design world, we often refer to human-centred design, an approach focused on empathy, usability, and solving real user problems, but in data science and engineering, users are often considered late in the process, typically as validators of inputs. While some teams still prioritise technical feasibility, ideally, system’s design should be driven by users’ needs and understanding.

This is often viewed in terms of the concept of technological determinism, which states that people, culture, and the economy evolve as a result of technological progress. In contrast, the social construction of technology suggests that people shape technology to cause change, rather than the other way around. However, in practice, technology development is often influenced by limitations such as data availability, system constraints, or organisational priorities, rather than solely by ideal user needs or societal goals. This framework is more in line with the social construction of technology, in which users actively shape the evolution of the system rather than passively adapt to it, which is necessary in strategic contexts.

General terms

Human-in-the-Loop (HITL): Humans actively participate in the decision-making process. Ask: “Does the user have to shape or approve the AI’s actions before they are taken?”

Human-on-the-Loop (HOTL): Humans provide feedback and oversight after the AI has made a decision. Ask: “Does the user monitor, interpret or improve the AI’s Output?”

Generative-AI (GenAI) and large language models (LLMs): Computational methods for generating new content such as text, images, code, video or audio from training data. Additionally they can serve as multi-agent assistants that can run multiple tasks at scale.

Double Diamond (DD): A clear, comprehensive, and visual description of the design process. It includes Discover, Define, Develop, Deliver phases.

AI systems that operate without human supervision can be powerful but also misleading, especially in high-stakes areas with strategic decisions (e.g., healthcare, finance). HOTL structures provide a vital balance between autonomy and accountability. Humans monitor AI operations, intervene when necessary, and ensure safety, ethical compliance, and continuous improvement of the system. Unlike fully automated systems, HOTL assigns humans an oversight role, such as monitoring, guiding, and intervening at critical moments, while still allowing AI to operate smoothly in the background.

Why it matters

Without proper controls and boundaries, AI can make biased or unintentionally harmful decisions. HOTL ensures that human experts provide qualitative judgments and contextual insights that may not be captured in the data or available to the system. This oversight not only reduces risk, but also gives humans nuanced judgments that AI cannot handle alone. Including human oversight increases trust by providing ongoing feedback that improves the AI over time.

Framework with key actions

1. Research and exploration

DD: Discover

HOTL focus: User insight informs model and tool design.

Before writing code or designing screens, immerse yourself in the user’s world. Their context, needs, environment, problems, and behaviours should drive your technical direction and data strategy.

  • Conduct qualitative and quantitative research.
  • Gather user context, behavioural patterns, and user experiences.
  • Map user journeys to surface pain points, needs, and motivations.
  • Integrate lightweight research tools into the platform (e. g., micro-surveys, feedback on the process).
  • Determine optimal feedback frequency.

2. Communications and interface design

DD: Define

HOTL: Interfaces guide and stimulate human interaction with models.

Clearly define problems and system structures for simulation tasks. Clarify interactions and demonstrate system behaviour to reduce cognitive load.

  • Translate research into actionable design decisions by empowering users to act, learn, and intervene in meaningful ways.
  • Translate research into clear problem definitions and system interaction patterns.
  • Design and implement 5 types of communications: ○ warnings (e.g., error alerts) ○ notices (e.g., completion alerts, email notifications) ○ status indicators (e.g., loading, success, pending) ○ training messages ○ policy messages.
  • Use visual signals to manage user cognitive load (colour, icons, animation) to indicate certainty, confidence or urgency.

2.1. Data inputs and trust mechanisms

DD: Define – Develop

2.1.1. Data preprocessing

HOTL: System alerts users, but no active decision-making is required.

  • Automate preprocessing where possible to reduce manual work.
  • Notify users of upload errors or invalid formats.
  • Create transparent model input guides.

2.1.2. Inputs that guide system behaviour (e.g., scenario parameters or decision boundaries)

HITL: Users shape inputs and set controls.

  • Segment users based on their roles and expertise level.
  • Enable expert users to manually set simulation variables.

Provide presets for beginners.

2.1.3. Output validator

HITL: Set approval thresholds for high-risk outcomes before taking any action.

  • Implement checkpoints where user approval is required.
  • Create audit trails for user decisions for different user segments (who approved what and when).
  • Create transparent model output guides.

2.1.4. Trust and security in the loop

HOTL: The team identifies anomalies and defines thresholds that require user review.

Define thresholds that trigger mandatory user validation.

Track validation errors over time to improve logic or the need to conduct and develop additional training.

3. Communication processing

DD: Develop – Deliver

HOTL: Feedback must be clear and applicable in practice.

Just showing results is not enough, it is important explaining users their meaning and what to do next. It is necessary to be aware of the user’s ability to understand the communication. Provide learning experiences where users can not only understand the message but also learn what to do in response to it.

  • Design messages with clarity, relevance, and actionability.
  • Include explanations or links for further help.
  • Conduct user testing on messaging effectiveness.
  • Track metrics for comprehension and appropriate user response.

4. Platform hierarchy and navigation

DD: Deliver

HOTL: Help users in complex environments.

As systems become more complex, intuitive navigation and clarity become critical. A clear information architecture allows users to know where to find content (e.g., intuitive layout, accessibility).

  • Develop intuitive navigation and content grouping.
  • Create a sitemap with sections (e.g., training, feedback, support).
  • Validate structure with card-sorting and usability tests.
  • Add a search function.
  • Maintain consistency across user journeys to achieve key user goals.

5.1. Training

DD: Deliver

HOTL: Onboarding and training support deeper engagement.

Users need confidence to deeply engage with models. This can be achieved through excellent training and reinforcement. It is necessary to create different types of training materials, for beginners and pros. This may include video content, successful case studies, and updates.

  • Create multi-level learning content (from introductory to advanced).
  • Record short tutorial videos, walkthroughs and showcase successful use cases.
  • Send notifications to users about new /additional content (via in-app messages, email, alerts).
  • Run usability tests to validate learning efficacy.

5.2. Knowledge application

Promote:

  • knowledge retention, when the user is able to remember a message when a situation arises in which that knowledge needs to be applied.
  • knowledge transfer, when the user is able to recognise situations in which a message is applicable and understand how to apply it.

Key actions:

  • Create mechanisms for continuous learning (e.g., built-in tips, contextual prompts and guidance).
  • Reinforce learning with periodic micro reminders.
  • Create practice scenarios.
  • Conduct user testing to ensure effectiveness.

6.1. Feedback loops

DD: Rediscover – Redefine

HOTL: User feedback helps improve the system continuously.

A living system evolves with its users, so feedback shouldn’t be an afterthought; it should be a function of the system. It is necessary to create a mechanism to ensure continuous improvement based on user interactions and feedback by organising the collection of post-implementation information, integrating system and human feedback loops (combining human + system input).

  • Create mechanisms for continuous collection of user feedback (e.g. post-task surveys, built-in tips and guidance).
  • Set up automated feedback collection based on system usage data (e.g. error trigger, automatically open a window to send feedback to the team).
  • Schedule periodic feedback reviews with cross-functional teams.
  • Set up dashboards for feedback analytics.
  • Update the system iteratively based on real world insights.

6.2. Continuous improvement

  • Document version changes with change logs visible to users.

6.3. Data modelling

HITL: Users adjust or select scenarios.

  • Integrate feedback loops where users correct, validate or select the best scenario.

7. Personalisation

DD: Rediscover

HOTL: Contextual relevance improves usability and comprehension.

Create personalised user experiences based on the user’s role. Tailored experiences increase adoption and satisfaction.

  • Segment users and deliver role-based UI elements.
  • Adapt training depth and frequency per user group.
  • Recommend content dynamically based on past actions, errors, feedback or questions asked.

8. Collaboration

DD: Redefine – Develop

HITL: Users are collaborators, not just validators.

Involve users and your team in the iterative improvement process. Their insights are critical to improving AI systems.

  • Set up feedback sessions and co-creation workshops with different user’s segments.
  • Encourage cross-disciplinary collaboration (design, engineering, data science, management).
  • Create and maintain a shared document for ongoing user insights and feedback.
  • Establish a framework for prioritising and implementing suggestions.

Conclusion

Human-centred design is more than just a workflow; it’s a philosophy that recognises users as co-creators of AI systems. By integrating design thinking and machine learning, this framework ensures that technology serves people, not the other way around. This means creating intuitive interfaces that empower users through incorporating transparency, trust, and continuous feedback.

In a nutshell:

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Bitcoin hits record high as Capitol Hill turns to 'crypto week'
Next Article All Best Buy stores will have Switch 2 consoles on July 17th
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

Two channels brought BACK to Virgin Media today and Sky Sports changes coming
News
Deepfake Detection: Neural Networks Are Fighting Themselves | HackerNoon
Computing
When Is the Home Run Derby 2025? Find Out When It Starts and How to Watch for Free
News
Elon Musk’s AI bot adds a ridiculous anime companion with ‘NSFW’ mode
News

You Might also Like

Computing

Deepfake Detection: Neural Networks Are Fighting Themselves | HackerNoon

9 Min Read
Computing

Grok 4 Claims “PhD‑level” Intelligence but at a Cost | HackerNoon

14 Min Read
Computing

Grok 4 Arrives With Impressive Coding Powers | HackerNoon

10 Min Read
Computing

How to Balance Investor Relations and Product Growth | HackerNoon

8 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?