Article: How managers can use AI to augment their capability Written by Jean Gan Tuesday 26 August 2025 Share Share to LinkedIn Share to Facebook Share via email Effective use of AI is about balancing technology and human judgement – but oversight must not be a box-ticking exercise AI is no longer a future possibility. It is already shaping how decisions are made, tasks are prioritised and performance is evaluated. Managers are not preparing for disruption. They are already managing it. While AI can generate, predict and optimise quickly, it cannot reflect, empathise or lead. Machines analyse data. Humans interpret meaning, apply judgement and remain accountable for outcomes. That distinction is essential to responsible leadership. This article offers guidance on how managers can balance the strengths of AI with the human capabilities that remain essential. When AI supports and when humans must lead AI is well suited to structured, rule-based tasks. Using it to summarise meetings, screen CVs or segment customers saves time and increases consistency. But most work involves nuance, incomplete data or shifting goals. AI may produce results that appear correct but lack depth. Take hiring. AI might filter CVs based on past patterns. But if those patterns contain bias, such as penalising career breaks or excluding certain groups, it can reinforce unfairness. A human manager must step in and assess with fairness in mind. Even in performance reviews, AI might highlight metrics, but it cannot detect low morale or interpersonal friction. Only a human leader can interpret the bigger picture and respond appropriately. AI should support. Humans must decide. Oversight must be active, not symbolic Having a ‘human in the loop’ is often used as a safeguard, but in practice this can become a passive approval step. Managers may be asked to sign off on AI outputs without the understanding or authority to question them. Meaningful oversight means knowing how the system works, recognising its limits and feeling confident to revise or reject results. This is especially important with AI hallucinations – false yet convincing information generated by the system. If accepted without scrutiny, these can undermine trust and decision-making. Keep reading: why governance must be led by managers Login or register below for Free Instant Access Login If you are already registered as a CMI Friend, Subscriber or Member, just login to view this article. Confirm your registration Login below to confirm your details and access this article. Sign in with email Email remember me remember Forget? Please confirm that you want to switch off the "Sign in with email" remember me feature. Yes No Register for Free Access Not yet a Member, Subscriber or Friend? Register as a CMI Friend for free, and get access to this and many other exclusive resources, as well as weekly updates straight to your inbox. You have successfully registered As a CMI Friend, you now have access to whole range of CMI Friendship benefits. Please login to the left to confirm your registration and access the article. Article Our extensive range of articles are designed to keep you in the loop with all the latest management and leadership best practice, research and news. Members See More CMI Members have access to thousands of online learning and CPD resources. Learn more about our membership benefits Join The Community CMI offers a variety of flexible membership solutions, tailored to your needs. Find out more and get involved in the CMI community today.