• Current students
      • Student centre
        Enrol on a course/exam
        My enrolments
        Exam results
        Mock exams
      • Course information
        Students FAQs
        Student induction
        Course enrolment information
        Key dates
        Book distribution
        Timetables
        FAE elective information
        CPA Ireland student
      • Exams
        CAP1 exam
        CAP2 exam
        FAE exam
        Access support/reasonable accommodation
        E-Assessment information
        Exam and appeals regulations/exam rules
        Timetables for exams & interim assessments
        Sample papers
        Practice papers
        Extenuating circumstances
        PEC/FAEC reports
        Information and appeals scheme
        Certified statements of results
        JIEB: NI Insolvency Qualification
      • Training and development
        Mentors: Getting started on the CA Diary
        CA Diary for Flexible Route FAQs
        Training Development Log
      • Admission to membership
        Joining as a reciprocal member
        Admission to Membership Ceremonies
        Admissions FAQs
      • Support & services
        Recruitment to and transferring of training contracts
        CASSI
        Student supports and wellbeing
        Audit qualification
        Diversity and Inclusion Committee
    • Students

      View all the services available for students of the Institute

      Read More
  • Becoming a student
      • About Chartered Accountancy
        The Chartered difference
        Student benefits
        Study in Northern Ireland
        Events
        Hear from past students
        Become a Chartered Accountant podcast series
      • Entry routes
        College
        Working
        Accounting Technicians
        School leavers
        Member of another body
        CPA student
        International student
        Flexible Route
        Training Contract
      • Course description
        CAP1
        CAP2
        FAE
        Our education offering
      • Apply
        How to apply
        Exemptions guide
        Fees & payment options
        External students
      • Training vacancies
        Training vacancies search
        Training firms list
        Large training firms
        Milkround
        Recruitment to and transferring of training contract
      • Support & services
        Becoming a student FAQs
        School Bootcamp
        Register for a school visit
        Third Level Hub
        Who to contact for employers
    • Becoming a
      student

      Study with us

      Read More
  • Members
      • Members Hub
        My account
        Member subscriptions
        Newly admitted members
        Annual returns
        Application forms
        CPD/events
        Member services A-Z
        District societies
        Professional Standards
        ACA Professionals
        Careers development
        Recruitment service
        Diversity and Inclusion Committee
      • Members in practice
        Going into practice
        Managing your practice FAQs
        Practice compliance FAQs
        Toolkits and resources
        Audit FAQs
        Practice Consulting services
        Practice News/Practice Matters
        Practice Link
      • In business
        Networking and special interest groups
        Articles
      • District societies
        Overseas members
      • Public sector
        Public sector presentations
      • Member benefits
        Member benefits
      • Support & services
        Letters of good standing form
        Member FAQs
        AML confidential disclosure form
        Institute Technical content
        TaxSource Total
        The Educational Requirements for the Audit Qualification
        Pocket diaries
        Thrive Hub
    • Members

      View member services

      Read More
  • Employers
      • Training organisations
        Authorise to train
        Training in business
        Manage my students
        Incentive Scheme
        Recruitment to and transferring of training contracts
        Securing and retaining the best talent
        Tips on writing a job specification
      • Training
        In-house training
        Training tickets
      • Recruitment services
        Hire a qualified Chartered Accountant
        Hire a trainee student
      • Non executive directors recruitment service
      • Support & services
        Hire members: log a job vacancy
        Firm/employers FAQs
        Training ticket FAQs
        Authorisations
        Hire a room
        Who to contact for employers
    • Employers

      Services to support your business

      Read More
☰
  • Find a firm
  • Jobs
  • Login
☰
  • Home
  • Knowledge centre
  • Professional development
  • About us
  • Shop
  • News
Search
View Cart 0 Item

Thought Leadership

☰
  • Home
  • Resources
  • Articles & insights
  • Events
  • Home/
  • Thought Leadership/
  • Articles & insights/
  • Articles items
☰
  • Resources
  • Position papers
    • The Next Financial Year: Reducing Red Tape
    • The Next Financial Year: Supporting SMEs
    • The Next Financial Year: Climate Goals
    • The Next Financial Year: Building Capacity
    • The Next Financial Year 2022
    • The Next Financial Year 2021
    • The Next Financial Year 2020
    • Irish Public Sector Accounting Reforms
  • Guides & reports
  • Podcasts & videos
    • Podcasts
      • Transcripts
    • Videos
  • Bookshop
  • Authors and lecturers
News
(?)

Geopolitical risk: the must-tackle issue for your board

Geopolitical uncertainty is reshaping boardroom priorities and acquiring the right expertise is crucial for strategic resilience, writes Dan Byrne Geopolitical risk: Is your board talking about it? If so, do they know how to handle it? The harsh reality is that many companies can’t do so properly. However, stakeholders are rarely patient when it comes to geopolitics. When something happens, they want a response from your corporate leadership.  The last thing your board needs to be is unaware of how to handle a situation, what to say, and how to adapt your strategy to changing global events. The challenge is processing that it’s all happening at once.  The news cycle is now dominated by the Israel-Hamas war. Before this started, the spotlight was on the Russian invasion of Ukraine and, before that, the chaotic US withdrawal from Afghanistan.  Meanwhile, we’ve got tensions between the West and China, the right-wing backlash against Brazilian and US elections, and unresolved Brexit issues – not to mention the protracted conflicts that are now so ingrained in the fabric of modern geopolitics. Every geopolitical crisis begins a new chapter of geopolitical pressure in corporate playbooks. The importance of geopolitical risk Assessing geopolitical risk is essential. It’s not going away and, depending on your company, it could be crucial to your strategy.  This doesn’t have to be direct – your company’s stance on a particular issue, for example. It can also be indirect – such as the businesses you work with within your supply chain. Many American companies have been shifting their manufacturing from China to other locations, such as Vietnam, out of fear that Chinese authorities could disrupt their business at the drop of a hat. Corporate leaders will be prodded by investors wanting to know if their company can survive through sanctions or consumers wanting to see their response to escalating conflict. The storm of questions will come; the challenge is how best to weather it. Expertise needed Experts in geopolitical risk will have the following skills: A deep understanding of corporate strategy and risk; Knowledge of global affairs, new or potential conflicts with global impacts, the intricacies of trade sanctions and the knock-on effects of government changes on international relations; and The ability to navigate through substantial geopolitical fallouts. The hard part is finding this expertise. Finding the right candidate to fill a board seat depends on multiple factors, like the availability of talent, training, networks, and an alignment of values. In some situations, this is a heavy ask.  It’s also worth noting that the market for geopolitical expertise is highly active right now as companies realise that they need to be prepared. Playing the long game Organisations should realise that the quest for geopolitical experience for your board may be a long game.  It can take time to find the talent that works well for your business – and it’s time that stakeholders may not always give you, pushing you for an answer and refusing to accept that you might need more time. That’s why it is essential to start now on geopolitical expertise if you haven’t already. If it feels like you’re playing catch-up, bear in mind that this won’t always be the case. Eventually, you will have the solid knowledge you need on your board to help you develop thorough answers to complex questions.  In reality, the world always moves faster than corporate governance is comfortable with, so it’s better to get ahead. Dan Byrne is a content writer at The Corporate Governance Institute

Nov 17, 2023
READ MORE
News
(?)

How digital leaders can unlock business success

Successful digital transformation requires strong leadership. Dave Vincent outlines his tips for successfully embedding innovation in business Growing up in Belfast in the seventies and eighties, I was convinced that by 2020, we’d all be travelling around on flying cars or hoverboards and have a host of robot servants looking after us. Fast forward to the nineties, and I can vividly remember sitting in my university computer lab wrestling with the logic and code required to help teach a hungry virtual monkey how to get their hands on a hidden bunch of bananas. In 2023, while software applications and systems are significantly more developed than in the nineties, the reality doesn’t quite match the vision of the eighties (the hoverboard being the biggest disappointment). I couldn’t have dreamed of many of the developments that have instead taken place, however. Since the term ‘artificial intelligence’ (AI) was first coined almost 70 years ago, we’ve seen wave after wave of technology-enabled innovation, from the rise of personal computing to the internet, mobile devices, augmented and virtual reality, the cloud, the metaverse, self-driving vehicles and now, generative AI. Each shift has captured the imagination, created new opportunities and raised further questions and challenges for business leaders. We are surrounded by technology, and every day, we can see that technology evolves and changes as it impacts how we live and do business. How can technology help? Some of the most frequent questions I hear from clients considering digital transformation are: “Where do I start?” “How do I create the most impact?” “What does success look like?” Rather than starting by asking or thinking about what a particular tool or technology can do, I prefer to reverse engineer the questions and ask: “What are you trying to do in your business or what problem would you like to solve, and how can a digital mindset or technology help?” As these new technologies continue to influence all areas of our business operations, customers and employees, companies need a new type of (digital) leader who can understand, interpret and navigate this digital transformation era. The digital leader Implementing new technology is challenging. The organisation seeking to embed the latest technology — and its staff — must unlearn old concepts and embrace the new systems. For digital leaders, this means adopting alternative leadership styles. In the past, leadership was about giving orders and making decisions. Digital leaders know that successful digital transformation is not just about adopting technology; it’s about transforming business and operating models, driving growth, enhancing competitive advantage and increasing business agility. Today’s leaders must be able to evaluate progress, priorities and business models continually and be prepared to change direction quickly. Digital leaders need to understand not only how systems and technology work but also how that technology will be received and used by staff, as well as how it will impact how employees work and the type of work they do. Digital leaders need to be able to effectively manage employees through shifts and changes to ensure that digital technology is used to deliver the best business outcomes. To make informed and pragmatic decisions about technology, digital leaders must be able to evaluate the impact technology can deliver for their organisations, use data to inform policy and decision-making, and proactively assess and manage risks related to data security. To drive digital innovation, leaders must be agile and flexible, creating a culture where innovation, collaboration and continuous learning can flourish and empower their teams to make data-driven decisions. To ensure focus and alignment, leaders must share a well-defined and compelling strategic vision, calling out what success looks like and showing the roadmap that will get there. It is also important to remember that leadership in the digital age is not just the responsibility of the nominated digital leaders and senior leadership teams, however. Every employee can be a digital leader. The future is digitally enabled Organisations can drive digital innovation and growth from the ground up by empowering staff at all levels to take ownership of, and show leadership in, their work.  The future of work is undoubtedly digitally enabled, and business leaders who are prepared to embrace this change and lead their teams effectively will be the ones to succeed. The digital age allows leaders to create more meaningful and purpose-driven work for their employees and promote innovation and growth for their organisations. Organisations can position themselves for continued success by investing in digital leadership development.  And maybe somebody will finally work out how to create the hoverboard of my dreams. Dave Vincent is a Director of Digital Transformation at Grant Thornton Northern Ireland

Oct 20, 2023
READ MORE
Ethics and Governance
(?)

Navigating the ethics of AI

Michael Diviney and Níall Fitzgerald explore the ethical challenges arising from artificial intelligence (AI), particularly ‘narrow’ AI, and highlight the importance of ethics and professional competence in its deployment Earlier this year, artificial intelligence (AI) industry leaders, leading researchers and influencers signed a succinct statement and warning: “Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.” Was this a publicity stunt? Well, probably not, as the generative AI ChatGPT was already the fastest-adopted application in history.  Was this an over-the-top, alarmist statement by a group possibly trying to steal a march on self-regulation of a rapidly emerging technology and growing industry?  Again, this is unlikely if one considers the warnings of pioneer thinkers like Nick Bostrom, Max Tegmark, Stephen Hawking and Astronomer Royal Martin Rees. They concur that there is an existential threat to humankind if human-level or ‘general’ AI is developed and the ‘singularity’ is reached when AI surpasses human intelligence.  Autonomous weapons and targeting are a clear risk, but more broadly, unless we can ensure that the goals of a future superintelligence are aligned and remain aligned with our goals, we may be considered superfluous and dispensable by that superintelligence.  As well as the extinction threat, general AI presents other potential ethical challenges.  For example, if AI attains subjective consciousness and is capable of suffering, does it then acquire rights? Do we have the right to interfere with these, including the right to attempt to switch it off and end its digital life?  Will AI become a legal entity and have property rights? After all, much of our economy is owned by companies, another form of artificial ‘person’. Ethical challenges from ‘narrow’ AI Until general AI is here, however – and there is informed scepticism about its possibility – the AI tools currently in use are weak or ‘narrow’ AI. They are designed to perform a specific task or a group of related tasks and rely on algorithms to process data on which they have been trained.  Narrow AI presents various ethical challenges:  Unfairness arising from bias and opacity (e.g. AI used in the initial screening of job candidates include a gender bias based on historical data – in the past more men were hired); The right to privacy (AI trained with data without the consent of the data subjects); Threats to physical safety (e.g. self-driving vehicles); Intellectual property and moral rights, plagiarism and passing-off issues in the use of generative AI like ChatGPT and Bard; and Threats to human dignity from the hollowing out of work and loss of purpose. Regulation vs. ethics Such issues arising from the use of AI, particularly related to personal data, mean that regulation is inevitable.  We can see this, for example, with the EU’s landmark AI Act, due to apply by the end of 2025, which aims to regulate AI’s potential to cause harm and to hold companies accountable for how their systems are used. However, as Professor Pat Barker explained at a recent Consultative Committee of Accountancy Bodies (CCAB) webinar, until such laws are in place, and in the absence of clear rules, ethics are required for deciding on the right way to use AI.  Even when the regulation is in place, there are likely to be cases and dilemmas that it has not anticipated or about which it is unclear. Legal compliance should not be assumed to have all the ethical issues covered, and as AI is evolving so quickly, new ethical issues and choices will inevitably emerge.  Ethics involves the application of a decision-making framework to a dilemma or choice about the right thing to do. While such a framework or philosophy can reflect one’s values, it must also be objective, considered, universalisable and not just based on an instinctual response or what may be expedient. Established ethics frameworks include: the consequentialist or utilitarian approach – in the case of AI, does it maximise benefits for the greatest number of people?; and the deontological approach, which is based on first principles, such as the inalienable rights of the individual (an underlying philosophy of the EU’s AI Act). (The Institute’s Ethics Quick Reference Guide, found on the charteredaccountants.ie website, outlines five steps to prepare for ethical dilemmas and decision-making.)  A practical approach While such philosophical approaches are effective for questions like “Should we do this?” and “Is it good for society”, as Reid Blackman argues in Harvard Business Review, businesses and professionals may need a more practical approach, asking: “Given that we are going to [use AI], how can we do it without making ourselves vulnerable to ethical risks?”  Clear protocols, policies, due diligence and an emphasis on ethical risk management and mitigation are required, for example responsible AI clauses in agreements with suppliers. In this respect, accountants have an arguably competitive advantage in being members of a profession; they can access and apply an existing ethical framework, which is evolving and adapting as the technology, its opportunities and challenges change.  The Code of Ethics The International Ethics Standards Board for Accountants (IESBA) recently revised the Code of Ethics for Professional Accountants (Code) to reflect the impact of technology, including AI, on the profession. The Chartered Accountants Ireland Code of Ethics will ultimately reflect these revisions.  IESBA has identified the two types of AI likely to have the most impact on the ethical behaviour of accountants:  Assisted intelligence or robotic process automation (RPA) in which machines carry out tasks previously done by humans, who continue to make decisions; and  Augmented intelligence, which involves collaboration between human and machine in decision-making. The revisions also include guidance on how accountants might address the risks presented by AI to ethical behaviour and decision-making in performing their role and responsibilities.  Professional competence and due care The Code requires an accountant to ensure they have an appropriate level of understanding relevant to their role and responsibilities and the work they undertake. The revisions acknowledge that the accountant’s role is evolving and that many of the activities they undertake can be impacted by AI.  The degree of competency required in relation to AI will be commensurate with the extent of an accountant’s use of and/or reliance on it. While programming AI may be beyond the competency of many accountants, they have the skill set to:  identify and articulate the problem the AI is being used to solve;  understand the type, source and integrity of the data required; and assess the utility and reasonableness of the output.  This makes accountants well placed to advise on aspects of the use of AI. The Code provides some examples of risks and considerations to be managed by professional accountants using AI, including: The data available might not be sufficient for the effective use of the AI tool. The accountant needs to consider the appropriateness of the source data (e.g. relevance, completeness and integrity) and other inputs, such as the decisions and assumptions being used as inputs by the AI. This includes identifying any underlying bias so that it can be addressed in final decision-making. The AI might not be appropriate for the purpose for which the organisation intends to use it. Is it the right tool for the job and designed for that particular purpose? Are users of the AI tool authorised and trained in its correct use within the organisation’s control framework? (One chief technology officer has suggested not only considering the capabilities of the AI tool but also its limitations to be better aware of the risks of something going wrong or where its use may not be appropriate.) The accountant may not have the ability, or have access to an expert with that ability, to understand and explain the AI and its appropriate use.  If the AI has been appropriately tested and evaluated for the purpose intended. The controls relating to the source data and the AI’s design, implementation and use, including user access. So, how does the accountant apply their skills and expertise in this context?  It is expected that accountants will use many of the established skills for which the profession is known to assess the input and interpret the output of an AI tool, including interpersonal, communication and organisational skills, but also technical knowledge relevant to the activity they are performing, whether it is an accounting, tax, auditing, compliance, strategic or operational business decision that is being made.  Data and confidentiality According to the Code, when an accountant receives or acquires confidential information, their duty of confidentiality begins. AI requires data, usually lots of it, with which it is trained. It also requires decisions by individuals in relation to how the AI should work (programming), when it should be used, how its use should be controlled, etc.  The use of confidential information with AI presents several confidentiality challenges for accountants. The Code includes several considerations for accountants in this regard, including: Obtaining authorisation from the source (e.g. clients or customers) for the use of confidential information, whether anonymised or otherwise, for purposes other than those for which it was provided. This includes whether the information can be used for training AI tools.  Considering controls to safeguard confidentiality, including anonymising data, encryption and access controls, and security policies to protect against data leaks.  Ensuring controls are in place for the coding and updating of the AI used in the organisation. Outdated code, bugs and irregular updates to the software can pose a security risk. Reviewing the security certification of the AI tool and ensuring it is up to date can offer some comfort.  Many data breaches result from human error, e.g. inputting confidential information into an open-access web-based application is a confidentiality breach if that information is saved, stored and later used by that application. Staff need to be trained in the correct use and purpose of AI applications and the safeguarding of confidential information. Dealing with complexity The Code acknowledges that technology, including AI, can help manage complexity.  AI tools can be particularly useful for performing complex analysis or financial modelling to inform decision-making or alerting the accountant to any developments or changes that require a re-assessment of a situation. In doing so, vast amounts of data are collected and used by AI, and the ability to check and verify the integrity of the data introduces another level of complexity.  The Code makes frequent reference to “relevancy” in relation to the analysis of information, scenarios, variables, relationships, etc., and highlights the importance of ensuring that data is relevant to the problem or issue being addressed. IESBA was mindful, when revising the Code, that there are various conceivable ways AI tools can be designed and developed to use and interpret data.  For example, objectivity can be challenged when faced with the complexity of divergent views supported by data, making it difficult to come to a decision. AI can present additional complexity for accountants, but the considerations set out in the Code are useful reminders of the essential skills necessary to manage complexity. Changing how we work As well as its hugely beneficial applications in, for example, healthcare and science, AI is proving to be transformative as a source of business value.  With a range of significant new tools launched daily, from personal effectiveness to analysis and process optimisation, AI is changing how we work. These are powerful tools, but with power comes responsibility. For the professional accountant, certain skills will be brought to the fore, including adaptability, change and risk management, and leadership amidst rapidly evolving work practices and business models. Accountants are well placed to provide these skills and support the responsible and ethical use of AI.  Rather than fearing being replaced by AI, accountants can prepare to meet expectations to provide added value and be at the helm of using AI tools for finance, management, strategic decision-making and other opportunities. Michael Diviney is Executive Head of Thought Leadership at Chartered Accountants Ireland Níall Fitzgerald is Head of Ethics and Governance at Chartered Accountants Ireland

Aug 02, 2023
READ MORE
...11121314151617181920...

Was this article helpful?

yes no

The latest news to your inbox

Please enter a valid email address You have entered an invalid email address.

Useful links

  • Current students
  • Becoming a student
  • Knowledge centre
  • Shop
  • District societies

Get in touch

Dublin HQ 

Chartered Accountants
House, 47-49 Pearse St,
Dublin 2, D02 YN40, Ireland

TEL: +353 1 637 7200
Belfast HQ

The Linenhall
32-38 Linenhall Street, Belfast,
Antrim, BT2 8BG, United Kingdom

TEL: +44 28 9043 5840

Contact us

Connect with us

Something wrong? Is the website not looking right/working right for you? Browser support
Chartered Accountants Worldwide homepage
Global Accounting Alliance homepage
CCAB-I homepage
Accounting Bodies Network homepage

© Copyright Chartered Accountants Ireland 2020. All Rights Reserved.

☰
  • Terms & conditions
  • Privacy statement
  • Event privacy statement
  • Privacy complaint
  • Sitemap
LOADING...

Please wait while the page loads.