Capgemini New Zealand https://www.capgemini.com/nz-en/ Get the future you want Fri, 23 Feb 2024 09:21:10 +0000 en-US hourly 1 https://wordpress.org/?v=6.3.3 https://www.capgemini.com/nz-en/wp-content/uploads/sites/37/2022/10/cropped-cropped-favicon.webp?w=32 Capgemini New Zealand https://www.capgemini.com/nz-en/ 32 32 209921165 How Gen AI will revolutionize Telecom Network Operations https://www.capgemini.com/nz-en/2024/02/23/how-gen-ai-will-revolutionize-telecom-network-operations/ https://www.capgemini.com/nz-en/2024/02/23/how-gen-ai-will-revolutionize-telecom-network-operations/#respond Fri, 23 Feb 2024 09:21:01 +0000 https://www.capgemini.com/nz-en/?p=631814&preview=true&preview_id=631814 The post How Gen AI will revolutionize Telecom Network Operations appeared first on Capgemini New Zealand .

]]>

How Gen AI will revolutionize Telecom Network Operations

Yannick Martel
Feb 1, 2024

The launch of ChatGPT in late 2022 propelled Generative AI to prominence, gaining significant visibility and popularity among both consumers and organizations. Telecom operators have been quick to experiment with this technology, exploring applications that would boost individual productivity and optimize industry-specific processes, like client interactions

Most recently we have seen a significant interest in applying Generative AI to Network Operations, which is an area in which CSPs (Communications Service Providers) have long sought efficiency gains to enable the management of an increasingly complex technology landscape. In this article, we explore how operators can leverage Gen AI to augment the employee experience and automate business processes, all in the name of smarter, faster, more resilient networks.

Network knowledge at your fingertips

Generative AI is a gamechanger when it comes to intelligent document querying and information retrieval. The retrieval augmented generation (RAG) pattern allows the interrogation of text indexes via a semantic representation of a query and has quickly become a standard. Based on retrieved sections of selected documents, Generative AI can easily produce summaries compliant with pre-defined templates.

This application is demonstrated with one of our clients, who is now providing network technicians with a tool that enables them to quickly access a summary of past incidents that have affected a specific node on the broadband access network. Thanks to Generative AI, the new tool can retrieve all incidents relevant to the current investigation and produce a formatted summary of past events – reducing the time for this task.

CSPs are also investigating the use of Generative AI for managing contracts, such as interconnect and roaming contracts, or cell tower lease agreements. Through the help of Gen AI-enabled tools, users can quickly search for specific clauses or ask direct questions, such as “What are the security procedures for accessing this site?” or “What is the average price for site rentals in Madrid?” This removes the need for extensive, complex searches for the most up-to-date and relevant contracts and amendments. It also reduces the risk of error, especially when navigating large, complex document repositories that can span several years and many geographies.

The rise of conversational interfaces

For the past 60 years, most traditional applications have been using either command-line or graphical, menu-based interfaces. While regular users have mastered these tools, occasional users struggle. This is why some customers still prefer calling the contact center instead of using a mobile app!

Generative AI gives a new dimension to user interfaces, moving from predefined, rigid dialogs, to more free-flowing and intuitive conversations. This is relevant for some of the interfaces used when operating networks, where the user experience can be much improved.

For instance, we are currently defining a proof-of-concept with one of our CSP clients to support field technicians who perform interventions at customer homes and/or network points of interest. These technicians frequently need specialized help from team leads or colleagues while in the field; if this help cannot be provided on demand, then the service agent may need to schedule a follow up intervention.

Generative AI allows the development of a conversational bot that enables the technician to get the information they need about the specific location/services and technologies that will enable them to successfully troubleshoot or deploy services. A voice bot or a chat bot makes the job of the technician easier and quicker, allowing for faster issue resolution and avoiding repeat and costly truck rolls.

These service tools can also be combined with augmented reality (AR) applications, such as using a mobile phone to scan physical devices and generate relevant information. A variety of new interfaces that enable this type of support is made possible by the emergence of multimodal models such as Google’s Gemini, OpenAI’s GPT-4 and Mistral AI’s Mistral 7B.

An additional application focuses on network configuration. Specific intents, like boosting radio capacity at a stadium ahead of a major event, are set up through dedicated interfaces that require expertise on network management applications. By employing a conversational interface, network engineers can effortlessly grasp available capacity and make configurations for upcoming activities through friendly conversations with an agent. This conversational agent doubles as an advisor, drawing insights from historical activities and the present network status.

Autonomous Networks monitoring with a human touch

Moving to a higher level of network automation is critical for network operators. In fact, this is the key to improving service quality within a more complex technological environment without adding additional staff. AI is key to moving from human-managed networks, supported by insights from data, to AI-managed networks. Generative AI can thus complement other AI models, such as anomaly detection and classification.

While the industry has settled on the term “Autonomous Networks,” the goal is not complete autonomy. Configuration of intent is essential, and ongoing network monitoring for compliance is crucial. Even with a significant level of automation, human oversight remains imperative to ensure safety and maintain the quality of service.

Generative AI can produce a human readable summary of the status and activity in the network, allowing human agents to understand if and how the intent is satisfied. Even when operating networks with a high level of automation, human agents must be able to investigate, ask specific questions and get replies.

In the same way, an Autonomous Network’s reaction to an alarm or an anomaly must be defined in advance by a network engineer. On older generations of solutions, scripts must be developed and tested, which requires strong expertise. With Generative AI, natural language could be used to define responses to alarms, going as far as to extract appropriate remediation procedures from process documents. This approach allows human experts to review and make necessary adjustments.

Leading the way in an AI revolution

Like traditional AI, there are many use cases for Generative AI in the Telecom industry. In addition to individual agent productivity tools, Generative AI can be used to refine and streamline operational processes to better manage networks.

At Capgemini, we are now experimenting with leading CSPs on how to augment or automate existing workflows through AI technologies, helping them create a smarter, faster, more resilient network.

TelcoInsights is a series of posts about the latest trends and opportunities in the telecommunications industry – powered by a community of global industry experts and thought leaders.

Meet the author

Yannick Martel

Telco Leader

    The post How Gen AI will revolutionize Telecom Network Operations appeared first on Capgemini New Zealand .

    ]]>
    https://www.capgemini.com/nz-en/2024/02/23/how-gen-ai-will-revolutionize-telecom-network-operations/feed/ 0 631814
    Open RAN needs to automate – and fast https://www.capgemini.com/nz-en/2024/02/23/open-ran-needs-to-automate-and-fast/ https://www.capgemini.com/nz-en/2024/02/23/open-ran-needs-to-automate-and-fast/#respond Fri, 23 Feb 2024 09:13:50 +0000 https://www.capgemini.com/nz-en/?p=631812&preview=true&preview_id=631812 The post Open RAN needs to automate – and fast appeared first on Capgemini New Zealand .

    ]]>

    OPEN RAN NEEDS TO AUTOMATE – AND FAST

    Arnab Das
    20 May 2022
    capgemini-engineering

    Without efficient automation, telcos are currently devoting upwards of 50% of their network operations to RAN. So how can automation accelerate and take its place at the heart of Open RAN?

    The case for automation is clear: it will increase efficiency, lower the total cost of ownership – the list goes on. Telcos know that they need to move towards an automated and efficient network if they want to support agile service innovation and delivery on a competitive level. But at present, automation has only reached varying levels of maturity across the network span. For many telcos, RAN automation is still limited to discrete trials with small groups of vendors; it’s mostly experimental, and its scalability remains unproven.

    Even a few years ago, the complexity of trying to link multiple software systems would have been unthinkable. Today disaggregated RAN is not only possible – it provides a very real competitive edge. The problem is, Open RAN requires a level of intelligent automation that is difficult to build completely in-house1. Let’s look at the specific challenges to automating Open RAN, and then see what solutions are available.

    No one said Open RAN would be simple…

    Radio networks are inherently complex. Add to that business requirements that mandate compatibility between new, next-gen networks and legacy technologies, and that complexity multiplies. Operators find themselves facing two options: automation that’s fairly easy to implement, but limited in scope, or automation that links entire networks, but must be custom built, which typically requires some help from outside software experts. Add to that the steadily increasing number of sites2, plus the need to keep software expenditures and OPEX in check, and that complexity becomes a serious obstacle.

    5G generates a flood of data that – for all the reasons listed above – creates some very real challenges for operators. This data needs to be classified and prioritized for effective network control and management to be possible. The solution? Classic automation is not enough. Open RAN depends on intelligent automation.

    Abstracted architecture, concrete benefits

    Capgemini solves the issue of multi-vendor CNF diversity through layered, abstracted architecture. In plain English, abstracted means that the architecture is not tied to any specific vendors’ software, but can be quickly tailored to incorporate multiple combinations. It’s the difference between a recipe for a cake, which needs to be followed precisely, and only works for one set of ingredients, and the skill of barbequing. A good grill master can swap out any number of meats or vegetables and adjust the technique slightly, without the need to find (or create) a new recipe each time. In the same way, an abstracted architecture makes it possible to automate a network end-to-end, without the difficulty and cost of a fully unique solution. That goes a long way to managing time, costs and complexity. But the most interesting piece is yet to come.

    The spark of intelligence

    The heart of Capgemini’s OpenRAN Operations Automation solution lies in a set of RAN applications driven by our NetAnticipate AI-Model platform. To address the issues of automation in real time – when millions of impulses are streaming through networks and each one must be routed correctly and immediately – something more than standard automation is required. The innovative solution we’ve created uses the near Real-Time RIC (nRT-RIC) model. This is based on the extendible, abstracted architecture described above, that enables easy integration of multi-vendor xAPPs on nRT-RIC. So whatever vendors an operator is working with, the same powerful AI is able to handle the traffic. RAN-specific AI models that learn with no supervision, make O-RAN NonRealTime RIC implementation possible. The result is a complex network that essentially runs on autopilot3.

    The challenges of Open RAN – the complexity, the constraining brownfield environment, the risk of cost overruns – these all come down to the need to intelligently manage information flow. By doing so, our OpenRAN Operations Automation solution opens the door to a range of benefits.

    The benefits of automated Open RAN

    Intelligent automation turns Open RAN from a resource-intensive challenge into a source of value. Some of the benefits include:

    • The ability to deploy in multi-vendor RAN environments consisting of complex multi-technology networks where the automation platform can create the greatest operational impact.
    • Lower operational costs thanks to the automation of network deployment and network operation, leveraging new automation rApps and xApps deployed over O-RAN SMO.
    • The ease of harnessing proven operational models of legacy RAN application by modernizing to cloud-native service on nRT and NRT RIC platform.
    • The ability to deploy RAN automation across multi-technology networks, using design patterns in alignment with the O-RAN Alliance, and providing future-proof flexible automation across varying technology and vendors. The level of automation can also be adjusted in the network’s constituent layers to varying degrees – for example a high level of automation on RAN compute and connectivity infra, and a medium level in radio resource management.

    Looking forward

    Solving the riddle of efficient Open RAN automation is only the beginning. With reliable RAN, telcos can create new innovative services like end-to-end network slicing and open and closed loop service assurance, thanks to Capgemini’s RAN automation solution portfolio. With the same underlying infrastructure, telcos can start to capitalize on new business opportunities as service providers for 5G consumer and enterprise services, as well as for mission-critical communications providers. With the right partner and the right solutions, the benefits are just around the corner. Contact me below to learn more.

    TelcoInsights is a series of posts about the latest trends and opportunities in the telecommunications industry – powered by a community of global industry experts and thought leaders.

    1 Such automation depends on specialized, real-time cloud network functions (CNF) software management and deployment automation

    2 Due to network densifications and real-time sensitive Radio Network traffic, which makes it very difficult for any third-party generic automation master controller to seamlessly schedule and operate cloud network functions (CNF), virtual network functions (VNF), or physical network functions (PNF) without a large degree of customization.

    3 It’s worth noting that the nRT-RIC does not have a clearly-defined open interface to xApps, so to enable developers from different expertise background to be able to create xAPPs in an Application Builder environment, Cagpemini has introduced a SDK (xAPP SDK) and plugin template approach to xAPP development.  This can easily be integrated over the underlying nRT-RIC.

    Author

    Arnab Das

    Vice President of Advanced Connectivity (5G, EDGE Compute, O-RAN, IoT, Telco Engineering R&D) at Capgemini

      The post Open RAN needs to automate – and fast appeared first on Capgemini New Zealand .

      ]]>
      https://www.capgemini.com/nz-en/2024/02/23/open-ran-needs-to-automate-and-fast/feed/ 0 631812
      Why sustainable IT is the backbone of a greener future https://www.capgemini.com/nz-en/2023/10/10/why-sustainable-it-is-the-backbone-of-a-greener-future/ https://www.capgemini.com/nz-en/2023/10/10/why-sustainable-it-is-the-backbone-of-a-greener-future/#respond Tue, 10 Oct 2023 11:25:10 +0000 https://www.capgemini.com/nz-en/?p=630983&preview=true&preview_id=630983 The post Why sustainable IT is the backbone of a greener future appeared first on Capgemini New Zealand .

      ]]>

      why sustainable it is the backbone of a greener future

      Ygor Zeliviansky
      Apr 11, 2023

      As organizations begin seriously considering their post-pandemic futures, they now face the challenge of walking the tight rope between meeting growth objectives and building sustainable businesses. 

      Over the last year, promises of long-term sustainability agendas have more than tripled, with pledges of zero carbon and carbon neutrality abounding. While many organizations are turning to technology to meet these targets and solve environmental issues, enterprises need to ensure their IT does not become a part of the problem.  

      The era of sustainable tech is on the rise. Companies are now leveraging innovative, data-driven technology to simultaneously streamline operations, cut carbon emissions, and reduce their carbon footprint. Often perceived as a savior rather than a sinner, the production, use, and disposal of technology has an often-overlooked carbon footprint: an estimated 57.4 million tons of e-waste were generated worldwide in 2021 alone. The total is growing by an average of 2 Mt a year. 

      Accelerated by the fiercely dynamic and competitive markets, more organizations are embracing digital transformation across their business. As a result, demand for computing power and data storage is on the rise and so too is the energy required to produce and run them. Curating sustainable enterprise-wide digital systems will be crucial for any business trying to balance the objectives of sustainable growth post-pandemic.   

      But with so little awareness on the matter, the road to successful and sustainable IT needs a clear and rigorous roadmap. Our research has found crucial factors to consider when building and implementing sustainable IT strategies – let’s have a look through each one.  

      •  Understanding the task  

      When it comes to strategy, half of firms have defined an enterprise-wide sustainability approach, yet less than one in five (18%) have a comprehensive, sustainable IT strategy with well-defined goals and target timelines.  

      Before a clear and robust framework can be rolled out, organizations need to get clued in on what they are dealing with. Our research revealed an alarming gap in awareness regarding the overall environmental impact of IT, with fewer than half of executives (43%) globally aware of their organization’s IT footprint. Many are confused about the true impact of IT. Only 34%, for example, know that the production of mobiles and laptops has a higher carbon footprint than the usage of these devices over their lifetime.  

      Getting a clear understanding of the issue is the first critical stage for firms looking to develop sustainable strategies. Once the baselines and benchmarks of an enterprise’s environmental footprint have been marked out, organizations can then look to establish, implement, and monitor key performance indicators, targets, and frameworks. 

      •  Engaged and informed employees  

      Employees and leaders who are engaged with sustainability agendas drive greater progress. Even the most thorough strategies can come undone when those involved are not committed to the cause. Taking things one step further by developing a specialist sustainable IT team can provide streamlined purpose and coherence. Organizations must adopt the same mindset with employees as with consumers. People want to buy from companies with sustainable products and services. Likewise, employees want to work for such organizations. People are a critical component of sustainability transformation. Therefore, you must foster a culture that celebrates and promotes environmentalism, while trusting and empowering your people to contribute their own ideas. Those who have made sustainability a pillar of the organizational culture have seen greater progress.
      Our research found that 60% of organizations are adopting sustainability to align with the demands of potential employees.

      •  Sustainable software architecture 

      Sustainability needs to be at the very center of an organization’s business. While emissions and output need to be carefully scrutinized, developing a sustainable software architecture is an imperative.  

      Understanding the environmental consequences of software deployment and making decisions based on the carbon cost of infrastructure will ingrain sustainability into the foundations of an enterprise. Once the architecture is available, specific software modules within the structural design must be viewed from a sustainability perspective. For instance, organizations should empower their developers to understand the carbon cost of their software modules and use green coding to produce algorithms that have minimal energy consumption, at all times.

      Upskilling developers in circular design will help product and design teams lessen their waste and thus their environmental footprints.

      Sustainable IT can play a critical role in creating a circular economy by reducing waste, maximizing resource efficiency, and promoting more sustainable production and consumption practices. By introducing sustainability into the company’s value chain, you will drive the whole organization toward new efficiencies and a circular economy.

      Moving forward, sustainability must be at the core of all our efforts. While many organizations have begun to focus on their overall sustainability agenda, the critical issue of sustainable IT has been overlooked. To give sustainable IT the attention it deserves, organizations need to understand the carbon cost of our digital world and accelerate the move to sustainable systems with engaged and dedicated teams. In this way, sustainable IT can play a central part in tackling climate change, promoting a circular economy, driving innovation, and moving the world to a more resilient and sustainable future.  

      Meet the author

      Ygor Zeliviansky

      Head of Global Portfolio, Cloud Infrastructure Services, Capgemini 
      I am a Solutions Consultant with a demonstrated history of working in the information technology and services industry and have delivered business value for global clients in service delivery, enterprise software, HP products, enterprise architecture, and storage.

        The post Why sustainable IT is the backbone of a greener future appeared first on Capgemini New Zealand .

        ]]>
        https://www.capgemini.com/nz-en/2023/10/10/why-sustainable-it-is-the-backbone-of-a-greener-future/feed/ 0 630983
        Growing demand for Sustainable IT despite a lack of maturity https://www.capgemini.com/nz-en/2023/10/09/growing-demand-for-sustainable-it-despite-a-lack-of-maturity/ https://www.capgemini.com/nz-en/2023/10/09/growing-demand-for-sustainable-it-despite-a-lack-of-maturity/#respond Mon, 09 Oct 2023 13:01:09 +0000 https://www.capgemini.com/nz-en/?p=630913&preview=true&preview_id=630913 The post Growing demand for Sustainable IT despite a lack of maturity appeared first on Capgemini New Zealand .

        ]]>

        Growing demand for Sustainable IT
        despite a lack of maturity

        Philippe Roque
        15 Nov 2022

        Despite its recent popularity, sustainability continues to be an exclusive domain of SMEs.

        A plethora of jargons like carbon-neutral, net-zero, GWP, CO2ee etc often create an entry barrier which even confuse senior business leaders. But the message from COP27 is loud and clear, it is time for implementation and we all need to play our parts in reducing global CO2e emissions. But you cannot transform what you don’t understand or cannot measure. Like the ‘e’ after CO2 stands for “equivalent” which refer to other Green House Gases responsible for global warming.

        Instead of getting overwhelmed with jargons, it’s now time for IT leaders to demystify sustainability and act decisively on this critical issue. I would recommend not to see it as an emerging topic but as embedded into everyday actions- from baking a cake to commuting to the office. A practical way to start learning about global warming is by baselining individual CO2e emissions leveraging any online carbon calculator. See how this contributes to the overall ambition of Paris Agreement [1]. According to Emissions Gap Report[2], per capita CO2e emissions should be around 2tonnes per year to contain global warming within 1.5degree Celsius by 2050, that is roughly one round trip between Paris and New York by flight.

        Similarly, IT leaders must see sustainable IT as embedded into their existing IT ecosystem and begin their transformation journey with an accurate baseline of their enterprise IT carbon footprint.

        Demystifying three common myths around sustainable IT

        But before embarking on their sustainable IT transformation journey, IT leaders must be careful to avoid common myths surrounding this topic. One common myth is measuring the CO2e emissions of IT only during its use or run phase, while manufacturing also has a significant carbon footprint. Since user devices, networking, and data center equipment constitute a large share of enterprise IT’s carbon footprint, any sustainable IT transformation strategy will need to consider the impact of manufacturing too. And to close the loop it is important to have an end-of-life strategy for all IT devices and equipment. Global IT leaders must acknowledge that there could be a wide variance in CO2e emissions depending on many factors like location etc and you cannot fit one sustainability strategy to all.

        A second common myth is to treat cloud as a panacea for all of IT’s CO2e emissions. Although moving applications to the cloud has a potential to reduce IT’s carbon footprint, it also has a scope 3 impact based on the hyperscaler’s operational carbon footprint like electricity. Electricity generation needs energy that varies across countries and impacts its CO2e emissions commonly referred to as carbon intensity[1]. Hence, a simple move to cloud when the data center is in a country with higher carbon intensity might increase the CO2e emissions. There is no doubt that hyperscalers have been trying to optimize their carbon footprint and reporting, but today we still lack a clear view of carbon footprint on public cloud and especially the manufacturing CO2e emissions of servers etc. As Gartner® says, “Sustainability metrics and workload placement tools are still immature and not always transparent, making it difficult for organizations to fully and accurately assess true sustainability impacts of their cloud usage today”[2].

        Also while discussing cloud in the context of sustainability, I would advise more caution as it can have an infinite effect. I have seen many clients add much more into their cloud than they need, which in turn can increase the total CO2e emissions even though the emissions per workload on cloud may be lower. Hence, IT leaders should always have sobriety as a guiding principle to shape their sustainable IT strategy and transformation roadmap.

        A third myth I come across often is limiting ecodesign to only code optimization. In reality it also includes user devices, infrastructure, network on which the software is running and foremost it is about efficiently addressing the business needs.

        Sustainability is more than climate change

        I certainly believe that we need to do all we can to reduce CO2e emissions to the minimum before looking at offsetting, and also ensure compliance with regulations on sustainability reporting. But let’s also be aware that sustainability is a broader topic covering ecological degradation and depletion of abiotic resources like precious metals. It is important for any long-term sustainable IT vision to also consider its impact on other areas of sustainability beyond CO2e emissions.

        We see a strong demand for sustainable IT!

        We see a clear urgency in the market to act on this. Most clients want us to help them build awareness on this topic internally, baseline their IT carbon footprint, and set-up governance to track their CO2e emissions.  Industries where IT is a major contributor to their CO2e emissions are ahead in the curve compared to others, but even here sustainable IT is sometimes seen as a quick win while they transform their core business model. While cost was and continues to be a priority for CIOs, going forward we might see sustainability metrics like units of electricity also being used to measure the sustainability ROI from IT transformations. Would you agree? I would love to hear how you demystify sustainability myths for your teams and build a vision for your sustainable IT transformation.

        For more than 11years I am leading Capgemini’s unique eAPM capability around the key principle of helping our clients make fact-based decisions on their IT transformation journey. Most recently, we launched its sustainable IT module leveraging our proprietary eAPM studio powered by an AI engine. Using this we can model a 360° view of enterprise IT’s carbon footprint at an application level and identify key emission hotspots. With our unique benchmarks we are then able to recommend actionable levers that can accelerate CO2e emissions reduction. We are on a continuous innovation journey in this space along with our clients who are keen to accelerate their sustainable IT transformation journey. Would you like to embark on this journey with us? I would like to learn more about your sustainable IT plans. Please reach me at philippe.roques@capgemini.com or connect with me on LinkedIn.

        Visit our website to learn more about how we help our clients’ sustainable IT journey with eAPM. I would like to thank my colleagues Claire Egu & Joy Bhattacharjee for their valuable contributions to this article.

        [1] https://unfccc.int/process-and-meetings/the-paris-agreement/the-paris-agreement
        [2] https://www.unep.org/resources/emissions-gap-report-2021
        [3] https://ourworldindata.org/grapher/carbon-intensity-electricity
        ® Gartner Press Release, Gartner Predicts Hyperscalers’ Carbon Emissions Will Drive Cloud Purchase Decisions by 2025, January 24, 2022

        Author

        Philippe Roques

        Global Head and founder of eAPM, Executive Vice President Capgemini
        Philippe Roques is an Executive Vice president and the global leader and founder of Capgemini’s eAPM approach. Over the past 11 years, Philippe had incubated, nurtured, and developed eAPM to become one of the best approaches for CIOs to make sound data-driven decisions on the future of their enterprise IT. A future that is delivered at speed and scale of large enterprises through the transformation expertise of the Capgemini Group.

          The post Growing demand for Sustainable IT despite a lack of maturity appeared first on Capgemini New Zealand .

          ]]>
          https://www.capgemini.com/nz-en/2023/10/09/growing-demand-for-sustainable-it-despite-a-lack-of-maturity/feed/ 0 630913
          Deep stupidity – or why stupid is more likely to destroy the world than smart AI https://www.capgemini.com/nz-en/2023/06/28/deep-stupidity-or-why-stupid-is-more-likely-to-destroy-the-world-than-smart-ai/ Wed, 28 Jun 2023 07:03:21 +0000 https://www.capgemini.com/au-en/?p=511850&preview=true&preview_id=511850 The hype in AI is about whether a truly intelligent AI is an existential risk to society. Are we heading for Skynet or The Culture? What will the future bring?

          The post Deep stupidity – or why stupid is more likely to destroy the world than smart AI appeared first on Capgemini New Zealand .

          ]]>

          Deep stupidity – or why stupid is more likely to destroy the world than smart AI

          Steve Jones
          7 Jun 2023

          The hype in AI is about whether a truly intelligent AI is an existential risk to society. Are we heading for Skynet or The Culture? What will the future bring?

          I’d argue that the larger and more realistic threat is from Deep Stupidity — the weaponization of Artificial General Intelligence to amplify misinformation and create distrust in society.

          Social media is the platform, AI is the weapon

          One of the depressing things about the internet is how its made conspiracy theories spread. Where before people were lone idiots, potentially subscribing to some bizarre magazine or conspiracy society in a given area, you really didn’t have the ability to industrial scale these things. Social media and the Internet has increased the spread of such ideas. So while some AI folks talk about the existential threat of AGI, personally I’m much more concerned about Artificial General Stupidity.

          So I thought it is worth looking at why it is much easier to build an AI that is a flat earther than it is to build a High School physics teacher, let alone a Stephen Hawking.

          It is easier being confidently wrong and not understanding

          LLMs are confidently wrong, that inability to actually understand is a great advantage when being a conspiracy theorist. Because when you understand stuff, then conspiracy theories are dumb.

          This means the training data set for our AI conspiracy theorist must be incomplete, what we need is not something that has access to a broad set of data, but actually something that has access to an incredibly small and specific set of data that repeats the same point over and over again.

          To be a conspiracy theorist means denying evidence and ignoring contradictions, this is much easier to learn and code for than actually receiving new information that challenges your current model and altering it.

          Small data set for a single topic

          So this is a massive advantage for LLMs when trying to create a conspiracy theorist. What we need is a limited set of data that repeats a given conclusion and continually lines up all evidence to that conclusion. We can apply this to lots of conspiracy theorists out there, for instance those folks who scream “false flag” after every single mass shooting incident in the US, in other words we have a small set of data, possibly only a few hundred data points that always result in the same conclusion. This means for our custom trained conspiracy theorist the association it always knows is “what ever the data, the answer is the conspiracy”.

          Now we could get fancy and have a number of conspiracies, but given very few of them are logically consistent with each other, let alone with reality, it is more effective to have a model per conspiracy and just switch between them. That a conspiracy theorist is inconsistent with what they’ve previously said isn’t a problem, but we don’t want inconsistencies between conspiracies on a single topic. What we need to add are the standard “rebuttals of reality” like “Water finds its level”, “We don’t see the curve”, “NASA is fake” or “Spurs are a top Premier League club”.

          Hallucinations help

          This small set of data really helps us take advantage of the largest flaw in LLMs, hallucinations, or when the LLM just makes stuff up either because it has no data on the topic, or because the actual answer is rare so the weightings bias it towards an invalid answer. This is where LLMs really can scale conspiracy theories, because the probabilities are weighted towards the conspiracy theory already (as that is the only “correct” answer within the model) then any information we are provided with is recast within that context. So if someone tells us that the Greeks proved the earth was round in the 2nd Century BC our LLM could easily reply:

          Context makes hallucinations doubly annoying

          Our LLMs can go beyond the average conspiracy theorist thanks to the context and hallucinations. While an average conspiracy person will only have a fixed set of talking points, and potentially be constrained at some level by reality, the hallucinations and context of the conversation enables our conspiracy LLM to keep building its conspiracy and adding elements to it. Because our LLM is unconstrained by reality and counter arguments, instead being able to reframe any counter argument by using a hallucination it will be significantly more maddening. It will also mean it will create new justifications for the conspiracy that have never been put forwards before. These will, of course, be total nonsense but new total nonsense is mana from heaven to other conspiracy theorists.

          Reset and start again

          The final piece that makes a conspiracy LLM much easier to create is that if the LLM goes truly bonkers and you need to reset… this is exactly what conspiracy theorists do today. So if our LLM is creating hallucinations that fail some form of basic test, or just every 20 responses, we can reset the conversation in a totally different direction. Making my generative LLM detect either a frustration or an “ah ha” moment from the person it is annoying, a trivial task, enables me to then have my conspiracy bot just jump to another topic, and to do so in a much smoother way than most conspiracy theorists do today.

          This is a much smoother transition for a flat earth conspiracy than you’ll hear on TikTok or YouTube.

          We have achieved AGS, that isn’t a good thing

          I’ve argued that the current generation of AIs aren’t close to genuinely passing the Turing test, let alone more modern tests. Turing set the bar of intelligence as the CEO of a Fortune 50 company, and made it have awareness of what it didn’t know.

          Some folks are concerned about a coming existential crisis where Artificial General Intelligence becomes a threat to humanity.

          But for me that is assuming the current generation of technologies are not a threat, and that intelligence is a greater threat than weaponized stupidity. Many people in AI are in fact arguing that GPT passes the Turing test, not because it replicates an intelligent human, but because either it can pass a multiple choice or formulaic example, or because it can convince people they are speaking to a not very bright person.

          We can today make an AI that is the equivalent of a conspiracy theorist, someone untethered to reality and disconnected from logic. This isn’t General Intelligence, but it is General Stupidity.

          Deep fakes and deep stupidity

          Where Deep Fakes can make us not trust sources, Deep Stupidity can amplify misinformation and constantly give it justification and explanation. Where Deep Fakes imitate a person or event, Deep Stupidity can imitate the response of the crowd to that event. Spinning up a million conspiracy theorists to amplify not just the Deep Fake but the creation of an alternative reality around it.

          The internet and particularly social media has proven a fertile ground for human created stupidity and conspiracy theories. Entire political movements and groups have been created based on internet created nonsense. These have succeeded in gaining significant mindshare without having the capacity to really generate either convincing material or convincing narratives.

          AIs today have the ability to change that.

          Stupidity and misinformation are today’s existential threat

          We need to stop talking about the challenge with AI being only when it becomes “intelligent”, because it is already sufficiently stupid to have massive negative consequences on society. It is madness to think that companies, and especially governments, aren’t looking at this technologies and how they can use them to achieve their ends, even if their ends are simply to sew chaos.

          Stupidity is the foundation for worrying about intelligence

          Worrying about an AI ‘waking up’ and threatening humanity is a philosophical problem, but addressing Artificial Stupidity would give us the framework to deal with that future challenge. Everything about controlling and managing AI in future can be mapped to controlling and avoiding AGS today.

          When we talk about frameworks for Trusted AI and legislation on things like Ethical Data Sourcing these are elements that apply to General Stupidity just as much as to intelligence. So we should stop worrying simply about some amorphous future threat and instead start worrying about how we avoid, detect and control Artificial General Stupidity, because in doing that we lay the platform for controlling AI overall.

          This article first appeared on Medium.

          The post Deep stupidity – or why stupid is more likely to destroy the world than smart AI appeared first on Capgemini New Zealand .

          ]]>
          630284
          ChatGPT and I have trust issues https://www.capgemini.com/nz-en/2023/03/30/chatgpt-and-i-have-trust-issues/ Thu, 30 Mar 2023 13:04:00 +0000 https://www.capgemini.com/?p=913268 Whether we are ready for it or not, we are currently in the era of generative AI, with the explosion of generative models such as DALL-e, GPT-3, and, notably, ChatGPT, which racked up one million users in one day.

          The post ChatGPT and I have trust issues appeared first on Capgemini New Zealand .

          ]]>

          ChatGPT and I have trust issues

          Tijana Nikolic
          30 March 2023

          Disclaimer: This blog was NOT written by ChatGPT, but by a group of human data scientists: Shahryar MasoumiWouter ZirkzeeAlmira PillaySven Hendrikx and myself.

          Stable diffusion generated image with prompt = “an illustration of a human having trust issues with generative AI technology”

          Whether we are ready for it or not, we are currently in the era of generative AI, with the explosion of generative models such as DALL-eGPT-3, and, notably, ChatGPT, which racked up one million users in one day. Recently, on March 14th, 2023, OpenAI released GPT-4, which caused quite a stir and thousands of people lining up to try it.

          Generative AI can be used as a powerful resource to aid us in the most complex tasks. But like with any powerful innovation, there are some important questions to be asked… Can we really trust these AI models? How do we know if the data used in model training is representative, unbiased, and copyright safe? Are the safety constraints implemented robust enough? And most importantly, will AI replace the human workforce?

          These are tough questions that we need to keep in mind and address. In this blog, we will focus on generative AI models, their trustworthiness, and how we can mitigate the risks that come with using them in a business setting.

          Before we lay out our trust issues, let’s take a step back and explain what this new generative AI era means. Generative models are deep learning models that create new data. Their predecessors are Chatbots, VAE, GANs, and transformer-based NLP models, they hold an architecture that can fantasize about and create new data points based on the original data that was used to train them — and today, we can do this all based on just a text prompt!

          The evolution of generative AI, with 2022 and 2023 bringing about many more generative models.

          We can consider chatbots as the first generative models, but looking back we’ve come very far since then, with ChatGPT and DALL-e being easily accessible interfaces that everyone can use in their day-to-day. It is important to remember these are interfaces with generative pre-trained transformer (GPT) models under the hood.

          The widespread accessibility of these two models has brought about a boom in the open-source community where we see more and more models being published, in the hopes of making the technology more user-friendly and enabling more robust implementations.

          But let’s not get ahead of ourselves just yet — we will come back to this in our next blog. What’s that infamous Spiderman quote again?

          With great power…

          The generative AI era has so much potential in moving us closer to artificial general intelligence (AGI) because these models are trained on understanding language but can also perform on a wide variety of other tasks, that in some cases even exceed human capability. This makes them very powerful in many business applications.

          Starting with the most common — text application, which is fueled by GPT and GAN models. Including everything from text generation to summarization and personalized content creation, these can be used in educationhealthcare, marketing, and day-to-day life. The conversational application component of text application is used in chatbots and voice assistants.

          Next, code-based applications are fueled by the same models, with GitHub’s Co-pilot as the most notable example. Here we can use generative AI to complete our code, review it, fix bugs, refactor, and write code comments and documentation.

          On the topic of visual applications, we can use DALL-eStable Diffusion, and Midjourney. These models can be used to create new or improved visual material for marketing, education, and design. In the health sector, we can use these models for semantic translation, where semantic images are taken as input and a realistic visual output is generated. 3D shape generation with GANs is another interesting application in the video game industry. Finally, text-to-video editing with natural language is a novel and interesting application for the entertainment industry.

          GANs and sequence-to-sequence automatic speech recognition (ASR) models (such as Whisper) are used in audio applications. Their text-to-speech application can be used in education and marketing. Speech-to-speech conversion and music generation have advantages for the entertainment and video game industry, such as game character voice generation.

          Some applications of generative AI in industries.

          Although powerful, such models also come with societal limitations and risks, which are crucial to address. For example, generative models are susceptible to unexplainable or faulty behavior, often because the data can have a variety of flaws, such as poor quality, bias, or just straight-up wrong information.

          So, with great power indeed comes great responsibility… and a few trust issues

          If we take a closer look at the risks regarding ethics and fairness in generative models, we can distinguish multiple categories of risk.

          The first major risk is bias, which can occur in different settings. An example of bias is the use of stereotypes such as race, gender, or sexuality. This can lead to discrimination and unjust or oppressive answers generated from the model. Another form of bias is the model’s word choice. Its answers should be formulated without toxic or vulgar content, and slurs.

          One example of a language model that learned a wrong bias is Tay, a Twitter bot developed by Microsoft in 2016. Tay was created to learn, by actively engaging with other Twitter users by answering, retweeting, or liking their posts. Through these interactions, the model swiftly learned wrong, racist, and unethical information, which it included in its own Twitter posts. This led to the shutdown of Tay, less than 24 hours after its initial release.

          Large language models (LLMs) like ChatGPT generate the most relevant answer based on the constraints, but it is not always 100% correct and can contain false information. Currently, such models provide their answers written as confident statements, which can be misleading as they may not be correct. Such events where a model confidently makes inaccurate statements are also called hallucinations.

          In 2023, Microsoft released a GPT-backed model to empower their Bing search engine with chat capabilities. However, there have already been multiple reports of undesirable behavior by this new service. It has threatened users with legal consequences or exposed their personal information. In another situation, it tried to convince a tech reporter he was not happily married and that he was in love with the chatbot (it also proclaimed their love for the reporter) and consequently should leave his wife (you see why we have trust issues now?!).

          Generative models are trained on large corpora of data, which in many cases, is scraped from the internet. This data can contain private information, causing a privacy risk as it can unintentionally be learned and memorized by the model. This private data not only contain people, but also project documents, code bases, and works of art. When using medical models to diagnose a patient, it could also include private patient data. This also ties into copyright when this private memorized data is used in a generated output. For example, there have even been cases where image diffusion models have included slightly altered signatures or watermarks it has learned from their training set.

          The public can also maliciously use generative models to harm/cheat others. This risk is linked with the other mentioned risks, except that it is intentional. Generative models can easily be used to create entirely new content with (purposefully) incorrect, private, or stolen information. Scarily, it doesn’t take much effort to flood the internet with maliciously generated content.

          Building trust takes time…and tests

          To mitigate these risks, we need to ensure the models are reliable and transparent through testing. Testing of AI models comes with some nuances when compared to testing of software, and they need to be addressed in an MLOps setting with data, model, and system tests.

          These tests are captured in a test strategy at the very start of the project (problem formulation). In this early stage, it is important to capture key performance indicators (KPIs) to ensure a robust implementation. Next to that, assessing the impact of the model on the user and society is a crucial step in this phase. Based on the assessment, user subpopulation KPIs are collected and measured against, in addition to the performance KPIs.

          An example of a subpopulation KPI is model accuracy on a specific user segment, which needs to be measured on data, model, and system levels. There are open-source packages that we can use to do this, like the AI Fairness 360 package.

          Data testing can be used to address bias, privacy, and false information (consistency) trust issues. We make sure these are mitigated through exploratory data analysis (EDA), with assessments on bias, consistency, and toxicity of the data sources.

          The data bias mitigation methods vary depending on the data used for training (images, text, audio, tabular), but they boil down to re-weighting the features of the minority group, oversampling the minority group, or under-sampling the majority group.

          These changes need to be documented and reproducible, which is done with the help of data version control (DVC). DVC allows us to commit versions of data, parameters, and models in the same way “traditional” version control tools such as git do.

          Model testing focuses on model performance metrics, which are assessed through training iterations with validated training data from previous tests. These need to be reproducible and saved with model versions. We can support this through open MLOPs packages like MLFlow.

          Next, model robustness tests like metamorphic and adversarial tests should be implemented. These tests help assess if the model performs well on independent test scenarios. The usability of the model is assessed through user acceptance tests (UAT). Lags in the pipeline, false information, and interpretability of the prediction are measured on this level.

          In terms of ChatGPT, a UAT could be constructed around assessing if the answer to the prompt is according to the user’s expectation. In addition, the explainability aspect is added — does the model provide sources used to generate the expected response?

          System testing is extremely important to mitigate malicious use and false information risks. Malicious use needs to be assessed in the first phase and system tests are constructed based on that. Constraints in the model are then programmed.

          OpenAI is aware of possible malicious uses of ChatGPT and have incorporated safety as part of their strategy. They have described how they try to mitigate some of these risks and limitations. In a system test, these constraints are validated on real-life scenarios, as opposed to controlled environments used in previous tests.

          Let’s not forget about model and data drift. These are monitored, and retraining mechanisms can be set up to ensure the model stays relevant over time. Finally, the human-in-the-loop (HIL) method is also used to provide feedback to an online model.

          ChatGPT and Bard (Google’s chatbot) have the possibility of human feedback through a thumbs up/down. Though simple, this feedback is used to effectively retrain and align the underlying models to users’ expectations, providing more relevant feedback in future iterations.

          To trust or not to trust?

          Just like the internet, truth and facts are not always given — and we’ve seen (and will continue to see) instances where ChatGPT and other generative AI models get it wrong. While it is a powerful tool, and we completely understand the hype, there will always be some risk. It should be standard practice to implement risk and quality control techniques to minimize the risks as much as possible. And we do see this happening in practice — OpenAI has been transparent about the limitations of their models, how they have tested them, and the governance that has been set up. Google also has responsible AI principles that they have abided by when developing Bard. As both organizations release new and improved models — they also advance their testing controls to continuously improve quality, safety, and user-friendliness.

          Perhaps we can argue that using generative AI models like ChatGPT doesn’t necessarily leave us vulnerable to misinformation, but more familiar with how AI works and its limitations. Overall, the future of generative AI is bright and will continue to revolutionize the industry if we can trust it. And as we know, trust is an ongoing process…

          In the next part of our Trustworthy Generative AI series, we will explore testing LLMs (bring your techie hat) and how quality LLM solutions lead to trust, which in turn, will increase adoption among businesses and the public.

          This article first appeared on SogetiLabs blog.

          The post ChatGPT and I have trust issues appeared first on Capgemini New Zealand .

          ]]>
          630294
          Hello world! https://www.capgemini.com/nz-en/2022/08/23/hello-world/ https://www.capgemini.com/nz-en/2022/08/23/hello-world/#comments Tue, 23 Aug 2022 12:22:20 +0000 https://www.capgemini.com/nz-en/?p=1 Welcome to www.capgemini.com. This is your first post. Edit or delete it, then start writing!

          The post Hello world! appeared first on Capgemini New Zealand .

          ]]>
          Welcome to www.capgemini.com. This is your first post. Edit or delete it, then start writing!

          The post Hello world! appeared first on Capgemini New Zealand .

          ]]>
          https://www.capgemini.com/nz-en/2022/08/23/hello-world/feed/ 1 1
          Happy Birthday Capgemini Paris 5G Lab! https://www.capgemini.com/nz-en/2022/04/14/happy-birthday-capgemini-paris-5g-lab/ https://www.capgemini.com/nz-en/2022/04/14/happy-birthday-capgemini-paris-5g-lab/#respond Thu, 14 Apr 2022 12:19:00 +0000 https://www.capgemini.com/?p=700687 The post Happy Birthday Capgemini Paris 5G Lab! appeared first on Capgemini New Zealand .

          ]]>

          Happy Birthday Capgemini Paris 5G Lab!

          Cédric Bourrely
          14 April 2022
          capgemini-invent

          It has been one year already since we launched our first 5G Lab dedicated to the Industries. And what a challenging and passionate year working on a breakthrough topic with this incredible team!

          By Cédric Bourrely, 5G Labs Program Director, Capgemini Invent

          The tremendous efforts and the pandemic constraints haven’t reduced the dynamism and the willingness to innovate of the people working on the program.

          One year later, time for a first summary. Let me share some insights on this epic run.

          So what have we done since the launch of the Paris Lab last year?

          Well first, we presented the Lab to our clients: more than 40 CxO meetings in multiple sectors (Manufacturing, Automotive, Defense, Transportation, Retail, Telecom…) on site in Paris, where we could measure the enthusiasm of people talking again together about innovation, projecting themselves in a post-Covid future. Such excitement!

          And not just clients, but also our technology partners. This year also allowed us to improve our portfolio of partners, sharing latest innovation trends and solutions to continuously improve our setup and market understanding.

          We also had the opportunity to present the Lab to the French “5G for Industry” government taskforce, explaining our challenges, our successes and most of all, our lessons learned after two years of “hands-on” 5G. For sure a great honor to be able to be part of such a strategic initiative, with international synergies.

          Talking about international, well, 5G Labs at MWC a few weeks back was also a highlight: this gave the 5G Labs significant international exposure to present our setup, our knowledge, our vision of the Industries transformation powered by “The Future of Technology”.

          Clients visits, for sure, have been the drumbeat of this past year, but we also focused on continuous improvement, testing new components, troubleshooting new features, and on-boarding new partners. Among our successes:

          • We complemented our Private Network setup with access to the 5G public network of Bouygues Telecom. The Labs is now completely able to provide both Private Stand Alone 5G (Release 16) as well as Public Non-Stand Alone 5G, and benchmark performances and usages.
          • We finalized the complete virtualization of the 5G Core & Multi-Access Edge Computing Platform, allowing us to gain in agility and operations efficiency.
          • We enriched the platform integration patterns with the establishment of dedicated connectivity with our Hyperscalers partners.
          • We integrated more devices to our setup, each time performing intense troubleshooting sessions to ensure compatibility with available spectrums and networks.
          • We started the implementation of breakthrough / long-awaited 5G features such as Dynamic Slicing
          • We developed new High-potential Industrial use cases, including Autonomous Intra-Logistics Vehicle for smart warehousing. We definitely made significant progress on technology and market understanding, while delivery value to key clients

          Our key learnings

          One year ago, when the lab was launched and made available for client visits, we already had significant experience in architecture design, modules implementations, and partners selection. We came into this project with solid knowledge. But that pales compared to where we stand now. The clients’ feedbacks, the experimentation, the connection to the 5G (& Edge) ecosystems pushed further our understanding of the market, and the key success factors to address while going for 5G. Here are some of the major findings:

          First, starting the 5G journey relies on knowledge and innovation:

          • The level of 5G-awareness is very heterogeneous within industry clients, but still, most of our clients understand that 5G is not just an extension of 4G, and that the new architecture patterns (link with Edge, link with Cloud) will make a difference.
          • And yet, finding use cases that can be truly accelerated by 5G (and Edge) vs existing connectivity solutions (4G, WIFI 6) is one of the most important topics to tackle for industrials…
          • … but the approaches taken vary from one geography to another. As a primary example, French industrial clients tend to prove ROI before investing into 5G, whereas German neighbors consider 5G as a foundation enabler among others for industrial innovation.

          Second, there are some key operational questions to answer, in order to maximize the chances of adoption:

          • Strategic choices: 5G is all about new architectures, complex and diverse ecosystems, implying new skills within the teams. Make / Buy / Partner strategic choice needs to be thought through early in the process.
          • Data privacy is one of the major concerns for industrial risks control or sensitive data handling (ex: military or health), and therefore 5G Private Networks are expected to provide significant benefits.
          • While the questions about public health no longer seem to be a major barrier, the impacts on sustainability still raise frequent questions. On this sensitive topic, while we know that the reduction of energy consumption by the antennas if one of the key 5G benefits, the overall balance between end-to-end carbon footprint of 5G and energy savings is still under evaluation and one of the major topics of interest for our clients. And finally, obviously, major topics linked to technologies evaluation and integration must be addressed
          • Most of our clients are convinced that Private/Hybrid networks will be of an interest for their industrial activities. Deployment models are key to tackling this topic.
          • Edge / Cloud / 5G represent a new paradigm shift as IT and Network technologies converge, creating deep need for architecture pattern creation, strategic partners alliances or resources upskilling.
          • 5G performances have been in the heart of the discussions, especially when we compare them to the original marketing promises. A clear path must be evaluated to find the 5G “sweet spot” for use cases. Big Bang revolution of 5G deployment won’t be applicable in most “brownfield” industrial operations and a tactical approach has to be set up.
          • As the ecosystem is deeply fragmented (cloud, devices makers, telco…), multiple technical options can be envisioned, taking into account local specificities and barriers to overcome. Finding the right alliances strategy is one major, time-consuming activity for which pre-integrated patterns and return on experiences are very valuable.
          • And finally, Cybersecurity. 5G has advantages, but also creates risks that must be anticipated. Key question relate to the global integration of 5G within corporate IT, in a global context where cyber threats are constantly

          What to expect for 2022?

          These great achievements and learnings will be enriched by our upcoming 2022 work.

          We will keep evolving our setup, continuing the exchange of great innovations and knowledge with our talented Mumbai Lab Team: new radio equipment, new features (Mission Critical MCX), new point of views (sustainability, cybersecurity…), and obviously, more fun and innovation (hello Metaverse, we’ll see each other soon…)

          In conclusion, I must say that I am very glad and deeply honored to be in charge of this program.

          These great achievements, past and upcoming, wouldn’t be possible without the effort of these incredible, passionate, super-skilled team members in France and India, and without the support of our visionary Top Management within Capgemini, still one step ahead of the upcoming digital innovations.

          Stay tuned for more!

          Author:

          Cédric BOURRELY – 5G Labs Program Lead, Capgemini Invent

          The post Happy Birthday Capgemini Paris 5G Lab! appeared first on Capgemini New Zealand .

          ]]>
          https://www.capgemini.com/nz-en/2022/04/14/happy-birthday-capgemini-paris-5g-lab/feed/ 0 629421
          A case for context awareness in AI https://www.capgemini.com/nz-en/2022/04/04/a-case-for-context-awareness-in-ai/ https://www.capgemini.com/nz-en/2022/04/04/a-case-for-context-awareness-in-ai/#respond Mon, 04 Apr 2022 07:15:00 +0000 https://www.capgemini.com/?p=676077 The post A case for context awareness in AI appeared first on Capgemini New Zealand .

          ]]>

          A case for context awareness in AI

          Robert Engels
          04 Apr 2022

          There have been catastrophic effects of AI use in self-driving cars, including road crashes, social media, and failures in critical infrastructures, making some ask: can we trust AI in production?

          Also, what can we do to make AI more robust while operating in dynamic surroundings and, most importantly, how can we make AI understand the real world?

          Does applied AI have the necessary insights to tackle even the slightest (unlearned or unseen) change in context of the world surrounding it? In discussions, AI often equals deep-learning models. Current deep-learning methods heavily depend on the presumption of “independent and identically distributed” data to learn from, something which has serious implications for the robustness and transferability of models. Despite very good results on classification tasks, regression, and pattern encoding, current deep-learning methods are failing to tackle the difficult and open problem of generalization and abstraction across problems. Both are prerequisites for general learning and explanation capabilities.

          There is great optimism that deep-learning algorithms, as a specific type of neural network, will be able to close in on “real AI” if only it is further developed and scaled up enough (Yoshua Bengio, 2018). Others feel that current AI approaches are merely a smart encoding of a general distribution into a deep-learning networks’ parameters, and regard it as insufficient to act independently within the real world. So, where are the real intelligent behaviors, as in the ability to recognize problems and plan for solving them and understand the physics, logic, causality, and analogy?

          “THERE IS A NEED FOR CONTEXTUAL KNOWLEDGE IN ORDER TO MAKE APPLIED AI MODELS TRUSTABLE AND ROBUST IN CHANGING ENVIRONMENTS.”

          Understanding the real world

          What is needed is a better understanding by machines of their context, as in the surrounding world and its inner workings. Only then can machines capture, interpret, and act upon previously unseen situations. This will require the following:

          • Understanding of logical constructs as causality (as opposed to correlation). If it rains, you put on a raincoat, but putting on a raincoat does not stop the rain. Current ML struggles to learn causality. Being able to represent and model causality will to a large extent facilitate better explanations and understanding of decisions made by ML models.
          • The ability to tackle counterfactuals, such as “if a crane has no counterweight, it will topple over.”
          • Transferability of learned “knowledge” across/between domains; current transfer learning only works on small tasks with large domain overlap between them, which means similar tasks in similar domains.
          • Withstand adverse attacks. Only small random changes made in the input data (deliberately or not) can make the results of connectionist models highly unreliable. Abstraction mechanisms might be a solution to this issue.
          • Reasoning on possible outcomes, finding problematic outcomes and
            a) plan for avoiding them while reaching the goal
            or b) if that is not possible, find alternative goals and try to reach those.

          In the first edition of this review, we already made the case for extending the context in which AI models are operating, using a specific type of model that can benefit from domain knowledge in the form of knowledge graphs. From the above, it follows that knowledge alone probably will not be enough. Higher-level abstraction and reasoning capabilities are also needed. Current approaches aim at combining “connectionist” approaches with logical theory.

          1. Some connectionists feel that abstraction capability will follow automatically from scaling up networks, adding computing power, and using more data. But it seems that deep-learning models cannot abstract or generalize more than learning general distributions. The output will at the most be a better encoding but still not deliver symbolic abstraction, causality, or showing reasoning capabilities.
          2. Symbolic AI advocates concepts as abstracted symbols, logic, and reasoning. Symbolic methods allow for learning and understanding humanmade social constructs like law, jurisprudence, country, state, religion, and culture. Could connectionist methods be “symbolized” to provide the capabilities as mentioned above?
          3. Several innovative directions can be found in trying to merge methods into hybrid approaches consisting of multiple layers or capabilities.
          • Intuition layer: Let deep-learning algorithms take care of the low-level modeling of intuition or tacit skills shown by people having performed tasks over a long time, like a good welder who can hardly explain how she makes the perfect weld after years of experience.
          • Rationality layer: The skill-based learning where explicit learning by conveying rules and symbols to a “learner” plays a role, as in a child told by her mother not to get too close to the edge. A single example, not even experienced, might be enough to learn for life. Assimilating such explicit knowledge can steer and guide execution cycles which, “through acting,” can create “tacit skills” within a different execution domain as the original layer.
          • Logical layer: Logics to represent causality, analogy, and providing explanations
          • Planning and problem-solving layer: A problem is understood, a final goal is defined, and the problem is divided into sub-domains/problems which lead to a chain of ordered tasks to be executed, monitored (with intuition and rationality), and adapted.

           In general, ML models that incorporate or learn structural knowledge of an environment have been shown to be more efficient and generalize better. Some great examples of applications are not difficult to find, with the Neuro-Symbolic AI by MIT-IBM Watson lab as a good demonstration of how hybrid approaches (like NSQA in this case) can be utilized for learning in a connectionist way while preserving and utilizing the benefits of full-order logics in enhanced query answering in knowledge-intensive domains like medicine. The NSQA system allows for complex query-answering, learns along, and understands relations and causality while being able to explain results.

          The latest developments in applied AI show that we get far by learning from observations and empirical data, but there is a need for contextual knowledge in order to make applied AI models trustable and robust in changing environments.

          INNOVATION TAKEAWAYS

          HYBRID APPROACHES are needed to model and use causality, counterfactual thinking, problem solving, and structural knowledge of context.
          NEURAL-SYMBOLIC PROCESSING combines the benefits of connectionist and symbolic approaches to solve issues of trust, proof, and explainability.
          CONTEXTUAL KNOWLEDGE AI needs modeling more of the world to be able to understand the physics and logic, causality, and analogy in the surrounding world.

          Interesting read?

          Data-powered Innovation Review | Wave 3 features 15 such articles crafted by leading Capgemini and partner experts in data, sharing their life-long experience and vision in innovation. In addition, several articles are in collaboration with key technology partners such as Google, Snowflake, Informatica, Altair, A21 Labs, and Zelros to reimagine what’s possible. 

          The post A case for context awareness in AI appeared first on Capgemini New Zealand .

          ]]>
          https://www.capgemini.com/nz-en/2022/04/04/a-case-for-context-awareness-in-ai/feed/ 0 630289