🏆
Verified Scholarly Platform
Serving Researchers Since 2012

Fundamentals of Artificial Intelligence: A Review

DOI : 10.17577/IJERTV15IS031420
Download Full-Text PDF Cite this Publication

Text Only Version

Fundamentals of Artificial Intelligence: A Review

Dr. Sandeepak Bhandari

Department of Informatics and Biotechnology Klaipeda State College, Lithuania

Sigute Ezerskiene

Department of Informatics and Biotechnology Klaipeda State College, Lithuania

Abstract From last two to three decades, Artificial Intelligence has gained so much attention from government, industry and academia. The technological and industrial revolution is accelerating by the widespread application of new generation information and communication technologies, namely AI. In this research article, a systematic literature review (SLR) is performed and discusses the artificial intelligence concept. It includes numerous definitions of AI by different researchers, a comprehensive life cycle for the design, development, and deployment of artificial intelligence (AI), distinct types and techniques of AI along with its challenges. The literature studies is perform from 82 research articles via well-known databases such as Scopus and Web of Science, Google Scholar, IEEE Explorer and. Science Direct. The outcome of this article can play a role in AI related research and should provide important insights for new entrants to the artificial intelligence area.

Keywords artificial intelligence, CDAC AI life cycle, machine learning, deep learning

  1. ‌Introduction

    The concept of "Artificial Intelligence" was first introduced by John McCarthy in 1955 as part of a proposal for the Dartmouth Summer Research Project on Artificial Intelligence, which he co-authored alongside Marvin L. Minsky, Nathaniel Rochester, and Claude E. Shannon [3]. AI is characterized by systems that exhibit intelligent behavior through environmental analysis and autonomous actions aimed at achieving specific objectives [1]. The discipline encompasses a diverse array of methodologies and tasks; some have been effectively solved through software applications, while others still necessitate extensive research efforts. AI is crucial for societal advancement, yielding transformative outcomes in enhancing labor productivity, lowering labor expenses, optimizing human resource structures, and generating new employment opportunities. The rest of the paper is structured as follows: distnict explanations of AI by various researchers, CDAC AI life cycle and types of AI is addressed in Section II. Section III describes AI techniques. Section IV illustrates the AI challenges and Section VI concludes the paper.

  2. ‌Literature Studies

    1. ‌Understanding the Artificial Intelligence

      Artificial intelligences integration in community requires to understand what is an artificial intelligence, in what ways has technology advanced and now, what is the current state of artificial intelligence. In [4], authors mentioned and discuss that defining the artificial intelligence is not easy as there is no commonly recognized definition of AI.

      Various different definitions of artificial intelligence are used by different authors in their research work which can easily lead to confusion. The sheer diversity of definitions of

      AI is not because of carelessness but inherent in the phenomenon of AI itself. So, it is very significant to illustrate the concept of artificial intelligence appropriately. For this, a systematic literature review is performed to discussed and presented the numerous distinct definitions of AI from different authors.

      The generous definition of AI is defined with algorithms. Algorithms predate AI and and have been broadly used outside the field of AI. The term algorithm is derived from the name of the ninth-century Persian mathematician Mohammed ibn Musa al-Kharizmi and algorithm defined as particular set of instructions to solve the particular problem or or performing a calculation. So, if AI will be defined as the use of algorithms, it would contains various other activities, for instance operations of a pocket calculator or even the instructions in a cookbook. In the strictest definition of AI, it defined as imitation by computers of the intelligence inherent in humans. Purists point out that various present applications are still relatively simple and therefore not true artificial intelligence. This compose this definition inaccurate too, the usage of this definition would imply that Artificial intelligence does not exist at present.

      A popular definition of artificial intelligence is a technology that allows machines to mimic a variety of complex human skills. However, this does not provide much information. In actuality, it merely

      uses different terminology to refer to "artificial intelligence.". The precise nature of artificial intelligence is still unknown as long as those "complex human skills" are not defined. The definition of artificial intelligence (AI) as the ability of computers to perform complex tasks in complex environments is also applicable.

      Other interpretations delve deeper into the explanation of these abilities and duties. For instance, computer scientist Nils John Nilsson characterizes a technology that operates suitably and with foresight in its environment [5]. In addition, some individuals refer to the capacity to sense, pursue objectives, initiate actions, and learn through a feedback mechanism [6].

      A comparable explanation has been provided by the High- Level Expert Group on Artificial Intelligence (AI HLEG) of the European Commission (EC): Systems that exhibit intelligent behavior by analyzing their surroundings and taking actionswith a certain level of autonomyto achieve specific goals [2]. These task-oriented definitions enhance our comprehension of what constitutes AI; however, they are not without their drawbacks. Phrases like a certain level of autonomy remain rather ambiguous. Furthermore, these descriptions tend to be excessively broad since they encompass phenomena that many would not typically categorize as AI. For example, Nilssons description is also applicable to a

      traditional thermostat. This device can perceive (measure room temperature), pursue goals (maintain the set temperature), initiate actions (adjust the thermostat), and learn from feedback (cease operation once the desired temperature is attained). Nonetheless, most individuals would not consider a thermostat to be an example of AI.

      AI is so hard to define precisely is not surprising. After all, it is a simulation or imitation of human intelligence, which we still do not fully comprehend [7, 8]. Psychologists, behavioral scientists, and neurologists, among others, have long studied this. Although our understanding of intelligence and the human brain is extensive, there is still disagreement over the precise definition of human intelligence. It is impossible to pinpoint exactly how that intelligence can be artificially replicated [9]. In the research paper [10], authors mentioned that modern technology does not provides us intelligence but only one of its components.

      Going a step further, philosopher Daniel Dennett says we shouldn't even model AI on humans. Instead of being artificial people, these are a brand-new kind of entity and one he compares with oracles: entities that can make predictions but, in contrast to humans, lack a personality, conscience, or feelings [11]. Stated differently, artificial intelligence (AI) does something different from what humans do. The question "Do submarines swim?" was used by Edsger Dijkstra to demonstrate this [12]. Though it would be incorrect to refer to what these boats do as swimming, it is comparabl. Artificial intelligence (AI) is capable of performing tasks that appear to be intelligent but are actually quite different.

      Lastly, the newest technology is frequently associated with artificial intelligence. AI has become extremely popular in the last few years. Progress in one particular area of the field, "machine learning" (ML), has been one of the main drivers of this. Innovation in this area has led to what is now known as "deep learning" (DL). This technology is responsible for recent innovations like computers that can play games like Go and recognize faces. In contrast to more conventional methods, which use fixed rules applied by computer systems, machine learning and deep learning algorithms are able to identify patterns in data. In this context, we also refer to "self-learning algorithms.". When many people discuss AI these days, they are really talking about these algorithms, and frequently DL in particular. It is crucial to concentrate on this technology because it raises a number of urgent AI-related issues, including explainability issues.

      In the [1], authors have settled on an open definition of AI. Two factors are relevant in this respect. Firstly, it would be inappropriate to restraint the definition of AI to a specific part of the technology. For instance, we were to confine ourselves to deep learning as discussed above, we would ignore the fact that many current issues also play a role in other AI domains, such as logical systems. Secondly, because of the nature of this scientific field, as was previously mentioned, the definition of artificial intelligence will inevitably evolve over time.

      Instead of viewing AI as a clearly defined discipline with simple definitions and established methodologies, it is more beneficial to regard it as a multifaceted and intricate field oriented towards a specific objective. This objective is represented by the aspiration to comprehend and replicate all

      human cognitive abilities. This aim is often referred to as artificial general intelligence (AGI), with alternative terms including strong AI and full AI. Nonetheless, it remains uncertain if this goal, characterized by such a broad definition of AI, will ever be achieved. The majority of specialists believe that we are still several decades away from its realizationif it can be realized at all [13].

      The authors [1] outline AI as encompassing the entire spectrum of applications that are currently being implemented or are anticipated in the near future. The definition provided by the AI HLEG allows for considerable flexibility in scope. By characterizing AI as systems that exhibit intelligent behavior through environmental analysis and action-takingwith a degree of autonomyto fulfill specific objectives, this definition covers all applications presently categorized as AI while also accommodating potential future modifications to that classification. In addition to advanced machine learning and deep learning technologies, this definition includes other methodologies, such as traditional approaches utilized by various governmental organizations. In summary, this definition is precise enough to differentiate AI from algorithms and digital technologies broadly while remaining sufficiently adaptable to incorporate future innovations.

    2. ‌CDAC Artificial Intelligence life cycle

      The next significant steps to understand the artificial intelligence is to grasp how the AI functioning. In the sytamatic literature studies, numerous sources presented and discussed the different number of stages to demonstrate the working of artificial intelligence. In this research article, the CDAC AI life cycle, a comprehensive life cycle for the design, development, and deployment of artificial intelligence (AI) systems and solutions is considered and discussed [15]. In addition to focusing on the difficulties of risk analysis of AI adoption, transferability of prebuilt models, the growing significance of ethics and governance, and the makeup, expertise, and knowledge of an AI team necessary for successful completion, it fills the gap in a practical and inclusive approach that goes beyond the technical constructs.

      The CDAC AI life cycle,is presented as the progression of an AI solution through its distinct phases namely design, develop, and deploy and 19 constituent stages from conception to production as applicable to any AI initiative as shown in Fig. 1.

      As shown in the same figure, each phase calls for a distinct set of human skills, such as design (AI/data scientist), development (AI/ML scientist), and deployment (AI/ML engineer).

      A senior position with several years of experience is usually held by the AI/data scientist assigned to the design phase. They should be able to identify the issue and then come up with a solution based on the body of knowledge already in existence as well as their prior experience working on a variety of AI projects. When they deliver a prescriptive problem formulation, solution description, and representative data to the AI/ML scientist in charge of the develop phase, they should also be able to identify the representative data, required data, and available data by working through the first five stages of the life cycle. A junior position that is more technical and less conceptual, this AI/ML scientist usually has

      extensive technical knowledge of AI algorithms, model development, and evaluation. The next seven steps will be

      1. Identify & formulate the problem

      2. Review data &AI ethics

      3. Review technical literature on AI algorithms application & pre trained

      4. Data preparation

        1. Design (AI/Data Scientist)

      5. Data exploration

      6. External data acquistion

      7. Data pre-processing

    8 Build initial AI model

    1. Develop (AI/ML Scientist)

      1. Data agumentation

      2. Develop a benchmark

      3. Build mutiple AI models

      4. Evaluate primary metrics

    2. Deploy (AI.ML Enginer)

    1. AI model explainability

    2. Evaluate secondary metrics

    3. AI model deployment & risk assessment

    4. Post-deployment review

    5. Operationalise using AI pipelines

    6. Hyperautomation of process & systems

    7. Monitor & Evaluate the performance

    Fig. 1. CDAC Artificial Intelligence Cycle

    used to turn the problem formulation into a prototype AI model. This prototype AI model is then further transformed by an AI/ML engineer during the deploy phase into a deployed service or solution that is standardized for access by all stakeholders and end users. Usually, the AI/ML engineer comes from the DevOps field. Although the skills needed for this phase are common, they must be combined with expertise in the subtleties of implementing AI models, as DevOps is a well-established practice in software development and IT operations. The AI/ML engineer will complete the last seven steps to produce an AI solution that is a component of a bigger process and can be automatically assessed for accuracy and quality using a number of metrics. It may be necessary to hire or contract for several of these positions, depending on the project's size, scope, and scale. Additional enabling roles that contribute to the AI endeavor's value, inclusivity, and quality include an ethicist (or ethics committee), a project manager, a group of domain experts, a participatory design group, a pilot study cohot, legal counsel on intellectual property law, and a steering/advisory committee with complete oversight. The type of project, project timelines, organizational data maturity, and AI expertise all affect how all 19 stages are carried out. To successfully complete an AI project, all stages must be carefully considered and formally documented, even if only a portion of them are completed.

    1. ‌Identify and formulate the problem

      Finding, clarification, and formulating the problem in terms of its typology, environment (or setting), expected goal, people (or stakeholders), systems, processes, and data is the first step in the life cycle. Typology can be broadly categorized as research, tactical, strategic, and operational.

    2. ‌Review data and AI ethics

      This step involves cross-examining and validating the problem formulation, chosen approach, possible solution,

      and necessary data sets for ethical and legal compliance as well as potential security risks (referring back to the preliminary risk assessment). Since the AI team has a very technical skill set, professional ethicists with the necessary theoretical and practical experience must conduct this ethics review.

    3. ‌Review technical literature on AI algorithms, applications, & pre-trained models

      The problem formulation offers the background information required to investigate and evaluate published studies, implemented systems, solutions, and libraries that have been used in comparable contexts [39]. Some of the more popular sources of information include research article search engines (e.g. A. Google Scholar), publications (e.g. A. medium), Q&A websites (e.g. A g. Stack Exchange), code archives (e.g. A g. cloud platform providers (e.g., GitHub).

      A. AWS, GCP, Azure), and social media (e.g. A. (Twitter). Literature reviews, commentary, letters, articles, op-eds, case studies, best practices, product/tool documentation, tutorials, demonstrations, API documentation, and answers, upvotes, and likes on Q&A forums are just a few of the many types of resources available. Furthermore, rather than creating a new AI model from scratch, recent advancements in open-access pre- trained models like AlexNet [16], ResNet [17], BERT[18], and

      GPT [19] should be examined to learn and investigate how they can be repurposed, retrained, or fine-tuned.

    4. ‌Data preparation

      Before developing AI models, it is advised to design and create a unified data repository, such as a datawarehouse [20, 21] datalake [22], or data lakehouse [23, 24, 39], which centralizes data access, ownership, stewardship, metadata, data ethics, governance, and regulations.

    5. ‌Data exploration

      Unlike the previous stage, where the focus was on the data structure, this stage is more concerned with the actual data. Usually, the stage starts with a comparison with algorithmic baselines and industry benchmarks that have been documented in the literature and where comparable issues have been solved. Real data is added to the data structure created during the preparation stage using methods like data visualization, correlation analysis, data granularity alignment, examining the connections between data points and attributes, managing outliers, and implementing data quality checks.

    6. ‌External data acquisition

      When combined, the stages of data preparation and exploration can reveal data limitations that render the development of AI models impractical. In this case, it is important to look into possibilities for obtaining data from outside sources. One common short-term tactic is to obtain the necessary data in aggregate or detailed form from data brokers and data vendors. These brokers and vendors gather vast amounts of data by using public records, credit scoring services, social media content, and third-party data sharing agreements. These data are then marketed as aggregate products, like health care or socio demographic profiles.

    7. ‌Data pre-processing

      With the least amount of compromise to accuracy, informational value, and data quality, the data pre-processing stage guarantees that all the data collected to construct the AI model or application can be accurately input into the AI algorithm. This includes formatting, transformation, and imput ation.

    8. ‌Build initial AI model

      By identifying an appropriate AI algorithm that reflects the AI capabilities corresponding to the AI application, this step starts the development of an AI model. One or more of the four capabilitiesprediction, classification, association, and optimization can be applied to all current real-world applications. It is advised to start with "applications" before moving on to "algorithms" via "capabilities" when deciding which algorithm to use to construct the model.

    9. ‌Data augmentation

      Data augmentation addresses limitations in the data set that impact the model's output, based on the output and devaluation of the original model. The original model will be run multiple times as the efficiency of the augmentation measures is assessed, in contrast to data transformation, which is specific to variables and includes class imbalance, feature engineering, and feature representation. Undersampling and oversampling techniques are commonly

      used to address class imbalance, [25, 26], whereas feature engineering and feature representation are broad topics that span various fields of study, including signal processing techniques, vector symbolic architectures, [27,28, 2931], and dynamic time warping [32, 33].

    10. ‌Develop a benchmark

      The development of the first AI model necessitates the consideration of an appropriate evaluation benchmark. A common-sense heuristic, such as human expertise in solving the same problem (for example, anomaly detection) or the success rate of the human expert in detecting anomalies using the same input data, is usually used to determine this benchmark. As an alternative, the industry sector or the literature related to the same algorithm may serve as the benchmark. The benchmark not only evaluates the model but also determines which input vectors or attributes do not fit the model. If the data are an incomplete representation of the problem domain, the benchmark returns to the data pre- processing stage or moves on to the next stage of building multiple models.

    11. ‌Build multiple AI models

      The maturity of the AI field, particularly with regard to build technologies like Jupyter notebooks and the availability of feature-oriented algorithms as code libraries on repositories like GitHub, has made it possible to build multiple AI models quickly and effectively. To guarantee that the developed application is legitimate and lawful, licensing for distribution and reuse, specifically in code repositories and any third-party content, must be followed and documented. Alternatively, these algorithms can be swiftly re-implemented without the limitations imposed by the code repository license via modern programming languages like Python.

    12. ‌Evaluate primary metrics

      The primary metrics build on the performance benchmark determined earlier. Understanding what the metric represents and how it is calculated is crucial, and this information can be found in a variety of literature, including case studies and API documentation. Accuracy, robustness, agnosticism, scalability, and interpretability are all important characteristics of an evaluation metric. Metrics like accuracy, precision, recall, F1 score, root mean-squared error, purity, and entropy are frequently utilized. The common misconception that "accuracy" is the only performance metric worth takig into consideration and choosing an appropriate metric when there are numerous options is addressed by the generalized formulation of a performance metric, outcome = model

      + error. The evaluation metrics are also helpful for determining the bias variance trade-off and comparing all models across default and fine-tuned parameter settings.

    13. ‌AI model explainability

      Explainable AI [XAI], also referred to as model interpretability or model explainability, is a relatively new development motivated by the ethical and legal need to make AI transparent. However, explainability has also benefited model development since AI scientists are better able to understand how model parameters, the learning process, and the intricacies of attributes contribute to the desired AI result.

      For complex models (like gradient boosting or neural networks), where the flow of information from input vectors to decision output is obscured by numerous layers of distributed (or partial) computations, AI transparency is necessary.

    14. ‌Evaluate secondary metrics

      At this point, the AI scientist gives the AI/ML engineer a prototypical and working AI model. The AI model must be computationally efficient in order to be deployed or operationalized for a wider audience, in addition to effectively demonstrating "intelligence" in completing the task. Consequently, this stage involves the calculation of a number of secondary metrics, including convergence metrics, time-complexity, memory performance, computational (CPU) performance, and ethical implications.

    15. ‌AI model deployment and risk assessment

      This stage, which is also referred to as model scoring, model production, or model serving, involves deploying the evaluated model for operational use. It is smaller in scale tooperationalization. Instead of granting access to the entire organization, deployment would usually involve a smaller group of specialists and users. The number of end users and types of applications, the expected output formats, the anticipated turnaround time, the frequency of use, and real-time versus batch use of the AI model are the main factors to take into account when deploying the mode

    16. ‌Post-deployment review

      An expert panel, steering committee, or regulatory body will perform a technical and ethical review of the entire project, from a datasets approach to an AI model to assess metrics and effectiveness, depending on the industry sector and project scope. This stage will also involve the administration of contracts, service level agreements, post- implementation documentation, compliance, and standardizatio

      n. Particularly in the field of healthcare, additional research will be carried out through observational studies, small-scale clinical trials, training, and user-acceptance exercises. At this stage, the team will also take into account the legal protection of intellectual property through patenting or the alternatives of trade secrecy, which offers ongoing protection (as patents expire) or defensive publications in scholarly journals that advance the field.

    17. ‌Operationalize using AI pipelines

      This stage, also referred to as MLOps or AI Ops, is an adaptation of the highly successful DevOps software automation capabilities into the deployment of AI models. DevOps, a product of the Agile movement, was defined by Dyck [34] as "a collaborative and multidisciplinary effort within an organization to automate continuous delivery of new software versions, while guaranteeing their Correctness and reliability.". Numerous survey articles have examined the benefits, drawbacks, and difficulties associated with DevOps [35, 36] and its integration with AI [37, 38].

    18. ‌Hyperautomation of processes and systems

      To provide hyperautomated systems and processes, an operationalized AI service will be linked to a process automation pipeline as part of the AI lifecycle. To spark interest in a hyperautomation pilot phase, process owners and stakeholders both upstream and downstream must be shown the

      capabilities of the AI solution. Following this pilot phase, the performance improvements in productivity, efficiency, and effictiveness must be measured and compared to earlier configurations to develop a business case for deployment at scale.

    19. ‌Monitor and evaluate performance

    This final stage of the lifecycle will involve monitoring and evaluation of the AI model that has either been deployed independently or integrated as a hyper-automation process. The primary evaluation criteria are value created by the technology, diverse people using it in a variety of contexts, and representation of the technology itself. The people

    from data observations) and improve their behaviour to accomplish a certain task, according to the definitions provided

    Machine Learning

    Robotics

    Planning

    criterion is measured by end-user activity, the technology is assessed using model drift and model staleness, and value generation is measured by return on investment (ROI). Model drift, which indicates a decline in the model's accuracy as a result of the data's changing nature can be addressed by retraining the model with more recently collected data points that capture these changes or drifts. On the other hand, model staleness is caused by changes in the problem or environment description underlying the model design and development. Addressing staleness requires a fundamental rethink of the model architecture, inputs, algorithm,and parameters.A stale model will trigger a new iteration of the full life cycle.

    Artificial Intellligence

    Natural Language Processing

    Expert systems

    Speech

    Vision

    Continuous monitoring of end-user activity is critical to nd out if and how the model is contributing toward organizational functions.The level of end user activity depends on each use case and metrics can be drawn from adoption, questions, frequency of use, use/revision of documentation, feedback,and requests for features. Finally, although ROI is not directly visible like most other knowledge work, it can be determined using several types of metrics, such as reduced costs (due to reductions in turn around time, human effort, human skill), increased revenue (due to new revenue streams, customer satisfaction, increased market share),and productivity gains (such as reduced errors, low employee turn over, increased agility of teams,and work ow.)

  3. ‌Artificial Intelligence Techniques

    AI contains numerous major scientific areas, it includes Machine learning, natural language processing (NLP), text and speech synthesis, computer vision, robotics, planning, and expert systems. Fig. 2 illustrates the AI domains that the authors created using resources [40]. Many AI applications are based on machine learning techniques, which apply the core concepts of AI.. ML is used to improve speech recognition [41] and speech emotions [42]. Economic planning [43] and factory control [44] employ a variety of machine learning techniques. According to Ref. [45], machine learning (ML) is a powerful data analysis technology that may be applied to a variety of expert systems. One of the primary focuses of robotics research at the moment is machine learning [46].

    The potential of artificial intelligence is greatly explored via machine learning. The development of adaptable, flexible, and "teachable" algorithms or computational techniques is the primary expectation related to machine learning. New features of programs and systems are thus made available. Machine learning (ML) is a subset of artificial intelligence techniques that enable computer systems to learn from past experience (i.e.,

    Fig. 2. Types of Artificial Intelligence

    in [47]. Suppor vector techniques, decision trees, Bayesian learning, k-means clustering, association rule learning, regression, neural networks, and many more are examples of machine learning.

    A subset of machine learning techniques, neural networks (NN) or artificial NNs have some indirect connections to biological neural networks. They tend to be defined as a collection of interconnected components arranged in layers and referred to as artificial neurones. A subset of NN called deep learning (DL) offers computation for multilayer NN. Deep neural networks (DNN), convolutional neural networks (CNN), recurrent neural networks (RNN), generative adversarial networks (GAN), and more are examples of common DL concepts. In [4851], machine learning model classification was examined. ML techniques split into five categories [52] (Fig. 3) namely Unsupervised learning (UL) or cluster analysis, Supervised learning (SL), Semi-supervised learning, Reinforcement learning, and Deep learning.

    1. ‌Supervised learning:

      The basis of this kind of machine learning is training the learni ng algorithm on labeled data.

      Since the data is composed of pairs

      an input that can be represented by a vector and its correspondi ng desired output, which can be defined as a supervisory signal it is referred to as labeled [53].

      Since the proper output is known, the learning algorithm tries t o predict it iteratively and is adjusted to minimize the variation gap between its predicted and actual output, which is why the l earning mechanism is referred to as supervised [41].

      The supervised learning algorithm can generate a function that is referred to as a regression function if the output is continuous

      and a classifier function if the output is discrete by analyzing th e training data [53].

      Regression

      Classification

      Dimensionality Reduction

      Supervised Learnimg

      Clustering

      Self Training

      Machine Learning

      Unsupervised Learnimg

      Semi-supervised learning

      Low density Separation Models

      Graph based Models

      Reinforcement Learning

      Dynamic Programming

      Deep Learning

      Heuristic Methods

      Monte Carlo methods

      Convolutional Neural Networks

      Recurrent Neural Networks

      Fig. 3. Types of Machine Learning

      The generated function predicts the output corresponding to ev ery given input since the learning method reasonably generaliz es patterns and characteristics found in the training data to new input data.The two primary types of supervised learning algorit hms are classification algorithms (discrete output) and regressi on algorithms (continuous output). The goal of regression algorithms is to find the function that matches the training dataset's points.

      The three primary categories of regression algorithms are poly nomial regression, multiple linear regression, and linear regres sion [53].

    2. ‌Unsupervised Learning:

      In contrast to supervised learning, this approach trains the l earning algorithm using an input dataset without any labeled o utputs.Unlike supervised learning, there is no human interventi on to make corrections or adjustments.

      Consequently, compared to supervised learning, unsupervised l earning is more subjective [54].

      Finding the basic structure or distribution patterns present in th e data itself is the primary objective of unsupervised learning. The algorithm learns on its own by reflecting the overall struct ure of input patterns and attempting to represent a specific iden tified input pattern.

      As a result, the various inputs are grouped according to the characteristics that were taken from each input item [54].

      Unlike supervised learning, this method uses an input datasets without any labeled outputs to train the learning algorithm.The method can create and distinguish between the r esulting clusters, even though it won't give them names. It can also use some of these clusters to assign new examples to othe clusters.

      When sufficient data is available for usage, this method which is based on input datacan function effectively.

      Social information filtering algorithms, like those employed by Amazon.com to suggest books to users, are one example of th at.Finding comparable groups of people and then adding new members to these groups is the foundation of these algorithms[ 54].The three primary types of algorithms used in unsupervise d learning are anomaly detection, dimensionality reduction, an d clustering [44].

    3. ‌Semi-supervised learning:

      In situations where we have a lot of input data, some of which are labeled and the remainder are not, this approach lies in the middle of supervised and unsupervised learning techniques. This branch of machine learning encompasses several real- world learning issues. This is because semi-supervised uses a lot of unlabeled data and relatively little labeled data, requiring minimal human interaction. It is more desirable to use less labeled datasets because they are more difficult to obtain, costly, and may need access to subject matter experts. Conversely, unlabeled datasets are more accessible and less expensive [55]. In semi-supervised learning, the learning algorithm can be trained using both supervised and unsupervised learning methods.

      The input dataset's hidden structures and patterns can be uncov ered using unsupervised learning techniques.

      On the other hand, supervised learning techniques can be used to make guess predictions on unlabeled data, feed the data bac k to the learning algorithm as training data, and use the inform ation acquired to make predictions on fresh data sets.

      Therefore, we may claim that predictions or hypotheses derive d from labeled data are modified or reprioritized using unlabel ed data [55]. All semi-

      supervised learning methods make at least one of the following assumptions in order to utilize the unlabeled training data [55]

      : manifold, cluster, and smoothness assumptions.

    4. ‌Reinforcement

      Learning through interaction with the issue environment is kno wn as reinforcement learning.

      Instead of being explicitly instructed on what to do, a reinforce ment learning agent learns from its own activities.

      It chooses what to do now based on current possibilities (explo ration) and past experiences (exploitation).

      It can therefore be characterized as a process of learning by tria l & error.

      The reinforcement learning agent receives a signal in the fo rm of a numerical reward value that indicates whether an actio n was successful.

      In order to maximize the numerical reward, the agent seeks to l earn how to choose actions [56].

      Actions can have an impact on subsequent situations and rewar d values in addition to the current condition and reward value. Learning agents often have predetermined objectives and are a ble to perceive the state of their surroundings to some degree.

      As a result, they can perform actions that influence the stat e and move it closer to the objectives.

      Based on how each approach acquires knowledge, reinforceme nt learning differs from supervised learning.

      The supervised learning approach uses examples from an outsi desupervisor.Reinforcement learning, on the other hand, gains

      knowledge through direct interactions with the issue environm ent [5

    5. ‌Deep Learning

    This area of machine learning has grown significantly since 2 006 and has been used in hundreds of studies since.

    Deep learning has been used to a variety of fields, including a rtificial intelligence and information processing.

    In order to create a model that depicts intricate relationships b etween data, deep learning is a branch of machine learning tha t relies on algorithms that learn from several levels.

    It is called "deep architecture" because there is a hierarchy of featuresthat defines high level characteristics in terms of lowe r level features.Unsupervised learning representations serve as the foundation for the majority of the models included in this class [57].

    In essence, deep learning is the meeting point of signal pro cessing, neural networks, graphical modeling, optimization, art ificial intelligence, and pattern recognition.

    Deep learning's popularity can be summed up as follows: it gre atly increased computer chips' processing power, made it possi ble to incorporate enormous amounts of training data, and was the driving force behind recent advancements in machine learn ing in the fields of information and signal processing [57].

  4. ‌Artificial Intelligence Challenges

    After applying the SLR, seven new challenges of AI in DT were identified [83]. These new challenges are the technical blockages, top management support, employees' lack of trust for AI and resistance to change, transparency and explainability, responsible use of predictive analytics, data governance, and intellectual property. The most frequently mentioned challenges were in the ethical category: high possibility of bias, data privacy considerations, and transparency and explainability.

    TABLE I. AI CHALLENGES

    Challenges of Artificial Intelligence

    References

    High cost of AI technology

    [60, 61, 62, 65, 66, 67]

    Data collection & data accessibility challenges

    [58, 68, 69]

    Poor data quality

    [58, 60, 66, 68, 69, 70, 71, 72]

    Technical blockages

    59, 60, 67, 68, 72

    Lack of skills/knowledge

    [59, 61, 68, 69, 73, 74, 75]

    Top management support

    60, 68, 72, 74, 76

    Employees' lack of trust in AI & resistance to change

    [64, 65, 66, 68, 72, 74, 76, 77,

    78]

    Over-dependency on technology

    63, 76, 77

    Lack of empathy

    63, 77, 79

    High possibility of bias

    58, 59, 63, 68, 69, 70, 75,76,

    77, 78, 79, 80, 81, 82

    Data privacy comsidernatins

    58, 59, 69, 70, 73, 74, 76, 77,

    47, 82

    Transparency and explainability

    [58, 59, 69, 70, 73, 74, 76, 77,

    78, 82]

    Responsible use of predictive analytics

    58, 63, 80, 82 , 52

    Data governance

    59, 68, 80

    Intellectual property

    77, 82

    Cyber security risks

    61, 66, 68, 69 71, 75, 81

  5. ‌CONCLUSIONS

Research focusing on Artificial intelligence inreasing from last two – three decades in government, industry and academia. Therefore, this article conducted a systematic literature review to explore the concept of AI, CDAC life cycle of AI, its types, AI techniques along with challenges of AI. The outcome of SLR is to assist the researchers or new entrants to understand the concept of artificial intelligence precisely, to grasp how the AI functioning via CDAC life cycle of AI, its techniques and challenges. Based on the outcome of system literature studies, firstly, AI can be defined as systems that display intelligent behaviour by analysing their environment and taking actions with some degree of autonomy to achieve specific goals. Secondly, The CDAC AI life cycle,is presented as the progression of an AI solution through its distinct phasesdesign, develop, and deploy and 19 constituent stages from conception to production as applicable to any AI initiative. Thirdly the distinct major AI techniques namely Machine learning, natural language processing (NLP), text and speech synthesis, computer vision, robotics and deep learning. At last the numerous challenges facing by artificial intelligence are technical blockages, top management support, employees lack of trust for AI and resistance to change, transparency and explainability, responsible use of predictive analytics, data governance, and many more.The study has some limitations like the number of databases that were used and also the use of only the research studies written in English language. Researcher and new entrants in the artificial intelligence could use the outcome of this study for better and precisely understanding of AI along with their own study and research.

‌REFERENCES

  1. H. Sheikh, C. Prins, and E. Schrijvers, Artificial Intelligence: Definition and Background, In: Mission AI. Research for Policy. Springer, Cham., Jan. 2023.

  2. High-Level Expert Group on Artificial Intelligence, A definition of AI: Main capabilities and scientific disciplines. [Online]. Available: https://ec.europa.eu/newsroom/dae/document.cfm?doc_id=56341Return to ref 2019 in article

  3. IBM, The history of artificial intelligence. [Online]. Availabile: https://www.ibm.com/think/topics/history-of-artificial-intelligence

  4. S. Russell, and P. Norvig, Artificial Intelligence: A modern approach, Fourth Edition, Pearso, United Kingdom, 2022.

  5. N. Nilsson, The quest for artificial intelligence, Cambridge University Press, 2009.

  6. DenkWerk, Artificial Intelligence in Nederland: Zelf Aan Het Stuur. [Online]

    Available: https://denkwerk.online/media/1029/artitificial_intelligence_ in_nederland_juli_2018.pdf

  7. N. Bostrom, Superintelligence: Paths, Dangers, Strategies, Oxford University Press, 2016.

  8. L. Floridi, The Fourth Revolution: How the Infosphere is reshaping human reality, Oxford University Press, 2014.

  9. H. Moravec, Mind Children: The future of robot and human intelligence, Harvard University Press, 1988.

  10. A. Agrawal, J.Gans, and A.Goldfarb,Prediction machines: The simple economics of Artificial Intelligence, Harvard Business Press, 2018.

  11. J. Brockman, Possible minds: Twenty-five ways of looking at AI, Penguin, pp. 4153, 2019..

  12. V. Dignum, Responsible Artificial Intelligence: How to develop and use AI in a responsible way, Springer, 2019.

  13. M. Ford, Architects of Intelligence, Packt Publishing, 2018.

  14. P. Domingos, The master algorithm: How the Quest for the ultimate learning machine will remake our world, Penguin Random House, 2017.

  15. D. D. Silva, and D. Alahakoon, An artificial intelligence life cycle: From conception to production, Patterns, vol. 3, Issue 6, 2022.

  16. A. Krizhevsky, I. Sutskever, and G.E. Hinton, Imagenet classification with deep convolutional neural networks,. Adv. Neural Inf. Process. Syst. 25, pp. 10971105, 2012.

  17. K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition,, In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770778, 2016.

  18. J. Devlin, M. -W. Chang, K. Lee, and K. Toutanova, Bert: pretraining of deep bidirectional transformers for language understanding,, Preprint at arXiv. 1810.04805, 2018.

  19. A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, Improving language understanding by generative pre-training,. Preprint at arXiv. cs.ubc.ca, 2018.

  20. R. Kimball, and M. Ross, The Data Warehouse Toolkit: The Complete Guide to Dimensional Modeling, John Wiley & Sons, 2011.

  21. W.H. Inmon, The data warehouse and data mining, Communications of the ACM, Volume 39, Issue 11, pp. 49-50, 1996.

  22. D. E. OLeary, Embedding AI and crowdsourcing in the big datalake, IEEE Intelligent Systems, pp. 7073, 2014.

  23. A. Halevy, F. Korn, N. F. Noy, C. Olston, N. Polyzotis, S. Roy, and S.

    E. Whang, Managing googles data lake: an overview of the goods system, IEEE Data Eng. Bull, pp. 514, 2016.

  24. A. Gorelik, The Enterprise Big Data Lake: Delivering the Promise of Big Data and Data Science , OReilly Media, 2019.

  25. B. W. et al., An application of oversampling, undersampling, bagging and boosting in handling imbalanced datasets, In Proceedings of the first international conference on advanced data and information engineering, (DaEng-2013), Springer, pp. 1322, 2013.

  26. R. Mohammed, J. Rawashdeh, and M. Abdullah, Machine learning with oversampling and undersampling techniques: overview study and experimental results, 11th International Conference on Information and Communication Systems (ICICS), pp. 243248, 2020.

  27. S. A. Kassam, and H. V. Poor, Robust techniques for signal processing: a survey, IEEE Proceedings, pp.433481, 1985.

  28. N. Yu, Z. Li, and Z. Yu,Survey on encoding schemes for genomic data representation and feature learningfrom signal processing to machine learning, Big Data Mining and Analytics , pp. 191210, 2018.

  29. P. Kanerva, Hyperdimensional computing: an introduction to computing in distributed representation with high-dimensional random vectors, Cogn. Comput. Pp. 139159, 2009.

  30. R. W. Gayler, Vector symbolic architectures answer jackendoffs challenges for cognitive neuroscience, Preprint at arXiv. cs/0412059, 2003.

  31. D. Kleyko, D, Pattern Recognition with Vector Symbolic architectures, PhD Thesis, Luleå University of Technology, 2016.

  32. E. J. Keogh, and M. J. Pazzani, Scaling up dynamic time warping for datamining applications, In Proceedings of the sixth ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 285289, 2000.

  33. H. Li et al., Adaptively Constrained Dynamic Time Warping for Time Series Classification and Clustering, Information Sciences, pp. 97116, 2020.

  34. A. Dyck, R. Penners, and H. Lichter, Towards definitions for release engineering and devops, Proceedings of the 3rd International Workshop on Release Engineering, pp. 3, 2015.

  35. C. Ebert, G. Gallardo, J. Hernantes, and N. Serrano, Devops, IEEE Software, pp. 94-100, 2016.

  36. L. Leite, C. Rocha, F. Kon, D. Milojicic, and P. Meirelles, A survey of devops concepts and challenges, ACM Compuing Surveys, 2019.

  37. S. Makinen, H. Skogstrom, E. Laaksonen, and T. Mikkonen, Who needs mlops: what data scientists seek to accomplish and how can mlops help?, Preprint at arXiv. 2103.08942, 2021.

  38. M. Treveil et al., Introducing MLOps, OReilly Media, 2020.

  39. MOR, How Does AI Work? Complete Guide For Businesses 2025 [Detailed], [Online]. Availabile: https://morsoftware.com/blog/how- does-ai-work

  40. R. I. Mukhamediev, A. Symagulov, Y. Kuchin K. Yakunin and M. Yelis, From Classical Machine Learning to Deep Neural Networks: A Simplified Scientometric Review, Applied Sciences, 2021.

  41. S. Watanabe et al., Espnet: End-to-end speech processing toolkit, Preprint at arXiv:1804.00015, 2018.

  42. L. Kerkeni et al., Automatic speech emotion recognition using machine learning, Social Media and Machine Learning, IntechOpen: London, UK, 2019.

  43. J. An, A. Mikhaylov and N. Sokolinskaya, Machine learning in economic planning: Ensembles of algorithms, Journal of Physics: Conference Series, 2019.

  44. J. P. Usuga Cadavid, S. Lamouri, B. Grabot, R. Pellerin, A. Fortin, Machine learning applied in production planning and control: A state- of-the-art in the era of industry 4.0, Journal of Intelligent Manufacturing, pp. 15311558, 2020.

  45. E. T. Ogidan, K. Dimililer, and Y. K. Ever, Machine learning for expert systems in data analysis, Proceedings of the 2018 2nd International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT), Ankara, Turkey, 2018.

  46. A. Mosavi, and A. Varkonyi, Learning in robotics, International Journal of Computer Applications, pp. 811, 2017.

  47. G. Nguyen et al., Machine learning and deep learning frameworks and libraries for large-scale data mining: A survey, Artificial Intelligence Review, pp. 77124, 2019.

  48. T. O. Ayodele, Types of machine learning algorithms, In book: New Advances in Machine Learning, pp. 1948, 2010.

  49. H. Awad, H. Ibrahim, S. Nor, A. Mohammed, and A. Babiker, Taxonomy of machine learning algorithms to classify real time interactive applications, International Journal of Computer Network and Wireless Communication, pp. 6973, 2012.

  50. R. Muhamedyev, Machine learning methods: An overview, Computer Modelling & New Technologies, pp. 1429, 2015.

  51. I. Goodfellow, Y. Bengio, and A. Courville, Deep Learning, MIT Press Cambridge: Massachusetts, MA, USA, 2016.

  52. A. B. Nassif, I. Shahin, I. Attili, M. Azzeh, and K. Shaalan, Speech recognition using deep neural networks: A systematic review, IEEE Access, pp. 1914319165, 2019.

  53. Wikipedia, Supervised Learning, [Online] Available: https://en.wikipedia.org/w/index.php?oldid=457924979

  54. M. Caza-Szoka, D. Massicotte, and F. Nougarou, Naive Bayesian learning for small training samples: Application on chronic low back pain diagnostic with sEMG sensors, In Proceedings of the 2015 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), pp. 470-475. IEEE, 2015.

  55. X. Zhu, Semi-Supervised Learning Literature Survey Contents, Department of Computer Sciences, University of WisconsinMadison, Madison, WI, USA, Technical Report. 1530, 2008.

  56. A. Mukherjee, A. Pal, and P. Misra, Data analytics in ubiquitous sensorbased health information systems, In Proceeding of 6th International Conference Next Generation Mobile Applications Services Technology (NGMAST), pp. 193198, 2012.

  57. Y. Cho and L. K. Saul, Kernel methods for deep learning, Advances in Neural Information Processing Systems, 2009.

  58. A. Aldoseri, K. N. Al-Khalifa, and A. M. Hamouda, AI-powered innovation in digital transformation: Key Pillars and Industry Impact, Sustainability 16 (5), 2024.

  59. N. A. Perifanis and F. Kitsios, Investigating the influence of artificial intelligence on Business Value in the digital era of strategy: A literature review, Information 14 (2), 2023.

  60. E. Golab-Andrzejak, AI-powered digital transformation: Tools, benefits, and challenges for marketers a case study of LPP, Procedia Computer Science 219, pp. 397-404, 2023.

  61. M . F. Ahmad et al., Digital Transformation: An exploring barriers and challenges practice of artificial intelligence in manufacturing firms in Malaysia, Journal of Advanced Research in Applied Sciences and Engineering Technology, pp. 110117, 2022.

  62. I. Klopov et al., Digital Transformation of education based on Artificial Intelligence, TEM Journal 12 (4), pp. 2625-2634, 2023.

  63. D. Mhlanga, Digital Transformation Education, opportunities, and challenges of the application of ChatGPT to emerging economies, Education Research International, pp. 1-13, 2023.

  64. F. Kitsios, and M. Kamariotou, Artificial Intelligence and Business Strategy towards Digital Transformation: A Research Agenda, Sustainability 13 (4), 2021.

  65. P. Li, A. Bastone, T. A. Mohamad, and F. Schiavone, F, How does artificial intelligence impact human resources performance. evidence from a healthcare institution in the United Arab Emirates. Journal of Innovation and Knowledge 8 (2), 2023.

  66. D. Sjödin, V. Parida, M. Palmié, and J. Wincent, How AI Capabilities Enable Business Model Innovation: Scaling AI through co-evolutionary processes and feedback loops, Journal of Business Research 134, pp. 574-587, 2021.

  67. V. Thokal, and P. R. Patil, Review and analysis of the literature: Artificial Intelligence-based digital transformation of automated customer onboarding, Engineering Proceedings, 2023.

  68. A. Aldoseri, K. N. Al-Khalifa, and A. M. Hamouda, Methodological approach to assessing the current state of organizations for AIbased Digital Transformation, Applied System Innovation 7 (1), 2024.

  69. B. Weber-Lewerenz, Corporate Digital Responsibility (CDR) in Construction Engineering-ethical guidelines for the application of digital transformation and artificial intelligence (AI) in User Practice,

    SN Applied Sciences 3, 2021

  70. N. Böhmer, and H. Schinnenburg,Critical exploration of AI-driven HRM to build up organizational capabilities, Employee Relations: The International Journal 45 (5), 2023.

  71. M. Kovai, M. Mutavdija, K. Buntak, and I. Pus, Using artificial intelligence for creating and managing organizational knowledge, Tehnicki vjesnik 29 (4), pp. 1413-1418, 2022.

  72. M. H. Philippart, Success factors to deliver organizational digital transformation: A Framework for Transformation Leadership, Journal of Global Information Management 30 (8), pp. 1-17, 2022.

  73. E. Peretz-Andersson, S. Tabares, P. Mikalef, V. Parida, Artificial Intelligence implementation in manufacturing SMEs: A resource orchestration approach, International Journal of Information Management 77, 2024.

  74. W. Reim, J. Åström, and O. Eriksson, Implementation of Artificial Intelligence (AI): A roadmap for business model innovation, AI 1 (2),

    pp. 180-191, 2020.

  75. Y. Xiong, Q. Ling, and J. Liu, Educational Missions and Tactics in the EU Artificial Intelligence Strategy in the Era of Digital Transformation: A Policy Analysis based on EU documents published during 2018 2022, In Proceedings of Applied Mathematics, Modeling and Computer Simulation third International conference (AMMCS 2023), 1 Oliver's Yard, 55 City Road, London, EC1Y 1SP: SAGE Publications Pvt. Ltd, 2023.

  76. X. Zhang, M. F. Antwi-Afari, Y. Zhang, and X. Xing, The impact of artificial intelligence on Organizational Justice and project performance: A systematic literature and science mapping review. Buildings 14 (1), 2024.

  77. B. George, and O. Wooden, Managing the strategic transformation of higher education through Artificial Intelligence, Administrative Sciences 13 (9), 2023.

  78. I. Sobrino-García, Artificial Intelligence Risks and Challenges in the Spanish Public Administration: An exploratory analysis through expert judgments. Administrative Sciences 11 (3), 2021.

  79. V. S.Athota et al.,Overcoming financial planners cognitive biases through digitalization: A qualitative study, Journal of Business Research 154, 2023.

  80. S. Akter, K. Michael, M. R. Uddin, G. McCarthy, and M. Rahman, Transforming Business Using Digital Innovations: The Application of AI, Blockchain, Cloud, and Data Analytics, Annals of Operations Research 308, 2022.

  81. F. Mihai, O. E.Aleca, and M. Gheorghe, Digital transformation based on AI Technologies in European Union organizations, Electronics 12 (11), 2023.

  82. R. F. Reier Forradellas, and L . M. Garay Gallastegui, Digital Transformation and artificial intelligence applied to business: Legal regulations, economic impact, and perspective, Laws 10 (3), 2021

  83. M. Kavak, and L. Rusu, Challenges and Opportunities of Artificial Intelligence in Digital Transformation: A Systematic Literature Review, Procedia Computer Science, pp. 369-377, 2025.