Webinar - Feb 12th - Harnessing Data and AI to Fight Financial Crime in Payments
Harnessing Data and AI to Fight Financial Crime in Payments🚀 Join us for an insightful webinar hosted by Microsoft, in collaboration with our partners Quantexa, iPiD, and EY. We'll explore the latest trends in financial crime within the payments industry, including progress on the EU's Verification of Payee (VOP) deadline and its implications. 🌐 Event Details: 📅 February 12th, 2025 🕒 11:00 AM – 11:45 AM CET Don't miss out on this golden opportunity to fortify your knowledge and network with industry leaders. Secure your spot now! 🔗 Sign Up Now: Harnessing Data and AI to Fight Financial Crime in Payments #Microsoft #MicrosoftVirtualBriefing #Data #Financial #Crime #Payments #Qalliancespowered11Views0likes0Comments📣Upcoming Webinar: The Biggest Challenges in Data Quality: How Far Can AI Go to Solve Them? 📣
In this webinar, Dan Onions, Global Head of Data Management at Quantexa, and Martin Maisey, Head of Data Management EMEA, will delve into the pressing question on every data professional's mind: "How can AI help me?" Unlock the full potential of your data strategy: As AI technologies, particularly LLMs, become increasingly integral to data management strategies, ensuring the quality and reliability of these systems' outputs is paramount. Our experts will explore the critical role of foundational data quality in harnessing AI effectively and responsibly, and address key challenges, such as achieving consistency and accuracy in AI-generated outputs and aligning them with regulatory standards already on the horizon. Attendees will gain insights into practical applications of AI in the real world, understanding how to make AI outputs on data trustworthy across the entire organization. Register Your Place Here: The Biggest Challenges in Data Quality: How Far Can AI Go to Solve Them? (quantexa.com)21Views1like0CommentsQuanCon24 On-Demand: Tackling Financial Crime and Fraud In The Era Of AI
About QuanCon24 In an era of disruption – from supply chain disruption, geopolitics, cost of living, and elections – organizations face a multitude of critical decisions that can lead to increased risk, missed opportunities, or dead ends. Organizations must evolve to make fast confident decisions to become efficient, resilient, and reach goals before the competition. But so much more is possible with the right data in the right context. Now, you can watch our key sessions on-demand. Watch now. About this session This expert panel, moderated by Quantexa’s Chief Marketing Officer Matt Hooper, focused on how macroeconomic shifts are impacting risk and governance across industries. It featured four industry trailblazers: Jocelyn Norval, Global TM and Screening Lead, ING Kai Schrimpf, Head of Monitoring and Screening Controls, Morgan Stanley Mark Cheeseman, Chief Executive Officer, Public Sector Fraud Authority Neil Jones, Head of 1st Line Financial Crime, Danske Bank There is a multitude of socioeconomic factors that are impacting the ecosystem of financial and economic crime activity: the reshaping of the governance, risk, and compliance landscape; an impactful year from a political perspective across the world; ongoing market volatility for the last few years – to name a few. But AI and innovative technologies hold some answers to overcoming these challenges and effectively tackling financial crime and fraud in a consistently evolving landscape. Join Matt Hooper, Chief Marketing Officer at Quantexa, and experts at private and public sector organizations from across the globe to discover how the transformative power of generative AI and new technologies can counter financial and economic crime. Key topics: The importance of taking a holistic and integrated approach to tackling economic crime and embracing AI How the experts are looking at technology to meet today’s demands and future-proofing approaches The innovative work around counter fraud and anti-money laundering and lessons learned from the experts’ decision intelligence journeys The potential and future vision for organizations to continue fighting economic crime, challenges and opportunities that may arise, and why public-private partnerships are integral to success Watch Tackling Financial Crime and Fraud In The Era Of AI on-demand and discover our key sessions from QuanCon24.21Views0likes0CommentsMLflow at Quantexa
In the dynamic landscape of Machine Learning (ML) development, the need for effective experiment tracking is paramount. This is especially true as organizations scale their operations. As the complexity of ML projects grows, so does the need for comprehensive tools to efficiently manage experimentation, iterations, and model versions. At Quantexa, we encountered this challenge head-on and sought a robust solution to streamline our ML workflows. We found a big part of the solution to these challenges was to use MLflow. MLflow is a powerful platform designed to simplify the end-to-end machine learning lifecycle. In this blog, we delve into how we leverage MLflow at Quantexa. What is MLflow? MLflow is an open-source platform for managing the end-to-end machine learning lifecycle. It's designed to help Data Scientists and Machine Learning Engineers with tracking experiments, packaging code into reproducible runs, and sharing and deploying models. It comes with a graphical user interface which makes it easy to use. At Quantexa, we primarily use MLflow for experiment tracking and reproducibility. This will be the focus of this post. If you are interested in learning more about how you can use MLflow for model deployment, we recommend you spend time reading collateral on the MLflow Website. What issues is MLflow solving? MLflow addresses several critical challenges in our machine learning workflow. This enhances our overall efficiency and effectiveness. These challenges will also be faced by any client building models on top of Quantexa data. Here's how it tackles three of our most key issues: Reproducibility and traceability At Quantexa, we have a considerable number of machine learning models. It is critical that the origin of all these models is well-documented and readily available. This is important from a compliance perspective but also for internal purposes. The team and clients should be aware of performance metrics and each model's strengths and weaknesses. The models are also valuable intellectual property for the company, and we need to be able to re-produce results. MLflow helps us with these issues. When models are trained, we record versions of code, model parameters, and the source of the original data. We can then use this information to perfectly recreate a model if ever needed. Furthermore, we have a record of any experiments which were used in selecting models. This means that a Data Scientist in the future can understand why certain decisions were made during the model prototyping phase. Collaborative insight We want Data Scientists to be able to collaborate easily. MLflow fosters collaboration by providing a centralized repository for tracking and sharing experiments. This is particularly useful when it comes to gaining insight from non-technical stakeholders. Stakeholders can use the MLflow UI to access all the information they need to assess and give feedback on models. This dramatically speeds up the feedback cycle and leads to better models. It also saves time for all involved. MLflow enables data scientists to collaborate better with one another. This encourages a culture of knowledge exchange and innovation within the team. Data correctness At Quantexa, we are dealing with very large datasets when running predictions using our models. This scale can make it difficult to assess the quality and correctness of data. Manual sense checks are always performed, but mistakes can be missed. This is particularly true if errors only affect a certain portion of the dataset (e.g. one country). To avoid errors slipping through, we produce lots of metrics and plots when we receive new data refreshes. We automatically compare this data to what we have seen previously. We log all of these metrics in MLflow. This makes it easy for the whole team to audit the data and spot mistakes such as missing data. Large differences are highlighted automatically even if only in one geographical region or data source. This information can then be easily forwarded to upstream teams to resolve issues. This process dramatically reduces the risk of erroneous predictions from incorrect data. How do we use it? In practice, we mainly use MLflow for three different workflows. Feature generation A pre-processing step transforms data produced by Quantexa Entity Resolution into a different tabular format. Our models then use this new format for training and prediction. We record aggregate information about data produced during this pre-processing step in the form of plots and CSVs. This aggregate information is compared to previous runs to highlight big discrepancies in easy-to-read files automatically. As mentioned, this reduces the risk of downstream tasks using erroneous data critical for model training and prediction. Code versions and parameters for this pipeline are recorded. This allows us always to re-create the pre-processing step. It also provides an understanding of the data source we are training our models on. Here is a list of some of the information we record at this stage: Mean, median, and variance of features Data drift scores for each feature Percentage of null values in a column A plot produced to visualize the top drifting features across countries. Country and feature names have been omitted. A plot produced to visualize the percentage of fields that have missing values in the Documents. These fields are used to produce features. Country and field names have been omitted. Model training We use MLflow to record model training. This can be to track multiple experiments during the prototyping phase of a project or be used as a reference for models in production. All the required information is recorded so that models can be re-produced if required. Here is a list of some of the information we record when training models Model hyperparameters Test metrics such as precision and recall Confusion matrix An ROC curve (receiver operating characteristic curve) produced against the test set during training. A cumulative gains curve produced against the test set during training. Staging model evaluation We often iterate on our production machine learning models to improve their performance. We do this by adding more labeled examples to their training datasets or by adding more features. When we train a new model that we think should replace the existing model, we want to ensure that the new model is superior to the model we are replacing. We can look at the performance of the model across a test set, but sometimes this does not tell the full story. This is particularly true if the amount of labeled data we have is limited. Before upgrading to any new model, we run an evaluation across the entire unlabeled dataset to see how the models differ in practice. This is important for understanding the real business impact of a change. This evaluation is recorded in MLflow such that the data scientists and any relevant stakeholders can review it. Here is some of the information we record during the evaluation: The average size of differences in model score between the old model and new model Individual examples with the biggest differences in the model score between the old model and the new model Individual examples with SHAP explainability plots Explainability plot produced when the model is used for inference across unlabelled data. Feature names have been omitted. Visualization of the different scores produced by a model currently in production compared to the model selected to replace it. Conclusion In the dynamic realm of machine learning, effective experiment tracking is indispensable, especially with growing organizational scale. At Quantexa, MLflow has emerged as our solution, simplifying the ML lifecycle and addressing critical challenges in reproducibility, collaboration, and data correctness. Leveraging MLflow enhances our internal workflows and empowers clients to build models on Quantexa data to navigate complexities seamlessly.671Views1like0CommentsKPMG: AI boosts supply chain agility
Artificial Intelligence (AI) has the potential to enhance supply chain agility, according to a new KPMG report titled “Raise Your AI Short Game to Boost Supply Chain Agility.” The report emphasizes that businesses are grappling with compound volatility, necessitating the need for supply chain flexibility. AI emerges as a technology that can help companies anticipate market volatility, make complex network decisions, and achieve business objectives in the short term. In fact, 57% of the executives who participated in the 2023 KPMG Global Tech survey, stated that AI will help them achieve their short-term goals. Read more here … https://outlookseries.com/A0769/Services/3395_KPMG_AI_boosts_supply_chain_agility.htm Quantexa's Donna Goodwin shares her thoughts …21Views0likes0CommentsQuantexa's AI Roundup - 2023
In July 2023, Quantexa announced a significant investment into its Artificial Intelligence (AI) capabilities (Quantexa Bringing Total Investment in AI R&D to over $250M by 2027). Since this announcement, there has been significant advancement in the AI space, and growth in some of the core AI capabilities at Quantexa. Alongside the significant growth of the NLP capability, Quantexa’s Analytical Innovation team have completed the MVPs of their three flagship products which are now released under experimental. These tools use Quantexa networks to uncover insights: the Entity Resolution AI suite; Q-Knowledge Graph and Shell Company Detection. In this round up post, we introduce the three products and demonstrate how they can add value to your Quantexa deployment. The Entity Resolution (ER) AI Suite The ER AI suite provides a series of tools for analysing the outputs of Quantexa’s ER product and provides suggestions for improving the configurations powering the ER using AI. In particular, the tool can detect overlinking and underlinking in Quantexa Entities and their root causes. The overlinking detection tool is powered by machine learning with features based on the qualities of the Entity’s constituent record-compound graph (read more about using the Entity Quality Overlinking tool for the first time). These graph-based features include the use of several complex graph algorithms (e.g., the Stoer-Wagner algorithm) to find shapes which are indicative of overlinking. Such shapes include ‘bridges’ in the network which incorrectly link Entities together, as well as graphs with very long paths. Statistical techniques can then be applied to determine what compounds or data points may be leading to this overlinking. The underlinking tool uses sophisticated graph algorithms to find ‘Super-Entities’ – Entities which should be formed of several existing Entities. This helps the user to identify template changes to merge such entities together in future ER runs. Q-Knowledge Graph Q-Knowledge Graph is a series of tools for analysing large-scale Quantexa Entity and Document graphs. It scales to billions of nodes and edges and uses sophisticated optimisation techniques to provide extremely fast implementations of core graph transformations and algorithms (including page rank). Not only does the tool provide access to commonly used graph algorithms out of the box (for example, PageRank) – it also provides a connection to common graph learning libraries such as PyG. This enables several use cases across Risk, KYC and MDM and has already been deployed for transactional use cases in a global bank. It will also be a core back-end component of a number of upcoming Quantexa AI products. Shell Company Detection The Shell company detection tool uses machine learning to identify shell companies, using characteristics of the local ego-networks of the companies. The model uses a combination of structural features (e.g., links to known shell directors); temporal features (e.g., patterns of director resignation) and static features (including the size of the corresponding corporate registry Document). For more information, see What can Network structure tell us about risk? The current model is built specifically for the UK and Singapore and can encapsulate some behaviours specific to shells in these jurisdictions. Models focused on other jurisdictions are coming this year. Upcoming AI releases The NLP team at Quantexa are also developing a machine learning pipeline called Text2Networks for working with unstructured data, which will be available in the next major release of Quantexa. The Text2Networks pipeline is a highly-configurable pipeline of ML models for mapping any unstructured textual data into a graph. The pipeline detects, labels and organizes people, places, and things in the real world – the supported Entity types include People, Locations, Companies and Geo-political organizations. With text2networks integrated into the core Quantexa product, our users will be able to incorporate any textual data source that is important for their business. Concrete example could include global news, intelligence reports, and Suspicious Activity Reports (SARs). There are several tools in development, including further tooling within the ER quality suite and Q-Knowledge Graph, as well as other risk models such as the SME detection tool which will be coming in later releases of Quantexa. To keep up with the latest releases, be sure to follow our Release Announcements topic.531Views1like0CommentsGen AI and the evolving role of marketing - Capgemini
The majority of marketers (62%) believe that generative AI will augment human creativity, enhancing unique human qualities such as intuition, emotion, and context understanding. Organizations already investing in generative AI for marketing dedicate 62% of their total marketing technology budget towards it, seeing this breakthrough technology as a catalyst for creativity and innovation in marketing. That’s according to Capgemini Research Institute’s latest report ‘Generative AI and the evolving role of marketing: A CMO’s Playbook’, which reveals that half of organizations have already set aside specific budgets and almost half (47%) have allocated teams for the implementation of generative AI in marketing. Quantexa's Matt Hooper weighs in: Capgemini: 60% Integrating Generative AI Into Marketing31Views0likes0CommentsGartner: 39% of Organizations Currently Use AI in the Finance Function
Sixty-eight percent of finance organizations are using AI or plan to use the technology, according to a recent survey of finance leaders by Gartner, Inc. A May 2023 survey of 133 finance leaders found that 39 percent of respondents are using AI/machine learning (ML), and an additional 29% of respondents said AI/ML is planned (see Figure 1). Read more here: Gartner: 39% of Organizations Currently Use AI in the Finance Function1View0likes0CommentsGoogle Banking Survey: C-Suites & Boards More Involved in Tech Decisions with AI
New research explored the sentiment towards generative AI (gen AI) in banking among North American banking executives and consumers. The study, based on a survey of 350 banking executives responsible for genAI decisioning and more than 2,000 banking consumers in the United States, found broad interest in gen AI technologies as a way to improve operations and the customer experience, while some barriers and risks remain. The majority (92%) of banking executives stated there is high demand for gen AI within the banking industry, with 95% stating it has the potential to transform the industry. Increased interest in gen AI is driving senior leadership, like C-suite executives and boards of directors, to get more involved in technology and IT decisions, according to almost all banking respondents (96%). Quantexa's Parsa Ghaffari weighs in: Google Banking Survey: C-Suites and Boards More Involved in Tech Decisions Due to Heightened Interest in Gen AI41Views1like0CommentsIDC Revenue for AI software Will Reach $279B in 2027
A recent forecast from International Data Corporation (IDC) shows that the worldwide artificial intelligence (AI) software market will grow from $64 billion in 2022 to nearly $251 billion in 2027 at a compound annual growth rate (CAGR) of 31.4%. The forecast for AI-centric software includes Artificial Intelligence Platforms, AI Applications, AI System Infrastructure Software (SIS), and AI Application Development and Deployment (AD&D) software (excluding AI platforms). However, it does not include Generative AI platforms and applications, which IDC recently forecast will generate revenues of $28.3 billion in 2027. A recent IDC survey found that, in the next 12 months, roughly a third of respondents believe that organizations will prefer to buy AI software from a vendor or use in-house support alongside vendor-supplied AI software for specific use cases or application areas. This indicates a growing demand for AI solutions and highlights the need for customized approaches based on individual business requirements. Quantexa's Parsa Ghaffari weighs in with his thoughts: IDC: Revenue for AI Software Will Reach $279B in 202751Views0likes0Comments