<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>SEI Blog | Artificial Intelligence Engineering</title><link>http://www.sei.cmu.edu/feeds/topic/</link><description>Updates on changes and additions to the                         SEI Blog for posts matching Artificial Intelligence Engineering</description><atom:link href="http://www.sei.cmu.edu/blog/feeds/topic/artificial-intelligence-engineering/rss/" rel="self"/><language>en-us</language><lastBuildDate>Wed, 06 May 2026 00:00:00 -0400</lastBuildDate><item><title>The ELM Library: An LLM Evaluation Toolset</title><link>https://www.sei.cmu.edu/blog/the-elm-library-an-llm-evaluation-toolset/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>To help teams meet the need for rigorous evaluation methods, researchers in SEI’s AI Division developed a library built on best practices for LLM evaluation and benchmarking.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Violet Turri, Natalie Schieber, Charles Loughin, Tyler Brooks</dc:creator><pubDate>Wed, 06 May 2026 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/the-elm-library-an-llm-evaluation-toolset/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>From Reality to Virtual Reality: The Impact of 3DGS on Training, Education, and Beyond</title><link>https://www.sei.cmu.edu/blog/from-reality-to-virtual-reality-the-impact-of-3dgs-on-training-education-and-beyond/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>This blog post describes cutting-edge method for creating digital models of the physical world called 3D Gaussian Splatting.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Roxxanne White, Matt Walsh, Dominic Ross, Richard Laughlin</dc:creator><pubDate>Wed, 25 Mar 2026 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/from-reality-to-virtual-reality-the-impact-of-3dgs-on-training-education-and-beyond/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>My AI System Works…But Is It Safe to Use?</title><link>https://www.sei.cmu.edu/blog/my-ai-system-worksbut-is-it-safe-to-use/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>This blog post introduce System Theoretic Process Analysis (STPA), a hazard analysis technique uniquely suitable for dealing with the complexity of AI systems.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">David Schulker, Matt Walsh, Emil Mathew</dc:creator><pubDate>Tue, 09 Sep 2025 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/my-ai-system-worksbut-is-it-safe-to-use/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>Artificial Intelligence in National Security: Acquisition and Integration</title><link>https://www.sei.cmu.edu/blog/artificial-intelligence-in-national-security-acquisition-and-integration/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>This blog post highlights practitioner insights from a recent AI Acquisition workshop, including challenges in differentiating AI systems, guidance on when to use AI, and matching AI tools to mission needs.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Paige Rishel, Carol Smith, Brigid O'Hearn, Rita Creel</dc:creator><pubDate>Tue, 05 Aug 2025 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/artificial-intelligence-in-national-security-acquisition-and-integration/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>Amplifying AI Readiness in the DoD Workforce</title><link>https://www.sei.cmu.edu/blog/amplifying-ai-readiness-in-the-dod-workforce/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>The SEI recently partnered with the Department of the Air Force Chief Data and AI Office to develop a strategy to identify and assess hidden workforce talent for data and AI work roles.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Eric Keylor, Robert Beveridge, Jonathan Frederick</dc:creator><pubDate>Mon, 23 Jun 2025 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/amplifying-ai-readiness-in-the-dod-workforce/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>Out of Distribution Detection: Knowing When AI Doesn't Know</title><link>https://www.sei.cmu.edu/blog/out-of-distribution-detection-knowing-when-ai-doesnt-know/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>How do we know when an AI system is operating outside its intended knowledge boundaries?</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Eric Heim, Cole Frank</dc:creator><pubDate>Mon, 09 Jun 2025 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/out-of-distribution-detection-knowing-when-ai-doesnt-know/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>10 Things Organizations Should Know About AI Workforce Development</title><link>https://www.sei.cmu.edu/blog/10-things-organizations-should-know-about-ai-workforce-development/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>This post outlines 10 recommendations developed in response to work with our mission partners in the Department of Defense.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Jonathan Frederick, Dominic Ross, Eric Keylor, Cole Frank, Intae Nam</dc:creator><pubDate>Mon, 28 Apr 2025 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/10-things-organizations-should-know-about-ai-workforce-development/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>DataOps: Towards More Reliable Machine Learning Systems</title><link>https://www.sei.cmu.edu/blog/dataops-towards-more-reliable-machine-learning-systems/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>Decisions based on ML models can have significant consequences, and managing the raw material—data—in ML systems is a challenge. This post explains DataOps, an area that focuses on the management and optimization of data throughout its lifecycle.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Daniel DeCapria</dc:creator><pubDate>Mon, 21 Apr 2025 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/dataops-towards-more-reliable-machine-learning-systems/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid><category>Artificial Intelligence Engineering</category><category>Machine Learning</category></item><item><title>Evaluating LLMs for Text Summarization: An Introduction</title><link>https://www.sei.cmu.edu/blog/evaluating-llms-for-text-summarization-introduction/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>Deploying LLMs without human supervision and evaluation can lead to significant errors. This post outlines the fundamentals of LLM evaluation for text summarization in high-stakes applications.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Shannon Gallagher, Swati Rallapalli, Tyler Brooks</dc:creator><pubDate>Mon, 07 Apr 2025 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/evaluating-llms-for-text-summarization-introduction/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid><category>Machine Learning</category></item><item><title>The Essential Role of AISIRT in Flaw and Vulnerability Management</title><link>https://www.sei.cmu.edu/blog/the-essential-role-of-aisirt-in-flaw-and-vulnerability-management/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>The SEI established the first Artificial Intelligence Security Incident Response Team (AISIRT) in 2023. This post discusses the role of AISIRT in coordinating flaws and vulnerabilities in AI systems.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Lauren McIlvenny, Vijay Sarvepalli</dc:creator><pubDate>Wed, 26 Mar 2025 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/the-essential-role-of-aisirt-in-flaw-and-vulnerability-management/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid><category>CERT/CC Vulnerabilities</category><category>Cybersecurity</category><category>AISIRT</category></item><item><title>Enhancing Machine Learning Assurance with Portend</title><link>https://www.sei.cmu.edu/blog/enhancing-machine-learning-assurance-with-portend/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>This post introduces Portend, a new open source toolset that simulates data drift in machine learning models and identifies the proper metrics to detect drift in production environments.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Jeffrey Hansen, Sebastián Echeverría, Lena Pons, Gabriel Moreno, Grace Lewis, Lihan Zhan</dc:creator><pubDate>Mon, 24 Mar 2025 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/enhancing-machine-learning-assurance-with-portend/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid><category>Software Assurance</category><category>Machine Learning</category></item><item><title>Introducing MLTE: A Systems Approach to Machine Learning Test and Evaluation</title><link>https://www.sei.cmu.edu/blog/introducing-mlte-systems-approach-to-machine-learning-test-and-evaluation/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>Machine learning systems are notoriously difficult to test. This post introduces Machine Learning Test and Evaluation (MLTE), a new process and tool to mitigate this problem and create safer, more reliable systems.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Alex Derr, Sebastián Echeverría, Katherine Maffey, Grace Lewis</dc:creator><pubDate>Mon, 17 Feb 2025 00:00:00 -0500</pubDate><guid>https://www.sei.cmu.edu/blog/introducing-mlte-systems-approach-to-machine-learning-test-and-evaluation/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid><category>Testing</category><category>Machine Learning</category></item><item><title>The Myth of Machine Learning Non-Reproducibility and Randomness for Acquisitions and Testing, Evaluation, Verification, and Validation</title><link>https://www.sei.cmu.edu/blog/the-myth-of-machine-learning-reproducibility-and-randomness-for-acquisitions-and-testing-evaluation-verification-and-validation/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>A reproducibility challenge faces machine learning (ML) systems today. This post explores  configurations that increase reproducibility and provides recommendations for these challenges.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Andrew Mellinger, Daniel Justice, Marissa Connor, Shannon Gallagher, Tyler Brooks</dc:creator><pubDate>Mon, 13 Jan 2025 00:00:00 -0500</pubDate><guid>https://www.sei.cmu.edu/blog/the-myth-of-machine-learning-reproducibility-and-randomness-for-acquisitions-and-testing-evaluation-verification-and-validation/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid><category>Acquisition Transformation</category><category>Testing</category><category>Machine Learning</category><category>Verification</category></item><item><title>Beyond Capable: Accuracy, Calibration, and Robustness in Large Language Models</title><link>https://www.sei.cmu.edu/blog/beyond-capable-accuracy-calibration-and-robustness-in-large-language-models/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>For any organization seeking to responsibly harness the potential of large language models, we present a holistic approach to LLM evaluation that goes beyond accuracy.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Matt Walsh, David Schulker, Shing-hon Lau</dc:creator><pubDate>Tue, 03 Dec 2024 00:00:00 -0500</pubDate><guid>https://www.sei.cmu.edu/blog/beyond-capable-accuracy-calibration-and-robustness-in-large-language-models/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>GenAI for Code Review of C++ and Java</title><link>https://www.sei.cmu.edu/blog/genai-for-code-review-of-c-and-java/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>Would ChatGPT-3.5 and ChatGPT-4o correctly identify errors in noncompliant code and correctly recognize compliant code as error-free?</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">David Schulker</dc:creator><pubDate>Mon, 18 Nov 2024 00:00:00 -0500</pubDate><guid>https://www.sei.cmu.edu/blog/genai-for-code-review-of-c-and-java/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>Introduction to MLOps: Bridging Machine Learning and Operations</title><link>https://www.sei.cmu.edu/blog/introduction-to-mlops-bridging-machine-learning-and-operations/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>Machine learning operations (MLOps) has emerged as a critical discipline in artificial intelligence and data science. This post introduces MLOps and its applications.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Daniel DeCapria</dc:creator><pubDate>Mon, 04 Nov 2024 00:00:00 -0500</pubDate><guid>https://www.sei.cmu.edu/blog/introduction-to-mlops-bridging-machine-learning-and-operations/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid><category>Artificial Intelligence Engineering</category><category>Machine Learning</category><category>Edge Computing</category></item><item><title>Measuring AI Accuracy with the AI Robustness (AIR) Tool</title><link>https://www.sei.cmu.edu/blog/measuring-ai-accuracy-with-the-ai-robustness-air-tool/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>Understanding your artificial intelligence (AI) system’s predictions can be challenging. In this post, SEI researchers discuss a new tool to help improve AI classifier performance.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Michael Konrad, Nicholas Testa, Linda Parker Gates, Crisanne Nolan, David Shepard, Julie Cohen, Andrew Mellinger, Suzanne Miller, Melissa Ludwick</dc:creator><pubDate>Mon, 30 Sep 2024 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/measuring-ai-accuracy-with-the-ai-robustness-air-tool/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid><category>Machine Learning</category><category>Artificial Intelligence</category></item><item><title>Weaknesses and Vulnerabilities in Modern AI: AI Risk, Cyber Risk, and Planning for Test and Evaluation</title><link>https://www.sei.cmu.edu/blog/weaknesses-and-vulnerabilities-in-modern-ai-ai-risk-cyber-risk-and-planning-for-test-and-evaluation/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>Modern AI systems pose consequential, poorly understood risks. This blog post explores strategies for framing test and evaluation practices based on a holistic approach to AI risk.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Bill Scherlis</dc:creator><pubDate>Mon, 12 Aug 2024 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/weaknesses-and-vulnerabilities-in-modern-ai-ai-risk-cyber-risk-and-planning-for-test-and-evaluation/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>Weaknesses and Vulnerabilities in Modern AI: Integrity, Confidentiality, and Governance</title><link>https://www.sei.cmu.edu/blog/weaknesses-and-vulnerabilities-in-modern-ai-integrity-confidentiality-and-governance/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>In the rush to develop AI, it is easy to overlook factors that increase risk. This post explores AI risk through the lens of confidentiality, governance, and integrity.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Bill Scherlis</dc:creator><pubDate>Mon, 05 Aug 2024 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/weaknesses-and-vulnerabilities-in-modern-ai-integrity-confidentiality-and-governance/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item><item><title>Weaknesses and Vulnerabilities in Modern AI: Why Security and Safety Are so Challenging</title><link>https://www.sei.cmu.edu/blog/weaknesses-and-vulnerabilities-in-modern-ai-why-security-and-safety-are-so-challenging/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</link><description>This post explores concepts of security and safety for neural-network-based AI, including ML and generative AI, as well as AI-specific challenges in developing safe and secure systems.</description><dc:creator xmlns:dc="http://purl.org/dc/elements/1.1/">Bill Scherlis</dc:creator><pubDate>Mon, 29 Jul 2024 00:00:00 -0400</pubDate><guid>https://www.sei.cmu.edu/blog/weaknesses-and-vulnerabilities-in-modern-ai-why-security-and-safety-are-so-challenging/?utm_source=blog&amp;utm_medium=rss&amp;utm_campaign=my_site_updates</guid></item></channel></rss>