LLMs Archives - gettectonic.com
ai trust layer

Gen AI Trust Layers

Addressing the Generative AI Production Gap with Trust Layers Despite the growing excitement around generative AI, only a small percentage of projects have successfully moved into production. A key barrier is the persistent concern over large language models (LLMs) generating hallucinations—responses that are inconsistent or completely disconnected from reality. To address these issues, organizations are increasingly adopting AI trust layers to enhance reliability and mitigate risk. Understanding the Challenge Generative AI models, like LLMs, are powerful tools trained on vast amounts of unstructured data, enabling them to answer questions and complete tasks based on text, documents, recordings, images, and videos. This capability has revolutionized the creation of chatbots, co-pilots, and even semi-autonomous agents. However, these models are inherently non-deterministic, meaning they don’t always produce consistent outputs. This lack of predictability leads to the infamous phenomenon of hallucination—what the National Institute of Standards and Technology (NIST) terms “confabulation.” While hallucination is a byproduct of how generative models function, its risks in mission-critical applications cannot be ignored. Implementing AI Trust Layers To address these challenges, organizations are turning to AI trust layers—frameworks designed to monitor and control generative AI behavior. These trust layers vary in implementation: Galileo: Building AI Trust from the Ground Up Galileo, founded in 2021 by Yash Sheth, Atindriyo Sanyal, and Vikram Chatterji, has emerged as a leader in developing AI trust solutions. Drawing on his decade of experience at Google building LLMs for speech recognition, Sheth recognized early on that non-deterministic AI systems needed robust trust frameworks to achieve widespread adoption in enterprise settings. The Need for Trust in Mission-Critical AI “Sheth explained: ‘Generative AI doesn’t give you the same answer every time. To mitigate risk in mission-critical tasks, you need a trust framework to ensure these models behave as expected in production.’ Enterprises, which prioritize privacy, security, and reputation, require this level of assurance before deploying LLMs at scale. Galileo’s Approach to Trust Layers Galileo’s AI trust layer is built on its proprietary foundation model, which evaluates the behavior of target LLMs. This approach is bolstered by metrics and real-time guardrails to block undesirable outcomes, such as hallucinations, data leaks, or harmful outputs. Key Products in Galileo’s Suite Sheth described the underlying technology: “Our evaluation foundation models are dependable, reliable, and scalable. They run continuously in production, ensuring bad outcomes are blocked in real time.” By combining these components, Galileo provides enterprises with a trust layer that gives them confidence in their generative AI applications, mirroring the reliability of traditional software systems. From Research to Real-World Impact Unlike vendors who quickly adapted traditional machine learning frameworks for generative AI, Galileo spent two years conducting research and developing its Generative AI Studio, launched in August 2023. This thorough approach has started to pay off: A Crucial Moment for AI Trust Layers As enterprises prepare to move generative AI experiments into production, trust layers are becoming essential. These frameworks address lingering concerns about the unpredictable nature of LLMs, allowing organizations to scale AI while minimizing risk. Sheth emphasized the stakes: “When mission-critical software starts becoming infused with AI, trust layers will define whether we progress or regress to the stone ages of software. That’s what’s holding back proof-of-concepts from reaching production.” With Galileo’s innovative approach, enterprises now have a path to unlock the full potential of generative AI—responsibly, securely, and at scale. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
pydanticai

Pydantic AI

The evaluation of agentic applications is most effective when integrated into the development process, rather than being an afterthought. For this to succeed, developers must be able to mock both internal and external dependencies of the agent being built. PydanticAI introduces a groundbreaking framework that supports dependency injection from the start, enabling developers to build agentic applications with an evaluation-driven approach. An architectural parallel can be drawn to the historic Krakow Cloth Hall, a structure refined over centuries through evaluation-driven enhancements. Similarly, PydanticAI allows developers to iteratively address challenges during development, ensuring optimal outcomes. Challenges in Developing GenAI Applications Developers of LLM-based applications face recurring challenges, which become significant during production deployment: To address non-determinism, developers must adopt evaluation-driven development, a method akin to test-driven development. This approach focuses on designing software with guardrails, real-time monitoring, and human oversight, accommodating systems that are only x% correct. The Promise of PydanticAI PydanticAI stands out as an agent framework that supports dependency injection, model-agnostic workflows, and evaluation-driven development. Its design is Pythonic and simplifies testing by allowing the injection of mock dependencies. For instance, in contrast to frameworks like Langchain, where dependency injection is cumbersome, PydanticAI streamlines this process, making the workflows more readable and efficient. Building an Evaluation-Driven Application with PydanticAI Example Use Case: Evaluating Mountain Data By employing tools like Wikipedia as a data source, the agent can fetch accurate mountain heights during production. For testing, developers can inject mocked responses, ensuring predictable outputs and faster development cycles. Advancing Agentic Applications with PydanticAI PydanticAI provides the building blocks for creating scalable, evaluation-driven GenAI applications. Its support for dependency injection, structured outputs, and model-agnostic workflows addresses core challenges, empowering developers to create robust and adaptive LLM-powered systems. This paradigm shift ensures that evaluation is seamlessly embedded into the development lifecycle, paving the way for more reliable and efficient agentic applications. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More

Einstein Service Agent

It’s been a little over a year since the global surge in GenAI chatbots, sparked by the excitement around ChatGPT. Since then, numerous vendors, both large and mid-sized, have invested heavily in the technology, and many users have already adopted AI-powered chatbots. The competition is intensifying, with CRM giant Salesforce releasing its own GenAI chatbot software, Einstein Service Agent. Einstein Service Agent, built on the Einstein 1 Platform, is Salesforce’s first fully autonomous AI agent. It interacts with large language models (LLMs) by analyzing the context of customer messages to determine the next actions. Utilizing GenAI, the agent generates conversational responses grounded in a company’s trusted business data, including Salesforce CRM data. Salesforce claims that service organizations can now significantly reduce the number of tedious inquiries that hinder productivity, allowing human agents to focus on more complex tasks. For customers, this means getting answers faster without waiting for human agents. Additionally, the service promises 24/7 availability for customer communication in natural language, with an easy handoff to human agents for more complicated issues. Businesses are increasingly turning to AI-based chatbots because, unlike traditional chatbots, they don’t rely on specific programmed queries and can understand context and nuance. Alongside Salesforce, other tech leaders like AWS and Google Cloud have released their own chatbots, such as Amazon Lex and Vertex AI, continuously enhancing their software. Recently, AWS updated its chatbot with the QnAIntent capability in Amazon Lex, allowing integration with a knowledge base in Amazon Bedrock. Similarly, Google released Vertex AI Agent Builder earlier this year, enabling organizations to build AI agents with no code, which can function together with one main agent and subagents. The AI arms race is just beginning, with more vendors developing software to meet market demands. For users, this means that while AI takes over many manual and tedious tasks, the primary challenge will be choosing the right vendor that best suits the needs and resources of their business. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More

From Generative AI to Agentic AI

Understanding the Coming Shift: From Generative AI to Agentic AI Large Language Models (LLMs), such as GPT, excel at generating text, answering questions, and supporting various tasks. However, they operate reactively, responding only to the input they receive based on learned patterns. LLMs cannot make decisions independently, adapt to new situations, or plan ahead. Agentic AI addresses these limitations. Unlike Generative AI, Agentic AI can set goals for itself, take initiative by itself, and learn from its experiences. It is proactive, capable of adjusting its actions over time, and can manage complex, evolving tasks that demand continuous problem-solving and decision-making. This transition from reactive to proactive AI unlocks exciting new possibilities across industries. In this insight, we will explore the differences between Agentic AI and Generative AI, examining their distinct impacts on technology and industries. Let’s begin by understanding what sets them apart. What is Agentic AI? Agentic AI refers to systems capable of autonomous decision-making and action to achieve specific goals. These systems go beyond generating content—they interact with their environments, respond to changes, and complete tasks with minimal human guidance. For example: What is Generative AI? Generative AI focuses on creating content—text, images, music, or video—by learning from large datasets to identify patterns, styles, or structures. For instance: Generative AI acts like a creative assistant, producing content based on what it has learned, but it remains reactive and task-specific. Key Differences in Workflows Agentic AI employs an iterative, cyclical workflow that includes stages like “Thinking/Research” and “Revision.” This adaptive process involves self-assessment, testing, and refinement, enabling the system to learn from each phase and tackle complex, evolving tasks effectively. Generative AI, in contrast, follows a linear, single-step workflow, moving directly from input to output without iterative improvements. While efficient for straightforward tasks, it lacks the ability to revisit or refine its results, limiting its effectiveness for dynamic or nuanced challenges. Characteristics of Agentic AI vs. Generative AI Feature Agentic AI Generative AI Autonomy Acts independently, making decisions and executing tasks. Requires human input to generate responses. Behavior Goal-directed, proactively working toward specific objectives. Task-oriented, reacting to immediate prompts. Adaptation and Learning Learns from experiences, adjusting actions dynamically. Operates based on pre-trained patterns, without learning. Decision-Making Handles complex decisions, weighing multiple outcomes. Makes basic decisions, selecting outputs based on patterns. Environmental Perception Understands and interacts with its surroundings. Lacks awareness of the physical environment. Case Study: Agentic Workflow in Action Andrew Ng highlighted the power of the Agentic Workflow in a coding task. Using the HumanEval benchmark, his team tested two approaches: This illustrates how iterative methods can enhance performance, even for older AI models. Conclusion As AI becomes increasingly integrated into our lives and workplaces, understanding the distinction between Generative AI and Agentic AI is essential. Generative AI has transformed tasks like content creation, offering immediate, reactive solutions. However, it remains limited to following instructions without true autonomy. Agentic AI represents a significant leap in technology. From chatbots to today. By setting goals, making decisions, and adapting in real-time, it can tackle complex, dynamic tasks without constant human oversight. Approaches like the Agentic Workflow further enhance AI’s capabilities, enabling iterative learning and continuous improvement. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Python-Based Reasoning

Python-Based Reasoning

Introducing a Python-Based Reasoning Engine for Deterministic AI As the demand for deterministic systems grows reviving foundational ideas for the age of large language models (LLMs) is here. The Challenge One of the critical issues with modern AI systems is establishing constraints around how they validate and reason about incoming data. As we increasingly rely on stochastic LLMs to process unstructured data, enforcing rules and guardrails becomes vital for ensuring reliability and consistency. The Solution Thus a company has developed a Python-based reasoning and validation framework inspired by Pydantic, designed to empower developers and non-technical domain experts to create sophisticated rule engines. The system is: By transforming Standard Operating Procedures (SOPs) and business guardrails into enforceable code, this symbolic reasoning framework addresses the need for structured, interpretable, and reliable AI systems. Key Features System Architecture The framework includes five core components: Types of Engines Case Studies 1. Validation Engine: Mining Company Compliance A mining company needed to validate employee qualifications against region-specific requirements. The system was configured to check rules such as minimum age and required certifications for specific roles. Input Example:Employee data and validation rules were modeled as JSON: jsonCopy code{ “employees”: [ { “name”: “Sarah”, “age”: 25, “documents”: [{ “type”: “safe_handling_at_work” }] }, { “name”: “John”, “age”: 17, “documents”: [{ “type”: “heavy_lifting” }] } ], “rules”: [ { “type”: “min_age”, “parameters”: { “min_age”: 18 } } ] } Output:Violations, such as “Minimum age must be 18,” were flagged immediately, enabling quick remediation. 2. Reasoning Engine: Solving the River Crossing Puzzle To showcase its capabilities, we modeled the classic river crossing puzzle, where a farmer must transport a wolf, a goat, and a cabbage across a river without leaving incompatible items together. Steps Taken: Enhanced Scenario:Adding a new rule—“Wolf cannot be left with a chicken”—created an unsolvable scenario. By introducing a compensatory rule, “Farmer can carry two items at once,” the system adapted and solved the puzzle with fewer moves. Developer Insights The system supports rapid iteration and debugging. For example, adding rules is as simple as defining Python classes: pythonCopy codeclass GoatCabbageRule(Rule): def evaluate(self, state): return not (state.goat == state.cabbage and state.farmer != state.goat) def get_description(self): return “Goat cannot be left alone with cabbage” Real-World Impact This framework accelerates development by enabling non-technical stakeholders to contribute to rule creation through natural language, with developers approving and implementing these rules. This process reduces development time by up to 5x and adapts seamlessly to varied use cases, from logistics to compliance. 🔔🔔 Follow us on LinkedIn 🔔🔔 Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
Salesforce Foundations

Salesforce Foundations

We are excited that Agentforce Service Agents are now live! Agentforce Service Agent is the autonomous conversational AI assistant to help your customers with their service and support needs. What does this mean for Foundations Customers?Salesforce Foundations is required for all customers in order to try or buy Agentforce. Additionally, customers who have Foundations can try Agentforce Agents for free with a limited number of credits to test a use case or deploy a proof of concept. Salesforce Foundations is not a product or add-on. It’s a multi-cloud feature set that will be added to Sales and Service Cloud — no integration needed, with no additional upfront cost for our customers. It includes foundational features from Sales, Service, Marketing, Commerce, and Data Cloud. Salesforce Foundations provides a 360-degree view of your customer relationships across sales, service, marketing, and commerce through integrated applications and unified data. It also boosts productivity with streamlined, visually friendly user interface improvements, that you can turn on or off per your requirements. If you’re a Salesforce Sales Cloud or Service Cloud customer, you’ve become accustomed to the power, convenience, and full-featured functionality of our trusted CRM. Adding the additional functionality and engagement capabilities of a new Salesforce Cloud is exciting, but it’s also a big change for your organization to consider when you’re not sure about the value it brings. So, what if you could use essential features in the most popular Salesforce Clouds and turn them on when you’re ready? Now you can with Salesforce Foundations. Salesforce Foundations is a new, no-cost addition to your existing CRM that equips you to expand your business reach. The suite gives Salesforce customers on Enterprise, Unlimited, and Einstein 1 editions the power of Data Cloud, and access to essential Salesforce sales, service, Agentforce, marketing, and commerce capabilities. This suite is built into your existing CRM, and provides new functionality to give you a more robust 360-degree view of your customers. This chart shows the Salesforce Foundations features you get with your current Sales Cloud or Service Cloud package. You get Sales for Salesforce Foundations You get Service for Salesforce Foundations You get Marketing for Salesforce Foundations You get Commerce for Salesforce Foundations You get Data Cloud for Salesforce Foundations You get Agentforce for Salesforce Foundations If you already have Sales Cloud * Yes Yes Yes Yes Yes If you already have Service Cloud Yes * Yes Yes Yes Yes If you already have Sales & Service Clouds * * Yes Yes Yes Yes *Your current Salesforce product. Benefits of Salesforce Foundations The features you get with Salesforce Foundations open doors to all sorts of new ways your teams can work more efficiently and engage with your customers on a more personal level. The benefits listed below are only a few of the ways Salesforce Foundations can help your business grow and thrive. Check out Discover Salesforce Foundations to see the full list of capabilities included with Salesforce Foundations. With Salesforce Foundations, your organization benefits from: Sales features that help you take care of your entire sales pipeline, from prospecting to closing. You can manage your leads, opportunities, accounts, and contacts in the preconfigured Sales Console. Service features that make it easy to provide proactive, personalized support to your customers through the preconfigured Service Console. Omni-channel case routing makes sure the most qualified agents work each case, Knowledge Management helps agents provide accurate and relevant help articles to customers, and macros help agents complete repetitive tasks with a single click. Agentforce brings the power of conversational AI to your business. Try out an intelligent, trusted, and customizable AI agent and help your users get more done with Salesforce. Agentforce’s autonomous apps use LLMs and context to assist customers and human agents. Marketing features that allow you to join data from disparate sources, better understand and analyze your customers, and choose how to connect with your audiences. You can create customized marketing campaigns powered by Salesforce Flows to send at the right time. Commerce features that help boost sales with a Direct to Customer (D2C) online storefront. You can define customer experiences like search, carts, and checkout. Pay Now lets you generate secure payment links for customers when opportunities close, so you get paid faster. Data Cloud functionality that creates unified profiles by aggregating data from all of your data sources into a single view so you can better understand your customers. Create customer segments to more accurately target campaigns, analyze your customers, and manage consent data. Data Cloud also powers features so you can send online store order confirmation emails and marketing messages. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
Agentforce Redefines Generative AI

Agentforce Redefines Generative AI

Agentforce: Redefining Generative AI in Salesforce Many Dreamforce attendees who expected to hear about Einstein Copilot were surprised when Salesforce introduced Agentforce just a week before the conference. While it might seem like a rebranding of Copilot, Agentforce marks a significant evolution by enabling more autonomous agents that go beyond summarizing or generating content to perform specific actions. Here’s a breakdown of the transition and what it means for Salesforce users: Key Vocabulary Updates How Agentforce Works Agents take user input, known as an “utterance,” and translate it into actionable steps based on predefined configurations. This allows the system to enhance performance over time while delivering responses tailored to user needs. Understanding Agentforce 1. Topics: Organizing Agent Capabilities Agentforce introduces “Topics,” a new layer of organization that categorizes actions by business function. When a user provides an utterance, the agent identifies the relevant topic first, then determines the best actions to address it. 2. Actions: What Agents Can Do Actions remain largely unchanged from Einstein Copilot. These are tasks agents perform to execute plans. 3. Prompts: The Key to Better Results LLMs rely on prompts to generate outputs, and crafting effective prompts is essential for reducing irrelevant responses and optimizing agent behavior. How Generative AI Enhances Salesforce Agentforce unlocks several benefits across productivity, personalization, standardization, and efficiency: Implementing Agentforce: Tips for Success Getting Started Start by using standard Agent actions. These out-of-the-box tools, such as opportunity summarization or close plan creation, provide a strong foundation. You can make minor adjustments to optimize their performance before diving into more complex custom actions. Testing and Iteration Testing AI agents is different from traditional workflows. Agents must handle various phrasing of the same user request (utterances) while maintaining consistency in responses. The Future of Salesforce with Agentforce As you gain expertise in planning, developing, testing, and deploying Agentforce actions, you’ll unlock new possibilities for transforming your Salesforce experience. With generative AI tools like Agentforce, Salesforce evolves from a traditional point-and-click interface into an intelligent, agent-driven platform with streamlined, conversational workflows. This isn’t just an upgrade — it’s the foundation for reimagining how businesses interact with their CRM in an AI-assisted world. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
Statement Accuracy Prediction based on Language Model Activations

Statement Accuracy Prediction based on Language Model Activations

When users first began interacting with ChatGPT, they noticed an intriguing behavior: the model would often reverse its stance when told it was wrong. This raised concerns about the reliability of its outputs. How can users trust a system that appears to contradict itself? Recent research has revealed that large language models (LLMs) not only generate inaccurate information (often referred to as “hallucinations”) but are also aware of their inaccuracies. Despite this awareness, these models proceed to present their responses confidently. Unveiling LLM Awareness of Hallucinations Researchers discovered this phenomenon by analyzing the internal mechanisms of LLMs. Whenever an LLM generates a response, it transforms the input query into a numerical representation and performs a series of computations before producing the output. At intermediate stages, these numerical representations are called “activations.” These activations contain significantly more information than what is reflected in the final output. By scrutinizing these activations, researchers can identify whether the LLM “knows” its response is inaccurate. A technique called SAPLMA (Statement Accuracy Prediction based on Language Model Activations) has been developed to explore this capability. SAPLMA examines the internal activations of LLMs to predict whether their outputs are truthful or not. Why Do Hallucinations Occur? LLMs function as next-word prediction models. Each word is selected based on its likelihood given the preceding words. For example, starting with “I ate,” the model might predict the next words as follows: The issue arises when earlier predictions constrain subsequent outputs. Once the model commits to a word, it cannot go back to revise its earlier choice. For instance: In another case: This mechanism reveals how the constraints of next-word prediction can lead to hallucinations, even when the model “knows” it is generating an incorrect response. Detecting Inaccuracies with SAPLMA To investigate whether an LLM recognizes its own inaccuracies, researchers developed the SAPLMA method. Here’s how it works: The classifier itself is a simple neural network with three dense layers, culminating in a binary output that predicts the truthfulness of the statement. Results and Insights The SAPLMA method achieved an accuracy of 60–80%, depending on the topic. While this is a promising result, it is not perfect and has notable limitations. For example: However, if LLMs can learn to detect inaccuracies during the generation process, they could potentially refine their outputs in real time, reducing hallucinations and improving reliability. The Future of Error Mitigation in LLMs The SAPLMA method represents a step forward in understanding and mitigating LLM errors. Accurate classification of inaccuracies could pave the way for models that can self-correct and produce more reliable outputs. While the current limitations are significant, ongoing research into these methods could lead to substantial improvements in LLM performance. By combining techniques like SAPLMA with advancements in LLM architecture, researchers aim to build models that are not only aware of their errors but capable of addressing them dynamically, enhancing both the accuracy and trustworthiness of AI systems. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More

Autonomy, Architecture, and Action

Redefining AI Agents: Autonomy, Architecture, and Action AI agents are reshaping how technology interacts with us and executes tasks. Their mission? To reason, plan, and act independently—following instructions, making autonomous decisions, and completing actions, often without user involvement. These agents adapt to new information, adjust in real time, and pursue their objectives autonomously. This evolution in agentic AI is revolutionizing how goals are accomplished, ushering in a future of semi-autonomous technology. At their foundation, AI agents rely on one or more large language models (LLMs). However, designing agents is far more intricate than building chatbots or generative assistants. While traditional AI applications often depend on user-driven inputs—such as prompt engineering or active supervision—agents operate autonomously. Core Principles of Agentic AI Architectures To enable autonomous functionality, agentic AI systems must incorporate: Essential Infrastructure for AI Agents Building and deploying agentic AI systems requires robust software infrastructure that supports: Agent Development Made Easier with Langflow and Astra DB Langflow simplifies the development of agentic applications with its visual IDE. It integrates with Astra DB, which combines vector and graph capabilities for ultra-low latency data access. This synergy accelerates development by enabling: Transforming Autonomy into Action Agentic AI is fundamentally changing how tasks are executed by empowering systems to act autonomously. By leveraging platforms like Astra DB and Langflow, organizations can simplify agent design and deploy scalable, effective AI applications. Start building the next generation of AI-powered autonomy today. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
Apple's Privacy Changes: A Call for Email Marketing Innovation

Liar Liar Apple on Fire

Apple Developing Update After AI System Generates Inaccurate News Summaries Apple is working on a software update to address inaccuracies generated by its Apple Intelligence system after multiple instances of false news summaries were reported. The BBC first alerted Apple in mid-December to significant errors in the system, including a fabricated summary that falsely attributed a statement to BBC News. The summary suggested Luigi Mangione, accused of killing United Healthcare CEO Brian Thompson, had shot himself, a claim entirely unsubstantiated. Other publishers, such as ProPublica, also raised concerns about Apple Intelligence producing misleading summaries. While Apple did not respond immediately to the BBC’s December report, it issued a statement after pressure mounted from groups like the National Union of Journalists and Reporters Without Borders, both of which called for the removal of Apple Intelligence. Apple assured stakeholders it is working to refine the technology. A Widespread AI Issue: Hallucinations Apple joins the ranks of other AI vendors struggling with generative AI hallucinations—instances where AI produces false or misleading information. In October 2024, Perplexity AI faced a lawsuit from Dow Jones & Co. and the New York Post over fabricated news content attributed to their publications. Similarly, Google had to improve its AI summaries after providing users with inaccurate information. On January 16, Apple temporarily disabled AI-generated summaries for news apps on iPhone, iPad, and Mac devices. The Core Problem: AI Hallucination Chirag Shah, a professor of Information Science at the University of Washington, emphasized that hallucination is inherent to the way large language models (LLMs) function. “The nature of AI models is to generate, synthesize, and summarize, which makes them prone to mistakes,” Shah explained. “This isn’t something you can debug easily—it’s intrinsic to how LLMs operate.” While Apple plans to introduce an update that clearly labels summaries as AI-generated, Shah believes this measure falls short. “Most people don’t understand how these headlines or summaries are created. The responsible approach is to pause the technology until it’s better understood and mitigation strategies are in place,” he said. Legal and Brand Implications for Apple The hallucinated summaries pose significant reputational and legal risks for Apple, according to Michael Bennett, an AI adviser at Northeastern University. Before launching Apple Intelligence, the company was perceived as lagging in the AI race. The release of this system was intended to position Apple as a leader. Instead, the inaccuracies have damaged its credibility. “This type of hallucinated summarization is both an embarrassment and a serious legal liability,” Bennett said. “These errors could form the basis for defamation claims, as Apple Intelligence misattributes false information to reputable news sources.” Bennett criticized Apple’s seemingly minimal response. “It’s surprising how casual Apple’s reaction has been. This is a major issue for their brand and could expose them to significant legal consequences,” he added. Opportunity for Publishers The incident highlights the need for publishers to protect their interests when partnering with AI vendors like Apple and Google. Publishers should demand stronger safeguards to prevent false attributions and negotiate new contractual clauses to minimize brand risk. “This is an opportunity for publishers to lead the charge, pushing AI companies to refine their models or stop attributing false summaries to news sources,” Bennett said. He suggested legal action as a potential recourse if vendors fail to address these issues. Potential Regulatory Action The Federal Trade Commission (FTC) may also scrutinize the issue, as consumers paying for products like iPhones with AI capabilities could argue they are not receiving the promised service. However, Bennett believes Apple will likely act to resolve the problem before regulatory involvement becomes necessary. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
Generative AI Energy Consumption Rises

Generative AI Tools

Generative AI Tools: A Comprehensive Overview of Emerging Capabilities The widespread adoption of generative AI services like ChatGPT has sparked immense interest in leveraging these tools for practical enterprise applications. Today, nearly every enterprise app integrates generative AI capabilities to enhance functionality and efficiency. A broad range of AI, data science, and machine learning tools now support generative AI use cases. These tools assist in managing the AI lifecycle, governing data, and addressing security and privacy concerns. While such capabilities also aid in traditional AI development, this discussion focuses on tools specifically designed for generative AI. Not all generative AI relies on large language models (LLMs). Emerging techniques generate images, videos, audio, synthetic data, and translations using methods such as generative adversarial networks (GANs), diffusion models, variational autoencoders, and multimodal approaches. Here is an in-depth look at the top categories of generative AI tools, their capabilities, and notable implementations. It’s worth noting that many leading vendors are expanding their offerings to support multiple categories through acquisitions or integrated platforms. Enterprises may want to explore comprehensive platforms when planning their generative AI strategies. 1. Foundation Models and Services Generative AI tools increasingly simplify the development and responsible use of LLMs, initially pioneered through transformer-based approaches by Google researchers in 2017. 2. Cloud Generative AI Platforms Major cloud providers offer generative AI platforms to streamline development and deployment. These include: 3. Use Case Optimization Tools Foundation models often require optimization for specific tasks. Enterprises use tools such as: 4. Quality Assurance and Hallucination Mitigation Hallucination detection tools address the tendency of generative models to produce inaccurate or misleading information. Leading tools include: 5. Prompt Engineering Tools Prompt engineering tools optimize interactions with LLMs and streamline testing for bias, toxicity, and accuracy. Examples include: 6. Data Aggregation Tools Generative AI tools have evolved to handle larger data contexts efficiently: 7. Agentic and Autonomous AI Tools Developers are creating tools to automate interactions across foundation models and services, paving the way for autonomous AI. Notable examples include: 8. Generative AI Cost Optimization Tools These tools aim to balance performance, accuracy, and cost effectively. Martian’s Model Router is an early example, while traditional cloud cost optimization platforms are expected to expand into this area. Generative AI tools are rapidly transforming enterprise applications, with foundational, cloud-based, and domain-specific solutions leading the way. By addressing challenges like accuracy, hallucination, and cost, these tools unlock new potential across industries and use cases, enabling enterprises to stay ahead in the AI-driven landscape. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
Scope of Generative AI

Exploring Generative AI

Like most employees at most companies, I wear a few different hats around Tectonic. Whether I’m building a data model, creating and scheduing an email campaign, standing up a platform generative AI is always at my fingertips. At my very core, I’m a marketer. Have been for so long I do it without eveven thinking. Or at least, everyuthing I do has a hat tip to its future marketing needs. Today I want to share some of the AI content generators I’ve been using, am looking to use, or just heard about. But before we rip into the insight, here’s a primer. Types of AI Content Generators ChatGPT, a powerful AI chatbot, drew significant attention upon its November 2022 release. While the GPT-3 language model behind it had existed for some time, ChatGPT made this technology accessible to nontechnical users, showcasing how AI can generate content. Over two years later, numerous AI content generators have emerged to cater to diverse use cases. This rapid development raises questions about the technology’s impact on work. Schools are grappling with fears of plagiarism, while others are embracing AI. Legal debates about copyright and digital media authenticity continue. President Joe Biden’s October 2023 executive order addressed AI’s risks and opportunities in areas like education, workforce, and consumer privacy, underscoring generative AI’s transformative potential. What is AI-Generated Content? AI-generated content, also known as generative AI, refers to algorithms that automatically create new content across digital media. These algorithms are trained on extensive datasets and require minimal user input to produce novel outputs. For instance, ChatGPT sets a standard for AI-generated content. Based on GPT-4o, it processes text, images, and audio, offering natural language and multimodal capabilities. Many other generative AI tools operate similarly, leveraging large language models (LLMs) and multimodal frameworks to create diverse outputs. What are the Different Types of AI-Generated Content? AI-generated content spans multiple media types: Despite their varied outputs, most generative AI systems are built on advanced LLMs like GPT-4 and Google Gemini. These multimodal models process and generate content across multiple formats, with enhanced capabilities evolving over time. How Generative AI is Used Generative AI applications span industries: These tools often combine outputs from various media for complex, multifaceted projects. AI Content Generators AI content generators exist across various media. Below are good examples organized by gen ai type: Written Content Generators Image Content Generators Music Content Generators Code Content Generators Other AI Content Generators These tools showcase how AI-powered content generation is revolutionizing industries, making content creation faster and more accessible. I do hope you will comment below on your favorites, other AI tools not showcased above, or anything else AI-related that is on your mind. Written by Tectonic’s Marketing Operations Director, Shannan Hearne. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
From Chatbots to Agentic AI

From Chatbots to Agentic AI

The transition from LLM-powered chatbots to agentic systems, or agentic AI, can be summed up by the old saying: “Less talk, more action.” Keeping up with advancements in AI can be overwhelming, especially when managing an existing business. The speed and complexity of innovation can make it feel like the first day of school all over again. This insight offers a comprehensive look at AI agents, their components, and key characteristics. The introductory section breaks down the elements that form the term “AI agent,” providing a clear definition. After establishing this foundation, we explore the evolution of LLM applications, particularly the shift from traditional chatbots to agentic systems. The goal is to understand why AI agents are becoming increasingly vital in AI development and how they differ from LLM-powered chatbots. By the end of this guide, you will have a deeper understanding of AI agents, their potential applications, and their impact on organizational workflows. For those of you with a technical background who prefer to get hands-on, click here for the best repository for AI developers and builders. What is an AI Agent? Components of AI Agents To understand the term “AI agent,” we need to examine its two main components. First, let’s consider artificial intelligence, or AI. Artificial Intelligence (AI) refers to non-biological intelligence that mimics human cognition to perform tasks traditionally requiring human intellect. Through machine learning and deep learning techniques, algorithms—especially neural networks—learn patterns from data. AI systems are used for tasks such as detection, classification, and prediction, with content generation becoming a prominent domain due to transformer-based models. These systems can match or exceed human performance in specific scenarios. The second component is “agent,” a term commonly used in both technology and human contexts. In computer science, an agent refers to a software entity with environmental awareness, able to perceive and act within its surroundings. A computational agent typically has the ability to: In human contexts, an agent is someone who acts on behalf of another person or organization, making decisions, gathering information, and facilitating interactions. They often play intermediary roles in transactions and decision-making. To define an AI agent, we combine these two perspectives: it is a computational entity with environmental awareness, capable of perceiving inputs, acting with tools, and processing information using foundation models backed by both long-term and short-term memory. Key Components and Characteristics of AI Agents From LLMs to AI Agents Now, let’s take a step back and understand how we arrived at the concept of AI agents, particularly by looking at how LLM applications have evolved. The shift from traditional chatbots to LLM-powered applications has been rapid and transformative. Form Factor Evolution of LLM Applications Traditional Chatbots to LLM-Powered Chatbots Traditional chatbots, which existed before generative AI, were simpler and relied on heuristic responses: “If this, then that.” They followed predefined rules and decision trees to generate responses. These systems had limited interactivity, with the fallback option of “Speak to a human” for complex scenarios. LLM-Powered Chatbots The release of OpenAI’s ChatGPT on November 30, 2022, marked the introduction of LLM-powered chatbots, fundamentally changing the game. These chatbots, like ChatGPT, were built on GPT-3.5, a large language model trained on massive datasets. Unlike traditional chatbots, LLM-powered systems can generate human-like responses, offering a much more flexible and intelligent interaction. However, challenges remained. LLM-powered chatbots struggled with personalization and consistency, often generating plausible but incorrect information—a phenomenon known as “hallucination.” This led to efforts in grounding LLM responses through techniques like retrieval-augmented generation (RAG). RAG Chatbots RAG is a method that combines data retrieval with LLM generation, allowing systems to access real-time or proprietary data, improving accuracy and relevance. This hybrid approach addresses the hallucination problem, ensuring more reliable outputs. LLM-Powered Chatbots to AI Agents As LLMs expanded, their abilities grew more sophisticated, incorporating advanced reasoning, multi-step planning, and the use of external tools (function calling). Tool use refers to an LLM’s ability to invoke specific functions, enabling it to perform more complex tasks. Tool-Augmented LLMs and AI Agents As LLMs became tool-augmented, the emergence of AI agents followed. These agents integrate reasoning, planning, and tool use into an autonomous, goal-driven system that can operate iteratively within a dynamic environment. Unlike traditional chatbot interfaces, AI agents leverage a broader set of tools to interact with various systems and accomplish tasks. Agentic Systems Agentic systems—computational architectures that include AI agents—embody these advanced capabilities. They can autonomously interact with systems, make decisions, and adapt to feedback, forming the foundation for more complex AI applications. Components of an AI Agent AI agents consist of several key components: Characteristics of AI Agents AI agents are defined by the following traits: Conclusion AI agents represent a significant leap from traditional chatbots, offering greater autonomy, complexity, and interactivity. However, the term “AI agent” remains fluid, with no universal industry standard. Instead, it exists on a continuum, with varying degrees of autonomy, adaptability, and proactive behavior defining agentic systems. Value and Impact of AI Agents The key benefits of AI agents lie in their ability to automate manual processes, reduce decision-making burdens, and enhance workflows in enterprise environments. By “agentifying” repetitive tasks, AI agents offer substantial productivity gains and the potential to transform how businesses operate. As AI agents evolve, their applications will only expand, driving new efficiencies and enabling organizations to leverage AI in increasingly sophisticated ways. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
ThoughtSpot AI agent Spotter enables conversational BI

ThoughtSpot AI agent Spotter enables conversational BI

ThoughtSpot Unveils Spotter: A Generative AI-Powered Data Agent ThoughtSpot, a leading analytics vendor, has launched Spotter, an advanced generative AI-powered agent designed to revolutionize how users interact with data. Spotter enables conversational data exploration, contextual understanding, and autonomous analysis, making it a significant leap forward in the analytics landscape. Spotter’s Role in ThoughtSpot’s Evolution Spotter replaces Sage, ThoughtSpot’s earlier generative AI-powered interface, which debuted in March 2023. Despite moving from private to public preview and gaining new capabilities, Sage never reached general availability. Spotter is now generally available for ThoughtSpot Analytics, while its embedded version is in beta testing. Unlike earlier AI tools that focused on question-and-answer interactions, such as Sage and Microsoft’s copilots, Spotter takes the concept further by integrating contextual awareness and autonomous decision-making. Spotter doesn’t just respond to queries; it suggests follow-up questions, identifies anomalies, and provides proactive insights, functioning more like a virtual analyst than a reactive chatbot. Key Features of Spotter Spotter is built to enhance productivity and insight generation through the following capabilities: Generative AI’s Growing Impact on BI ThoughtSpot has long aimed to make analytics accessible to non-technical users through natural language search. However, previous NLP tools often required users to learn specific vocabularies, limiting widespread adoption. Generative AI bridges this gap. By leveraging extensive vocabularies and LLM technology, tools like Spotter enable users of all skill levels to access and analyze data effortlessly. Spotter stands out with its ability to deliver proactive insights, identify trends, and adapt to user behavior, enhancing the decision-making process. Expert Perspectives on Spotter Donald Farmer, founder of TreeHive Strategy, highlighted Spotter’s autonomy as a game-changer: “Spotter is a big move forward for ThoughtSpot and AI. The natural language interface is more conversational, but the key advantage is its autonomous analysis, which identifies trends and insights without users needing to ask.” Mike Leone, an analyst at TechTarget’s Enterprise Strategy Group, emphasized Spotter’s ability to adapt to users: “Spotter’s ability to deliver personalized and contextually relevant responses is critical for organizations pursuing generative AI initiatives. This goes a long way in delivering unique value across a business.” Farmer also pointed to Spotter’s embedded capabilities, noting its growing appeal as an embedded analytics solution integrated with productivity tools like Salesforce and ServiceNow. Competitive Positioning Spotter aligns ThoughtSpot with other vendors embracing agentic AI in analytics. Google recently introduced Conversational Analytics in Looker, and Salesforce’s Tableau platform now includes Tableau Agent. ThoughtSpot’s approach builds on its core strength in search-based analytics while expanding into generative AI-driven capabilities. Leone observed: “ThoughtSpot is right in line with the market in delivering an agentic experience and is laying the groundwork for broader AI functionality over time.” A Step Toward the Future of Analytics With Spotter, ThoughtSpot is redefining the role of AI in business intelligence. The tool combines conversational ease, proactive insights, and seamless integration, empowering users to make data-driven decisions more efficiently. As generative AI continues to evolve, tools like Spotter demonstrate how businesses can unlock the full potential of their data. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
gettectonic.com