Large Language Model Archives - gettectonic.com

From Generative AI to Agentic AI

Understanding the Coming Shift: From Generative AI to Agentic AI Large Language Models (LLMs), such as GPT, excel at generating text, answering questions, and supporting various tasks. However, they operate reactively, responding only to the input they receive based on learned patterns. LLMs cannot make decisions independently, adapt to new situations, or plan ahead. Agentic AI addresses these limitations. Unlike Generative AI, Agentic AI can set goals for itself, take initiative by itself, and learn from its experiences. It is proactive, capable of adjusting its actions over time, and can manage complex, evolving tasks that demand continuous problem-solving and decision-making. This transition from reactive to proactive AI unlocks exciting new possibilities across industries. In this insight, we will explore the differences between Agentic AI and Generative AI, examining their distinct impacts on technology and industries. Let’s begin by understanding what sets them apart. What is Agentic AI? Agentic AI refers to systems capable of autonomous decision-making and action to achieve specific goals. These systems go beyond generating content—they interact with their environments, respond to changes, and complete tasks with minimal human guidance. For example: What is Generative AI? Generative AI focuses on creating content—text, images, music, or video—by learning from large datasets to identify patterns, styles, or structures. For instance: Generative AI acts like a creative assistant, producing content based on what it has learned, but it remains reactive and task-specific. Key Differences in Workflows Agentic AI employs an iterative, cyclical workflow that includes stages like “Thinking/Research” and “Revision.” This adaptive process involves self-assessment, testing, and refinement, enabling the system to learn from each phase and tackle complex, evolving tasks effectively. Generative AI, in contrast, follows a linear, single-step workflow, moving directly from input to output without iterative improvements. While efficient for straightforward tasks, it lacks the ability to revisit or refine its results, limiting its effectiveness for dynamic or nuanced challenges. Characteristics of Agentic AI vs. Generative AI Feature Agentic AI Generative AI Autonomy Acts independently, making decisions and executing tasks. Requires human input to generate responses. Behavior Goal-directed, proactively working toward specific objectives. Task-oriented, reacting to immediate prompts. Adaptation and Learning Learns from experiences, adjusting actions dynamically. Operates based on pre-trained patterns, without learning. Decision-Making Handles complex decisions, weighing multiple outcomes. Makes basic decisions, selecting outputs based on patterns. Environmental Perception Understands and interacts with its surroundings. Lacks awareness of the physical environment. Case Study: Agentic Workflow in Action Andrew Ng highlighted the power of the Agentic Workflow in a coding task. Using the HumanEval benchmark, his team tested two approaches: This illustrates how iterative methods can enhance performance, even for older AI models. Conclusion As AI becomes increasingly integrated into our lives and workplaces, understanding the distinction between Generative AI and Agentic AI is essential. Generative AI has transformed tasks like content creation, offering immediate, reactive solutions. However, it remains limited to following instructions without true autonomy. Agentic AI represents a significant leap in technology. From chatbots to today. By setting goals, making decisions, and adapting in real-time, it can tackle complex, dynamic tasks without constant human oversight. Approaches like the Agentic Workflow further enhance AI’s capabilities, enabling iterative learning and continuous improvement. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more Top Ten Reasons Why Tectonic Loves the Cloud The Cloud is Good for Everyone – Why Tectonic loves the cloud You don’t need to worry about tracking licenses. Read more

Read More
Python-Based Reasoning

Python-Based Reasoning

Introducing a Python-Based Reasoning Engine for Deterministic AI As the demand for deterministic systems grows reviving foundational ideas for the age of large language models (LLMs) is here. The Challenge One of the critical issues with modern AI systems is establishing constraints around how they validate and reason about incoming data. As we increasingly rely on stochastic LLMs to process unstructured data, enforcing rules and guardrails becomes vital for ensuring reliability and consistency. The Solution Thus a company has developed a Python-based reasoning and validation framework inspired by Pydantic, designed to empower developers and non-technical domain experts to create sophisticated rule engines. The system is: By transforming Standard Operating Procedures (SOPs) and business guardrails into enforceable code, this symbolic reasoning framework addresses the need for structured, interpretable, and reliable AI systems. Key Features System Architecture The framework includes five core components: Types of Engines Case Studies 1. Validation Engine: Mining Company Compliance A mining company needed to validate employee qualifications against region-specific requirements. The system was configured to check rules such as minimum age and required certifications for specific roles. Input Example:Employee data and validation rules were modeled as JSON: jsonCopy code{ “employees”: [ { “name”: “Sarah”, “age”: 25, “documents”: [{ “type”: “safe_handling_at_work” }] }, { “name”: “John”, “age”: 17, “documents”: [{ “type”: “heavy_lifting” }] } ], “rules”: [ { “type”: “min_age”, “parameters”: { “min_age”: 18 } } ] } Output:Violations, such as “Minimum age must be 18,” were flagged immediately, enabling quick remediation. 2. Reasoning Engine: Solving the River Crossing Puzzle To showcase its capabilities, we modeled the classic river crossing puzzle, where a farmer must transport a wolf, a goat, and a cabbage across a river without leaving incompatible items together. Steps Taken: Enhanced Scenario:Adding a new rule—“Wolf cannot be left with a chicken”—created an unsolvable scenario. By introducing a compensatory rule, “Farmer can carry two items at once,” the system adapted and solved the puzzle with fewer moves. Developer Insights The system supports rapid iteration and debugging. For example, adding rules is as simple as defining Python classes: pythonCopy codeclass GoatCabbageRule(Rule): def evaluate(self, state): return not (state.goat == state.cabbage and state.farmer != state.goat) def get_description(self): return “Goat cannot be left alone with cabbage” Real-World Impact This framework accelerates development by enabling non-technical stakeholders to contribute to rule creation through natural language, with developers approving and implementing these rules. This process reduces development time by up to 5x and adapts seamlessly to varied use cases, from logistics to compliance. 🔔🔔 Follow us on LinkedIn 🔔🔔 Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Agentforce Redefines Generative AI

Agentforce Redefines Generative AI

Agentforce: Redefining Generative AI in Salesforce Many Dreamforce attendees who expected to hear about Einstein Copilot were surprised when Salesforce introduced Agentforce just a week before the conference. While it might seem like a rebranding of Copilot, Agentforce marks a significant evolution by enabling more autonomous agents that go beyond summarizing or generating content to perform specific actions. Here’s a breakdown of the transition and what it means for Salesforce users: Key Vocabulary Updates How Agentforce Works Agents take user input, known as an “utterance,” and translate it into actionable steps based on predefined configurations. This allows the system to enhance performance over time while delivering responses tailored to user needs. Understanding Agentforce 1. Topics: Organizing Agent Capabilities Agentforce introduces “Topics,” a new layer of organization that categorizes actions by business function. When a user provides an utterance, the agent identifies the relevant topic first, then determines the best actions to address it. 2. Actions: What Agents Can Do Actions remain largely unchanged from Einstein Copilot. These are tasks agents perform to execute plans. 3. Prompts: The Key to Better Results LLMs rely on prompts to generate outputs, and crafting effective prompts is essential for reducing irrelevant responses and optimizing agent behavior. How Generative AI Enhances Salesforce Agentforce unlocks several benefits across productivity, personalization, standardization, and efficiency: Implementing Agentforce: Tips for Success Getting Started Start by using standard Agent actions. These out-of-the-box tools, such as opportunity summarization or close plan creation, provide a strong foundation. You can make minor adjustments to optimize their performance before diving into more complex custom actions. Testing and Iteration Testing AI agents is different from traditional workflows. Agents must handle various phrasing of the same user request (utterances) while maintaining consistency in responses. The Future of Salesforce with Agentforce As you gain expertise in planning, developing, testing, and deploying Agentforce actions, you’ll unlock new possibilities for transforming your Salesforce experience. With generative AI tools like Agentforce, Salesforce evolves from a traditional point-and-click interface into an intelligent, agent-driven platform with streamlined, conversational workflows. This isn’t just an upgrade — it’s the foundation for reimagining how businesses interact with their CRM in an AI-assisted world. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Agentic AI is Here

On Premise Gen AI

In 2025, enterprises transitioning generative AI (GenAI) into production after years of experimentation are increasingly considering on-premises deployment as a cost-effective alternative to the cloud. Since OpenAI ignited the AI revolution in late 2022, organizations have tested large language models powering GenAI services on platforms like AWS, Microsoft Azure, and Google Cloud. These experiments demonstrated GenAI’s potential to enhance business operations while exposing the substantial costs of cloud usage. To avoid difficult conversations with CFOs about escalating cloud expenses, CIOs are exploring on-premises AI as a financially viable solution. Advances in software from startups and packaged infrastructure from vendors such as HPE and Dell are making private data centers an attractive option for managing costs. A survey conducted by Menlo Ventures in late 2024 found that 47% of U.S. enterprises with at least 50 employees were developing GenAI solutions in-house. Similarly, Informa TechTarget’s Enterprise Strategy Group reported a rise in enterprises considering on-premises and public cloud equally for new applications—from 37% in 2024 to 45% in 2025. This shift is reflected in hardware sales. HPE reported a 16% revenue increase in AI systems, reaching $1.5 billion in Q4 2024. During the same period, Dell recorded a record .6 billion in AI server orders, with its sales pipeline expanding by over 50% across various customer segments. “Customers are seeking diverse AI-capable server solutions,” noted David Schmidt, senior director of Dell’s PowerEdge server line. While heavily regulated industries have traditionally relied on on-premises systems to ensure data privacy and security, broader adoption is now driven by the need for cost control. Fortune 2000 companies are leading this trend, opting for private infrastructure over the cloud due to more predictable expenses. “It’s not unusual to see cloud bills exceeding 0,000 or even million per month,” said John Annand, an analyst at Info-Tech Research Group. Global manufacturing giant Jabil primarily uses AWS for GenAI development but emphasizes ongoing cost management. “Does moving to the cloud provide a cost advantage? Sometimes it doesn’t,” said CIO May Yap. Jabil employs a continuous cloud financial optimization process to maximize efficiency. On-Premises AI: Technology and Trends Enterprises now have alternatives to cloud infrastructure, including as-a-service solutions like Dell APEX and HPE GreenLake, which offer flexible pay-per-use pricing for AI servers, storage, and networking tailored for private data centers or colocation facilities. “The high cost of cloud drives organizations to seek more predictable expenses,” said Tiffany Osias, vice president of global colocation services at Equinix. Walmart exemplifies in-house AI development, creating tools like a document summarization app for its benefits help desk and an AI assistant for corporate employees. Startups are also enabling enterprises to build AI applications with turnkey solutions. “About 80% of GenAI requirements can now be addressed with push-button solutions from startups,” said Tim Tully, partner at Menlo Ventures. Companies like Ragie (RAG-as-a-service) and Lamatic.ai (GenAI platform-as-a-service) are driving this innovation. Others, like Squid AI, integrate custom AI agents with existing enterprise infrastructure. Open-source frameworks like LangChain further empower on-premises development, offering tools for creating chatbots, virtual assistants, and intelligent search systems. Its extension, LangGraph, adds functionality for building multi-agent workflows. As enterprises develop AI applications internally, consulting services will play a pivotal role. “Companies offering guidance on effective AI tool usage and aligning them with business outcomes will thrive,” Annand said. This evolution in AI deployment highlights the growing importance of balancing technological innovation with financial sustainability. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Statement Accuracy Prediction based on Language Model Activations

Statement Accuracy Prediction based on Language Model Activations

When users first began interacting with ChatGPT, they noticed an intriguing behavior: the model would often reverse its stance when told it was wrong. This raised concerns about the reliability of its outputs. How can users trust a system that appears to contradict itself? Recent research has revealed that large language models (LLMs) not only generate inaccurate information (often referred to as “hallucinations”) but are also aware of their inaccuracies. Despite this awareness, these models proceed to present their responses confidently. Unveiling LLM Awareness of Hallucinations Researchers discovered this phenomenon by analyzing the internal mechanisms of LLMs. Whenever an LLM generates a response, it transforms the input query into a numerical representation and performs a series of computations before producing the output. At intermediate stages, these numerical representations are called “activations.” These activations contain significantly more information than what is reflected in the final output. By scrutinizing these activations, researchers can identify whether the LLM “knows” its response is inaccurate. A technique called SAPLMA (Statement Accuracy Prediction based on Language Model Activations) has been developed to explore this capability. SAPLMA examines the internal activations of LLMs to predict whether their outputs are truthful or not. Why Do Hallucinations Occur? LLMs function as next-word prediction models. Each word is selected based on its likelihood given the preceding words. For example, starting with “I ate,” the model might predict the next words as follows: The issue arises when earlier predictions constrain subsequent outputs. Once the model commits to a word, it cannot go back to revise its earlier choice. For instance: In another case: This mechanism reveals how the constraints of next-word prediction can lead to hallucinations, even when the model “knows” it is generating an incorrect response. Detecting Inaccuracies with SAPLMA To investigate whether an LLM recognizes its own inaccuracies, researchers developed the SAPLMA method. Here’s how it works: The classifier itself is a simple neural network with three dense layers, culminating in a binary output that predicts the truthfulness of the statement. Results and Insights The SAPLMA method achieved an accuracy of 60–80%, depending on the topic. While this is a promising result, it is not perfect and has notable limitations. For example: However, if LLMs can learn to detect inaccuracies during the generation process, they could potentially refine their outputs in real time, reducing hallucinations and improving reliability. The Future of Error Mitigation in LLMs The SAPLMA method represents a step forward in understanding and mitigating LLM errors. Accurate classification of inaccuracies could pave the way for models that can self-correct and produce more reliable outputs. While the current limitations are significant, ongoing research into these methods could lead to substantial improvements in LLM performance. By combining techniques like SAPLMA with advancements in LLM architecture, researchers aim to build models that are not only aware of their errors but capable of addressing them dynamically, enhancing both the accuracy and trustworthiness of AI systems. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More

Salesforce and AWS-Agentic Enterprise

Salesforce and AWS: Driving the Future of the Agentic Enterprise As AI-powered agents redefine the way businesses operate, strategic partnerships are playing a pivotal role in harnessing the power of data and artificial intelligence. Salesforce and AWS, two industry leaders, have taken significant steps toward building a smarter, agentic enterprise through their expanded collaboration. One year into this strategic partnership, their joint efforts are delivering transformative AI and data solutions, helping customers like Buyers Edge Platform unlock new efficiencies and capabilities. A Partnership Fueling Agentic AI Salesforce and AWS are aligning their AI and data initiatives to pave the way for advanced agentic systems—autonomous AI agents designed to enhance business operations and customer experiences. Among their notable achievements over the past year are: These innovations are creating an ecosystem that supports the delivery of agentic AI, enabling businesses to streamline operations and tap into new value from their data. “By integrating data and AI capabilities across our platforms, Salesforce and AWS are building a strong foundation for the future of agentic systems,” said Brian Landsman, EVP of Global Business Development and Technology Partnerships at Salesforce. “With a majority of large companies planning to implement agents by 2027, organizations need trusted partners to help them achieve their vision of a smarter enterprise.” Making AI More Accessible Salesforce is simplifying access to AI technology through the AWS Marketplace, offering customers an integrated solution that includes Agentforce—the agentic layer of the Salesforce platform. Agentforce enables businesses to deploy autonomous AI agents across various operations, streamlining workflows and delivering measurable results. Available in 23 countries, Salesforce’s presence on AWS Marketplace offers customers key advantages, including: By removing barriers to adoption, Salesforce and AWS empower companies to focus on leveraging technology for growth rather than navigating complex procurement systems. A New Era of Enterprise Efficiency As businesses increasingly rely on data and AI to remain competitive, the Salesforce-AWS partnership is setting the stage for enterprises to achieve more with agentic systems. These systems allow companies to execute complex tasks with unprecedented efficiency, maximizing ROI on technology investments. “Our partnership with Salesforce empowers mutual customers to realize the full potential of their data and AI investments,” said Chris Grusz, Managing Director of Technology Partnerships at AWS. “Together, we’re delivering immediate, actionable insights with agentic AI, enabling organizations to automate strategically and unlock more value across their operations.” Looking Ahead By seamlessly integrating data and AI capabilities, Salesforce and AWS are not just building technology solutions—they’re reshaping how enterprises operate and thrive in the digital age. As agentic AI becomes an essential part of business strategy, this partnership provides a blueprint for leveraging technology to drive smarter, more agile, and more effective enterprises. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More

Autonomy, Architecture, and Action

Redefining AI Agents: Autonomy, Architecture, and Action AI agents are reshaping how technology interacts with us and executes tasks. Their mission? To reason, plan, and act independently—following instructions, making autonomous decisions, and completing actions, often without user involvement. These agents adapt to new information, adjust in real time, and pursue their objectives autonomously. This evolution in agentic AI is revolutionizing how goals are accomplished, ushering in a future of semi-autonomous technology. At their foundation, AI agents rely on one or more large language models (LLMs). However, designing agents is far more intricate than building chatbots or generative assistants. While traditional AI applications often depend on user-driven inputs—such as prompt engineering or active supervision—agents operate autonomously. Core Principles of Agentic AI Architectures To enable autonomous functionality, agentic AI systems must incorporate: Essential Infrastructure for AI Agents Building and deploying agentic AI systems requires robust software infrastructure that supports: Agent Development Made Easier with Langflow and Astra DB Langflow simplifies the development of agentic applications with its visual IDE. It integrates with Astra DB, which combines vector and graph capabilities for ultra-low latency data access. This synergy accelerates development by enabling: Transforming Autonomy into Action Agentic AI is fundamentally changing how tasks are executed by empowering systems to act autonomously. By leveraging platforms like Astra DB and Langflow, organizations can simplify agent design and deploy scalable, effective AI applications. Start building the next generation of AI-powered autonomy today. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Apple's Privacy Changes: A Call for Email Marketing Innovation

Liar Liar Apple on Fire

Apple Developing Update After AI System Generates Inaccurate News Summaries Apple is working on a software update to address inaccuracies generated by its Apple Intelligence system after multiple instances of false news summaries were reported. The BBC first alerted Apple in mid-December to significant errors in the system, including a fabricated summary that falsely attributed a statement to BBC News. The summary suggested Luigi Mangione, accused of killing United Healthcare CEO Brian Thompson, had shot himself, a claim entirely unsubstantiated. Other publishers, such as ProPublica, also raised concerns about Apple Intelligence producing misleading summaries. While Apple did not respond immediately to the BBC’s December report, it issued a statement after pressure mounted from groups like the National Union of Journalists and Reporters Without Borders, both of which called for the removal of Apple Intelligence. Apple assured stakeholders it is working to refine the technology. A Widespread AI Issue: Hallucinations Apple joins the ranks of other AI vendors struggling with generative AI hallucinations—instances where AI produces false or misleading information. In October 2024, Perplexity AI faced a lawsuit from Dow Jones & Co. and the New York Post over fabricated news content attributed to their publications. Similarly, Google had to improve its AI summaries after providing users with inaccurate information. On January 16, Apple temporarily disabled AI-generated summaries for news apps on iPhone, iPad, and Mac devices. The Core Problem: AI Hallucination Chirag Shah, a professor of Information Science at the University of Washington, emphasized that hallucination is inherent to the way large language models (LLMs) function. “The nature of AI models is to generate, synthesize, and summarize, which makes them prone to mistakes,” Shah explained. “This isn’t something you can debug easily—it’s intrinsic to how LLMs operate.” While Apple plans to introduce an update that clearly labels summaries as AI-generated, Shah believes this measure falls short. “Most people don’t understand how these headlines or summaries are created. The responsible approach is to pause the technology until it’s better understood and mitigation strategies are in place,” he said. Legal and Brand Implications for Apple The hallucinated summaries pose significant reputational and legal risks for Apple, according to Michael Bennett, an AI adviser at Northeastern University. Before launching Apple Intelligence, the company was perceived as lagging in the AI race. The release of this system was intended to position Apple as a leader. Instead, the inaccuracies have damaged its credibility. “This type of hallucinated summarization is both an embarrassment and a serious legal liability,” Bennett said. “These errors could form the basis for defamation claims, as Apple Intelligence misattributes false information to reputable news sources.” Bennett criticized Apple’s seemingly minimal response. “It’s surprising how casual Apple’s reaction has been. This is a major issue for their brand and could expose them to significant legal consequences,” he added. Opportunity for Publishers The incident highlights the need for publishers to protect their interests when partnering with AI vendors like Apple and Google. Publishers should demand stronger safeguards to prevent false attributions and negotiate new contractual clauses to minimize brand risk. “This is an opportunity for publishers to lead the charge, pushing AI companies to refine their models or stop attributing false summaries to news sources,” Bennett said. He suggested legal action as a potential recourse if vendors fail to address these issues. Potential Regulatory Action The Federal Trade Commission (FTC) may also scrutinize the issue, as consumers paying for products like iPhones with AI capabilities could argue they are not receiving the promised service. However, Bennett believes Apple will likely act to resolve the problem before regulatory involvement becomes necessary. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Generative AI Energy Consumption Rises

Generative AI Tools

Generative AI Tools: A Comprehensive Overview of Emerging Capabilities The widespread adoption of generative AI services like ChatGPT has sparked immense interest in leveraging these tools for practical enterprise applications. Today, nearly every enterprise app integrates generative AI capabilities to enhance functionality and efficiency. A broad range of AI, data science, and machine learning tools now support generative AI use cases. These tools assist in managing the AI lifecycle, governing data, and addressing security and privacy concerns. While such capabilities also aid in traditional AI development, this discussion focuses on tools specifically designed for generative AI. Not all generative AI relies on large language models (LLMs). Emerging techniques generate images, videos, audio, synthetic data, and translations using methods such as generative adversarial networks (GANs), diffusion models, variational autoencoders, and multimodal approaches. Here is an in-depth look at the top categories of generative AI tools, their capabilities, and notable implementations. It’s worth noting that many leading vendors are expanding their offerings to support multiple categories through acquisitions or integrated platforms. Enterprises may want to explore comprehensive platforms when planning their generative AI strategies. 1. Foundation Models and Services Generative AI tools increasingly simplify the development and responsible use of LLMs, initially pioneered through transformer-based approaches by Google researchers in 2017. 2. Cloud Generative AI Platforms Major cloud providers offer generative AI platforms to streamline development and deployment. These include: 3. Use Case Optimization Tools Foundation models often require optimization for specific tasks. Enterprises use tools such as: 4. Quality Assurance and Hallucination Mitigation Hallucination detection tools address the tendency of generative models to produce inaccurate or misleading information. Leading tools include: 5. Prompt Engineering Tools Prompt engineering tools optimize interactions with LLMs and streamline testing for bias, toxicity, and accuracy. Examples include: 6. Data Aggregation Tools Generative AI tools have evolved to handle larger data contexts efficiently: 7. Agentic and Autonomous AI Tools Developers are creating tools to automate interactions across foundation models and services, paving the way for autonomous AI. Notable examples include: 8. Generative AI Cost Optimization Tools These tools aim to balance performance, accuracy, and cost effectively. Martian’s Model Router is an early example, while traditional cloud cost optimization platforms are expected to expand into this area. Generative AI tools are rapidly transforming enterprise applications, with foundational, cloud-based, and domain-specific solutions leading the way. By addressing challenges like accuracy, hallucination, and cost, these tools unlock new potential across industries and use cases, enabling enterprises to stay ahead in the AI-driven landscape. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Scope of Generative AI

Exploring Generative AI

Like most employees at most companies, I wear a few different hats around Tectonic. Whether I’m building a data model, creating and scheduing an email campaign, standing up a platform generative AI is always at my fingertips. At my very core, I’m a marketer. Have been for so long I do it without eveven thinking. Or at least, everyuthing I do has a hat tip to its future marketing needs. Today I want to share some of the AI content generators I’ve been using, am looking to use, or just heard about. But before we rip into the insight, here’s a primer. Types of AI Content Generators ChatGPT, a powerful AI chatbot, drew significant attention upon its November 2022 release. While the GPT-3 language model behind it had existed for some time, ChatGPT made this technology accessible to nontechnical users, showcasing how AI can generate content. Over two years later, numerous AI content generators have emerged to cater to diverse use cases. This rapid development raises questions about the technology’s impact on work. Schools are grappling with fears of plagiarism, while others are embracing AI. Legal debates about copyright and digital media authenticity continue. President Joe Biden’s October 2023 executive order addressed AI’s risks and opportunities in areas like education, workforce, and consumer privacy, underscoring generative AI’s transformative potential. What is AI-Generated Content? AI-generated content, also known as generative AI, refers to algorithms that automatically create new content across digital media. These algorithms are trained on extensive datasets and require minimal user input to produce novel outputs. For instance, ChatGPT sets a standard for AI-generated content. Based on GPT-4o, it processes text, images, and audio, offering natural language and multimodal capabilities. Many other generative AI tools operate similarly, leveraging large language models (LLMs) and multimodal frameworks to create diverse outputs. What are the Different Types of AI-Generated Content? AI-generated content spans multiple media types: Despite their varied outputs, most generative AI systems are built on advanced LLMs like GPT-4 and Google Gemini. These multimodal models process and generate content across multiple formats, with enhanced capabilities evolving over time. How Generative AI is Used Generative AI applications span industries: These tools often combine outputs from various media for complex, multifaceted projects. AI Content Generators AI content generators exist across various media. Below are good examples organized by gen ai type: Written Content Generators Image Content Generators Music Content Generators Code Content Generators Other AI Content Generators These tools showcase how AI-powered content generation is revolutionizing industries, making content creation faster and more accessible. I do hope you will comment below on your favorites, other AI tools not showcased above, or anything else AI-related that is on your mind. Written by Tectonic’s Marketing Operations Director, Shannan Hearne. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
From Chatbots to Agentic AI

From Chatbots to Agentic AI

The transition from LLM-powered chatbots to agentic systems, or agentic AI, can be summed up by the old saying: “Less talk, more action.” Keeping up with advancements in AI can be overwhelming, especially when managing an existing business. The speed and complexity of innovation can make it feel like the first day of school all over again. This insight offers a comprehensive look at AI agents, their components, and key characteristics. The introductory section breaks down the elements that form the term “AI agent,” providing a clear definition. After establishing this foundation, we explore the evolution of LLM applications, particularly the shift from traditional chatbots to agentic systems. The goal is to understand why AI agents are becoming increasingly vital in AI development and how they differ from LLM-powered chatbots. By the end of this guide, you will have a deeper understanding of AI agents, their potential applications, and their impact on organizational workflows. For those of you with a technical background who prefer to get hands-on, click here for the best repository for AI developers and builders. What is an AI Agent? Components of AI Agents To understand the term “AI agent,” we need to examine its two main components. First, let’s consider artificial intelligence, or AI. Artificial Intelligence (AI) refers to non-biological intelligence that mimics human cognition to perform tasks traditionally requiring human intellect. Through machine learning and deep learning techniques, algorithms—especially neural networks—learn patterns from data. AI systems are used for tasks such as detection, classification, and prediction, with content generation becoming a prominent domain due to transformer-based models. These systems can match or exceed human performance in specific scenarios. The second component is “agent,” a term commonly used in both technology and human contexts. In computer science, an agent refers to a software entity with environmental awareness, able to perceive and act within its surroundings. A computational agent typically has the ability to: In human contexts, an agent is someone who acts on behalf of another person or organization, making decisions, gathering information, and facilitating interactions. They often play intermediary roles in transactions and decision-making. To define an AI agent, we combine these two perspectives: it is a computational entity with environmental awareness, capable of perceiving inputs, acting with tools, and processing information using foundation models backed by both long-term and short-term memory. Key Components and Characteristics of AI Agents From LLMs to AI Agents Now, let’s take a step back and understand how we arrived at the concept of AI agents, particularly by looking at how LLM applications have evolved. The shift from traditional chatbots to LLM-powered applications has been rapid and transformative. Form Factor Evolution of LLM Applications Traditional Chatbots to LLM-Powered Chatbots Traditional chatbots, which existed before generative AI, were simpler and relied on heuristic responses: “If this, then that.” They followed predefined rules and decision trees to generate responses. These systems had limited interactivity, with the fallback option of “Speak to a human” for complex scenarios. LLM-Powered Chatbots The release of OpenAI’s ChatGPT on November 30, 2022, marked the introduction of LLM-powered chatbots, fundamentally changing the game. These chatbots, like ChatGPT, were built on GPT-3.5, a large language model trained on massive datasets. Unlike traditional chatbots, LLM-powered systems can generate human-like responses, offering a much more flexible and intelligent interaction. However, challenges remained. LLM-powered chatbots struggled with personalization and consistency, often generating plausible but incorrect information—a phenomenon known as “hallucination.” This led to efforts in grounding LLM responses through techniques like retrieval-augmented generation (RAG). RAG Chatbots RAG is a method that combines data retrieval with LLM generation, allowing systems to access real-time or proprietary data, improving accuracy and relevance. This hybrid approach addresses the hallucination problem, ensuring more reliable outputs. LLM-Powered Chatbots to AI Agents As LLMs expanded, their abilities grew more sophisticated, incorporating advanced reasoning, multi-step planning, and the use of external tools (function calling). Tool use refers to an LLM’s ability to invoke specific functions, enabling it to perform more complex tasks. Tool-Augmented LLMs and AI Agents As LLMs became tool-augmented, the emergence of AI agents followed. These agents integrate reasoning, planning, and tool use into an autonomous, goal-driven system that can operate iteratively within a dynamic environment. Unlike traditional chatbot interfaces, AI agents leverage a broader set of tools to interact with various systems and accomplish tasks. Agentic Systems Agentic systems—computational architectures that include AI agents—embody these advanced capabilities. They can autonomously interact with systems, make decisions, and adapt to feedback, forming the foundation for more complex AI applications. Components of an AI Agent AI agents consist of several key components: Characteristics of AI Agents AI agents are defined by the following traits: Conclusion AI agents represent a significant leap from traditional chatbots, offering greater autonomy, complexity, and interactivity. However, the term “AI agent” remains fluid, with no universal industry standard. Instead, it exists on a continuum, with varying degrees of autonomy, adaptability, and proactive behavior defining agentic systems. Value and Impact of AI Agents The key benefits of AI agents lie in their ability to automate manual processes, reduce decision-making burdens, and enhance workflows in enterprise environments. By “agentifying” repetitive tasks, AI agents offer substantial productivity gains and the potential to transform how businesses operate. As AI agents evolve, their applications will only expand, driving new efficiencies and enabling organizations to leverage AI in increasingly sophisticated ways. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Agentic AI is Here

Agentic AI Revolution

The Agentic AI Revolution: Lead, Follow, or Get Out of the Way The era of agentic AI is here, and the message is clear—if you’re not leading the charge, you’re falling behind. Companies like Wiley and OpenTable are reshaping their industries with autonomous AI agents that don’t just assist but also analyze, strategize, and execute tasks with unparalleled efficiency. As these organizations demonstrate, the key to AI success lies in rewriting the rules of your industry rather than playing catch-up. Rewriting Industry Standards with Agentic AI Wiley: The education giant leveraged Agentforce, a digital labor platform for deploying autonomous AI agents, to transform its customer service operations. By onboarding representatives 50% faster and improving case resolution by 40%, Wiley streamlined its processes in just a few weeks. AI agents now handle registration and payment inquiries, directing students to resources and reducing the workload on human representatives. OpenTable: As the go-to reservation platform for 1.7 billion diners annually, OpenTable deploys AI agents to manage reservation changes and loyalty points. This allows employees to focus on customer relationships. Even a two-minute efficiency gain per interaction translates to massive operational savings. Salesforce Help Site: With over 60 million annual visits, the Salesforce Help site integrated Agentforce to resolve 83% of queries without human involvement. In just weeks, Agentforce doubled its capacity, handling over 32,000 automated conversations. These examples showcase a new era of digital labor where AI agents orchestrate high-value, multistep tasks, working tirelessly to deliver results. Far from replacing humans, they supercharge productivity and innovation, enabling companies to do more than ever before. How to Empower Your Workforce with AI Empowering your workforce for the next wave of AI doesn’t require months of preparation or millions of dollars. You don’t need to build or train your own large language model (LLM). Instead, integrating AI with existing data, automation, and workflows is the key to success, as demonstrated by leaders like Wiley and OpenTable. Here’s how to get started: 1. Real-Time Data Access AI thrives on real-time, high-quality data. Platforms like Salesforce Data Cloud unify structured and unstructured data, connecting it seamlessly to the LLM. Techniques such as retrieval-augmented generation (RAG) and semantic search ensure AI agents can access the most relevant data for any task. 2. Advanced Reasoning AI agents aren’t just about answering queries—they execute complex, multistep tasks. For example, they can process returns, reorder items, and even flag anomalies. Powered by reasoning engines, these agents draw data from systems like CRM, refine plans, and adapt dynamically until the task is completed correctly. 3. Built-In Security AI agents must operate within clear guardrails, knowing their limits and handing tasks off to humans when necessary. Strong permissions and security protocols are essential to ensure data protection and prevent unauthorized actions. 4. Action-Oriented Workflows Generative AI’s real value lies in action. By integrating tools like Salesforce Flow for task automation and MuleSoft APIs for system connectivity, AI agents can execute business workflows such as fraud detection, customer outreach, and case management. 5. Human-AI Collaboration The future of work isn’t AI replacing humans—it’s AI and humans working together. While agents handle data-intensive and repetitive tasks, humans bring strategic thinking, empathy, and creativity. This synergy leads to smarter decisions and redefines workflows across industries. Why Training Your Own LLM May Not Be the Answer Many companies assume training a proprietary LLM will give them a competitive edge. In reality, this process is costly, time-intensive, and requires constant updates to remain accurate. An LLM trained on static data quickly becomes outdated, much like a GPS that fails after the first detour. Instead, companies are turning to out-of-the-box AI solutions that integrate seamlessly with their existing systems. These tools offer the flexibility to scale quickly and adapt in real time, enabling businesses to stay competitive without the heavy lift of building from scratch. Scaling AI for the Future Many organizations remain stuck in pilot phases with AI due to data quality issues and a limited understanding of use cases. Companies like Wiley and OpenTable, however, have cracked the code: integrating prebuilt AI systems with robust data flows, automation, and workflows. By embracing agentic AI, forward-thinking organizations are creating digital labor forces that unlock new efficiencies, enhance customer experiences, and position themselves for long-term success. The trillion-dollar AI opportunity awaits—will you lead or trail behind? Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More

Why Build a General-Purpose Agent?

A general-purpose LLM agent serves as an excellent starting point for prototyping use cases and establishing the foundation for a custom agentic architecture tailored to your needs. What is an LLM Agent? An LLM (Large Language Model) agent is a program where execution logic is governed by the underlying model. Unlike approaches such as few-shot prompting or fixed workflows, LLM agents adapt dynamically. They can determine which tools to use (e.g., web search or code execution), how to use them, and iterate based on results. This adaptability enables handling diverse tasks with minimal configuration. Agentic Architectures Explained:Agentic systems range from the reliability of fixed workflows to the flexibility of autonomous agents. For instance: Your architecture choice will depend on the desired balance between reliability and flexibility for your use case. Building a General-Purpose LLM Agent Step 1: Select the Right LLM Choosing the right model is critical for performance. Evaluate based on: Model Recommendations (as of now): For simpler use cases, smaller models running locally can also be effective, but with limited functionality. Step 2: Define the Agent’s Control Logic The system prompt differentiates an LLM agent from a standalone model. This prompt contains rules, instructions, and structures that guide the agent’s behavior. Common Agentic Patterns: Starting with ReAct or Plan-then-Execute patterns is recommended for general-purpose agents. Step 3: Define the Agent’s Core Instructions To optimize the agent’s behavior, clearly define its features and constraints in the system prompt: Example Instructions: Step 4: Define and Optimize Core Tools Tools expand an agent’s capabilities. Common tools include: For each tool, define: Example: Implementing an Arxiv API tool for scientific queries. Step 5: Memory Handling Strategy Since LLMs have limited memory (context window), a strategy is necessary to manage past interactions. Common approaches include: For personalization, long-term memory can store user preferences or critical information. Step 6: Parse the Agent’s Output To make raw LLM outputs actionable, implement a parser to convert outputs into a structured format like JSON. Structured outputs simplify execution and ensure consistency. Step 7: Orchestrate the Agent’s Workflow Define orchestration logic to handle the agent’s next steps after receiving an output: Example Orchestration Code: pythonCopy codedef orchestrator(llm_agent, llm_output, tools, user_query): while True: action = llm_output.get(“action”) if action == “tool_call”: tool_name = llm_output.get(“tool_name”) tool_params = llm_output.get(“tool_params”, {}) if tool_name in tools: try: tool_result = tools[tool_name](**tool_params) llm_output = llm_agent({“tool_output”: tool_result}) except Exception as e: return f”Error executing tool ‘{tool_name}’: {str(e)}” else: return f”Error: Tool ‘{tool_name}’ not found.” elif action == “return_answer”: return llm_output.get(“answer”, “No answer provided.”) else: return “Error: Unrecognized action type from LLM output.” This orchestration ensures seamless interaction between tools, memory, and user queries. When to Consider Multi-Agent Systems A single-agent setup works well for prototyping but may hit limits with complex workflows or extensive toolsets. Multi-agent architectures can: Starting with a single agent helps refine workflows, identify bottlenecks, and scale effectively. By following these steps, you’ll have a versatile system capable of handling diverse use cases, from competitive analysis to automating workflows. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more Top Ten Reasons Why Tectonic Loves the Cloud The Cloud is Good for Everyone – Why Tectonic loves the cloud You don’t need to worry about tracking licenses. Read more

Read More
Python-Based Reasoning Engine

Python-Based Reasoning Engine

Introducing a Python-Based Reasoning Engine for Deterministic AI In the age of large language models (LLMs), there’s a growing need for deterministic systems that enforce rules and constraints while reasoning about information. We’ve developed a Python-based reasoning and validation framework that bridges the gap between traditional rule-based logic and modern AI capabilities, inspired by frameworks like Pydantic. This approach is designed for developers and non-technical experts alike, making it easy to build complex rule engines that translate natural language instructions into enforceable code. Our fine-tuned model automates the creation of rules while ensuring human oversight for quality and conflict detection. The result? Faster implementation of rule engines, reduced developer overhead, and flexible extensibility across domains. The Framework at a Glance Our system consists of five core components: To analogize, this framework operates like a game of chess: Our framework supports two primary use cases: Key Features and Benefits Case Studies Validation Engine: Ensuring Compliance A mining company needed to validate employee qualifications based on age, region, and role. Example Data Structure: jsonCopy code{ “employees”: [ { “name”: “Sarah”, “age”: 25, “role”: “Manager”, “documents”: [“safe_handling_at_work”, “heavy_lifting”] }, { “name”: “John”, “age”: 17, “role”: “Laborer”, “documents”: [“heavy_lifting”] } ] } Rules: jsonCopy code{ “rules”: [ { “type”: “min_age”, “parameters”: { “min_age”: 18 } }, { “type”: “dozer_operator”, “parameters”: { “document_type”: “dozer_qualification” } } ] } Outcome:The system flagged violations, such as employees under 18 or missing required qualifications, ensuring compliance with organizational rules. Reasoning Engine: Solving the River Crossing Puzzle The classic river crossing puzzle demonstrates the engine’s reasoning capabilities. Problem Setup:A farmer must ferry a goat, a wolf, and a cabbage across a river, adhering to specific constraints (e.g., the goat cannot be left alone with the cabbage). Steps: Output:The engine generated a solution in 0.0003 seconds, showcasing its efficiency in navigating complex logic. Advanced Features: Dynamic Rule Expansion The system supports real-time rule adjustments. For instance, adding a “wolf cannot be left with a chicken” constraint introduces a conflict. By extending rules (e.g., allowing the farmer to carry two items), the engine dynamically resolves previously unsolvable scenarios. Sample Code Snippet: pythonCopy codeclass CarryingCapacityRule(Rule): def evaluate(self, state): items_moved = sum(1 for item in [‘wolf’, ‘goat’, ‘cabbage’, ‘chicken’] if getattr(state, item) == state.farmer) return items_moved <= 2 def get_description(self): return “Farmer can carry up to two items at a time” Result:The adjusted engine solved the puzzle in three moves, down from seven, while maintaining rule integrity. Collaborative UI for Rule Creation Our user interface empowers domain experts to define rules without writing code. Developers validate these rules, which are then seamlessly integrated into the system. Visual Workflow: Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more Top Ten Reasons Why Tectonic Loves the Cloud The Cloud is Good for Everyone – Why Tectonic loves the cloud You don’t need to worry about tracking licenses. Read more

Read More
Standards in Healthcare Cybersecurity

Deploying Large Language Models in Healthcare

Study Identifies Cost-Effective Strategies for Deploying Large Language Models in Healthcare Efficient deployment of large language models (LLMs) at scale in healthcare can streamline clinical workflows and reduce costs by up to 17 times without compromising reliability, according to a study published in NPJ Digital Medicine by researchers at the Icahn School of Medicine at Mount Sinai. The research highlights the potential of LLMs to enhance clinical operations while addressing the financial and computational hurdles healthcare organizations face in scaling these technologies. To investigate solutions, the team evaluated 10 LLMs of varying sizes and capacities using real-world patient data. The models were tested on chained queries and increasingly complex clinical notes, with outputs assessed for accuracy, formatting quality, and adherence to clinical instructions. “Our study was driven by the need to identify practical ways to cut costs while maintaining performance, enabling health systems to confidently adopt LLMs at scale,” said Dr. Eyal Klang, director of the Generative AI Research Program at Icahn Mount Sinai. “We aimed to stress-test these models, evaluating their ability to manage multiple tasks simultaneously and identifying strategies to balance performance and affordability.” The team conducted over 300,000 experiments, finding that high-capacity models like Meta’s Llama-3-70B and GPT-4 Turbo 128k performed best, maintaining high accuracy and low failure rates. However, performance began to degrade as task volume and complexity increased, particularly beyond 50 tasks involving large prompts. The study further revealed that grouping tasks—such as identifying patients for preventive screenings, analyzing medication safety, and matching patients for clinical trials—enabled LLMs to handle up to 50 simultaneous tasks without significant accuracy loss. This strategy also led to dramatic cost savings, with API costs reduced by up to 17-fold, offering a pathway for health systems to save millions annually. “Understanding where these models reach their cognitive limits is critical for ensuring reliability and operational stability,” said Dr. Girish N. Nadkarni, co-senior author and director of The Charles Bronfman Institute of Personalized Medicine. “Our findings pave the way for the integration of generative AI in hospitals while accounting for real-world constraints.” Beyond cost efficiency, the study underscores the potential of LLMs to automate key tasks, conserve resources, and free up healthcare providers to focus more on patient care. “This research highlights how AI can transform healthcare operations. Grouping tasks not only cuts costs but also optimizes resources that can be redirected toward improving patient outcomes,” said Dr. David L. Reich, co-author and chief clinical officer of the Mount Sinai Health System. The research team plans to explore how LLMs perform in live clinical environments and assess emerging models to determine whether advancements in AI technology can expand their cognitive thresholds. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more Alphabet Soup of Cloud Terminology As with any technology, the cloud brings its own alphabet soup of terms. This insight will hopefully help you navigate Read more

Read More
gettectonic.com