LLMs Archives - gettectonic.com - Page 3
Where LLMs Fall Short

LLM Economies

Throughout history, disruptive technologies have been the catalyst for major social and economic revolutions. The invention of the plow and irrigation systems 12,000 years ago sparked the Agricultural Revolution, while Johannes Gutenberg’s 15th-century printing press fueled the Protestant Reformation and helped propel Europe out of the Middle Ages into the Renaissance. In the 18th century, James Watt’s steam engine ushered in the Industrial Revolution. More recently, the internet has revolutionized communication, commerce, and information access, shrinking the world into a global village. Similarly, smartphones have transformed how people interact with their surroundings. Now, we stand at the dawn of the AI revolution. Large Language Models (LLMs) represent a monumental leap forward, with significant economic implications at both macro and micro levels. These models are reshaping global markets, driving new forms of currency, and creating a novel economic landscape. The reason LLMs are transforming industries and redefining economies is simple: they automate both routine and complex tasks that traditionally require human intelligence. They enhance decision-making processes, boost productivity, and facilitate cost reductions across various sectors. This enables organizations to allocate human resources toward more creative and strategic endeavors, resulting in the development of new products and services. From healthcare to finance to customer service, LLMs are creating new markets and driving AI-driven services like content generation and conversational assistants into the mainstream. To truly grasp the engine driving this new global economy, it’s essential to understand the inner workings of this disruptive technology. These posts will provide both a macro-level overview of the economic forces at play and a deep dive into the technical mechanics of LLMs, equipping you with a comprehensive understanding of the revolution happening now. Why Now? The Connection Between Language and Human Intelligence AI did not begin with ChatGPT’s arrival in November 2022. Many people were developing machine learning classification models in 1999, and the roots of AI go back even further. Artificial Intelligence was formally born in 1950, when Alan Turing—considered the father of theoretical computer science and famed for cracking the Nazi Enigma code during World War II—created the first formal definition of intelligence. This definition, known as the Turing Test, demonstrated the potential for machines to exhibit human-like intelligence through natural language conversations. The test involves a human evaluator who engages in conversations with both a human and a machine. If the evaluator cannot reliably distinguish between the two, the machine is considered to have passed the test. Remarkably, after 72 years of gradual AI development, ChatGPT simulated this very interaction, passing the Turing Test and igniting the current AI explosion. But why is language so closely tied to human intelligence, rather than, for example, vision? While 70% of our brain’s neurons are devoted to vision, OpenAI’s pioneering image generation model, DALL-E, did not trigger the same level of excitement as ChatGPT. The answer lies in the profound role language has played in human evolution. The Evolution of Language The development of language was the turning point in humanity’s rise to dominance on Earth. As Yuval Noah Harari points out in his book Sapiens: A Brief History of Humankind, it was the ability to gossip and discuss abstract concepts that set humans apart from other species. Complex communication, such as gossip, requires a shared, sophisticated language. Human language evolved from primitive cave signs to structured alphabets, which, along with grammar rules, created languages capable of expressing thousands of words. In today’s digital age, language has further evolved with the inclusion of emojis, and now with the advent of GenAI, tokens have become the latest cornerstone in this progression. These shifts highlight the extraordinary journey of human language, from simple symbols to intricate digital representations. In the next post, we will explore the intricacies of LLMs, focusing specifically on tokens. But before that, let’s delve into the economic forces shaping the LLM-driven world. The Forces Shaping the LLM Economy AI Giants in Competition Karl Marx and Friedrich Engels argued that those who control the means of production hold power. The tech giants of today understand that AI is the future means of production, and the race to dominate the LLM market is well underway. This competition is fierce, with industry leaders like OpenAI, Google, Microsoft, and Facebook battling for supremacy. New challengers such as Mistral (France), AI21 (Israel), and Elon Musk’s xAI and Anthropic are also entering the fray. The LLM industry is expanding exponentially, with billions of dollars of investment pouring in. For example, Anthropic has raised $4.5 billion from 43 investors, including major players like Amazon, Google, and Microsoft. The Scarcity of GPUs Just as Bitcoin mining requires vast computational resources, training LLMs demands immense computing power, driving a search for new energy sources. Microsoft’s recent investment in nuclear energy underscores this urgency. At the heart of LLM technology are Graphics Processing Units (GPUs), essential for powering deep neural networks. These GPUs have become scarce and expensive, adding to the competitive tension. Tokens: The New Currency of the LLM Economy Tokens are the currency driving the emerging AI economy. Just as money facilitates transactions in traditional markets, tokens are the foundation of LLM economics. But what exactly are tokens? Tokens are the basic units of text that LLMs process. They can be single characters, parts of words, or entire words. For example, the word “Oscar” might be split into two tokens, “os” and “car.” The performance of LLMs—quality, speed, and cost—hinges on how efficiently they generate these tokens. LLM providers price their services based on token usage, with different rates for input (prompt) and output (completion) tokens. As companies rely more on LLMs, especially for complex tasks like agentic applications, token usage will significantly impact operational costs. With fierce competition and the rise of open-source models like Llama-3.1, the cost of tokens is rapidly decreasing. For instance, OpenAI reduced its GPT-4 pricing by about 80% over the past year and a half. This trend enables companies to expand their portfolio of AI-powered products, further fueling the LLM economy. Context Windows: Expanding Capabilities

Read More
RAGate

RAGate

RAGate: Revolutionizing Conversational AI with Adaptive Retrieval-Augmented Generation Building Conversational AI systems is challenging.It’s not just feasible; it’s complex, resource-intensive, and time-consuming. The difficulty lies in creating systems that can not only understand and generate human-like responses but also adapt effectively to conversational nuances, ensuring meaningful engagement with users. Retrieval-Augmented Generation (RAG) has already transformed Conversational AI by combining the internal knowledge of large language models (LLMs) with external knowledge sources. By leveraging RAG with business data, organizations empower their customers to ask natural language questions and receive insightful, data-driven answers. The challenge?Not every query requires external knowledge. Over-reliance on external sources can disrupt conversational flow, much like consulting a book for every question during a conversation—even when internal knowledge is sufficient. Worse, if no external knowledge is available, the system may respond with “I don’t know,” despite having relevant internal knowledge to answer. The solution?RAGate — an adaptive mechanism that dynamically determines when to use external knowledge and when to rely on internal insights. Developed by Xi Wang, Procheta Sen, Ruizhe Li, and Emine Yilmaz and introduced in their July 2024 paper on Adaptive Retrieval-Augmented Generation for Conversational Systems, RAGate addresses this balance with precision. What Is Conversational AI? At its core, conversation involves exchanging thoughts, emotions, and information, guided by tone, context, and subtle cues. Humans excel at this due to emotional intelligence, socialization, and cultural exposure. Conversational AI aims to replicate these human-like interactions by leveraging technology to generate natural, contextually appropriate, and engaging responses. These systems adapt fluidly to user inputs, making the interaction dynamic—like conversing with a human. Internal vs. External Knowledge in AI Systems To understand RAGate’s value, we need to differentiate between two key concepts: Limitations of Traditional RAG Systems RAG integrates LLMs’ natural language capabilities with external knowledge retrieval, often guided by “guardrails” to ensure responsible, domain-specific responses. However, strict reliance on external knowledge can lead to: How RAGate Enhances Conversational AI RAGate, or Retrieval-Augmented Generation Gate, adapts dynamically to determine when external knowledge retrieval is necessary. It enhances response quality by intelligently balancing internal and external knowledge, ensuring conversational relevance and efficiency. The mechanism: Traditional RAG vs. RAGate: An Example Scenario: A healthcare chatbot offers advice based on general wellness principles and up-to-date medical research. This adaptive approach improves response accuracy, reduces latency, and enhances the overall conversational experience. RAGate Variants RAGate offers three implementation methods, each tailored to optimize performance: Variant Approach Key Feature RAGate-Prompt Uses natural language prompts to decide when external augmentation is needed. Lightweight and simple to implement. RAGate-PEFT Employs parameter-efficient fine-tuning (e.g., QLoRA) for better decision-making. Fine-tunes the model with minimal resource requirements. RAGate-MHA Leverages multi-head attention to interactively assess context and retrieve external knowledge. Optimized for complex conversational scenarios. RAGate Varients How to Implement RAGate Key Takeaways RAGate represents a breakthrough in Conversational AI, delivering adaptive, contextually relevant, and efficient responses by balancing internal and external knowledge. Its potential spans industries like healthcare, education, finance, and customer support, enhancing decision-making and user engagement. By intelligently combining retrieval-augmented generation with nuanced adaptability, RAGate is set to redefine the way businesses and individuals interact with AI. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Salesforce AI Research Introduces LaTRO

Salesforce AI Research Introduces LaTRO

Salesforce AI Research Introduces LaTRO: A Breakthrough in Enhancing Reasoning for Large Language Models Large Language Models (LLMs) have revolutionized tasks such as answering questions, generating content, and assisting with workflows. However, they often struggle with advanced reasoning tasks like solving complex math problems, logical deduction, and structured data analysis. Salesforce AI Research has addressed this challenge by introducing LaTent Reasoning Optimization (LaTRO), a groundbreaking framework that enables LLMs to self-improve their reasoning capabilities during training. The Need for Advanced Reasoning in LLMs Reasoning—especially sequential, multi-step reasoning—is essential for tasks that require logical progression and problem-solving. While current models excel at simpler queries, they often fall short in tackling more complex tasks due to a reliance on external feedback mechanisms or runtime optimizations. Enhancing reasoning abilities is therefore critical to unlocking the full potential of LLMs across diverse applications, from advanced mathematics to real-time data analysis. Existing techniques like Chain-of-Thought (CoT) prompting guide models to break problems into smaller steps, while methods such as Tree-of-Thought and Program-of-Thought explore multiple reasoning pathways. Although these techniques improve runtime performance, they don’t fundamentally enhance reasoning during the model’s training phase, limiting the scope of improvement. Salesforce AI Research Introduces LaTRO: A Self-Rewarding Framework LaTRO shifts the paradigm by transforming reasoning into a training-level optimization problem. It introduces a self-rewarding mechanism that allows models to evaluate and refine their reasoning pathways without relying on external feedback or supervised fine-tuning. This intrinsic approach fosters continual improvement and empowers models to solve complex tasks more effectively. How LaTRO Works LaTRO’s methodology centers on sampling reasoning paths from a latent distribution and optimizing these paths using variational techniques. Here’s how it works: This self-rewarding cycle ensures that the model continuously refines its reasoning capabilities during training. Unlike traditional methods, LaTRO’s framework operates autonomously, without the need for external reward models or costly supervised feedback loops. Key Benefits of LaTRO Performance Highlights LaTRO’s effectiveness has been validated across various datasets and models: Applications and Implications LaTRO’s ability to foster logical coherence and structured reasoning has far-reaching applications in fields requiring robust problem-solving: By enabling LLMs to autonomously refine their reasoning processes, LaTRO brings AI closer to achieving human-like cognitive abilities. The Future of AI with LaTRO LaTRO sets a new benchmark in AI research by demonstrating that reasoning can be optimized during training, not just at runtime. This advancement by Salesforce AI Research highlights the potential for self-evolving AI models that can independently improve their problem-solving capabilities. Salesforce AI Research Introduces LaTRO As the field of AI progresses, frameworks like LaTRO pave the way for more autonomous, intelligent systems capable of navigating complex reasoning tasks across industries. LaTRO represents a significant leap forward, moving AI closer to achieving true autonomous reasoning. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
copilots and agentic ai

Copilots and Agentic AI

Agentic AI vs. Copilots: Defining the Future of Generative AI Artificial Intelligence has rapidly evolved, progressing from simple automation to generative models, to copilots. But now, a new player—Agentic AI—has emerged, promising to redefine the AI landscape. Is Agentic AI the next logical step, or will it coexist alongside copilots, each serving distinct roles? Copilots and Agentic AI. Generative AI: Creativity with a Human Touch Since the launch of ChatGPT, generative AI has dominated tech priorities, offering businesses the ability to generate content—text, images, videos, and more—from pre-defined data. However, while revolutionary, generative AI still relies heavily on human input to guide its output, making it a powerful collaborator rather than an autonomous actor. Enter Agentic AI: Autonomy Redefined Agentic AI represents a leap forward, offering systems that possess autonomy and the ability to act independently to achieve pre-defined goals. Unlike generative AI copilots that respond to human prompts, Agentic AI makes decisions, plans actions, and learns from experience. Think of it as Siri or Alexa—enhanced with autonomy and learning capabilities. Gartner recently spotlighted Agentic AI as its top technology trend for 2025, predicting that by 2028, at least 15% of day-to-day work decisions will be made autonomously, up from virtually none today. Agentforce and the Third Wave of AI Salesforce’s “Agentforce,” unveiled at Dreamforce, is a prime example of Agentic AI’s potential. These autonomous agents are designed to augment employees by handling tasks across sales, service, marketing, and commerce. Salesforce CEO Mark Benioff described it as the “Third Wave of AI,” going beyond copilots to deliver intelligent agents deeply embedded into customer workflows. Salesforce aims to empower one billion AI agents by 2025, integrating Agentforce into every aspect of customer success. Benioff took a swipe at competitors’ bolt-on generative AI solutions, emphasizing that Agentforce is deeply embedded for maximum value. The Role of Copilots: Collaboration First While Agentic AI gains traction, copilots like Microsoft’s Copilot Studio and SAP’s Joule remain critical for businesses focused on intelligent augmentation. Copilots act as productivity boosters, working alongside humans to optimize processes, enhance creativity, and provide decision-making support. SAP’s Joule, for example, integrates seamlessly into existing systems to optimize operations while leaving strategic decision-making in human hands. This collaborative model aligns well with businesses prioritizing agility and human oversight. Agentic AI: Opportunities and Challenges Agentic AI’s autonomy offers significant potential for streamlining complex processes, reducing human intervention, and driving productivity. However, it also comes with risks. Eleanor Watson, AI ethics engineer at Singularity University, warns that Agentic AI systems require careful alignment of values and goals to avoid unintended consequences like dangerous shortcuts or boundary violations. In contrast, copilots retain human agency, making them particularly suited for creative and knowledge-based roles where human oversight remains essential. Copilots and Agentic AI The choice between Agentic AI and copilots hinges on an organization’s priorities and risk tolerance. For simpler, task-specific applications, copilots excel by providing assistance without removing human input. Agentic AI, on the other hand, shines in complex, multi-task scenarios where autonomy is key. Dom Couldwell, head of field engineering EMEA at DataStax, emphasizes the importance of understanding when to deploy each model. “Use a copilot for specific, focused tasks. Use Agentic AI for complex, goal-oriented processes involving multiple tasks. And leverage Retrieval Augmented Generation (RAG) in both to provide context to LLMs.” The Road Ahead: Coexistence or Dominance? As AI evolves, Agentic AI and copilots may coexist, serving complementary roles. Businesses seeking full automation and scalability may gravitate toward Agentic AI, while those prioritizing augmented intelligence and human collaboration will continue to rely on copilots. Ultimately, the future of AI will be defined not by one model overtaking the other, but by how well each aligns with the specific needs, goals, and challenges of the organizations adopting them. Like1 Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
AI Research Agents

AI Research Agents

AI Research Agents: Transforming Knowledge Discovery by 2025 (Plus the Top 3 Free Tools) The research world is on the verge of a groundbreaking shift, driven by the evolution of AI research agents. By 2025, these agents are expected to move beyond being mere tools to becoming transformative assets for knowledge discovery, revolutionizing industries such as marketing, science, and beyond. Human researchers are inherently limited—they cannot scan 10,000 websites in an hour or analyze data at lightning speed. AI agents, however, are purpose-built for these tasks, providing efficiency and insights far beyond human capabilities. Here, we explore the anticipated impact of AI research agents and highlight three free tools redefining this space (spoiler alert: it’s not ChatGPT or Perplexity!). AI Research Agents: The New Era of Knowledge Exploration By 2030, the AI research market is projected to skyrocket from .1 billion in 2024 to .1 billion. This explosive growth represents not just advancements in AI but a fundamental transformation in how knowledge is gathered, analyzed, and applied. Unlike traditional AI systems, which require constant input and supervision, AI research agents function more like dynamic research assistants. They adapt their approach based on outcomes, handle vast quantities of data, and generate actionable insights with remarkable precision. Key Differentiator: These agents leverage advanced Retrieval Augmented Generation (RAG) technology, ensuring accuracy by pulling verified data from trusted sources. Equipped with anti-hallucination algorithms, they maintain factual integrity while citing their sources—making them indispensable for high-stakes research. The Technology Behind AI Research Agents AI research agents stand out due to their ability to: For example, an AI agent can deliver a detailed research report in 30 minutes, a task that might take a human team days. Why AI Research Agents Matter Now The timing couldn’t be more critical. The volume of data generated daily is overwhelming, and human researchers often struggle to keep up. Meanwhile, Google’s focus on Experience, Expertise, Authoritativeness, and Trustworthiness (EEAT) has heightened the demand for accurate, well-researched content. Some research teams have already reported time savings of up to 70% by integrating AI agents into their workflows. Beyond speed, these agents uncover perspectives and connections often overlooked by human researchers, adding significant value to the final output. Top 3 Free AI Research Tools 1. Stanford STORM Overview: STORM (Synthesis of Topic Outlines through Retrieval and Multi-perspective Question Asking) is an open-source system designed to generate comprehensive, Wikipedia-style articles. Learn more: Visit the STORM GitHub repository. 2. CustomGPT.ai Researcher Overview: CustomGPT.ai creates highly accurate, SEO-optimized long-form articles using deep Google research or proprietary databases. Learn more: Access the free Streamlit app for CustomGPT.ai. 3. GPT Researcher Overview: This open-source agent conducts thorough research tasks, pulling data from both web and local sources to produce customized reports. Learn more: Visit the GPT Researcher GitHub repository. The Human-AI Partnership Despite their capabilities, AI research agents are not replacements for human researchers. Instead, they act as powerful assistants, enabling researchers to focus on creative problem-solving and strategic thinking. Think of them as tireless collaborators, processing vast amounts of data while humans interpret and apply the findings to solve complex challenges. Preparing for the AI Research Revolution To harness the potential of AI research agents, researchers must adapt. Universities and organizations are already incorporating AI training into their programs to prepare the next generation of professionals. For smaller labs and institutions, these tools present a unique opportunity to level the playing field, democratizing access to high-quality research capabilities. Looking Ahead By 2025, AI research agents will likely reshape the research landscape, enabling cross-disciplinary breakthroughs and empowering researchers worldwide. From small teams to global enterprises, the benefits are immense—faster insights, deeper analysis, and unprecedented innovation. As with any transformative technology, challenges remain. But the potential to address some of humanity’s biggest problems makes this an AI revolution worth embracing. Now is the time to prepare and make the most of these groundbreaking tools. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
healthcare Can prioritize ai governance

Healthcare Can Prioritize AI Governance

As artificial intelligence gains momentum in healthcare, it’s critical for health systems and related stakeholders to develop robust AI governance programs. AI’s potential to address challenges in administration, operations, and clinical care is drawing interest across the sector. As this technology evolves, the range of applications in healthcare will only broaden.

Read More
Where LLMs Fall Short

Where LLMs Fall Short

Large Language Models (LLMs) have transformed natural language processing, showcasing exceptional abilities in text generation, translation, and various language tasks. Models like GPT-4, BERT, and T5 are based on transformer architectures, which enable them to predict the next word in a sequence by training on vast text datasets. How LLMs Function LLMs process input text through multiple layers of attention mechanisms, capturing complex relationships between words and phrases. Here’s an overview of the process: Tokenization and Embedding Initially, the input text is broken down into smaller units, typically words or subwords, through tokenization. Each token is then converted into a numerical representation known as an embedding. For instance, the sentence “The cat sat on the mat” could be tokenized into [“The”, “cat”, “sat”, “on”, “the”, “mat”], each assigned a unique vector. Multi-Layer Processing The embedded tokens are passed through multiple transformer layers, each containing self-attention mechanisms and feed-forward neural networks. Contextual Understanding As the input progresses through layers, the model develops a deeper understanding of the text, capturing both local and global context. This enables the model to comprehend relationships such as: Training and Pattern Recognition During training, LLMs are exposed to vast datasets, learning patterns related to grammar, syntax, and semantics: Generating Responses When generating text, the LLM predicts the next word or token based on its learned patterns. This process is iterative, where each generated token influences the next. For example, if prompted with “The Eiffel Tower is located in,” the model would likely generate “Paris,” given its learned associations between these terms. Limitations in Reasoning and Planning Despite their capabilities, LLMs face challenges in areas like reasoning and planning. Research by Subbarao Kambhampati highlights several limitations: Lack of Causal Understanding LLMs struggle with causal reasoning, which is crucial for understanding how events and actions relate in the real world. Difficulty with Multi-Step Planning LLMs often struggle to break down tasks into a logical sequence of actions. Blocksworld Problem Kambhampati’s research on the Blocksworld problem, which involves stacking and unstacking blocks, shows that LLMs like GPT-3 struggle with even simple planning tasks. When tested on 600 Blocksworld instances, GPT-3 solved only 12.5% of them using natural language prompts. Even after fine-tuning, the model solved only 20% of the instances, highlighting the model’s reliance on pattern recognition rather than true understanding of the planning task. Performance on GPT-4 Temporal and Counterfactual Reasoning LLMs also struggle with temporal reasoning (e.g., understanding the sequence of events) and counterfactual reasoning (e.g., constructing hypothetical scenarios). Token and Numerical Errors LLMs also exhibit errors in numerical reasoning due to inconsistencies in tokenization and their lack of true numerical understanding. Tokenization and Numerical Representation Numbers are often tokenized inconsistently. For example, “380” might be one token, while “381” might split into two tokens (“38” and “1”), leading to confusion in numerical interpretation. Decimal Comparison Errors LLMs can struggle with decimal comparisons. For example, comparing 9.9 and 9.11 may result in incorrect conclusions due to how the model processes these numbers as strings rather than numerically. Examples of Numerical Errors Hallucinations and Biases Hallucinations LLMs are prone to generating false or nonsensical content, known as hallucinations. This can happen when the model produces irrelevant or fabricated information. Biases LLMs can perpetuate biases present in their training data, which can lead to the generation of biased or stereotypical content. Inconsistencies and Context Drift LLMs often struggle to maintain consistency over long sequences of text or tasks. As the input grows, the model may prioritize more recent information, leading to contradictions or neglect of earlier context. This is particularly problematic in multi-turn conversations or tasks requiring persistence. Conclusion While LLMs have advanced the field of natural language processing, they still face significant challenges in reasoning, planning, and maintaining contextual accuracy. These limitations highlight the need for further research and development of hybrid AI systems that integrate LLMs with other techniques to improve reasoning, consistency, and overall performance. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
healthcare Can prioritize ai governance

AI Data Privacy and Security

Three Key Generative AI Data Privacy and Security Concerns The rise of generative AI is reshaping the digital landscape, introducing powerful tools like ChatGPT and Microsoft Copilot into the hands of professionals, students, and casual users alike. From creating AI-generated art to summarizing complex texts, generative AI (GenAI) is transforming workflows and sparking innovation. However, for information security and privacy professionals, this rapid proliferation also brings significant challenges in data governance and protection. Below are three critical data privacy and security concerns tied to generative AI: 1. Who Owns the Data? Data ownership is a contentious issue in the age of generative AI. In the European Union, the General Data Protection Regulation (GDPR) asserts that individuals own their personal data. In contrast, data ownership laws in the United States are less clear-cut, with recent state-level regulations echoing GDPR’s principles but failing to resolve ambiguity. Generative AI often ingests vast amounts of data, much of which may not belong to the person uploading it. This creates legal risks for both users and AI model providers, especially when third-party data is involved. Cases surrounding intellectual property, such as controversies involving Slack, Reddit, and LinkedIn, highlight public resistance to having personal data used for AI training. As lawsuits in this arena emerge, prior intellectual property rulings could shape the legal landscape for generative AI. 2. What Data Can Be Derived from LLM Output? Generative AI models are designed to be helpful, but they can inadvertently expose sensitive or proprietary information submitted during training. This risk has made many wary of uploading critical data into AI models. Techniques like tokenization, anonymization, and pseudonymization can reduce these risks by obscuring sensitive data before it is fed into AI systems. However, these practices may compromise the model’s performance by limiting the quality and specificity of the training data. Advocates for GenAI stress that high-quality, accurate data is essential to achieving the best results, which adds to the complexity of balancing privacy with performance. 3. Can the Output Be Trusted? The phenomenon of “hallucinations” — when generative AI produces incorrect or fabricated information — poses another significant concern. Whether these errors stem from poor training, flawed data, or malicious intent, they raise questions about the reliability of GenAI outputs. The impact of hallucinations varies depending on the context. While some errors may cause minor inconveniences, others could have serious or even dangerous consequences, particularly in sensitive domains like healthcare or legal advisory. As generative AI continues to evolve, ensuring the accuracy and integrity of its outputs will remain a top priority. The Generative AI Data Governance Imperative Generative AI’s transformative power lies in its ability to leverage vast amounts of information. For information security, data privacy, and governance professionals, this means grappling with key questions, such as: With high stakes and no way to reverse intellectual property violations, the need for robust data governance frameworks is urgent. As society navigates this transformative era, balancing innovation with responsibility will determine whether generative AI becomes a tool for progress or a source of new challenges. While generative AI heralds a bold future, history reminds us that groundbreaking advancements often come with growing pains. It is the responsibility of stakeholders to anticipate and address these challenges to ensure a safer and more equitable AI-powered world. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more Health Cloud Brings Healthcare Transformation Following swiftly after last week’s successful launch of Financial Services Cloud, Salesforce has announced the second installment in its series Read more

Read More
Consider AI Agents Personas

Consider AI Agents Personas

Treating AI Agents as Personas: Introducing the Era of Agent-Computer Interaction The UX landscape is evolving. While the design community has quickly adopted Large Language Models (LLMs) as tools, we’ve yet to fully grasp their transformative potential. With AI agents now deeply embedded in digital products, they are shifting from tools to active participants in our digital ecosystems. This change demands a new design paradigm—one that views AI agents not just as extensions of human users but as independent personas in their own right. The Rise of Agent-Computer Interaction AI agents represent a new class of users capable of navigating interfaces autonomously and completing complex tasks. This marks the dawn of Agent-Computer Interaction (ACI)—a paradigm where user experience design encompasses the needs of both human users and AI agents. Humans still play a critical role in guiding and supervising these systems, but AI agents must now be treated as distinct personas with unique goals, abilities, and requirements. This shift challenges UX designers to consider how these agents interact with interfaces and perform their tasks, ensuring they are equipped with the information and resources necessary to operate effectively. Understanding AI Agents AI agents are intelligent systems designed to reason, plan, and work across platforms with minimal human intervention. As defined during Google I/O, these agents retain context, anticipate needs, and execute multi-step processes. Advances in AI, such as Anthropic’s Claude and its ability to interact with graphical interfaces, have unlocked new levels of agency. Unlike earlier agents that relied solely on APIs, modern agents can manipulate graphical user interfaces much like human users, enabling seamless interaction with browser-based applications. This capability creates opportunities for new forms of interaction but also demands thoughtful design choices. Two Interaction Approaches for AI Agents Design teams must evaluate these methods based on the task’s complexity and transparency requirements, striking the right balance between efficiency and oversight. Designing Experiences Considering AI Agents Personas As AI agents transition into active users, UX design must expand to accommodate their specific needs. Much like human personas, AI agents require a deep understanding of their capabilities, limitations, and workflows. Creating AI Agent Personas Developing personas for AI agents involves identifying their unique characteristics: These personas inform interface designs that optimize agent workflows, ensuring both agents and humans can collaborate effectively. New UX Research Methodologies UX teams should embrace innovative research techniques, such as A/B testing interfaces for agent performance and monitoring their interaction patterns. While AI agents lack sentience, they exhibit behaviors—reasoning, planning, and adapting—that require careful study and design consideration. Shaping the AI Mind AI agents derive their reasoning capabilities from Large Language Models (LLMs), but their behavior and effectiveness are shaped by UX design. Designers have a unique role in crafting system prompts and developing feedback loops that refine LLM behavior over time. Key Areas for Designer Involvement: This work positions UX professionals as co-creators of AI intelligence, shaping not just interfaces but the underlying behaviors that drive agent interactions. Keeping Humans in the Loop Despite the rise of AI agents, human oversight and control remain essential. UX practitioners must prioritize transparency and trust in agent-driven systems. Key Considerations: Using tools like agentic experience maps—blueprints that visualize the interactions between humans, agents, and products—designers can ensure AI systems remain human-centered. A New Frontier for UX The emergence of AI agents heralds a shift as significant as the transition from desktop to mobile. Just as mobile devices unlocked new opportunities for interaction, AI agents are poised to redefine digital experiences in ways we can’t yet fully predict. By embracing Agent-Computer Interaction, UX designers have an unprecedented opportunity to shape the future of human-AI collaboration. Those who develop expertise in designing for these intelligent agents will lead the way in creating systems that are not only powerful but also deeply human-centered. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
being ai-driven

Being AI-Driven

Imagine a company where every decision, strategy, customer interaction, and routine task is enhanced by AI. From predictive analytics uncovering market insights to intelligent automation streamlining operations, this AI-driven enterprise represents what a successful business could look like. Does this company exist? Not yet, but the building blocks for creating it are already here. To envision a day in the life of such an AI enterprise, let’s fast forward to the year 2028 and visit Tectonic 5.0, a fictional 37-year-old mid-sized company in Oklahoma that provides home maintenance services. After years of steady sales and profit growth, the 2,300-employee company has hit a rough patch. Tectonic 5.0’s revenue grew just 3% last year, and its 8% operating margin is well below the industry benchmark. To jumpstart growth, Tectonic 5.0 has expanded its product portfolio and decided to break into the more lucrative commercial real estate market. But Tectonic 5.0 needs to act fast. The firm must quickly bring its new offerings to market while boosting profitability by eliminating inefficiencies and fostering collaboration across teams. To achieve these goals, Tectonic 5.0 is relying on artificial intelligence (AI). Here’s how each department at Tectonic 5.0 is using AI to reach these objectives. Spot Inefficiencies with AI With a renewed focus on cost-cutting, Tectonic 5.0 needed to identify and eliminate inefficiencies throughout the company. To assist in this effort, the company developed a tool called Jenny, an AI agent that’s automatically invited to all meetings. Always listening and analyzing, Jenny spots problems and inefficiencies that might otherwise go unnoticed. For example, Jenny compares internal data against industry benchmarks and historical data, identifying opportunities for optimization based on patterns in spending and resource allocation. Suggestions for cost-cutting can be offered in real time during meetings or shared later in a synthesized summary. AI can also analyze how meeting time is spent, revealing if too much time is wasted on non-essential issues and suggesting ways to have more constructive meetings. It does this by comparing meeting summaries against the company’s broader objectives. Tectonic 5.0’s leaders hope that by highlighting inefficiencies and communication gaps with Jenny’s help, employees will be more inclined to take action. In fact, it has already shown considerable promise, with employees being five times more likely to consider cost-cutting measures suggested by Penny. Market More Effectively with AI With cost management underway, Tectonic 5.0’s next step in its transformation is finding new revenue sources. The company has adopted a two-pronged approach: introducing a new lineup of products and services for homeowners, including smart home technology, sustainable living solutions like solar panels, and predictive maintenance on big-ticket systems like internet-connected HVACs; and expanding into commercial real estate maintenance. Smart home technology is exactly what homeowners are looking for, but Tectonic 5.0 needs to market it to the right customers, at the right time, and in the right way. A marketing platform with built-in AI capabilities is essential for spreading the word quickly and effectively about its new products. To start, the company segments its audience using generative AI, allowing marketers to ask the system, in natural language, to identify tech-savvy homeowners between the ages of 30 and 60 who have spent a certain amount on home maintenance in the last 18 months. This enables more precise audience targeting and helps marketing teams bring products to market faster. Previously, segmentation using legacy systems could take weeks, with marketing teams relying on tech teams for an audience breakdown. Now, Tectonic 5.0 is ready to reach out to its targeted customers. Using predictive AI, it can optimize personalized marketing campaigns. For example, it can determine which customers prefer to be contacted by text, email, or phone, the best time of day to reach out, and how often. The system also identifies which messaging—focused on cost savings, environmental impact, or preventative maintenance—will resonate most with each customer. This intelligence helps Tectonic 5.0 reach the optimal customer quickly in a way that speaks to their specific needs and concerns. AI also enables marketers to monitor campaign performance for red flags like decreasing open rates or click-through rates and take appropriate action. Sell More, and Faster, with AI With interested buyers lined up, it’s now up to the sales team to close deals. Generative AI for sales, integrated into CRM, can speed up and personalize the sales process for Tectonic 5.0 in several ways. First, it can generate email copy tailored to products and services that customers are interested in. Tectonic 5.0’s sales reps can prompt AI to draft solar panel prospecting emails. To maximize effectiveness, the system pulls customer info from the CRM, uncovering which emails have performed well in the past. Second, AI speeds up data analysis. Sales reps spend a significant amount of time generating, pulling, and analyzing data. Generative AI can act like a digital assistant, uncovering patterns and relationships in CRM data almost instantaneously, guiding Tectonic 5.0’s reps toward high-value deals most likely to close. Machine learning increases the accuracy of lead scoring, predicting which customers are most likely to buy based on historical data and predictive analytics. Provide Better Customer Service with AI Tectonic 5.0’s new initiatives are progressing well. Costs are starting to decrease, and sales of its new products are growing faster than expected. However, customer service calls are rising as well. Tectonic 5.0 is committed to maintaining excellent customer service, but smart home technology presents unique challenges. It’s more complex than analog systems, and customers often need help with setup and use, raising the stakes for Tectonic 5.0’s customer service team. The company knows that customers have many choices in home maintenance providers, and one bad experience could drive them to a competitor. Tectonic 5.0’s embedded AI-powered chatbots help deliver a consistent and delightful autonomous customer service experience across channels and touchpoints. Beyond answering common questions, these chatbots can greet customers, serve up knowledge articles, and even dispatch a field technician if needed. In the field, technicians can quickly diagnose and fix problems thanks to LLMs like xGen-Small, which

Read More
AI Agents

AI Agents Interview

In the rapidly evolving world of large language models and generative AI, a new concept is gaining momentum: AI agents. AI Agents Interview explores. AI agents are advanced tools designed to handle complex tasks that traditionally required human intervention. While they may be confused with robotic process automation (RPA) bots, AI agents are much more sophisticated, leveraging generative AI technology to execute tasks autonomously. Companies like Google are positioning AI agents as virtual assistants that can drive productivity across industries. In this Q&A, Jason Gelman, Director of Product Management for Vertex AI at Google Cloud, shares insights into Google’s vision for AI agents and some of the challenges that come with this emerging technology. AI Agents Interview How does Google define AI agents? Jason Gelman: An AI agent is something that acts on your behalf. There are two key components. First, you empower the agent to act on your behalf by providing instructions and granting necessary permissions—like authentication to access systems. Second, the agent must be capable of completing tasks. This is where large language models (LLMs) come in, as they can plan out the steps to accomplish a task. What used to require human planning is now handled by the AI, including gathering information and executing various steps. What are current use cases where AI agents can thrive? Gelman: AI agents can be useful across a wide range of industries. Call centers are a common example where customers already expect AI support, and we’re seeing demand there. In healthcare, organizations like Mayo Clinic are using AI agents to sift through vast amounts of information, helping professionals navigate data more efficiently. Different industries are exploring this technology in unique ways, and it’s gaining traction across many sectors. What are some misconceptions about AI agents? Gelman: One major misconception is that the technology is more advanced than it actually is. We’re still in the early stages, building critical infrastructure like authentication and function-calling capabilities. Right now, AI agents are more like interns—they can assist, but they’re not yet fully autonomous decision-makers. While LLMs appear powerful, we’re still some time away from having AI agents that can handle everything independently. Developing the technology and building trust with users are key challenges. I often compare this to driverless cars. While they might be safer than human drivers, we still roll them out cautiously. With AI agents, the risks aren’t physical, but we still need transparency, monitoring, and debugging capabilities to ensure they operate effectively. How can enterprises balance trust in AI agents while acknowledging the technology is still evolving? Gelman: Start simple and set clear guardrails. Build an AI agent that does one task reliably, then expand from there. Once you’ve proven the technology’s capability, you can layer in additional tasks, eventually creating a network of agents that handle multiple responsibilities. Right now, most organizations are still in the proof-of-concept phase. Some companies are using AI agents for more complex tasks, but for critical areas like financial services or healthcare, humans remain in the loop to oversee decision-making. It will take time before we can fully hand over tasks to AI agents. AI Agents Interview What is the difference between Google’s AI agent and Microsoft Copilot? Gelman: Microsoft Copilot is a product designed for business users to assist with personal tasks. Google’s approach with AI agents, particularly through Vertex AI, is more focused on API-driven, developer-based solutions that can be integrated into applications. In essence, while Copilot serves as a visible assistant for users, Vertex AI operates behind the scenes, embedded within applications, offering greater flexibility and control for enterprise customers. The real potential of AI agents lies in their ability to execute a wide range of tasks at the API level, without the limitations of a low-code/no-code interface. Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
Trends in AI for CRM

Trends in AI for CRM

Nearly half of customer service teams, over 40% of salespeople, and a third of marketers have fully implemented artificial intelligence (AI) to enhance their work. However, 77% of business leaders report persistent challenges related to trusted data and ethical concerns that could stall their AI initiatives, according to Salesforce research released today. The Trends in AI for CRM report analyzed data from multiple studies, revealing that companies are worried about missing out on the opportunities generative AI presents if the data powering large language models (LLMs) isn’t rooted in their own trusted customer records. At the same time, respondents expressed ongoing concerns about the lack of clear company policies governing the ethical use of AI, as well as the complexity of a vendor landscape where 80% of enterprises are currently using multiple LLMs. Salesforce’s Four Keys to Enterprise AI Success Why it matters: AI is one of the most transformative technologies in generations, with projections forecasting a net gain of over trillion in new business revenues by 2028 from Salesforce and its network of partners alone. As enterprises across industries develop their AI strategies, leaders in customer-facing departments such as sales, service, and marketing are eager to leverage AI to drive internal efficiencies and revolutionize customer experiences. Key Findings from the Trends in AI for CRM Report Expert Perspective “This is a pivotal moment as business leaders across industries look to AI to unlock growth, efficiency, and customer loyalty,” said Clara Shih, CEO of Salesforce AI. “But success requires much more than an LLM. Enterprise deployments need trusted data, user access control, vector search, audit trails and citations, data masking, low-code builders, and seamless UI integration. Salesforce brings all of these components together with our Einstein 1 Platform, Data Cloud, Slack, and dozens of customizable, turnkey prompts and actions offered across our clouds.” Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched a new service called AppExchange OEM Edition, aimed at non-CRM service providers. Read more The Salesforce Story In Marc Benioff’s own words How did salesforce.com grow from a start up in a rented apartment into the world’s Read more Salesforce Jigsaw Salesforce.com, a prominent figure in cloud computing, has finalized a deal to acquire Jigsaw, a wiki-style business contact database, for Read more Service Cloud with AI-Driven Intelligence Salesforce Enhances Service Cloud with AI-Driven Intelligence Engine Data science and analytics are rapidly becoming standard features in enterprise applications, Read more

Read More
AI Assistants Using LangGraph

AI Assistants Using LangGraph

In the evolving world of AI, retrieval-augmented generation (RAG) systems have become standard for handling straightforward queries and generating contextually relevant responses. However, as demand grows for more sophisticated AI applications, there is a need for systems that move beyond simple retrieval tasks. Enter AI agents—autonomous entities capable of executing complex, multi-step processes, maintaining state across interactions, and dynamically adapting to new information. LangGraph, a powerful extension of the LangChain library, is designed to help developers build these advanced AI agents, enabling stateful, multi-actor applications with cyclic computation capabilities. AI Assistants Using LangGraph. In this insight, we’ll explore how LangGraph revolutionizes AI development and provide a step-by-step guide to building your own AI agent using an example that computes energy savings for solar panels. This example will demonstrate how LangGraph’s unique features enable the creation of intelligent, adaptable, and practical AI systems. What is LangGraph? LangGraph is an advanced library built on top of LangChain, designed to extend Large Language Model (LLM) applications by introducing cyclic computational capabilities. While LangChain allows for the creation of Directed Acyclic Graphs (DAGs) for linear workflows, LangGraph enhances this by enabling the addition of cycles—essential for developing agent-like behaviors. These cycles allow LLMs to continuously loop through processes, making decisions dynamically based on evolving inputs. LangGraph: Nodes, States, and Edges The core of LangGraph lies in its stateful graph structure: LangGraph redefines AI development by managing the graph structure, state, and coordination, allowing for the creation of sophisticated, multi-actor applications. With automatic state management and precise agent coordination, LangGraph facilitates innovative workflows while minimizing technical complexity. Its flexibility enables the development of high-performance applications, and its scalability ensures robust and reliable systems, even at the enterprise level. Step-by-step Guide Now that we understand LangGraph’s capabilities, let’s dive into a practical example. We’ll build an AI agent that calculates potential energy savings for solar panels based on user input. This agent can function as a lead generation tool on a solar panel seller’s website, providing personalized savings estimates based on key data like monthly electricity costs. This example highlights how LangGraph can automate complex tasks and deliver business value. Step 1: Import Necessary Libraries We start by importing the essential Python libraries and modules for the project. pythonCopy codefrom langchain_core.tools import tool from langchain_community.tools.tavily_search import TavilySearchResults from langchain_core.prompts import ChatPromptTemplate from langchain_core.runnables import Runnable from langchain_aws import ChatBedrock import boto3 from typing import Annotated from typing_extensions import TypedDict from langgraph.graph.message import AnyMessage, add_messages from langchain_core.messages import ToolMessage from langchain_core.runnables import RunnableLambda from langgraph.prebuilt import ToolNode Step 2: Define the Tool for Calculating Solar Savings Next, we define a tool to calculate potential energy savings based on the user’s monthly electricity cost. pythonCopy code@tool def compute_savings(monthly_cost: float) -> float: “”” Tool to compute the potential savings when switching to solar energy based on the user’s monthly electricity cost. Args: monthly_cost (float): The user’s current monthly electricity cost. Returns: dict: A dictionary containing: – ‘number_of_panels’: The estimated number of solar panels required. – ‘installation_cost’: The estimated installation cost. – ‘net_savings_10_years’: The net savings over 10 years after installation costs. “”” def calculate_solar_savings(monthly_cost): cost_per_kWh = 0.28 cost_per_watt = 1.50 sunlight_hours_per_day = 3.5 panel_wattage = 350 system_lifetime_years = 10 monthly_consumption_kWh = monthly_cost / cost_per_kWh daily_energy_production = monthly_consumption_kWh / 30 system_size_kW = daily_energy_production / sunlight_hours_per_day number_of_panels = system_size_kW * 1000 / panel_wattage installation_cost = system_size_kW * 1000 * cost_per_watt annual_savings = monthly_cost * 12 total_savings_10_years = annual_savings * system_lifetime_years net_savings = total_savings_10_years – installation_cost return { “number_of_panels”: round(number_of_panels), “installation_cost”: round(installation_cost, 2), “net_savings_10_years”: round(net_savings, 2) } return calculate_solar_savings(monthly_cost) Step 3: Set Up State Management and Error Handling We define utilities to manage state and handle errors during tool execution. pythonCopy codedef handle_tool_error(state) -> dict: error = state.get(“error”) tool_calls = state[“messages”][-1].tool_calls return { “messages”: [ ToolMessage( content=f”Error: {repr(error)}n please fix your mistakes.”, tool_call_id=tc[“id”], ) for tc in tool_calls ] } def create_tool_node_with_fallback(tools: list) -> dict: return ToolNode(tools).with_fallbacks( [RunnableLambda(handle_tool_error)], exception_key=”error” ) Step 4: Define the State and Assistant Class We create the state management class and the assistant responsible for interacting with users. pythonCopy codeclass State(TypedDict): messages: Annotated[list[AnyMessage], add_messages] class Assistant: def __init__(self, runnable: Runnable): self.runnable = runnable def __call__(self, state: State): while True: result = self.runnable.invoke(state) if not result.tool_calls and ( not result.content or isinstance(result.content, list) and not result.content[0].get(“text”) ): messages = state[“messages”] + [(“user”, “Respond with a real output.”)] state = {**state, “messages”: messages} else: break return {“messages”: result} Step 5: Set Up the LLM with AWS Bedrock We configure AWS Bedrock to enable advanced LLM capabilities. pythonCopy codedef get_bedrock_client(region): return boto3.client(“bedrock-runtime”, region_name=region) def create_bedrock_llm(client): return ChatBedrock(model_id=’anthropic.claude-3-sonnet-20240229-v1:0′, client=client, model_kwargs={‘temperature’: 0}, region_name=’us-east-1′) llm = create_bedrock_llm(get_bedrock_client(region=’us-east-1′)) Step 6: Define the Assistant’s Workflow We create a template and bind the tools to the assistant’s workflow. pythonCopy codeprimary_assistant_prompt = ChatPromptTemplate.from_messages( [ ( “system”, ”’You are a helpful customer support assistant for Solar Panels Belgium. Get the following information from the user: – monthly electricity cost Ask for clarification if necessary. ”’, ), (“placeholder”, “{messages}”), ] ) part_1_tools = [compute_savings] part_1_assistant_runnable = primary_assistant_prompt | llm.bind_tools(part_1_tools) Step 7: Build the Graph Structure We define nodes and edges for managing the AI assistant’s conversation flow. pythonCopy codebuilder = StateGraph(State) builder.add_node(“assistant”, Assistant(part_1_assistant_runnable)) builder.add_node(“tools”, create_tool_node_with_fallback(part_1_tools)) builder.add_edge(START, “assistant”) builder.add_conditional_edges(“assistant”, tools_condition) builder.add_edge(“tools”, “assistant”) memory = MemorySaver() graph = builder.compile(checkpointer=memory) Step 8: Running the Assistant The assistant can now be run through its graph structure to interact with users. python import uuidtutorial_questions = [ ‘hey’, ‘can you calculate my energy saving’, “my montly cost is $100, what will I save”]thread_id = str(uuid.uuid4())config = {“configurable”: {“thread_id”: thread_id}}_printed = set()for question in tutorial_questions: events = graph.stream({“messages”: (“user”, question)}, config, stream_mode=”values”) for event in events: _print_event(event, _printed) Conclusion By following these steps, you can create AI Assistants Using LangGraph to calculate solar panel savings based on user input. This tutorial demonstrates how LangGraph empowers developers to create intelligent, adaptable systems capable of handling complex tasks efficiently. Whether your application is in customer support, energy management, or other domains, LangGraph provides the Like Related Posts Salesforce OEM AppExchange Expanding its reach beyond CRM, Salesforce.com has launched

Read More
Scaling Generative AI

Scaling Generative AI

Many organizations follow a hybrid approach to AI infrastructure, combining public clouds, colocation facilities, and on-prem solutions. Specialized GPU-as-a-service vendors, for instance, are becoming popular for handling high-demand AI computations, helping businesses manage costs without compromising performance. Business process outsourcing company TaskUs, for example, focuses on optimizing compute and data flows as it scales its gen AI deployments, while Cognizant advises that companies distinguish between training and inference needs, each with different latency requirements.

Read More
gettectonic.com