GPU Archives - gettectonic.com
Building the Intelligent Enterprise Network

Building the Intelligent Enterprise Network

Blueprint for the Agentic AI Era: Building the Intelligent Enterprise Network The Next Frontier: Agentic AI Demands a New Network Paradigm At Cisco Live 2024, company executives unveiled a strategic vision for enterprise AI that goes beyond today’s generative capabilities. As Jeetu Patel, Cisco’s Chief Product Officer, stated: “We’re witnessing one of the most consequential technological shifts in history—the move from reactive AI assistants to autonomous agentic systems that execute complex workflows.” This transition requires fundamental changes to enterprise infrastructure. Where generative AI focused on content creation, agentic AI introduces self-directed software agents that:✅ Operate autonomously across systems✅ Make real-time decisions without human intervention✅ Coordinate multi-step business processes Cisco’s Three Pillars for Agentic AI Success 1. Simplified Network Operations with AI Cisco is unifying its Catalyst and Meraki platforms into a single AI-powered management console featuring: “The future isn’t just AI-assisted ops—it’s agentic ops where AI systems autonomously maintain network health,” noted DJ Sampath, SVP of AI Platform at Cisco. 2. AI-Optimized Hardware Infrastructure New product releases specifically designed for AI workloads:🔹 Catalyst 9800-X Series – 400Gbps switches with AI-optimized ASICs🔹 Silicon One G200 Routers – Built-in NGFW and SD-WAN for distributed AI🔹 Wi-Fi 7 Access Points – 320MHz channels for high-density AI agent traffic 3. Security-Infused Network Fabric Cisco’s “Zero Trust by Design” approach incorporates: Why Networking is AI’s Make-or-Break Factor Patel highlighted a critical insight: “GPUs are only as good as their data pipelines. An idle GPU waiting for packets is like burning cash.” Cisco’s internal benchmarks show: 📉 30% GPU utilization on poorly configured networks📈 92% utilization on Cisco’s AI-optimized infrastructure The difference comes from: The Agentic AI Future: Beyond Hype to Transformation While some dismiss AI as overhyped, Cisco executives argue the true revolution is just beginning: “Agentic AI won’t just answer questions—it will create original insights and solve problems we couldn’t approach before. But this requires rethinking every layer of infrastructure.”— Jeetu Patel, EVP & Chief Product Officer, Cisco Early adopters are already seeing results: Preparing Your Enterprise Cisco recommends three immediate actions: “The companies that win will be those that build networks where AI agents thrive as first-class citizens,” Patel concluded. Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more Tectonic’s Successful Salesforce Track Record Salesforce Technology Services Integrator – Tectonic has successfully delivered Salesforce in a variety of industries including Public Sector, Hospitality, Manufacturing, Read more

Read More
DXP - Digital Experience Platform

DXP – Digital Experience Platform

A Digital Experience Platform (DXP) is a set of integrated technologies that help organizations create, manage, and deliver personalized digital experiences across various touchpoints. DXPs aim to provide a central hub for managing a company’s digital ecosystem, enabling consistent and engaging customer interactions. They often include features like content management, e-commerce, personalization, and experimentation.  Key aspects of a DXP: Benefits of using a DXP: Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Marketing Cloud Transactional Emails Salesforce Marketing Cloud Transactional Emails are immediate, automated, non-promotional messages crucial to business operations and customer satisfaction, such as order Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more

Read More
unpatched ai

Scrape the Web for Training Data

Do AI Companies Have the Right to Scrape the Web for Training Data? For the past two years, generative AI companies have faced lawsuits—some from high-profile authors and publishers—while simultaneously striking multi-million-dollar data licensing deals. Despite the legal battles, the political tide seems to be shifting in favor of AI firms. Both the European Union and the UK appear to be leaning toward an “opt-out” model, where web scraping is permitted unless content owners explicitly forbid it. But critical questions remain: How exactly does “opting out” work? And do creators and publishers truly have a fair chance to do so? Data as the New Oil The most valuable asset in AI isn’t GPUs or data centers—it’s the training data itself. Without the vast troves of text, images, videos, and artwork produced over decades (or even centuries), there would be no ChatGPT, Gemini, or Claude. Web scraping is nothing new. Search engines like Google have relied on crawlers for decades, indexing the web to deliver search results. But the rules of the game have changed. Old Conventions, New Conflicts Historically, website owners welcomed search engine crawlers to boost visibility while others (especially news publishers) saw them as competitors. The Robots Exclusion Standard (robots.txt) emerged as a gentleman’s agreement—a way for sites to signal which pages could be crawled. While robots.txt isn’t legally binding, reputable search engines like Google and Bing generally respect it. The arrangement was symbiotic: websites got traffic, and search engines got data. But AI crawlers operate differently. They don’t drive traffic—they consume content to generate competing products, often commercializing it via AI services. Will AI companies play fair? Nick Clegg, former UK deputy PM and current Meta executive, bluntly stated that requiring permission from artists would “kill” the AI industry. If unfettered data access is seen as existential, can we expect AI firms to respect opt-outs? Can Websites Really Block AI Crawlers? Theoretically, yes—by blocking AI user agents or monitoring suspicious traffic. But this is a game of whack-a-mole, requiring constant vigilance. And what about offline content? Books, research papers, and proprietary datasets aren’t protected by robots.txt. Some AI companies have allegedly bypassed ethical scraping altogether, sourcing data from shadowy corners of the internet—like torrent sites—as revealed in a recent lawsuit against Meta. The Transparency Problem Even if content owners could opt out, how would they know if their data was already used? Why resist transparency? Only two explanations make sense: Neither is a good look. Beyond Copyright: The Bigger Questions This debate isn’t just about copyright—it’s about: And what happens when Google replaces traditional search with AI summaries? Websites may face an impossible choice: Allow AI training or disappear from search results altogether. The Future of the Open Web If AI companies continue scraping indiscriminately, the open web could shrink further, with more content locked behind paywalls and logins. Ironically, the very ecosystem AI relies on may be destroyed by its own hunger for data. The question isn’t just whether AI firms have the right to scrape the web—but whether the web as we know it will survive their appetite. Footnotes Key Takeaways ✅ AI companies are winning the legal/political battle for web scraping rights.⚠️ Opt-out mechanisms (like robots.txt) may be ignored.🔍 Transparency is lacking—many AI firms won’t disclose training data sources.🌐 Indiscriminate scraping could kill the open web, pushing content behind paywalls. Would love to hear your thoughts—should AI companies have free rein over web data, or do content creators deserve more control? Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more Tectonic’s Successful Salesforce Track Record Salesforce Technology Services Integrator – Tectonic has successfully delivered Salesforce in a variety of industries including Public Sector, Hospitality, Manufacturing, Read more

Read More
Far Beyond Keywords

Far Beyond Keywords

Far Beyond Keywords: The Next Era of Intelligent Search with NLP & Vector Embeddings Traditional search has served us well—scalable systems can scan structured data in seconds using keywords, tags, or schemas. But 90% of enterprise data is unstructured: emails, support tickets, PDFs, audio, and video. Keyword search fails here because human language is nuanced—we use metaphors, synonyms, and context that rigid keyword matching can’t grasp. To search unstructured data effectively, we need AI-powered semantic understanding—not just pattern matching. How Neural Networks Understand Language Modern NLP models rely on neural networks (NNs), which aren’t magic—they’re pattern-recognition engines trained on vast text datasets. Here’s how they learn: From Words to Semantic Search To search entire documents, we: Why It’s Better Than Keyword Search ✅ Finds conceptually related content (e.g., “sustainability” matches “eco-friendly initiatives”).✅ Ignores exact phrasing—understands intent.✅ Faster at scale—vector math outperforms text scanning. Scaling Semantic Search with Vector Databases Storing millions of vectors requires specialized vector databases (e.g., Pinecone, Milvus), optimized for: 🔹 Low-latency retrieval – Nearest-neighbor search in milliseconds.🔹 Horizontal scaling – Partition data across clusters.🔹 Incremental updates – Only re-embed modified text.🔹 GPU acceleration – 2-3x faster queries vs. CPU. Real-World Impact Frameworks like AgoraWiki apply these principles to deliver: The Future of Search As NLP advances, semantic search will become smarter, faster, and more contextual—transforming how enterprises unlock insights from unstructured data. Ready to move beyond keywords? Explore AI-powered search solutions today. Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more Tectonic’s Successful Salesforce Track Record Salesforce Technology Services Integrator – Tectonic has successfully delivered Salesforce in a variety of industries including Public Sector, Hospitality, Manufacturing, Read more

Read More
Snowpark Container Services

Snowpark Container Services

Snowpark Container Services (SPCS) is a fully managed container service within Snowflake that allows you to deploy and manage containerized applications and services directly within the Snowflake environment. It enables you to run code, process data, and deploy machine learning models without moving data out of Snowflake.  Here’s a more detailed breakdown: In essence, SPCS extends the capabilities of Snowflake by providing a managed container runtime where you can run custom applications and services alongside your data, without the need to manage the underlying infrastructure.  Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more Tectonic’s Successful Salesforce Track Record Salesforce Technology Services Integrator – Tectonic has successfully delivered Salesforce in a variety of industries including Public Sector, Hospitality, Manufacturing, Read more

Read More
Whoever cracks reliable, scalable atomic power first could gain an insurmountable edge in the AI arms race.

The Nuclear Power Revival

The Nuclear Power Revival: How Big Tech is Fueling AI with Small Modular Reactors From Meltdowns to Megawatts: Nuclear’s Second Act Following two catastrophic nuclear accidents—Three Mile Island (1979) and Chernobyl (1986)—public trust in atomic energy plummeted. But today, an unlikely force is driving its resurgence: artificial intelligence. As generative AI explodes in demand, tech giants face an unprecedented energy crisis. Data centers, already consuming 2-3% of U.S. electricity, could devour 9% by 2030 (Electric Power Research Institute). With aging power grids struggling to keep up, cloud providers are taking matters into their own hands—by turning to small modular reactors (SMRs). Why AI Needs Nuclear Power The Energy Crisis No One Saw Coming Enter Small Modular Reactors (SMRs) The global SMR market for data centers is projected to hit 8M by 2033, growing at 48.72% annually (Research and Markets). The Big Four Tech Players Going Nuclear 1. Microsoft: Reviving Three Mile Island 2. Google: Betting on Next-Gen SMRs 3. Amazon: Three-Pronged Nuclear Push 4. Oracle: Plans Under Wraps The Startups Building Tomorrow’s Nuclear Tech Company Backer/Notable Feature Innovation Oklo Sam Altman (OpenAI) Rural SMRs targeting 2027 launch TerraPower Bill Gates Sodium-cooled fast reactors NuScale First U.S.-approved SMR design Factory-built, modular light-water reactors Last Energy 80+ microreactors planned in Europe/Texas 20MW units for data centers Deep Atomic Swiss startup MK60 reactor with dedicated cooling power Valar Atomics “Gigasite” assembly lines On-site SMR production Newcleo Lead-cooled fast reactors Higher safety via liquid metal cooling Challenges Ahead The Bottom Line As AI’s hunger for power grows exponentially, Big Tech is bypassing traditional utilities to build its own nuclear future. While risks remain, SMRs offer a scalable, clean solution—potentially rewriting energy economics in the AI era. The race is on: Whoever cracks reliable, scalable atomic power first could gain an insurmountable edge in the AI arms race. Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more Tectonic’s Successful Salesforce Track Record Salesforce Technology Services Integrator – Tectonic has successfully delivered Salesforce in a variety of industries including Public Sector, Hospitality, Manufacturing, Read more

Read More

Grok 3 Model Explained

Grok 3 Model Explained: Everything You Need to Know xAI has introduced its latest large language model (LLM), Grok 3, expanding its capabilities with advanced reasoning, knowledge retrieval, and text summarization. In the competitive landscape of generative AI (GenAI), LLMs and their chatbot services have become essential tools for users and organizations. While OpenAI’s ChatGPT (powered by the GPT series) pioneered the modern GenAI era, alternatives like Anthropic’s Claude, Google Gemini, and now Grok (developed by Elon Musk’s xAI) offer diverse choices. The term grok originates from Robert Heinlein’s 1961 sci-fi novel Stranger in a Strange Land, meaning to deeply understand something. Grok is closely tied to X (formerly Twitter), where it serves as an integrated AI chatbot, though it’s also available on other platforms. What Is Grok 3? Grok 3 is xAI’s latest LLM, announced on February 17, 2025, in a live stream featuring CEO Elon Musk and the engineering team. Musk, known for founding Tesla, SpaceX, and acquiring Twitter (now X), launched xAI on March 9, 2023, with the mission to “understand the universe.” Grok 3 is the third iteration of the model, built using Rust and Python. Unlike Grok 1 (partially open-sourced under Apache 2.0), Grok 3 is proprietary. Key Innovations in Grok 3 Grok 3 excels in advanced reasoning, positioning it as a strong competitor against models like OpenAI’s o3 and DeepSeek-R1. What Can Grok 3 Do? Grok 3 operates in two core modes: 1. Think Mode 2. DeepSearch Mode Core Capabilities ✔ Advanced Reasoning – Multi-step problem-solving with self-correction.✔ Content Summarization – Text, images, and video summaries.✔ Text Generation – Human-like writing for various use cases.✔ Knowledge Retrieval – Accesses real-time web data (especially in DeepSearch mode).✔ Mathematics – Strong performance on benchmarks like AIME 2024.✔ Coding – Writes, debugs, and optimizes code.✔ Voice Mode – Supports spoken responses. Previous Grok Versions Model Release Date Key Features Grok 1 Nov. 3, 2023 Humorous, personality-driven responses. Grok 1.5 Mar. 28, 2024 Expanded context (128K tokens), better problem-solving. Grok 1.5V Apr. 12, 2024 First multimodal version (image understanding). Grok 2 Aug. 14, 2024 Full multimodal support, image generation via Black Forest Labs’ FLUX. Grok 3 vs. GPT-4o vs. DeepSeek-R1 Feature Grok 3 GPT-4o DeepSeek-R1 Release Date Feb. 17, 2025 May 24, 2024 Jan. 20, 2025 Developer xAI (USA) OpenAI (USA) DeepSeek (China) Reasoning Advanced (Think mode) Limited Strong Real-Time Data DeepSearch (web access) Training data cutoff Training data cutoff License Proprietary Proprietary Open-source Coding (LiveCodeBench) 79.4 72.9 64.3 Math (AIME 2024) 99.3 87.3 79.8 How to Use Grok 3 1. On X (Twitter) 2. Grok.com 3. Mobile App (iOS/Android) Same subscription options as Grok.com. 4. API (Coming Soon) No confirmed release date yet. Final Thoughts Grok 3 is a powerful reasoning-focused LLM with real-time search capabilities, making it a strong alternative to GPT-4o and DeepSeek-R1. With its DeepSearch and Think modes, it offers advanced problem-solving beyond traditional chatbots. Will it surpass OpenAI and DeepSeek? Only time—and benchmarks—will tell.  Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more Tectonic’s Successful Salesforce Track Record Salesforce Technology Services Integrator – Tectonic has successfully delivered Salesforce in a variety of industries including Public Sector, Hospitality, Manufacturing, Read more

Read More
Neuro-symbolic AI

Neuro-symbolic AI

Neuro-Symbolic AI: Bridging Neural Networks and Symbolic Processing for Smarter AI Systems Neuro-symbolic AI integrates neural networks with rules-based symbolic processing to enhance artificial intelligence systems’ accuracy, explainability, and precision. Neural networks leverage statistical deep learning to identify patterns in large datasets, while symbolic AI applies logic and rules-based reasoning common in mathematics, programming languages, and expert systems. The Balance Between Neural and Symbolic AIThe fusion of neural and symbolic methods has revived debates in the AI community regarding their relative strengths. Neural AI excels in deep learning, including generative AI, by distilling patterns from data through distributed statistical processing across interconnected neurons. However, this approach often requires significant computational resources and may struggle with explainability. Conversely, symbolic AI, which relies on predefined rules and logic, has historically powered applications like fraud detection, expert systems, and argument mining. While symbolic systems are faster and more interpretable, their reliance on manual rule creation has been a limitation. Innovations in training generative AI models now allow more efficient automation of these processes, though challenges like hallucinations and poor mathematical reasoning persist. Complementary Thinking ModelsPsychologist Daniel Kahneman’s analogy of System 1 and System 2 thinking aptly describes the interplay between neural and symbolic AI. Neural AI, akin to System 1, is intuitive and fast—ideal for tasks like image recognition. Symbolic AI mirrors System 2, engaging in slower, deliberate reasoning, such as understanding the context and relationships in a scene. Core Concepts of Neural NetworksArtificial neural networks (ANNs) mimic the statistical connections between biological neurons. By modeling patterns in data, ANNs enable learning and feature extraction at different abstraction levels, such as edges, shapes, and objects in images. Key ANN architectures include: Despite their strengths, neural networks are prone to hallucinations, particularly when overconfident in their predictions, making human oversight crucial. The Role of Symbolic ReasoningSymbolic reasoning underpins modern programming languages, where logical constructs (e.g., “if-then” statements) drive decision-making. Symbolic AI excels in structured applications like solving math problems, representing knowledge, and decision-making. Algorithms like expert systems, Bayesian networks, and fuzzy logic offer precision and efficiency in well-defined workflows but struggle with ambiguity and edge cases. Although symbolic systems like IBM Watson demonstrated success in trivia and reasoning, scaling them to broader, dynamic applications has proven challenging due to their dependency on manual configuration. Neuro-Symbolic IntegrationThe integration of neural and symbolic AI spans a spectrum of techniques, from loosely coupled processes to tightly integrated systems. Examples of integration include: History of Neuro-Symbolic AIBoth neural and symbolic AI trace their roots to the 1950s, with symbolic methods dominating early AI due to their logical approach. Neural networks fell out of favor until the 1980s when innovations like backpropagation revived interest. The 2010s saw a breakthrough with GPUs enabling scalable neural network training, ushering in today’s deep learning era. Applications and Future DirectionsApplications of neuro-symbolic AI include: The next wave of innovation aims to merge these approaches more deeply. For instance, combining granular structural information from neural networks with symbolic abstraction can improve explainability and efficiency in AI systems like intelligent document processing or IoT data interpretation. Neuro-symbolic AI offers the potential to create smarter, more explainable systems by blending the pattern-recognition capabilities of neural networks with the precision of symbolic reasoning. As research advances, this synergy may unlock new horizons in AI capabilities. Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more Tectonic’s Successful Salesforce Track Record Salesforce Technology Services Integrator – Tectonic has successfully delivered Salesforce in a variety of industries including Public Sector, Hospitality, Manufacturing, Read more

Read More
AI Agents and Consumer Trust

AI Agents Next AI Evolution

AI agents are being hailed as the next big leap in artificial intelligence, but there’s no universally accepted definition of what they are—or what they should do. Even within the tech community, there’s debate about what constitutes an AI agent. At its core, an AI agent can be described as software powered by artificial intelligence that performs tasks once handled by human roles, such as customer service agents, HR representatives, or IT help desk staff. However, their potential spans much further. These agents don’t just answer questions—they take action, often working across multiple systems. For example, Perplexity recently launched an AI agent to assist with holiday shopping, while Google introduced Project Mariner, an agent that helps users book flights, find recipes, and shop for household items. While the idea seems straightforward, it’s muddied by inconsistent definitions. For Google, AI agents are task-based assistants tailored to specific roles, like coding help for developers or troubleshooting issues for IT professionals. In contrast, Asana views agents as digital co-workers that take on assigned tasks, and Sierra—a startup led by former Salesforce co-CEO Bret Taylor—envisions agents as sophisticated customer experience tools that surpass traditional chatbots by tackling complex problems. This lack of consensus adds to the uncertainty around what AI agents can truly achieve. Rudina Seseri, founder and managing partner at Glasswing Ventures, explains this ambiguity stems from the technology’s infancy. She describes AI agents as intelligent systems capable of perceiving their environment, reasoning, making decisions, and taking actions to achieve specific goals autonomously. These agents rely on a mix of AI technologies, including natural language processing, machine learning, and computer vision, to operate in dynamic environments. Optimists, like Box CEO Aaron Levie, believe AI agents will improve rapidly as advancements in GPU performance, model efficiency, and AI frameworks create a self-reinforcing cycle of innovation. However, skeptics like MIT robotics pioneer Rodney Brooks caution against overestimating progress, noting that solving real-world problems—especially those involving legacy systems with limited API access—can be far more challenging than anticipated. David Cushman of HFS Research likens current AI agents to assistants rather than fully autonomous entities, with their capabilities limited to helping users complete specific tasks within pre-defined boundaries. True autonomy, where AI agents handle contingencies and perform at scale without human oversight, remains a distant goal. Jon Turow, a partner at Madrona Ventures, emphasizes the need for dedicated infrastructure to support the development of AI agents. He envisions a tech stack that allows developers to focus on product differentiation while leaving scalability and reliability to the platform. This infrastructure would likely involve multiple specialized models working together under a routing layer, rather than relying on a single large language model (LLM). Fred Havemeyer of Macquarie US Equity Research agrees, noting that the most effective AI agents will combine various models to handle complex tasks. He imagines a future where agents act like autonomous supervisors, delegating tasks and reasoning through multi-step processes to achieve abstract goals. While this vision is compelling, the current state of AI agents suggests we’re still in a transitional phase. The progress so far is promising, but several breakthroughs are needed before agents can operate as envisioned—truly autonomous, multi-functional, and capable of seamless collaboration across diverse systems. This story, originally published on July 13, 2024, has been updated to reflect new developments from Perplexity and Google. Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Marketing Cloud Transactional Emails Salesforce Marketing Cloud Transactional Emails are immediate, automated, non-promotional messages crucial to business operations and customer satisfaction, such as order Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more

Read More
Where LLMs Fall Short

LLM Economies

Throughout history, disruptive technologies have been the catalyst for major social and economic revolutions. The invention of the plow and irrigation systems 12,000 years ago sparked the Agricultural Revolution, while Johannes Gutenberg’s 15th-century printing press fueled the Protestant Reformation and helped propel Europe out of the Middle Ages into the Renaissance. In the 18th century, James Watt’s steam engine ushered in the Industrial Revolution. More recently, the internet has revolutionized communication, commerce, and information access, shrinking the world into a global village. Similarly, smartphones have transformed how people interact with their surroundings. Now, we stand at the dawn of the AI revolution. Large Language Models (LLMs) represent a monumental leap forward, with significant economic implications at both macro and micro levels. These models are reshaping global markets, driving new forms of currency, and creating a novel economic landscape. The reason LLMs are transforming industries and redefining economies is simple: they automate both routine and complex tasks that traditionally require human intelligence. They enhance decision-making processes, boost productivity, and facilitate cost reductions across various sectors. This enables organizations to allocate human resources toward more creative and strategic endeavors, resulting in the development of new products and services. From healthcare to finance to customer service, LLMs are creating new markets and driving AI-driven services like content generation and conversational assistants into the mainstream. To truly grasp the engine driving this new global economy, it’s essential to understand the inner workings of this disruptive technology. These posts will provide both a macro-level overview of the economic forces at play and a deep dive into the technical mechanics of LLMs, equipping you with a comprehensive understanding of the revolution happening now. Why Now? The Connection Between Language and Human Intelligence AI did not begin with ChatGPT’s arrival in November 2022. Many people were developing machine learning classification models in 1999, and the roots of AI go back even further. Artificial Intelligence was formally born in 1950, when Alan Turing—considered the father of theoretical computer science and famed for cracking the Nazi Enigma code during World War II—created the first formal definition of intelligence. This definition, known as the Turing Test, demonstrated the potential for machines to exhibit human-like intelligence through natural language conversations. The test involves a human evaluator who engages in conversations with both a human and a machine. If the evaluator cannot reliably distinguish between the two, the machine is considered to have passed the test. Remarkably, after 72 years of gradual AI development, ChatGPT simulated this very interaction, passing the Turing Test and igniting the current AI explosion. But why is language so closely tied to human intelligence, rather than, for example, vision? While 70% of our brain’s neurons are devoted to vision, OpenAI’s pioneering image generation model, DALL-E, did not trigger the same level of excitement as ChatGPT. The answer lies in the profound role language has played in human evolution. The Evolution of Language The development of language was the turning point in humanity’s rise to dominance on Earth. As Yuval Noah Harari points out in his book Sapiens: A Brief History of Humankind, it was the ability to gossip and discuss abstract concepts that set humans apart from other species. Complex communication, such as gossip, requires a shared, sophisticated language. Human language evolved from primitive cave signs to structured alphabets, which, along with grammar rules, created languages capable of expressing thousands of words. In today’s digital age, language has further evolved with the inclusion of emojis, and now with the advent of GenAI, tokens have become the latest cornerstone in this progression. These shifts highlight the extraordinary journey of human language, from simple symbols to intricate digital representations. In the next post, we will explore the intricacies of LLMs, focusing specifically on tokens. But before that, let’s delve into the economic forces shaping the LLM-driven world. The Forces Shaping the LLM Economy AI Giants in Competition Karl Marx and Friedrich Engels argued that those who control the means of production hold power. The tech giants of today understand that AI is the future means of production, and the race to dominate the LLM market is well underway. This competition is fierce, with industry leaders like OpenAI, Google, Microsoft, and Facebook battling for supremacy. New challengers such as Mistral (France), AI21 (Israel), and Elon Musk’s xAI and Anthropic are also entering the fray. The LLM industry is expanding exponentially, with billions of dollars of investment pouring in. For example, Anthropic has raised $4.5 billion from 43 investors, including major players like Amazon, Google, and Microsoft. The Scarcity of GPUs Just as Bitcoin mining requires vast computational resources, training LLMs demands immense computing power, driving a search for new energy sources. Microsoft’s recent investment in nuclear energy underscores this urgency. At the heart of LLM technology are Graphics Processing Units (GPUs), essential for powering deep neural networks. These GPUs have become scarce and expensive, adding to the competitive tension. Tokens: The New Currency of the LLM Economy Tokens are the currency driving the emerging AI economy. Just as money facilitates transactions in traditional markets, tokens are the foundation of LLM economics. But what exactly are tokens? Tokens are the basic units of text that LLMs process. They can be single characters, parts of words, or entire words. For example, the word “Oscar” might be split into two tokens, “os” and “car.” The performance of LLMs—quality, speed, and cost—hinges on how efficiently they generate these tokens. LLM providers price their services based on token usage, with different rates for input (prompt) and output (completion) tokens. As companies rely more on LLMs, especially for complex tasks like agentic applications, token usage will significantly impact operational costs. With fierce competition and the rise of open-source models like Llama-3.1, the cost of tokens is rapidly decreasing. For instance, OpenAI reduced its GPT-4 pricing by about 80% over the past year and a half. This trend enables companies to expand their portfolio of AI-powered products, further fueling the LLM economy. Context Windows: Expanding Capabilities

Read More
Snowflake Security and Development

Snowflake Security and Development

Snowflake Unveils AI Development and Enhanced Security Features At its annual Build virtual developer conference, Snowflake introduced a suite of new capabilities focused on AI development and strengthened security measures. These enhancements aim to simplify the creation of conversational AI tools, improve collaboration, and address data security challenges following a significant breach earlier this year. AI Development Updates Snowflake announced updates to its Cortex AI suite to streamline the development of conversational AI applications. These new tools focus on enabling faster, more efficient development while ensuring data integrity and trust. Highlights include: These features address enterprise demands for generative AI tools that boost productivity while maintaining governance over proprietary data. Snowflake aims to eliminate barriers to data-driven decision-making by enabling natural language queries and easy integration of structured and unstructured data into AI models. According to Christian Kleinerman, Snowflake’s EVP of Product, the goal is to reduce the time it takes for developers to build reliable, cost-effective AI applications: “We want to help customers build conversational applications for structured and unstructured data faster and more efficiently.” Security Enhancements Following a breach last May, where hackers accessed customer data via stolen login credentials, Snowflake has implemented new security features: These additions come alongside existing tools like the Horizon Catalog for data governance. Kleinerman noted that while Snowflake’s previous security measures were effective at preventing unauthorized access, the company recognizes the need to improve user adoption of these tools: “It’s on us to ensure our customers can fully leverage the security capabilities we offer. That’s why we’re adding more monitoring, insights, and recommendations.” Collaboration Features Snowflake is also enhancing collaboration through its new Internal Marketplace, which enables organizations to share data, AI tools, and applications across business units. The Native App Framework now integrates with Snowpark Container Services to simplify the distribution and monetization of analytics and AI products. AI Governance and Competitive Position Industry analysts highlight the growing importance of AI governance as enterprises increasingly adopt generative AI tools. David Menninger of ISG’s Ventana Research emphasized that Snowflake’s governance-focused features, such as LLM observability, fill a critical gap in AI tooling: “Trustworthy AI enhancements like model explainability and observability are vital as enterprises scale their use of AI.” With these updates, Snowflake continues to compete with Databricks and other vendors. Its strategy focuses on offering both API-based flexibility for developers and built-in tools for users seeking simpler solutions. By combining innovative AI development tools with robust security and collaboration features, Snowflake aims to meet the evolving needs of enterprises while positioning itself as a leader in the data platform and AI space. Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more Tectonic’s Successful Salesforce Track Record Salesforce Technology Services Integrator – Tectonic has successfully delivered Salesforce in a variety of industries including Public Sector, Hospitality, Manufacturing, Read more

Read More
Scaling Generative AI

Scaling Generative AI

Many organizations follow a hybrid approach to AI infrastructure, combining public clouds, colocation facilities, and on-prem solutions. Specialized GPU-as-a-service vendors, for instance, are becoming popular for handling high-demand AI computations, helping businesses manage costs without compromising performance. Business process outsourcing company TaskUs, for example, focuses on optimizing compute and data flows as it scales its gen AI deployments, while Cognizant advises that companies distinguish between training and inference needs, each with different latency requirements.

Read More
LLMs and AI

LLMs and AI

Large Language Models (LLMs): Revolutionizing AI and Custom Solutions Large Language Models (LLMs) are transforming artificial intelligence by enabling machines to generate and comprehend human-like text, making them indispensable across numerous industries. The global LLM market is experiencing explosive growth, projected to rise from $1.59 billion in 2023 to $259.8 billion by 2030. This surge is driven by the increasing demand for automated content creation, advances in AI technology, and the need for improved human-machine communication. Several factors are propelling this growth, including advancements in AI and Natural Language Processing (NLP), large datasets, and the rising importance of seamless human-machine interaction. Additionally, private LLMs are gaining traction as businesses seek more control over their data and customization. These private models provide tailored solutions, reduce dependency on third-party providers, and enhance data privacy. This guide will walk you through building your own private LLM, offering valuable insights for both newcomers and seasoned professionals. What are Large Language Models? Large Language Models (LLMs) are advanced AI systems that generate human-like text by processing vast amounts of data using sophisticated neural networks, such as transformers. These models excel in tasks such as content creation, language translation, question answering, and conversation, making them valuable across industries, from customer service to data analysis. LLMs are generally classified into three types: LLMs learn language rules by analyzing vast text datasets, similar to how reading numerous books helps someone understand a language. Once trained, these models can generate content, answer questions, and engage in meaningful conversations. For example, an LLM can write a story about a space mission based on knowledge gained from reading space adventure stories, or it can explain photosynthesis using information drawn from biology texts. Building a Private LLM Data Curation for LLMs Recent LLMs, such as Llama 3 and GPT-4, are trained on massive datasets—Llama 3 on 15 trillion tokens and GPT-4 on 6.5 trillion tokens. These datasets are drawn from diverse sources, including social media (140 trillion tokens), academic texts, and private data, with sizes ranging from hundreds of terabytes to multiple petabytes. This breadth of training enables LLMs to develop a deep understanding of language, covering diverse patterns, vocabularies, and contexts. Common data sources for LLMs include: Data Preprocessing After data collection, the data must be cleaned and structured. Key steps include: LLM Training Loop Key training stages include: Evaluating Your LLM After training, it is crucial to assess the LLM’s performance using industry-standard benchmarks: When fine-tuning LLMs for specific applications, tailor your evaluation metrics to the task. For instance, in healthcare, matching disease descriptions with appropriate codes may be a top priority. Conclusion Building a private LLM provides unmatched customization, enhanced data privacy, and optimized performance. From data curation to model evaluation, this guide has outlined the essential steps to create an LLM tailored to your specific needs. Whether you’re just starting or seeking to refine your skills, building a private LLM can empower your organization with state-of-the-art AI capabilities. For expert guidance or to kickstart your LLM journey, feel free to contact us for a free consultation. Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Marketing Cloud Transactional Emails Salesforce Marketing Cloud Transactional Emails are immediate, automated, non-promotional messages crucial to business operations and customer satisfaction, such as order Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more

Read More
GPUs and AI Development

GPUs and AI Development

Graphics processing units (GPUs) have become widely recognized due to their growing role in AI development. However, a lesser-known but critical technology is also gaining attention: high-bandwidth memory (HBM). HBM is a high-density memory designed to overcome bottlenecks and maximize data transfer speeds between storage and processors. AI chipmakers like Nvidia rely on HBM for its superior bandwidth and energy efficiency. Its placement next to the GPU’s processor chip gives it a performance edge over traditional server RAM, which resides between storage and the processing unit. HBM’s ability to consume less power makes it ideal for AI model training, which demands significant energy resources. However, as the AI landscape transitions from model training to AI inferencing, HBM’s widespread adoption may slow. According to Gartner’s 2023 forecast, the use of accelerator chips incorporating HBM for AI model training is expected to decline from 65% in 2022 to 30% by 2027, as inferencing becomes more cost-effective with traditional technologies. How HBM Differs from Other Memory HBM shares similarities with other memory technologies, such as graphics double data rate (GDDR), in delivering high bandwidth for graphics-intensive tasks. But HBM stands out due to its unique positioning. Unlike GDDR, which sits on the printed circuit board of the GPU, HBM is placed directly beside the processor, enhancing speed by reducing signal delays caused by longer interconnections. This proximity, combined with its stacked DRAM architecture, boosts performance compared to GDDR’s side-by-side chip design. However, this stacked approach adds complexity. HBM relies on through-silicon via (TSV), a process that connects DRAM chips using electrical wires drilled through them, requiring larger die sizes and increasing production costs. According to analysts, this makes HBM more expensive and less efficient to manufacture than server DRAM, leading to higher yield losses during production. AI’s Demand for HBM Despite its manufacturing challenges, demand for HBM is surging due to its importance in AI model training. Major suppliers like SK Hynix, Samsung, and Micron have expanded production to meet this demand, with Micron reporting that its HBM is sold out through 2025. In fact, TrendForce predicts that HBM will contribute to record revenues for the memory industry in 2025. The high demand for GPUs, especially from Nvidia, drives the need for HBM as AI companies focus on accelerating model training. Hyperscalers, looking to monetize AI, are investing heavily in HBM to speed up the process. HBM’s Future in AI While HBM has proven essential for AI training, its future may be uncertain as the focus shifts to AI inferencing, which requires less intensive memory resources. As inferencing becomes more prevalent, companies may opt for more affordable and widely available memory solutions. Experts also see HBM following the same trajectory as other memory technologies, with continuous efforts to increase bandwidth and density. The next generation, HBM3E, is already in production, with HBM4 planned for release in 2026, promising even higher speeds. Ultimately, the adoption of HBM will depend on market demand, especially from hyperscalers. If AI continues to push the limits of GPU performance, HBM could remain a critical component. However, if businesses prioritize cost efficiency over peak performance, HBM’s growth may level off. Like Related Posts Who is Salesforce? Who is Salesforce? Here is their story in their own words. From our inception, we’ve proudly embraced the identity of Read more Salesforce Unites Einstein Analytics with Financial CRM Salesforce has unveiled a comprehensive analytics solution tailored for wealth managers, home office professionals, and retail bankers, merging its Financial Read more AI-Driven Propensity Scores AI plays a crucial role in propensity score estimation as it can discern underlying patterns between treatments and confounding variables Read more Tectonic’s Successful Salesforce Track Record Salesforce Technology Services Integrator – Tectonic has successfully delivered Salesforce in a variety of industries including Public Sector, Hospitality, Manufacturing, Read more

Read More
gettectonic.com