Category: Platforms

  • A Report Revealed the Winners and Losers in the New AI Landscape

    A Report Revealed the Winners and Losers in the New AI Landscape

    IBL News | New York

    AI spending surged to $13.8 billion in 2024 from $2.3 billion in 2023 as enterprises embed AI at the core of their business strategies and daily work, according to a study conducted by Menlo Ventures.

    This research, titled “2024 State of Generative AI in the Enterprise Report,” done after surveying 600 U.S. enterprise IT decision-makers, points out that we are still in the early stages of a large-scale transformation.

    This spending will continue: 72% of decision-makers anticipate broader adoption of generative AI tools soon.

    Investments in the LLM foundation model still dominate spending, but the application layer segment to optimize workflows is now growing faster.

    These app layer companies—mostly in highly verticalized sectors—leverage LLM’s capabilities across domains to unlock new efficiencies. Enterprise buyers will invest $4.6 billion in generative AI applications in 2024, an 8x increase from the $600 million invested in 2023.

    The use cases that deliver the most ROI through enhanced productivity or operational efficiency are:

    • Code copilots, such as GitHub Copilot, Cursor, Codeium, Harness, and All Hands.

    • Support knowledge-based chatbots for employees, customers, and contact centers. Aisera, Decagon, Sierra, and Observe AI are some of the examples.

    • Enterprise search, retrieval, data extraction, and transformation to unlock the knowledge hidden within data silos. Solutions like Glean and Sana connect to emails, messengers, and document stores, enabling unified semantic search across systems.

    • Meeting summarization to automate note-taking and takeaways. Examples are Fireflies.ai, Otter.ai, Fathom, and Eleos Health.

    AI-powered autonomous agents capable of managing complex, end-to-end workflow processes are emerging and can transform human-led industries. Forge, Sema4, and Clay are some tools.

    When deciding to build or buy, 47% of solutions are developed in-house, while 53% are sourced from vendors. Often, organizations discover too late that they have underestimated the difficulty of technical integration, scalability, and ongoing support.

    Most customers (64%) prefer buying from established vendors, citing trust.

    The leading vertical AI applications are:

    Healthcare, with examples like AbridgeAmbienceHeidi, Eleos Health, Notable, SmarterDxCodametrix, Adonis, and Rivet.

    Legal, with examples like Everlaw, Harvey, Spellbook, EvenUp, Garden, Manifest, and Eve.

    Financial Services, with examples like Numeric, Klarity, Arkifi, Rogo, Arch, Orby, Sema4, Greenlite, and Norm AI.

    Media and entertainment, with examples like Runway, CaptionsDescript, Black Forest LabsHiggsfield, IdeogramMidjourney, and Pika.

    Rather than relying on a single provider, enterprises have adopted a multi-model approach, typically deploying three or more LLM in their AI stacks, routing to different models depending on the use case or results.

    To date, close-source solutions underpin the vast majority of usage, with Meta’s Llama 3 holding at 19%, according to the Menlo Ventures research.

    Regarding architectures for building efficient and scalable AI systems, RAG (retrieval-augmented generation) dominates with 51% adoption, while fine-tuning of production molded is only 9%. Agentic architectures, which debuted this year, power 12% of implementations.

    Databases and data pipelines are needed to power RAG. Traditional databases like Postgres and MongoDB remain common, while AI-native vector databases like Pinecone gain ground.

    Menlo Ventures made three predictions for what lies ahead:

    1. Agentic automation will drive the next wave of transformation, tackling complex, multi-step tasks beyond the current systems of content generation and knowledge retrieval. Examples are platforms like Clay and Forge

    2. More incumbents will fall. Chegg saw 85% of its market cap vanish, while Stack Overflow’s web traffic halved. IT outsourcing firms like Cognizant, legacy automation players like UiPath, and even software giants like Salesforce and Autodesk will face AI-native challengers.

    3. The AI talent drought will intensify. AI-skilled enterprise architects will notably increase their salaries. 

    Squint, Typeface

  • New Research Suggest How AI Should Be Integrated on Learning Environments, Research, Administrative, and Campus Operations

    New Research Suggest How AI Should Be Integrated on Learning Environments, Research, Administrative, and Campus Operations

    IBL News | New York

    AI’s integration into learning environments, research, administrative functions, and campus operations reshapes how institutions operate, faculty teach, students learn, and staff perform their roles.

    It’s not about blindly accepting AI in higher education or banning its use.

    It is crucial to thoughtfully examine AI’s impact on higher education, specifically on student success, financial sustainability, accountability, and equity.

    This is the main conclusion of researcher Joe Sabado, who shared research titled “AI in Higher Education—Frameworks for Critical Inquiry and Innovation.”

    This document, created using AI, guides institutions through AI’s transformative process, helping them leverage this technology. It provides ten frameworks, offering valuable insights for all stakeholders: educators, administrators, policymakers, students, staff, and journalists.

    AI in Higher Education – Frameworks for Inquiry and Innovation (PDF)

  • Open Courses on AI

    Open Courses on AI

    IBL News | New York

    • Microsoft Generative AI for Beginners

    00
    Course Setup

    01
    Introduction to Generative AI and LLMs

    02
    Exploring and comparing different LLMs

    03
    Using Generative AI Responsibly

    04
    Understanding Prompt Engineering Fundamentals

    05
    Creating Advanced Prompts

    06
    Building Text Generation Applications

    07
    Building Chat Applications

    08
    Building Search Apps Vector Databases

    09
    Building Image Generation Applications

    10
    Building Low Code AI Applications

    11
    Integrating External Applications with Function Calling

    12
    Designing UX for AI Applications

    13
    Securing Your Generative AI Applications

    14
    The Generative AI Application Lifecycle

    15
    Retrieval Augmented Generation (RAG) and Vector Databases

    16
    Open Source Models and Hugging Face

    17
    AI Agents

    18
    Fine-Tuning LLMs

    • Anthropic Courses: API fundamentals and Prompt engineering

    • J.P. Morgan Chase: Training on Python

  • Top Open Source Apps and Tools for AI

    Top Open Source Apps and Tools for AI

    July 2024

    Groqbook: Generate entire books in seconds using Groq and Llama3 | Story at IBL News

    • Vanderbilt University’s Amplify GenAI platform | Story at IBL News

    LLM models | Story at IBL News | Five Top Open Source LLMs

    Abacus AI Open Source

  • NVIDIA Released Eight Free Courses on Generative AI

    NVIDIA Released Eight Free Courses on Generative AI

    IBL News | New York

    NVIDIA released eight free AI courses this month. Five are hosted at NVIDIA’s Deep Learning Institute (DLI) platform, two on Coursera, and one on YouTube.

    1. Generative AI Explained
    2. Building A Brain in 10 Minutes
    3. Augment your LLM with Retrieval Augmented Generation
    4. AI in the Data Center
    5. Accelerate Data Science Workflows with Zero Code Changes
    6. Mastering Recommender Systems
    7. Networking Introduction
    8. Building RAG Agents with LLMs

    [Disclosure: IBL works for NVIDIA by powering its learning platform]

  • AI Agents, the Second Phase of Generative AI

    AI Agents, the Second Phase of Generative AI

    IBL News | New York

    After the release of the bot ChatGPT a year ago, the second phase of personalized, autonomous AI agents is emerging.

    These agents can perform complex tasks, such as sending emails, scheduling meetings, booking flights or tables in a restaurant, or even complex tasks like buying presents for family members or negotiating a raise.

    Personalized chatbots, programmed for specific tasks, that GPT creators will be able to release through the upcoming OpenAI’s GPT Store, are a prelude.

    For now, these custom GPTs are easy to build without knowing how to code.

    Users just answer a few simple questions about their bot — its name, its purpose, the tone used to respond — and the bot builds itself in just a few seconds. Users can upload PDF documents they want to use as reference material or easily look up Q&A. They can also connect the bot to other apps or edit its instructions.

    Although these custom chatbots are far from working perfectly, they can be useful tools for answering repetitive questions in customer service departments.

    Some AI safety researchers fear that giving bots more autonomy could lead to disaster, The New York Times reported. The Center for AI Safety, a nonprofit research organization, listed autonomous agents as one of its “catastrophic AI risks” this year, saying that “malicious actors could intentionally create rogue AI with dangerous goals.”

    For now, these agents look harmless and limited in their scope.

    Its development seems to be dependent on gradual iterative deployment, that is, small improvements at a fast pace rather than a big leap.

    In the last OpenAI developer conference, Sam Atman built on stage a “start-up mentor” chatbot to give advice to aspiring founders, based on an uploaded file of a speech he had given years earlier.

    The San Francisco-based research lab envisions a world where AI agents will be extensions of us, gathering information and taking action on our behalf.
    .

    For now, OpenAI’s bots are limited to simple, well-defined tasks, and can’t handle complex planning or long sequences of actions.
    After a day care’s handbook was uploaded to OpenAI’s GPT creator tool, a chatbot could easily look up answers to questions about it.
    A screenshot of a GPT “Day Care Helper” conversation between the author and the chatbot about circle time.

     

  • Legal and Compliance Risks that ChatGPT Presents to Organizations, According to Gartner

    Legal and Compliance Risks that ChatGPT Presents to Organizations, According to Gartner

    IBL News | New York

    The output generated by ChatGPT and other LLMs presents legal and compliance risks that every organization has to face or face dire consequences, according to the consultancy firm Gartner, Inc, which has identified six areas.

    “Failure to do so could expose enterprises to legal, reputational, and financial consequences,” said Ron Friedmann, Senior Director Analyst at Gartner Legal & Compliance Practice.

    • Risk 1: Fabricated and Inaccurate Answers

    ChatGPT is also prone to ‘hallucinations,’ including fabricated answers that are wrong, and nonexistent legal or scientific citations,” said Friedmann.

    Only accurate training of the robot with limited sources will mitigate this tendency to provide incorrect information.

    •  Risk 2. Data Privacy and Confidentiality

    Sensitive, proprietary, or confidential information used in prompts may become a part of its training dataset and incorporated into responses for users outside the enterprise if chat history is not disabled,

    “Legal and compliance need to establish a compliance framework and clearly prohibit entering sensitive organizational or personal data into public LLM tools,” said Friedmann.

    • Risk 3. Model and Output Bias

    “Complete elimination of bias is likely impossible, but legal and compliance need to stay on top of laws governing AI bias and make sure their guidance is compliant,” said Friedmann.

    “This may involve working with subject matter experts to ensure output is reliable and with audit and technology functions to set data quality controls,” he added.

    • Risk 4.  Intellectual Property (IP) and Copyright risks

    As ChatGPT is trained on a large amount of internet data that likely includes copyrighted material, its outputs – which do not offer source references – have the potential to violate copyright or IP protection.

    “Legal and compliance leaders should keep a keen eye on any changes to copyright law that apply to ChatGPT output and require users to scrutinize any output they generate to ensure it doesn’t infringe on copyright or IP rights.”

    • Risk 5. Cyber Fraud Risks

    Bad actors are already using ChatGPT to generate false information at scale, like fake reviews, for instance.

    Moreover, applications that use LLM models, including ChatGPT, are also susceptible to prompt injection, a hacking technique in which

    A hacking technique known as “prompt injection” brings criminals to write malware codes or develop phishing sites that resemble well-known sites.

    “Legal and compliance leaders should coordinate with owners of cyber risks to explore whether or when to issue memos to company cybersecurity personnel on this issue,” said Friedmann.

    • Risk 6. Consumer Protection Risks

    Businesses that fail to disclose that they are using ChatGPT as a customer support chatbot run the risk of being charged with unfair practices under various laws and face the risk of losing their customers’ trust.

    For instance, the California chatbot law mandates that in certain consumer interactions, organizations must disclose that a consumer is communicating with a bot.

    Legal and compliance leaders need to ensure their organization’s use complies with regulations and laws.
    .

  • Critical Factors When Orchestrating an Optimized Large Language Model (LLM)

    Critical Factors When Orchestrating an Optimized Large Language Model (LLM)

    IBL News | New York

    When choosing and orchestrating an LLM, there are many critical technical factors, such as training data, dataset filtering, fine-tuning process, capabilities, latency, technical requirements, and price.

    Experts state that implementing an LLM API, like GPT-4 or others, is not the only option.

    As a paradigm-shifting technology and with the pace of innovation moving really fast, the LLMs and Natural Language Processing market is projected to reach $91 billion by 2030 growing at a CAGR of 27%.

    Beyond the parameter count, recent findings showed that smaller models trained on more data are just as effective, and can even lead to big gains in latency and a significant reduction in hardware requirements. In other words, the largest parameter count is not what matters.

    Training data should include conversations, games, and immersive experiences related to the subject rather than creating general-purpose models that knew a little about everything. For example, a model whose training data is 90% medical papers performs better on medical tasks than a much larger model where medical papers only make up 10% of its dataset.

    In terms of dataset filtering, certain kinds of content have to be removed to reduce toxicity and bias. OpenAI recently confirmed that for example erotic content has been filtered.

    It’s also important to create vocabularies based on how commonly words appear, removing colloquial conversation and common slang datasets.

    Models have to be fine-tuned intend to ensure the accuracy of the information and avoid false information in the dataset.

    LLMs are not commoditized, and some models have unique capabilities. GPT-4 accepts multimodal inputs like video and photos and writes up 25,000 words at a time while maintaining context. Google’s PaLM can generate text, images, code, videos, audio, etc.

    Other models can provide facial expressions and voice.

    Inference latency is higher in models with more parameters, adding extra milliseconds between query and response, which significantly impacts real-time applications.

    Google’s research found that just half a second of added latency cause traffic to drop by 20%.

    For low or real-time latency, many use cases, such as financial forecasting or video games, can’t be fulfilled by a standalone LLM. It’s required the orchestration of multiple models, specialized features, or additional automation, for text-to-speech, automatic speech recognition (ASR), machine vision, memory, etc.