Author: IBL News

  • Hugging Face, ServiceNow, and Nvidia Released ‘StarCoder2’, a Free Code-Generating Model

    Hugging Face, ServiceNow, and Nvidia Released ‘StarCoder2’, a Free Code-Generating Model

    IBL News | New York

    The BigCode project, an open-scientific collaboration focused on the development of LLMs for Code (Code LLMs), released this week StarCoder2, an AI-powered open-source code generator with a less restrictive license than GitHub Copilot, Amazon CodeWhisperer, and Meta’s Code Llama.

    Like most other code generators, StarCoder2 can suggest ways to complete unfinished lines of code as well as summarize and retrieve snippets of code when asked in natural language.

    All StarCoder2 variants are trained on The Stack v2, a new large and high-quality code dataset. StarCoder 2 is a family of open LLMs for code and comes in three different sizes with 3B trained by ServiceNow, 7B trained by Hugging Face, and 15B parameter trained by NVIDIA with NVIDIA NeMo and trained on NVIDIA accelerated infrastructure.

    The last one, StarCoder2-15B, was trained on over 4+ trillion tokens and 600+ programming languages from The Stack v2.

    BigCode released all models, datasets, and the processing, as well as the training code, as explained in a paper.

    In the project, at least these U.S. universities participated: Northeastern University, University of Illinois Urbana-Champaign, Johns Hopkins University, Leipzig University, Monash University, University of British Columbia, MIT, Technical University of Munich, Technion – Israel Institute of Technology, University of Notre Dame, Princeton University, Wellesley College, University College London, UC San Diego, Cornell University, and UC Berkeley.

    Beyond Academia, the project gathered Kaggle, Roblox 12Sea AI Lab 13, CSIRO’s Data61, Mazzuma, Contextual AI, Cohere, and Salesforce.

    StarCoder 2 can be fine-tuned in a few hours using a GPU like the Nvidia A100 on first- or third-party data to create apps such as chatbots and personal coding assistants. And, because it was trained on a larger and more diverse data set than the original StarCoder (~619 programming languages), StarCoder 2 can make more accurate, context-aware predictions — at least hypothetically.

    Harm de Vries, head of ServiceNow’s StarCoder 2 development team, told TechCrunch in an interview, that “with StarCoder2, developers can use its capabilities to make coding more efficient without sacrificing speed or quality.”

    A recent Stanford study found that engineers who use code-generating systems are more likely to introduce security vulnerabilities in the apps they develop. Moreover, a poll from Sonatype, the cybersecurity firm, shows that the majority of developers are concerned about the lack of insight into how code from code generators is produced and “code sprawl” from generators producing too much code to manage.

    StarCoder 2’s license might also prove to be a roadblock for some, according to TechCrunch.

    “StarCoder 2 is licensed under the BigCode Open RAIL-M 1.0, which aims to promote responsible use by imposing “light touch” restrictions on both model licensees and downstream users. While less constraining than many other licenses, RAIL-M isn’t truly “open” in the sense that it doesn’t permit developers to use StarCoder 2 for every conceivable application (medical advice-giving apps are strictly off limits, for example). Some commentators say RAIL-M’s requirements may be too vague to comply with in any case — and that RAIL-M could conflict with AI-related regulations like the EU AI Act.”

    ServiceNow has already used StarCoder to create Now LLM, a product for code generation fine-tuned for ServiceNow workflow patterns, use cases, and processes. Hugging Face, which offers model implementation consulting plans, is providing hosted versions of the StarCoder 2 models on its platform. Nvidia, which is making StarCoder 2 available through an API and web front-end.
    .

  • The George Washington University Forms AI Advisory Councils From Each School

    The George Washington University Forms AI Advisory Councils From Each School

    IBL News | New York

    The George Washington University (GWU) formed this year an advisory council of faculty from each school to improve the use of AI tools in their teaching and student learning and provide best practices.

    Managed by the Instructional Core within the Libraries & Academic Innovation (LAI) office of GW, led by Geneva Henry., these advisory councils provide input on AI resources and training for professors aligning with the guidelines released by the Office of the Provost in April 2023.

    These guidelines state that using AI to study is permissible, but submitting AI-generated material for an assignment or using AI during an assessment is cheating.

    The decision of whether or not to allow AI in courses is individual to each professor.

    Douglas Crawford, a member of the council and an assistant professor of interior architecture, said to The GW Hatchet that the goal of the council is to act as a knowledge base for faculty looking to implement or restrict AI use in their courses.

    He said he encourages his students to use AI as a starting point for projects because it can provide more tailored inspiration than platforms like Google Images or Pinterest.

    John Helveston, a member of the council and an assistant professor of engineering management and systems engineering, said the introduction of AI into classrooms has pushed educators to rethink what they want to accomplish in classrooms and how they organize course content in order to make students think critically.

    Lorena Barba, a member of the council and a professor of mechanical and aerospace engineering, said the council had a “grassroots” origin that bodes well for faculty’s willingness to collaborate in discussions surrounding AI in a way that hasn’t been commonly seen in universities.

    “GW has a unique opportunity to be at the forefront, and many members of the AI Advisory Council are courageously embracing the challenge,” Lorena Barba said.
    .

  • McKinsey: “Gen AI Will Unleash the Next Wave of Productivity”

    McKinsey: “Gen AI Will Unleash the Next Wave of Productivity”

    IBL News | New York

    Generative AI is poised to unleash the next wave of productivity, stated McKinsey in a research titled “The economic potential of generative AI: The next productivity frontier,” released last month.

    The ability of generative AI applications, built using foundation models, to write text, compose music, and create digital art has persuaded consumers and households to experiment on their own.

    AI trained on these models can perform several functions; it can classify, edit, summarize, answer questions, and draft new content, among other tasks.

    McKinsey’s research considers that we are at the beginning of a journey to understand generative AI’s power, reach, and capabilities.

    “Deep learning has powered many of the recent advances in AI, but the foundation models powering generative AI applications are a step-change evolution within deep learning. Unlike previous deep learning models, they can process extremely large and varied sets of unstructured data and perform more than one task.”

    Its research suggests that generative AI is poised to transform roles and boost performance across functions such as sales and marketing, customer operations, and software development.

    In the process, it could unlock trillions of dollars in value across sectors from banking to life sciences.

    Foundation models have enabled new capabilities and vastly improved existing ones across a broad range of modalities, including images, video, audio, and computer code. AI trained on these models can perform several functions; it can classify, edit, summarize, answer questions, and draft new content, among other tasks.
    .

     

  • Mistral Launches ‘Mistral Large’, a New, Non-Open-Source LLM

    Mistral Launches ‘Mistral Large’, a New, Non-Open-Source LLM

    IBL News | New York

    Paris-based Mistral AI — which is trying to build an alternative to OpenAI’s GPT-4 and Anthropic’s Claude 2 — released yesterday a new LLM named Mistral Large. The model was not released under an open-source license.

    In addition, the French start-up is launching its alternative to ChatGPT with a new service called Le Chat, now available in beta. Le Chat can’t access the web.

    Founded by Mensch, Timothée Lacroix, and Guillaume Lample, a trio of former Meta and Google researchers, has now a valuation of €2 billion. [People working at Mistral in the picture above.]

    It supports context windows of 32k tokens (generally more than 20,000 words in English) in English, French, Spanish, German, and Italian.

    As a comparison, GPT-4 Turbo has a 128k-token context window.

    Mistral AI made Mistral Large available through Microsoft’s Azure, its first distribution partner, after signing a “multi-year partnership.”

    As part of the deal, Microsoft said it would invest in Mistral, although the financial details were not disclosed. The partnership will include a research and development collaboration to build applications for governments across Europe.

    Microsoft has already invested about $13 billion in San Francisco-based OpenAI, which is estimated to be worth $86 billion.
    .
  • Huge Decrease of Jobs in Writing, Customer Service and Translation

    Huge Decrease of Jobs in Writing, Customer Service and Translation

    IBL News | San Diego

    Since the Release of ChatGPT Upwork freelance website’s postings data — publicly available in the form of an RSS feed — show an increase in the number of jobs since ChatGPT was released in November 2022, according to the analysis of an expert posted at Bloomberry.com.

    However, three categories showed a large decline in jobs: writing, translation, and customer service jobs. The number of writing jobs declined by 33%, translation jobs declined by 19%, and customer service jobs declined by 16%.

    Since November 2022, video editing/production jobs were up 39%, graphic design jobs 8%, and web design jobs 10%. Software development jobs were also up, with backend development jobs at 6% and frontend/web development jobs at 4%.

    “Generative AI tools are already good enough to replace many writing tasks, whether it’s writing an article or a social media post. But they’re not polished enough for other jobs like video and image generation,” said Henley Wing, the author of the analysis.

    Jobs like generating AI content, developing AI agents, integrating OpenAI/ChatGPT APIs, and developing chatbots and AI apps are becoming the norm.

    However, the vast majority of companies are not yet developing their own LLM models or tuning them with training data. They seem to be integrating OpenAI’s API into their existing products and developing chatbots to replace their customer service agents.

  • Google Meet Detects When the User Raises His Hand

    Google Meet Detects When the User Raises His Hand

    IBL News | New York

    Google Meet announced last month an upcoming gesture detection feature called “Raise hand” for Q&A sessions and other moderation.

    This is a button in the toolbar available for enterprise accounts on Google Workplace, which lets people know when you have something to say, as the webcam recognizes hands up.

    The feature turns off gesture detection when someone is an active speaker.

    Until now, raising your hand to ask a question in Google Meet was done by clicking the hand-raise icon.

    This feature is toggled off by default and can be enabled from More Options > Reactions > Hand Raise Gesture.

    It’s similar to how Google Camera in Pixel phones can start a selfie timer when you raise your palm.
    .

  • Adobe Launches an AI Assistant for PDF Documents

    Adobe Launches an AI Assistant for PDF Documents

    IBL News | San Diego

    Adobe introduced this week an AI assistant in beta in Reader and Acrobat, which instantly generates summaries and insights from long PDF documents. It also recommends and answers questions based on a PDF’s content through an intuitive conversational interface.

    The AI assistant generates citations with the source, and as an output, it formats the information for sharing in emails, reports, and presentations. Clickable links help quickly find information in long documents.

    This feature will be sold through a new add-on subscription plan when AI Assistant is out of beta.

    “Our AI Assistant is bringing generative AI to the masses, unlocking new value from the information inside the approximately 3 trillion PDFs in the world,” stated Adobe.

    This assistant leverages the same AI and machine learning models behind Acrobat Liquid Mode, the technology that supports responsive reading experiences for PDFs on mobile.

    PDF was invented by Adobe thirty years ago Adobe, and today remains the standard for reading, editing, and transforming PDFs.

    Currently, the new AI Assistant features are available in beta for Acrobat Standard and Pro Individual and Teams subscription plans on desktop and web in English, with features coming to Reader desktop customers in English over the next few weeks – all at no additional cost. Other languages will follow. A private beta is available for enterprise customers.

    Adobe: How people are using AI Assistant (YouTube videos)

  • Google Open-Sources a Small Model of Gemini

    Google Open-Sources a Small Model of Gemini

    IBL News | San Diego

    Google released yesterday Gemma 2B and 7B, two lightweight, pre-trained open-source AI models, mostly suitable for small developments such as simple chatbots or summarizations.

    It also lets developers use the research and technology used to create the Gemini closed models.

    They are available via Kaggle, Hugging Face, Nvidia’s NeMo, and Google’s Vertex AI. It’s designed with Google’s AI Principles at the forefront.

    Gemma supports multi-framework Keras 3.0, native PyTorch, JAX, and Hugging Face Transformers.

    Developers and researchers can work with Gemma using free access in Kaggle, a free tier for Colab notebooks, and $300 in credits for first-time Google Cloud users. Researchers can also apply for Google Cloud credits of up to $500,000 to accelerate their projects.

    Each size of Gemma is available at ai.google.dev/gemma.

    Google is also providing toolchains for inference and supervised fine-tuning (SFT) across all major frameworks: JAX, PyTorch, and TensorFlow through native Keras 3.0.

    Google’s Gemini comes in several weights, including Gemini Nano, Gemini Pro, and Gemini Ultra.

    Last week, Google announced a faster Gemini 1.5 intended for business users and developers.

  • D2L Brightspace Pilots Generative AI Tools for Teachers

    D2L Brightspace Pilots Generative AI Tools for Teachers

    IBL News | New York

    D2L Brightspace presented to selected customers new generative AI tools to generate practice and quiz questions as part of its LMS platform.

    With this new program, which will be in testing mode through the summer of 2024, teachers are able to set quiz questions before making them available to students, giving them more safety and control oversight.

    The beta program is based on D2L’s “Responsible AI Principles” document.

    “This automated question generation capability can make it easier for instructors to assess learners in the moment. It is the initial step in expanding our product roadmap with cutting-edge generative AI to help change the way the world learns,” said Stephen Laster, D2L president.
    .

     

  • Khanmigo Struggles with Basic Math, Showed a Report

    Khanmigo Struggles with Basic Math, Showed a Report

    IBL News | New York

    Khanmigo, Khan Academy’s ChatGPT-powered tutoring bot, makes frequent calculation errors, The Wall Street Journal reported after testing it. “We tested an AI tutor for kids. It struggled with basic math,” wrote the paper.

    Last year, Educator Sal Khan promised to “give every student on the planet an artificially intelligent but amazing personal tutor.”

    “Asking ChatGPT to do math is sort of like asking a goldfish to ride a bicycle—it’s just not what ChatGPT is for,” said Tom McCoy, a professor at Yale University who studies AI.

    According to the paper, “Khanmigo frequently made basic arithmetic errors, miscalculating subtraction problems such as 343 minus 17. It also didn’t consistently know how to round answers or calculate square roots. Khanmigo typically didn’t correct mistakes when asked to double-check solutions.”

    Now being piloted by about 65,000 students in 44 school districts, Khanmigo emphasizes to students and teachers that it is imperfect.

    Sal Khan said he expects “a million or two million” students to be using it by next school year at a price to schools of $35 a student.

    Unlike ChatGPT, Khanmigo is trained not to give students the right answer but to guide them through problems. It offers tutoring in third grade and up in math, language arts, history and science. It can give feedback on student essays, engage in simulated dialogue as famous literary characters and debate contemporary issues.

    In testing the product, the WSJ asked Khanmigo for help finding the length of the third side of a right triangle, a problem that students would likely encounter in eighth-grade math.

    Khanmigo correctly identified the Pythagorean theorem, a2 + b2 = c2, as crucial to finding the answer. When asked for the solution the bot offered responses such as: “I’m here to help you learn, not just give answers!”

    But Khanmigo struggled with math operations. When trying to solve a right triangle with a hypotenuse of 27 units and a side of 17, a reporter offered the wrong answer (430 rather than 440) to 272 minus 172. “Excellent!” Khanmigo responded. Later, it accepted the incorrect answer to the square root of 440.

    In another instance, Khanmigo constructed its own triangle problem with a hypotenuse of 15 units and a leg of nine. But when a reporter correctly said that 152 minus 92 equals 144, Khanmigo suggested the response was wrong. “I see where you’re coming from, but let’s take another look at the subtraction,” it said.
    .