Author: IBL News

  • Databrick Releases an Improved Open Source LLM Licensed For Reuse and Commercial Use

    Databrick Releases an Improved Open Source LLM Licensed For Reuse and Commercial Use

    IBL News | New York

    Databricks released yesterday an improved version of its open-source, free-to-commercialize, large language model (LLM) with 12 billion parameters, called Dolly 2.0.

    Based on the EleutherAI pythia model family, Dolly 2.0 has been “fine-tuned exclusively on a new, high-quality human-generated instruction following dataset, crowdsourced among Databricks 5,000 employees during March and April of 2023.,” according to the company.

    “We are open-sourcing the entirety of Dolly 2.0, including the training code, the dataset, and the model weights, all suitable for commercial use. This means that any organization can create, own, and customize powerful LLMs that can talk to people, without paying for API access or sharing data with third parties.”

    Under the licensing terms (Creative Commons Attribution-ShareAlike 3.0 Unported License), anyone can use, modify, or extend this dataset for any purpose, including commercial applications.

    databricks-dolly-15k on GitHub contains 15,000 high-quality human-generated prompt / response pairs specifically designed for instruction tuning large language models.

    The first version of Dolly was trained using a dataset created by the Stanford Alpaca team created using the OpenAI API. That dataset contained output from ChatGPT, and that prevented commercial use, as it would compete with OpenAI.

    “As far as we know, all the existing well-known instruction-following models (AlpacaKoalaGPT4AllVicuna) suffer from this limitation, prohibiting commercial use. To get around this conundrum, we started looking for ways to create a new dataset not “tainted” for commercial use.”

    Databricks said that “it doesn’t expect Dolly to be state-of-the-art in terms of effectiveness.”

    “However, we do expect Dolly and the open source dataset will act as the seed for a multitude of follow-on works, which may serve to bootstrap even more powerful language models.”

    • Download Dolly 2.0 model weights at Databricks Hugging Face page
    • Dolly repo on databricks-labs with databricks-dolly-15k dataset.

  • The AI Apps Associated Themselves with ChatGPT Generated Hefty Revenue

    The AI Apps Associated Themselves with ChatGPT Generated Hefty Revenue

    IBL News | New York

    The top 10 AI mobile apps have already pulled in over $14 million to date, according to analytics provider data.ai.

    These apps are mostly dubious ChatGPT apps claiming to be associated with OpenAI in order to charge hefty subscription prices for accessing ChatGPT — a service that’s been free via the web.

    This phenomenon is an indication of how much ground OpenAI has ceded in the mobile app market by not having its own official mobile app available.

    For the most part, none of this group of apps are trying to establish their own brand and identity. Instead, they’re keyword-stuffing their apps’ titles to match the search terms of people who are looking for ChatGPT or similar, which are mostly “AI,” “Chat,” or “Chatbot” and “Assistant” will help drive.

    The majority of the top 10 apps saw little consumer spending before ChatGTP’s addition, in December 2022.

    The only exception is the app Pixelcut AI Photo Editor, a “magic writer” copywriter tool that uses GPT. Its creator AI Photo Editor generated $19.8 million in 2022.

    In addition to Pixelcut, the other nine earning apps are: Genie – AI Chatbot; AI Chat – Chatbot AI Assistant; AI Chatbot – Open Chat Writer; Apo – AI Personal Assistant; Chat AI Bot – Writing Assistant; ChatOn – AI Chatbot Assistant; AI Chat – Ask Anything; Chat AI – Ask Anything; and GoatChat.

    Of these, Genie has generated the most revenue this year, with $3.2 million in global consumer spending so far in 2023.

  • Startup Tome, Which Uses Generative AI to Create PowerPoint-Style Presentations, Valued at $300M

    Startup Tome, Which Uses Generative AI to Create PowerPoint-Style Presentations, Valued at $300M

    IBL News | New York

    AI PowerPoint-style storytelling startup Tome recently raised $43 million in Series B funding.

    Lightspeed Venture Partners led the round while individual investors included Coatue, Greylock, Stability.ai CEO Emad Mostaque, and former Google CEO Eric Schmidt.

    While still in a pre-revenue stage, the company is now valued at $300 million post-investment, up from $175 million at its 2021 Series A round, according to an article in Forbes.

    San Francisco-based Tome says it’s the fastest productivity software maker to ever reach one million users since its September release.

    Tome’s AI software rethinks the PowerPoint system organizing almost real-time presentations in PowerPoint or Google Slides, with slides organized by a table of contents.

    “We set out to build a company that can help anyone tell a compelling story,” the Cofounders of the company, Keith Peiris and Henri Liriani, said. “Storytelling is the elementary building block of productivity for humanity, from cave drawings to stories around the fire to PowerPoint.”

    Tome uses large-language models to make queries behind the scenes to fulfill a user’s presentation prompt. First, its software generates an output framework; next, smaller queries fill out images and text. Users can then fine-tune the results or ask the AI to generate more pages via additional prompts.

    Tome plans to charge about $10 per user for a monthly subscription when it rolls out its enterprise version this year.

    Three decades after its debut, PowerPoint is still a force. Microsoft has invested heavily in OpenAI with a plan to integrate its models across its products. Collaboration software unicorns like Canva, Coda, and Notion are currently announcing AI features on their products.

     

  • Anthropic’s Plans to Raise Billions to Take on OpenAI

    Anthropic’s Plans to Raise Billions to Take on OpenAI

    IBL News | New York

    AI startup Anthropic aims to raise $5 billion over the next two years to take on rival OpenAI and enter over a dozen major industries, according to TechCrunch.

    Founded in 2021 by former OpenAI researchers as a public benefit corporation, Anthropic says in a pitch for Series C fundraising says that it requires a billion dollars in spending over the next 18 months to build a model with tens of thousands of GPUs, ten times more capable than today’s most powerful AI.

    This model — which would be a successor to Claude, Anthropic’s current chatbot — is described as a “next-gen algorithm for AI self-teaching,” making reference to an AI training technique it developed called “constitutional AI,” which aligns with human intentions.

    “These models could begin to automate large portions of the economy,” the pitch deck reads.

    Other competing startups are Cohere and AI21 Labs in the AI systems space.

    Google is also among Anthropic’s investors, having pledged $300 million in Anthropic for a 10% stake.

    Other Anthropic backers include James McClave, Facebook and Asana co-founder Dustin Moskovitz, former Google CEO Eric Schmidt, and founding Skype engineer Jaan Tallinn.

  • Cerebras Releases as Open Source Seven Large LLMs with 13 Billion Parameters

    Cerebras Releases as Open Source Seven Large LLMs with 13 Billion Parameters

    IBL News | New York

    Silicon Valley–based maker of a dedicated AI computer and the world’s largest computer chip, Cerebras Systems released a series of seven GPT large language models (LLMs), methodology, training weights, and a recipe for open use via the permissive industry-standard Apache 2.0 license. This solution, called Cerebras-GPT, means that these models can be used for research or commercial ventures without royalties.

    The company used non-Nvidia GPU-based systems to train LLMs up to 13 billion parameters. All seven models were trained on the sixteen CS-2 systems in the Cerebras Andromeda AI supercomputer using the Chinchilla formula.

    “These are the highest accuracy models for a computing budget and are available today open-source,” said the company.

    In a first among AI hardware companies, Cerebras researchers trained a series of seven GPT models with 111M, 256M, 590M, 1.3B, 2.7B, 6.7B, and 13B parameters.

    “Typically a multi-month undertaking, this work was completed in a few weeks thanks to the incredible speed of the Cerebras CS-2 systems that make up Andromeda, and the ability of Cerebras’ weight streaming architecture to eliminate the pain of distributed computing. These results demonstrate that Cerebras’ systems can train the largest and most complex AI workloads today.”

    • “The training weights provide a highly accurate pre-trained model for fine-tuning. By applying a modest amount of custom data, anyone can create powerful, industry-specific applications with minimal work.”
    • “The models’ various sizes and their accompanying checkpoints allow AI researchers to create and test new optimizations and workflows that broadly benefit the community.”

    Traditional LLM training on GPUs requires a complex amalgam of pipeline, model, and data parallelism techniques. Cerebras’ weight streaming architecture is a data-parallel-only model that requires no code or model modification to scale to arbitrarily large models.

    “We’ve worked to make this task easier with releases such as the Pile and the Eval Harness, and we are very excited to see Cerebras build on our work to produce a family of open models that will be useful to researchers around the world,” said Stella Biderman, Executive Director at EleutherAI.

    All seven Cerebras-GPT models are available on Hugging Face and Cerebras Model Zoo on GitHub. The Andromeda AI supercomputer used to train these models is available on-demand in this URL.

    Cerebras published a technical blog post with the details of the seven models and the scaling laws that they produce. A research paper will be released shortly.

    The company posted not just the programs’ source, in Python and TensorFlow format, but also the details of the training regimen by which the programs were brought to a developed state of functionality.

    Currently, a handful of companies hold the keys to LLMs. OpenAI is closed, with GTP-4 operating as a black box for the public. Meta’s LLAMA is closed to for-profit organizations, and Google is closed to a varying degree.

    Cerebras, echoing the researchers’ community, says that AI needs to be open and reproducible for it to broadly benefit humanity.


    • ZDNet: AI pioneer Cerebras opens up generative AI where OpenAI goes dark

    .

  • International Baccalaureate Assessment System Allows Students to Use ChatGPT

    International Baccalaureate Assessment System Allows Students to Use ChatGPT

    IBL News | New York

    International Baccalaureate (IB), which offers an alternative qualification to A-levels and Highers, said that their students can use ChatGPT in their essays. But they must be clear when they were quoting its responses.

    “Content created by this chatbot must be treated like any other source and attributed when used,” said IB.

    IB is taken by thousands of children every year in the UK at more than 120 schools and all over Europe.

    ChatGPT has become a sensation since its public release in November 2022, due to its ability to produce plausible responses to text prompts, including requests to write essays.

    However, ChatGPT-based cheating capabilities have alarmed teachers and the academic profession.

    Matt Glanville, the IB’s Head of Assessment Principles and Practice, said the chatbot should be embraced as “an extraordinary opportunity”.

    “The clear line between using ChatGPT and providing original work is exactly the same as using ideas taken from other people or the internet. As with any quote or material adapted from another source, it must be credited in the body of the text and appropriately referenced in the bibliography,” he added.

    “When AI can essentially write an essay at the touch of a button, we need our pupils to master different skills, such as understanding if the essay is any good or if it has missed context, has used biased data or if it is lacking in creativity. These will be far more important skills than writing an essay, so the assessment tasks we set will need to reflect this.”

  • Bloomberg Introduces a 50-Billion Parameter LLM Built For Finance

    Bloomberg Introduces a 50-Billion Parameter LLM Built For Finance

    IBL News | New York

    Bloomberg released this week a research paper introducing BloombergGPT, a new large-language (LLM) AI model with 50 billion parameter built from scratch for finance.

    The company said that BloombergGPT, that has been specifically trained on a wide range of financial data, outperforms similarly-sized models by significant margins (as shown in the table below).

    “It represents the first step in the development and application of this new technology for the financial industry,” said the company.

    “This model will assist Bloomberg in improving existing financial NLP tasks, such as sentiment analysis, named entity recognition, news classification, and question answering, among others. Furthermore, BloombergGPT will unlock new opportunities for marshalling the vast quantities of data available on the Bloomberg Terminal.”

    Bloomberg researchers pioneered a mixed approach that combines both finance data with general-purpose datasets to train a model that achieves best-in-class results on financial benchmarks, while also maintaining competitive performance on general-purpose LLM benchmarks.

    Bloomberg’s data analysts collected financial language documents over the span of forty years, pulled from this extensive archive to create a comprehensive 363 billion token dataset consisting of English financial documents.

    This data was augmented with a 345 billion token public dataset to create a large training corpus with over 700 billion tokens. Using a portion of this training corpus, the team trained a 50-billion parameter decoder-only causal language model.

     

  • 2U Sues U.S. Department of Education Over “New Regulation that Overreaches Its Authority”

    2U Sues U.S. Department of Education Over “New Regulation that Overreaches Its Authority”

    IBL News | New York

    OPM (Online Program Management) provider 2U Inc, the owner of edX.org, sued the U.S. Department of Education in federal court this Tuesday over guidance it issued in February governing the relationships between colleges and third-party contractors that perform key services for them.

    Hundreds of U.S. colleges use OPM services to start and run online programs, often trading upfront capital from the companies for a portion of their programs’ revenue.

    In the filed suit against the Department of Education and its Secretary Miguel Cardona, 2U says that the agency has overreached its authority.

    Under the Education Department’s new definition, OPMs that provide colleges with recruiting and retention services, as well as educational content, like 2U, will broadly be considered third-party services.

    2U’s lawsuit alleges the department overstepped its power by independently rewriting the Higher Education Act’s definition of a third-party servicer. The suit was filed in U.S. District Court for the District of Columbia.

    The agency is particularly focused on entities that receive a share of tuition revenue in exchange for their services, arguing that “it can drive up the price of higher education and draw students to low-value academic programs at subpar institutions.”

    “2U cares deeply about our partnerships with leading non-profit colleges and universities across the nation,” Matthew Norden, chief legal officer at 2U, said in a statement. “We believe this recent action by the Department of Education will not only impinge on our ability to serve their students but also ultimately hurt their quality of education.”

    According to the lawsuit, 2U would face substantial and irreparable harm if it is classified as a third-party servicer in the eyes of the department, being forced to amend current contracts, undergo “burdensome and intrusive” audits, and pay nonrefundable compliance costs.

    The company would also be forced to cut off its South African subsidiary due to the guidance’s ban on foreign-owned and foreign-based subcontractors.

     

  • The OpenAI’s CEO Envisions a Universal Income Society to Compensate Jobs Replaced by AI

    The OpenAI’s CEO Envisions a Universal Income Society to Compensate Jobs Replaced by AI

    IBL News | New York

    Sam Altman, the CEO of OpenAI, an organization that has moved at record speed from a small research nonprofit into a multibillion-dollar company, with the help of Microsoft, showed his contradictions in an interview with The Wall Street Journal last week.

    He is featured as an entrepreneur who made a fortune investing in young startups, owner of the three mansions in California, and a family office now employing dozens to manage those properties along with investments in companies such as Worldcoin, Helion Energy, and Retro.

    Sam Altman said he fears what could happen if AI is rolled out into society recklessly and argues that is uniquely dangerous to have profits be the main driver of developing powerful AI models.

    Meanwhile he says that his ultimate mission is to build AGI (artificial general intelligence) while stating a goal of forging a new world order in which machines free people to pursue more creative work. In his vision, universal basic income will help compensate for jobs replaced by AI and humanity will love AI so much that an advanced chatbot could represent “an extension of your will.”

    In the long run, he wants to set up a global governance structure that would oversee decisions about the future of AI and gradually reduce the power OpenAI’s executive team has over its technology.

    “Backers say his brand of social-minded capitalism makes him the ideal person to lead OpenAI. Others, including some who’ve worked for him, say he’s too commercially minded and immersed in Silicon Valley thinking to lead a technological revolution that is already reshaping business and social life,” writes The Wall Street Journal.

    “OpenAI’s headquarters — with 400 employees —, in San Francisco’s Mission District, evoke an affluent New Age utopia more than a nonprofit trying to save the world. Stone fountains are nestled amid succulents and ferns in nearly all of the sun-soaked rooms.”

    Elon Musk, one of OpenAI’s critics who co-founded the nonprofit in 2015 but parted ways in 2018 after a dispute over its control and direction, said that OpenAI had been founded as an open-source nonprofit “to serve as a counterweight to Google, but now it has become a closed source, maximum-profit company effectively controlled by Microsoft.” 

    Billionaire venture capitalist Peter Thiel, a close friend of Mr. Altman’s and an early donor to the nonprofit, has long been a proponent of the idea that humans and machines will one day merge.

    Behind OpenAI there is a for-profit arm, OpenAI LP, that reports to the nonprofit parent.

    According to some employees, the partnership of Sam Altman with Satya Nadella, the Microsoft CEO, started in 2019, contradicted OpenAI’s initial pledge to develop artificial intelligence outside the corporate world. They saw the deal as a Faustian bargain.

    Microsoft initially invested $1 billion in OpenAI and obtained exclusivity using Microsoft’s giant computer servers, via its Azure cloud service, to train its AI models, giving the tech giant the sole right to license OpenAI’s technology for future products.

    Altman’s other projects include Worldcoin, a company he co-founded that seeks to give cryptocurrency to every person on earth.

    He has put almost all his liquid wealth in recent years in two companies. He has put $375 million into Helion Energy, which is seeking to create carbon-free energy from nuclear fusion and is close to creating “legitimate net-gain energy in a real demo,” Mr. Altman said.

    He has also put $180 million into Retro, which aims to add 10 years to the human lifespan through “cellular reprogramming, plasma-inspired therapeutics and autophagy,” or the reuse of old and damaged cell parts.

     

  • Artificial Intelligence Enters a New Phase of Corporate Dominance

    Artificial Intelligence Enters a New Phase of Corporate Dominance

    IBL News | New York

    The 2023 AI Index [read in full here] — compiled by researcher from Stanford University as well as AI companies including Google, Anthropic, McKinsey, LinkedIn, and Hugging Face — suggests that AI is entering an era of corporate control, with industry players dominating over academia and government in deploying and safeguarding AI applications.

    Decisions about how to deploy this technology and how to balance risk and opportunity lie firmly in the hands of corporate players, as we’ve seen over the past years with AI tools, like ChatGPT, Bing, and image-generating software Midjourney, going mainstream.

    The report, released today, states: “Until 2014, most significant machine learning models were released by academia. Since then, industry has taken over. In 2022, there were 32 significant industry-produced machine learning models compared to just three produced by academia. Building state-of-the-art AI systems increasingly requires large amounts of data, compute, and money, resources that industry actors inherently possess in greater amounts compared to nonprofits and academia.”

    Many experts in the AI world, mentioned by The Verge, worry that the incentives of the business world will also lead to dangerous outcomes as companies rush out products and sideline safety concerns.

    As AI tools become more widespread, the number of errors and malicious use cases are increasing. Such incidents might include fatalities involving Tesla’s self-driving software; the use of audio deepfakes in corporate scams; the creation of nonconsensual deepfake nudes; and numerous cases of mistaken arrests caused by faulty facial recognition software.