AI — weekly megathread!
AI — weekly megathread!

AI — weekly megathread!

News provided by aibrews.com

  1. Meta AI presents Emu, a quality-tuned latent diffusion model for generating highly aesthetic images. Emu significantly outperforms SDXLv1.0 on visual appeal [Paper].
  2. Meta AI researchers present a series of long-context LLMs with context windows of up to 32,768 tokens. LLAMA 2 70B variant surpasses gpt-3.5-turbo-16k’s overall performance on a suite of long-context tasks [Paper].
  3. Abacus AI released a larger 70B version of Giraffe. Giraffe is a family of models that are finetuned from base Llama 2 and have a larger context length of 32K tokens [Details].
  4. Meta announced [Details]:
    1. Meta AI - a new AI assistant users can interact with on WhatsApp, Messenger and Instagram. Will also be available on Ray-Ban Meta smart glasses and Quest 3, Meta’s mixed reality headset.
    2. AI stickers that enable users to generate customized stickers for chats and stories using text. Powered by Llama 2 and the new foundational model for image generation, Emu.
    3. 28 AI characters, each with a unique personality that users can message on WhatsApp, Messenger, and Instagram.
    4. New AI editing tools, restyle and backdrop in Instagram.
    5. AI Studio - a platform that supports the creation of custom AIs by coders and non-coders alike.
  5. Cerebras and Opentensor released Bittensor Language Model, ‘BTLM-3B-8K’, a new 3 billion parameter open-source language model with an 8k context length trained on 627B tokens of SlimPajama. It outperforms models trained on hundreds of billions more tokens and achieves comparable performance to open 7B parameter models. The model needs only 3GB of memory with 4-bit precision and takes 2.5x less inference compute than 7B models and is available with an Apache 2.0 license for commercial use [Details].
  6. OpenAI is rolling out, over the next two weeks, new voice and image capabilities in ChatGPT enabling ChatGPT to understand images, understand speech and speak. The new voice capability is powered by a new text-to-speech model, capable of generating human-like audio from just text and a few seconds of sample speech. [Details].
  7. Mistral AI, a French startup, released its first 7B-parameter model, Mistral 7B, which outperforms all currently available open models up to 13B parameters on all standard English and code benchmarks. Mistral 7B is released in Apache 2.0, making it usable without restrictions anywhere [Details].
  8. OpenAI has returned the ChatGPT browsing feature for Plus subscribers, enabling ChatGPT to access internet for current information. It was disabled earlier as users were able to deploy it to bypass the paywalls of leading news publishers [Details].
  9. Microsoft has released AutoGen - an open-source framework that enables development of LLM applications using multiple agents that can converse with each other to solve a task. Agents can operate in various modes that employ combinations of LLMs, human inputs and tools [Details].
  10. LAION released LeoLM, the first open and commercially available German foundation language model built on Llama-2 [Details]
  11. Researchers from Google and Cornell University present and release code for DynIBaR (Neural Dynamic Image-Based Rendering) - a novel approach that generates photorealistic renderings from complex, dynamic videos taken with mobile device cameras, overcoming fundamental limitations of prior methods and enabling new video effects [Details].
  12. Cloudflare launched Workers AI (an AI inference as a service platform), Vectorize (a vector Database) and AI Gateway with tools to cache, rate limit and observe AI deployments. Llama2 is available on Workers AI [Details].
  13. Amazon announced the general availability of Bedrock, its service that offers a choice of generative AI models from Amazon itself and third-party partners through an API [Details].
  14. Google announced it’s giving website publishers a way to opt out of having their data used to train the company’s AI models while remaining accessible through Google Search [Details].
  15. Spotify has launched a pilot program for AI-powered voice translations of podcasts in other languages - in the podcaster’s voic. It uses OpenAI’s newly released voice generation model [Details].
  16. Getty Images has launched a generative AI image tool, ‘Generative AI by Getty Images’, that is ‘commercially‑safe’. It’s powered by Nvidia Picasso, a custom model trained exclusively using Getty’s images library [Details].
  17. Optimus, Tesla’s humanoid robot, can now sort objects autonomously and do yoga. Its neural network is trained fully end-to-end [Link].
  18. Amazon will invest up to $4 billion in Anthropic. Developers and engineers will be able to build on top of Anthropic’s models via Amazon Bedrock [Details].
  19. Google Search indexed shared Bard conversational links into its search results pages. Google says it is working on a fix [Details].
  20. Pika Labs' text-to-video tool now lets users encrypt a message in a video [Twitter Link].

🔦 Weekly Spotlight

  1. How AI-powered echoes are making waves in the fight against heart failure [Link].
  2. AI language models can exceed PNG and FLAC in lossless compression, says study [Link].
  3. Everyone is above average. Is AI a Leveler, King Maker, or Escalator? [Link].
  4. What Builders Talk About When They Talk About AI [Link].
  5. The Llama Ecosystem: Past, Present, and Future [Link].

- - -

Welcome to the r/artificial weekly megathread. This is where you can discuss Artificial Intelligence - talk about new models, recent news, ask questions, make predictions, and chat other related topics.

Click here for discussion starters for this thread or for a separate post.

Self-promo is allowed in these weekly discussions. If you want to make a separate post, please read and go by the rules or you will be banned.

Previous Megathreads & Subreddit revamp and going forward

submitted by /u/jaketocake
[link] [comments]