AI — weekly megathread!
AI — weekly megathread!

AI — weekly megathread!

News provided by aibrews.com

  1. Meta AI releases Code Llama, a large language model for coding that is built on top of Llama 2. Code Llama Code outperformed state-of-the-art publicly available LLMs on code tasks. It is free for research and commercial use. You can try it on Fireworks AI and Perplexity Labs [Details].
  2. Meta AI released SeamlessM4T (Massive Multilingual Multimodal Machine Translation) - the first all-in-one, multilingual multimodal translation model. SeamlessM4T can perform multiple tasks across speech and text: speech-to-text, speech-to-speech, text-to-speech, text-to-text translation, and speech recognition. It supports 100 languages for input (speech + text), 100 languages for text output and 35 languages (plus English) for speech output [Details | Demo | Hugging Face |GitHub].
  3. Researchers from UC San Francisco and UC Berkeley have developed new brain-computer technology (BCI) that enables a stroke survivor to speak with facial expressions for first time in 18 years via a digital avatar. It is the first time that either speech or facial expressions have been synthesized from brain signals [Details].
  4. Hugging Face released IDEFICS, an open-access 80 billion parameters multimodal model that accepts sequences of images and texts as input and generates coherent text as output. It is reproduction of Flamingo (developed by DeepMind) and is comparable in performance with the original closed-source model across various image-text understanding benchmarks. IDEFICS is built solely on publicly available data and models (LLaMA v1 and OpenCLIP) [Details].
  5. Allen Institute for AI has released Dolma, the largest open dataset of 3 trillion tokens from a diverse mix of web content, academic publications, code, books, and encyclopedic materials. [HuggingFace Hub].
  6. Open AI is now letting developers fine-tune GPT-3.5 Turbo. Fine-tuning for GPT-4 coming this fall. Early tests have shown that fine-tuned GPT-3.5 Turbo can match or exceed GPT-4 on certain narrow tasks [Details | Guide].
  7. ElevenLabs released Eleven Multilingual v2 - a new Foundational AI speech model for nearly 30 languages. ElevenLabs is now out of beta [Details].
  8. Hugging Face announced SafeCoder - a code assistant solution built for the enterprise [Details].
  9. Midjourney released 'Vary Region’, an ‘inpainting’ feature to regenerate specific parts of an upscaled image [Details].
  10. Stability AI is collaborating with Nvidia for improvement in the speed and efficiency of Stable Diffusion XL by integrating NVIDIA TensorRT, a high-performance optimization framework [Details | Hugging face].
  11. OpenAI partners with Scale to provide support for enterprises fine-tuning models [Details].
  12. YouTube is collaborating with Universal Music Group to launch Music AI Incubator [Details].
  13. IBM has built a new, state-of-the-art generative AI code model to transform legacy COBOL programs to enterprise Java [Details].
  14. A US federal judge gave a ruling that a piece of art created by AI is not open to protection [Details].
  15. ElevenLabs has teamed up with the open-access video platform ScienceCast, allowing users to generate instant narrated summaries of scientific papers [Details].
  16. Google announced a number of security-related enhancements to Google Workspace products, including GMail and Drive, some of which will take advantage of AI to automate certain tasks [Details].
  17. ChatGPT custom instructions are now live in the EU and UK [Link].
  18. HuggingChat now supports Amazon SageMaker deployment which allows organizations to build ChatGPT-like experiences fully within AWS [GitHub].
  19. Meta AI presents Shepherd - a language model specifically tuned to critique model responses & suggest refinements. It goes beyond the capabilities of untuned models to identify diverse errors & suggest improvements [Paper].
  20. Adobe Express adds generative AI features powered by Adobe Firefly to its free plan, enabling generation of images and text effects using text prompts [Link].
  21. Project Jupyter released Jupyter AI - generative artificial intelligence in Jupyter notebooks. Users can generate code, ask questions about their local files, and generate entire notebooks from natural language prompts [Link].
  22. Nvidia released the code for Neuralangelo, which can turn regular videos into highly detailed 3D models of both objects and large-scale indoor/outdoor scenes.[GitHub].

🔦 Weekly Spotlight

  1. Jailbreaking wrist watch into a real-life second brain [Link].
  2. I Made Stable Diffusion XL Smarter by Finetuning it on Bad AI-Generated Images [Link].
  3. DoctorGPT: an open-source LLM that can pass the US Medical Licensing Exam. It works offline and is cross-platform [Link].
  4. Llama-2-7B-32K-Instruct — and fine-tuning for Llama-2 models with Together API [Link].
  5. A MIT-licensed JS starter kit by a16z, for building and customizing your own version of AI town - a virtual town where AI characters live, chat and socialize [Link].

—-------

Welcome to the r/artificial weekly megathread. This is where you can discuss Artificial Intelligence - talk about new models, recent news, ask questions, make predictions, and chat other related topics.

Click here for discussion starters for this thread or for a separate post.

Self-promo is allowed in these weekly discussions. If you want to make a separate post, please read and go by the rules or you will be banned.

Previous Megathreads & Subreddit revamp and going forward

submitted by /u/jaketocake
[link] [comments]