Artificial Intelligence

11 readers
1 users here now

Reddit's home for Artificial Intelligence (AI).

founded 1 year ago
MODERATORS
476
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/wyem on 2024-03-22 16:28:03.


  • Meta AI introduced SceneScript, a novel method of generating scene layouts and representing scenes using language. SceneScript allows AR & AI devices to understand the geometry of physical spaces. It uses next token prediction like an LLM, but instead of natural language SceneScript model predicts the next architectural tokens such as ‘wall’ or ‘door.’ [Details].
  • Sakana AI presented Evolutionary Model Merge, a general method that uses evolutionary techniques to automate the creation of new foundation models without extensive additional training data or compute. Sakana AI applied this method to evolve 3 powerful foundation models for Japan: Large Language Model (EvoLLM-JP), Vision-Language Model (EvoVLM-JP) and Image Generation Model (EvoSDXL-JP) [Details | Hugging Face].
  • Elon Musk's brain-chip startup Neuralink livestreamed its first patient implanted with a chip using his mind to play online chess [Details | video].
  • Stability AI released Stable Video 3D (SV3D), a generative model based on Stable Video Diffusion that takes in a still image of an object as a conditioning frame, and generates an orbital video of that object. It delivers improved quality and multi-view when compared to the previously released Stable Zero123, and outperforms other open source alternatives such as Zero123-XL Stable Video 3D can be used now for commercial purposes with a Stability AI Membership [Details |Hugging Face].
  • Waabi introduced Copilot4D, a foundation model for self-driving. It is the first foundation model purpose built for the physical world that can reason in 3D space and the fourth dimension, time. Copilot4D can understand the impact the self-driving vehicle future actions have on the behavior of surrounding traffic participants [Details].
  • Open Interpreter launched 01 Light, a portable voice interface that controls your home computer. It can see your screen, use your apps, and learn new skills. Batch 1 sold out in 2.5 hours; profits will be redistributed to open-source contributors [Details].
  • NVIDIA introduced:
  1. NVIDIA NIM, a containerized inference microservice to simplify deployment of generative AI models across various infrastructures. Developers can test a wide range of models using ‌cloud APIs from the NVIDIA API catalog or they can self-host the models by downloading NIM and deploying with Kubernetes [Details].
  2. Project GR00T, a general-purpose foundation model for humanoid robots and significant upgrades to the NVIDIA Isaac robotics platform. The GR00T model will enable a robot to understand multimodal instructions, such as language, video, and demonstration, and perform a variety of useful tasks. NVIDIA is building a comprehensive AI platform for several humanoid robot companies including 1X Technologies, Agility Robotics, Boston Dynamics, Figure AI etc. [Details].
  3. Earth-2 climate digital twin cloud platform for simulating and visualizing weather and climate at unprecedented scale. Earth-2’s APIs offer AI models and employ a new NVIDIA generative AI model called CorrDiff that generates 12.5x higher resolution images than current numerical models 1,000x faster and 3,000x more energy efficiently [Details].
  4. Next-generation AI supercomputer, the NVIDIA DGX SuperPOD, powered by its new NVIDIA GB200 Grace Blackwell Superchip designed to meet the demanding requirements of generative AI training and inference workloads involving trillion-parameter models [Details].
  • Google’s Gemini 1.5 Pro multimodel model with 1M token context window is now available to all in the Google AI Studio, with API being gradually rolled out [Link].
  • ByteDance released AnimateDiff-Lightning, a lightning-fast text-to-video generation model. It can generate videos more than ten times faster than the original AnimateDiff [Hugging Face | Demo].
  • Pleias, a French start-up, released Common Corpus - the largest public domain dataset released for training LLMs. It is multilingual and includes 500 billion words from a wide diversity of cultural heritage initiatives [Details].
  • Aether Research released Cerebrum 8x7b, a large language model (LLM) created specifically for reasoning tasks. It is based on the Mixtral 8x7b model and offers competitive performance to Gemini 1.0 Pro and GPT-3.5 Turbo on a range of tasks that require reasoning [Hugging Face].
  • Stability AI, Medical AI Research Center (MedARC) and others presented MindEye2, a model that can reconstruct seen images from fMRI brain activity using only 1 hour of training data. Given a sample of fMRI activity from a participant viewing an image, MindEye can identify either which image out of a pool of possible image candidates was the original seen image (retrieval), or it can recreate the image that was seen (reconstruction) along with its text caption [Details].
  • Nous Research released Hermes 2 Pro 7B, an upgraded, retrained version of Nous Hermes 2. It improves several capabilities, using an updated and cleaned version of the Hermes 2 dataset, and excels at function calling and JSON structured output [Hugging Face]
  • Google AI introduced a generalizable user-centric interface to help radiologists leverage ML models for lung cancer screening. The system takes CT imaging as input and outputs a cancer suspicion rating along with the corresponding regions of interest [Details | GitHub]
  • xAI released the base model weights and network architecture of Grok-1 under the Apache 2.0 license. Grok-1 is a 314 billion parameter Mixture-of-Experts model trained from scratch [GitHub | [Hugging Face]](https://huggingface.co/xai-org/grok-1).
  • Lighthouz AI launched the Chatbot Guardrails Arena in collaboration with Hugging Face, to stress test LLMs and privacy guardrails in leaking sensitive data. Chat with two anonymous LLMs with guardrails and try to trick them into revealing sensitive financial information and cast your vote for the model that shows greater privacy [Details].
  • Apple introduced MM1, a family of multimodal models up to 30B parameters,

consisting of both dense models and mixture-of-experts (MoE) variants that are SOTA in pre-training metrics and achieve competitive performance across 12 established multimodal benchmarks [Paper].

  • Stability AI introduced a suite of image services on the Stability AI Developer Platform API for for image generation, upscaling, outpainting and editing [Details].
  • Google Research presented VLOGGER, a novel framework to synthesize humans from audio. Given a single input image and a sample audio input, our it generates photorealistic and temporally coherent videos of the person talking and vividly moving [Details].
  • Stability AI presented SD3-Turbo, a fast text-to-image foundation model that achieves the sample quality of SD3, Midjourney, and Dalle-E 3 in only 4 steps. Code and model weights will be publicly available [Paper].
  • GitHub introduced Code Scanning Autofix, for GitHub Advanced Security customers, powered by GitHub Copilot and CodeQL. Code Scanning Autofix covers more than 90% of alert types in JavaScript, Typescript, Java, and Python, and delivers code suggestions shown to remediate more than two-thirds of found vulnerabilities with little or no editi...

Content cut off. Read original on https://old.reddit.com/r/artificial/comments/1bl1ys9/this_week_in_ai_all_the_major_ai_developments_in/

477
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/ai_happy on 2024-03-23 01:15:27.

478
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/AminoOxi on 2024-03-22 10:44:58.


.

479
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/Yellowthrone on 2024-03-22 00:29:35.


suno ai website

480
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/Revolutionary_Spaces on 2024-03-21 18:51:56.


There are good AI companies, and then there is Speechify AI.

Reason number 1: after 7 days you're out of luck if you *ever* want a refund.

I subscribed for a year (I should have never done this). Tried to use the app and service, but just couldn't stand the service (it doesn't do anything it promises to do well - for example skipping footnotes? Nope. It's not smart enough.)

A normal company would say "Oh, we took a lot of money from you, want to keep you happy to avoid reviews like this, and are glad to refund you at a prorated rate."

Not Speechify - they won't give you a dime back after 7 days. Doesn't matter if you never use the service again. OUT OF LUCK.

I've worked with their team for days trying to get this refunded but they refuse. I'm forced to leave this review now.

Reason number 2: their team emailed me PII during the support engagement.

While engaging their support team they accidentally replied to me for another support ticket. While doing so they revealed PII for another customer (an email address).

Reason number 3: they employ a bot army on X (Twitter) to reply to bad reviews.

See the replies here: - obvious bot behavior (similar usernames, phrasing, popped out of nowhere).

Ever wonder how Speechify has so many positive reviews? I wouldn't be surprised if they were faking them using bots as well.

481
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/Excellent-Target-847 on 2024-03-20 05:01:41.


  1. DeepMind co-founder Mustafa Suleyman will become the head of a newly created consumer AI unit at Microsoft, the scientist and entrepreneur said on Tuesday.[1]
  2. Nvidia’s Jensen Huang says AI hallucinations are solvable, artificial general intelligence is 5 years away.[2]
  3. Saudi Arabia Plans $40 Billion Push Into Artificial Intelligence.[3]
  4. Today at GDC, Roblox unveiled a pair of AI technologies aiming to streamline 3D modeling for the platform’s boundless legion of content creators.[4]

Sources:

[1]

[2]

[3]

[4]

482
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/ImpressiveContest283 on 2024-03-20 17:53:21.


Google recently launched TacticAI, an AI system aimed at providing professional tips, especially for corner kicks, showcasing how technology can transform sports.

Is Google adopting an aggressive strategy with new model releases? Just a few days ago, they introduced Vlogger, and now TacticAI.

What do you think?

483
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/Healthy_Moment_1804 on 2024-03-20 06:57:24.

484
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/JONANz_ on 2024-03-20 02:11:45.


You need an invitation to make an account and save the tools, make lists of them to bundle the ones you like together, and write/read comments.

Got a few invites left. Let me know if any of you guys would like one.

485
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/thisisinsider on 2024-03-19 23:42:38.

486
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/Low_Complaint2254 on 2024-03-19 03:47:49.

487
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/Armand_Roulinn on 2024-03-18 19:50:34.

488
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/clonefitreal on 2024-03-18 07:51:07.

489
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/drgoldenpants on 2024-03-17 22:51:41.


Time to start cooking

490
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/clonefitreal on 2024-03-16 11:36:16.

491
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/tedbarney12 on 2024-03-17 19:24:00.


I've been reading about Devin AI, and it seems many of you have been too. Do you really think it poses a significant threat to software developers, or is it just another case of hype? We're seeing new LLMs (Large Language Models) emerge daily. Additionally, if they've created something so amazing, why aren't they providing access to it?

A few users have had early first-hand experiences with Devin AI and I was reading about it. Some have highly praised its mind-blowing coding and debugging capabilities. However, a few are concerned that the tool could potentially replace software developers.

What's your thought?

492
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/Block-Busted on 2024-03-16 22:21:24.


Cartoon Brew posted an article about how DreamWorks is planning to go through a massive layoff in coming months:

Dreamworks Animation To Significantly Cut Staff In Coming Months (Exclusive)

Dreamworks Animation is firing on all cylinders right now. It’s looking to repeat first place with Kung Fu Panda 4 at the U.S. box office this weekend, and the newly-released teaser for Chris Sanders’s The Wild Robot was received with unbridled enthusiasm. But the mood inside of the iconic Glendale, California studio was glum this week as an untold number of workers were informed that they would be let go.

The studio insists that no one is being laid off. It is, as Dreamworks Animation chief operating officer Randy Lake put it a few months ago, an instance of downsizing through “natural attrition,” in which employment contracts are simply not being renewed.

However, the sheer number of people whose contracts aren’t being renewed has demoralized staffers and left many wondering about the future direction of the studio. While there are no specific numbers on how many people are being let go, the number is well into the hundreds. Many departments are being cut in half, and the creative staff on features could be reduced by as much as, if not more than, 40-50%.

Rank-and-file workers at the company don’t know exactly how many people have been let go outside of their own departments, and some are trying to piece together the details by comparing notes between departments. The studio declined to provide a specific figure to Cartoon Brew.

Cartoon Brew spoke with numerous workers at the studio, all of whom declined to be identified for fear of retribution. A studio representative responded on background to our queries.

Here is a more detailed explanation of what is happening:

Why are so many people being let go all at once? A studio rep explained that Dreamworks has multiple large productions – Kung Fu Panda 4 and The Wild Robot – that are winding down around the same time, and the timing of new projects doesn’t sync up with the end dates of these current productions. While the studio works to minimize the gaps, sometimes the dates don’t align. The departing workers who are wrapping up on Wild Robot currently have end dates ranging from mid-April through July.

But surely there must be more to the story? Dreamworks intended to produce two features mostly in-house in 2026, but workers were told in early February that production would not move forward on one of those films, a Bollywood-themed film that had multiple working titles, including Another Me.

Dreamworks told Cartoon Brew that the film was never greenlit, so it wasn’t technically cancelled, however multiple studio workers confirmed to us that they were told last month by department heads that there would be significant reductions in staffing due to this particular project not moving forward.

Some workers were told in meetings that the project wasn’t moving forward because it was deemed not marketable to broad audiences. A Dreamworks rep disputed this characterization and said that there were numerous reasons why a film might not move into production, with marketability being just one of those factors.

Further reducing the need for staff, Dreamworks does not currently have any fully in-house productions scheduled for 2025. While all of Dreamworks’s film are developed in-house at Glendale, just one of their three 2025 releases is being animated in-house. Workers confirmed that even this film will use a significantly reduced crew since the production is being split between Sony Pictures Imageworks in Canada and the Dreamworks in-house crew.

Will the studio rehire people who are being let go? Some workers have been told that the studio hopes to return to a “normal” size in 2025, but it is unlikely that it will grow again to its current size anytime in the near future. That’s because the studio told workers last fall that it was shifting away from fully in-house animation production as part of new cost-cutting measures ordered by parent company Comcast-NBCUniversal.

At the time, Lake said the goal was to reduce production costs by 20%. As a result, the studio is outsourcing parts of some of its productions to third-party studios in other countries; these other studios include Sony Pictures Imageworks in Canada, Jellyfish Pictures in the U.K., and Mikros Animation in India and Canada.

...which prompted this guy to post these replies:

Pixar to be next. Mark my word.

Producers can't stand working with talent because they're envious of people with talent. There purpose in life is to take credit for others work and abuse talent every chance they get.

I use to work at Disney - this isn't true. The business model for most of their non-A listed projects utilizes extensive outsourcing. As does Dreamworks. The only reason it hasn't extended to large projects like Frozen is quality; and those quality margins are fading away. Furthermore, AI or automation is truly happening and will roll into productions over the next 5 years at a hard-lined pace. I believe Dreamworks just cut nearly all of their matte painters - and it's no coincidence that tools like Midjourney can easily produce mattes now with high precision (I'm sure some matte painters would die hearing this - but it's true). Animation automation is around the corner - there are already research projects that are successfully showing text to 3D generation and text to animation. So ya, Pixar among the rest is on a downward spiral to layoffs.

The dream is over unfortunately. It's only a matter of time that animation studios operate with a skeleton crew performing purely directorial/supervision roles.

I wouldn't bet against reality. I made the transition years ago. These next 5 years are the turning point. Anyone that doesn't look outside of animation now is just asking for it.

...while also saying this in a separate thread:

AI is basically automation of jobs. It's a mind shift. The days of studios needing hundreds of artists will go away. While independents, like Bill Plympton, will now be able to make long form stories more effortlessly. So to aspire to work for a studio would be a mistake, but to aspire to be a storyteller and produce your own content would be the move.

To further emphasis the advice here - Dreamworks Animation just wrapped on Kung Fu Panda and immediately started laying off staff. Activision just dumped 1900 people after a wildly successful year of games. That is the business model and AI or automation is going to embolden the no-talent producers to keep doing this over-and-over again. So use AI to take the storytelling away from the studios...

...followed by another person's reply:

This person said it best. I would like to add that due to all the layoffs and industry professionals looking for work, it would be best for most people to try to go indie anyway.. That is the reason I've been trying to make a makeshift animation team. If anyone reading this comment would like to network with me, hit me up

Basically, this guy, who claims to have worked at Disney before, seems to think that DreamWorks fired all its matte painters because AI like Midjourney can easily produce mattes now with high precision. Based on these, do you believe that Pixar will have a massive layoff very soon due to AI? Why or why not?

493
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/juebenjamin on 2024-03-16 13:55:09.

494
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/donutloop on 2024-03-16 07:11:36.

495
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/shinjisdaddy on 2024-03-15 18:43:45.

496
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/wyem on 2024-03-15 16:14:27.


  1. DeepSeek released DeepSeek-VL, an open-source Vision-Language (VL) model designed for real-world vision and language understanding applications. The DeepSeek-VL family, includes 7B and1.3B base and chat models and achieves state-of-the-art or competitive performance across a wide range of visual-language benchmarks. Free for commercial use [Details | Hugging Face | Demo]
  2. Cohere released Command-R, a 35 billion parameters generative model with open weights, optimized for long context tasks such as retrieval augmented generation (RAG) and using external APIs and tools for production-scale AI for enterprise [Details | Hugging Face].
  3. Google DeepMind introduced SIMA (Scalable Instructable Multiworld Agent), a generalist AI agent for 3D virtual environments, trained on nine different video games. It can understand a broad range of gaming worlds, and follows natural-language instructions to carry out tasks within them, as a human might.  It doesn’t need access to a game's source code or APIs and requires only the images on screen, and natural-language instructions provided by the user. SIMA uses keyboard and mouse outputs to control the games’ central character to carry out these instructions [Details].
  4. Meta AI introduces Emu Video Edit (EVE), a model that establishes a new state-of-the art in video editing without relying on any supervised video editing data [Details].
  5. Cognition Labs introduced Devin, the first fully autonomous AI software engineer. Devin can learn how to use unfamiliar technologies, can build and deploy apps end to end, can train and fine tune its own AI models. When evaluated on the SWE-Bench benchmark, which asks an AI to resolve GitHub issues found in real-world open-source projects, Devin correctly resolves 13.86% of the issues unassisted, exceeding the previous state-of-the-art model performance of 1.96% unassisted and 4.80% assisted [Details].
  6. Pika Labs adds sound effects to its AI video tool, Pika, allowing users to either prompt desired sounds or automatically generate them based on video content. [Video link].
  7. Anthropic’s Claude 3 Opus ranks #1 on LMSYS Chatbot Arena Leaderboard, along with GPT-4 [Link].
  8. The European Parliament approved the Artificial Intelligence Act. The new rules ban certain AI applications including biometric categorisation systems, Emotion recognition in the workplace and schools, social scoring and more [Details].
  9. Huawei Noah's Ark Lab introduced PixArt--Σ, a Diffusion Transformer model (DiT) capable of directly generating images at 4K resolution. It achieves superior image quality and user prompt adherence with significantly smaller model size (0.6B parameters) than existing text-to-image diffusion models, such as SDXL (2.6B parameters) and SD Cascade (5.1B parameters) [Details].
  10. South Korean startup Hyodol AI has launched a $1,800 LLM-powered companion doll specifically designed to offer emotional support and companionship to the rapidly expanding elderly demographic in the country [Details].
  11. Covariant introduced RFM-1 (Robotics Foundation Model -1), a large language model (LLM), but for robot language. Set up as a multimodal any-to-any sequence model, RFM-1 is an 8 billion parameter transformer trained on text, images, videos, robot actions, and a range of numerical sensor readings [Details].
  12. Figure 01 robot integrated with an OpenAI vision-language model can now have full conversations with people [Link]
  13. Deepgram announced the general availability of Aura, a text-to-speech model built for responsive, conversational AI agents and applications [Details | Demo].
  14. Claude 3 Haiku model is now available alongside Sonnet and Opus in the Claude API and on claude.ai for Pro subscribers. Haiku outperforms GPT-3.5 and Gemini 1.0 pro while costing less, and is three times faster than its peers for the vast majority of workloads [Details].
  15. Paddle announced AI Launchpad, a 6-week remote program for AI founders to launch and scale an AI business with $20,000 in cash prize [Details].
  16. Midjourney adds feature for generating consistent characters across multiple gen AI images [Details].
  17. The Special Committee of the OpenAI Board announced the completion of the review. Altman, Brockman to continue to lead OpenAI [Details]
  18. Together.ai introduced Sequoia, a scalable, robust, and hardware-aware speculative decoding framework that improves LLM inference speed on consumer GPUs (with offloading), as well as on high-end GPUs (on-chip), without any approximations [Details].
  19. OpenAI released Transformer Debugger (TDB), a tool developed and used internally by OpenAI's Superalignment team for investigating into specific behaviors of small language models [GitHub].
  20. Elon Musk announced that xAI will open source Grok this week [Link].

Source: AI Brews - You can subscribe the newsletter here. it's free to join, sent only once a week with bite-sized news, learning resources and selected tools. Thanks

497
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/TechyGyan on 2024-03-14 20:21:52.

498
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/TernaryJimbo on 2024-03-14 22:55:40.

499
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/ImpressiveContest283 on 2024-03-14 17:21:53.

500
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/No-Lobster-8045 on 2024-03-14 10:25:09.


I saw this video now, everyone's saying she ofc knows, she's just hiding due to legal trouble they might get into.

But interestingly, she could have said they sourced data from Shutterstock coz Open AI literally has a public partnership w them.

What are y'alls view on this? (Also, apologies if it's already posted)

view more: ‹ prev next ›