Comments

You must log in or register to comment.

Shelfrock77 OP t1_j1027uk wrote

The next breakthrough to take the AI world by storm might be 3D model generators. This week, OpenAI open sourced Point-E, a machine learning system that creates a 3D object given a text prompt. According to a paper published alongside the code base, Point-E can produce 3D models in one to two minutes on a single Nvidia V100 GPU.

Point-E doesn’t create 3D objects in the traditional sense. Rather, it generates point clouds, or discrete sets of data points in space that represent a 3D shape — hence the cheeky abbreviation. (The “E” in Point-E is short for “efficiency,” because it’s ostensibly faster than previous 3D object generation approaches.) Point clouds are easier to synthesize from a computational standpoint, but they don’t capture an object’s fine-grained shape or texture — a key limitation of Point-E currently.

To get around this limitation, the Point-E team trained an additional AI system to convert Point-E’s point clouds to meshes. (Meshes — the collections of vertices, edges and faces that define an object — are commonly used in 3D modeling and design.) But they note in the paper that the model can sometimes miss certain parts of objects, resulting in blocky or distorted shapes.

Image Credits: OpenAI

Outside of the mesh-generating model, which stands alone, Point-E consists of two models: a text-to-image model and an image-to-3D model. The text-to-image model, similar to generative art systems like OpenAI’s own DALL-E 2 and Stable Diffusion, was trained on labeled images to understand the associations between words and visual concepts. The image-to-3D model, on the other hand, was fed a set of images paired with 3D objects so that it learned to effectively translate between the two.

When given a text prompt — for example, “a 3D printable gear, a single gear 3 inches in diameter and half inch thick” — Point-E’s text-to-image model generates a synthetic rendered object that’s fed to the image-to-3D model, which then generates a point cloud.

After training the models on a dataset of “several million” 3D objects and associated metadata, Point-E could produce colored point clouds that frequently matched text prompts, the OpenAI researchers say. It’s not perfect — Point-E’s image-to-3D model sometimes fails to understand the image from the text-to-image model, resulting in a shape that doesn’t match the text prompt. Still, it’s orders of magnitude faster than the previous state-of-the-art — at least according to the OpenAI team.

16

FuturologyBot t1_j106v6d wrote

The following submission statement was provided by /u/Shelfrock77:


The next breakthrough to take the AI world by storm might be 3D model generators. This week, OpenAI open sourced Point-E, a machine learning system that creates a 3D object given a text prompt. According to a paper published alongside the code base, Point-E can produce 3D models in one to two minutes on a single Nvidia V100 GPU.

Point-E doesn’t create 3D objects in the traditional sense. Rather, it generates point clouds, or discrete sets of data points in space that represent a 3D shape — hence the cheeky abbreviation. (The “E” in Point-E is short for “efficiency,” because it’s ostensibly faster than previous 3D object generation approaches.) Point clouds are easier to synthesize from a computational standpoint, but they don’t capture an object’s fine-grained shape or texture — a key limitation of Point-E currently.

To get around this limitation, the Point-E team trained an additional AI system to convert Point-E’s point clouds to meshes. (Meshes — the collections of vertices, edges and faces that define an object — are commonly used in 3D modeling and design.) But they note in the paper that the model can sometimes miss certain parts of objects, resulting in blocky or distorted shapes.

Image Credits: OpenAI

Outside of the mesh-generating model, which stands alone, Point-E consists of two models: a text-to-image model and an image-to-3D model. The text-to-image model, similar to generative art systems like OpenAI’s own DALL-E 2 and Stable Diffusion, was trained on labeled images to understand the associations between words and visual concepts. The image-to-3D model, on the other hand, was fed a set of images paired with 3D objects so that it learned to effectively translate between the two.

When given a text prompt — for example, “a 3D printable gear, a single gear 3 inches in diameter and half inch thick” — Point-E’s text-to-image model generates a synthetic rendered object that’s fed to the image-to-3D model, which then generates a point cloud.

After training the models on a dataset of “several million” 3D objects and associated metadata, Point-E could produce colored point clouds that frequently matched text prompts, the OpenAI researchers say. It’s not perfect — Point-E’s image-to-3D model sometimes fails to understand the image from the text-to-image model, resulting in a shape that doesn’t match the text prompt. Still, it’s orders of magnitude faster than the previous state-of-the-art — at least according to the OpenAI team.


Please reply to OP's comment here: https://old.reddit.com/r/Futurology/comments/zqv8a7/openai_releases_pointe_an_ai_that_generates_3d/j1027uk/

1

Cheapskate-DM t1_j108dbu wrote

I suspect these tools are going to combine with and reinforce the 2D art generators - and possibly break into animation.

Having a stock model with known dimensions and posability, which can then be overlaid with a 2D illustration pass, could solve the wonkiness AI models currently have with things like hands and proportions. Plug that into a frame-by-frame animation, and you've just table-flipped the anime industry.

56

nycguy30 t1_j10a07v wrote

OpenAI making every small business and employees shit their pants

197

Robot-Candy t1_j10jch2 wrote

These look like they’re all made out of floam, an odd choice of rendering methods.

3

CheckMateFluff t1_j10jm46 wrote

It's not there yet but man this would be cool. I could generate a mesh from the text and clean up the topology. If I need simple basic things like lamps, cups, plates. etc this would save a lot of time.

1

JohnGabin t1_j10k7j4 wrote

Cool, because ChatGTP built me a really bad spaceship with a Blender Python script.

23

69_A_Porcupine t1_j10s5lm wrote

Is this the part where it automatically starts printing robot parts and takes over the power grid

6

UX-Edu t1_j10uivv wrote

Ooo! Can they rig the models? Because rigging is a goddamn pain in the ass.

77

mythoughtsforapenny t1_j10uyo9 wrote

Strengthen animation AI to the point that it can animate fully realistic humans, other animals, and objects, combine it with improved language composition AI, maybe with a program specialized in generating narrative structures, and you've just flipped over the entire film, television, and gaming industries, but also people won't be able to rely on video evidence anymore.

17

Hotporkwater t1_j1162qu wrote

"Oh no! What if this puts 3D modelers out of work?? Ban it!" - half the people on Reddit

−5

JohnnySasaki20 t1_j11666k wrote

That would certainly make developing video games a lot faster and easier. I imagine they could eventually make an AI write the code for the game as well. At some point you wouldn't have to do much of anything except tell the AI what to do.

24

fox-mcleod t1_j11cpxh wrote

Given it generates point clouds from descriptions, I wonder if the best application is VR rather than 3D printing.

2

hendrix320 t1_j11giww wrote

I’m starting to wonder if when I used dall-e to make thousands of corgi images if that messed with its algorithm to think corgi’s are the best dog. Corgi’s shows up on their website quite often

9

AfrikaCorps t1_j11h4d6 wrote

Big business! Small business benefit from democratization of tech.

I remember when my brother started a 3D printer farm in 2013 and undercut this huge company making specialized parts

6

norbertus t1_j11iyjy wrote

> I suspect these tools are going to combine with and reinforce the 2D art generators - and possibly break into animation.

It's already happening. Stability Diffusion has been integrated into GIMP and Blender. I.e., auto-texture 3D models.

5

TITANDERP t1_j11ozpw wrote

I for one would like to write a light novel and watch it slowly come to life. People are scared that media will be overtaken by ai junk but what they don't realize is the tools are likely to be marketed too. Imagine being able to write up an entire show, watch and share it, and vice versa. I understand it's likely an optimistic outlook and unlikely to happen in that fashion. That and the overall demeanor of how as of now people seeing it as soulless, which, fair enough.

12

Rhawk187 t1_j11pdfa wrote

Teaching my Game Engine Design class next semester. This will be a fun one for an assignment.

3

micktalian t1_j120upi wrote

Alright yall, you know what this means. This has been decades I'm the making and we should rejoice. All we have to do is feed the AI warhammer models to train off of, give it a prompt of the specific faction/unit/pose you want, and bam GW needs to change their business model.

31

bodden3113 t1_j1218u7 wrote

Actually, You ever heard of or experienced lucid dreaming? If AI is capable of generation at that level. We would be able to have dream like experiences on demand.

Think about it. When you talk to someone in your dream are you controlling it directly? Or controlling the environment you find yourself in, deliberately, sounds or music you hear? Are controlling each aspect consciously. No your subconscious is, and it's generating the entire phenomenon real-time with or without your input, most of the time your just along for the ride. AI media could potentially get to that level of real-time generation. And it seems it could get there much sooner rather then later by how fast and how good google deep dream and img generation tech got. It's literally a dream come true, LITERALLY?

−1

LegendaryPlayboy t1_j123w1o wrote

We will soon create infinite AR/VR worlds out of our own words. And movies. AI cinema should be a reality within a couple of years, more or less. Point-E is another good step forward for the whole picture.

3

SadcoreEmpire168 t1_j12d464 wrote

I’m genuinely getting surprised that stable diffusion is getting more advanced & introduced into making detailed images like this

3

MeatisOmalley t1_j12di0u wrote

There are already AI coding assitants, they dramatically speed up workflow. You don't describe your entire game in a single prompt, you describe aspects of the game, then the ai writes the code to program that aspect.

1

Kalwasky t1_j12jh23 wrote

To anyone wondering, this is largely an iterative work over facebook’s prior work. As far as I’ve been able to tell there is little going on that’s groundbreaking, think of it as the difference between a small GPT model and a large one.

1

Sugar_bytes t1_j12jpbk wrote

Omniverse has some interesting stuff for AI 3D modeling. Worth looking at for rigging solutions on the horizon as NVIDIA seems to want to eliminate the busy work.

1

pathego t1_j12n3ev wrote

More posts like this! The bot looking question posts about AI of late on this sub have devastated the quality of the sub. It has also revealed an interesting way to kill Reddit as more AI content buries the valuable stuff we come here for.

−1

ExternaJudgment t1_j12r2d4 wrote

Bullshitters getting priced out because the REAL value of their bullshit is now becoming clear.

I see it as progress, I would never pay them before but now prices might actually be reasonable.

−6

baelrog t1_j12t5mb wrote

It's been two weeks since ChatGPT comes out. I've already changed how I do free lance translation.

I feed every sentence I want to translate in the bot, and unlike Google translate where I have to edit every single time, with ChatGPT I only have to edit half of the time.

4

Themasterofcomedy209 t1_j12uyun wrote

For a while it’s going to be better for a lot of people, at first. AI will take a very long time to get to the point where you can tell it to generate like “long haired male elf” and it pulls up a professional and usable elf model. It’ll be helpful for a lot of people so they don’t have to spend 400 hours on crates and can focus on the stuff AI won’t be able to do for a while, which is the stuff most artists like doing the most anyway

2

AfrikaCorps t1_j12v77o wrote

Not here in Mexico, still good business, if you know spanish look up prices, it's stupid, we talking a print that takes 5 hours being like $20 or so.

Now, about your point which I will not ignore: that's even harder democratization because now some dudes can undercut small businesses and sell parts for less, making it a side hustle, it's positive in a way.

3

bnogal t1_j133e54 wrote

I had enough, please, some should implement a politician AI to setup an AI party

2

[deleted] t1_j136gwr wrote

the thing is, there will be an ai that deals with customers, and that ai will have infinite patience to cater to the specific needs while at the same time use powerful tools to guide the customer to the most profitable option.

7

NLwino t1_j13iu6e wrote

At some level, this is true. But it will greatly increase the productivity of someone that has the knowledge needed. Someone not using AI will simply not be able to compete on the market anymore.

2

NLwino t1_j13jmq5 wrote

The problem with this is that you need train the AI with a bunch of data. So where are going to find a lot of data of GOOD politicians?

2

ChronoFish t1_j13kk6d wrote

Maybe most profitable, but the ability to capture correct sentiment from a person is something that humans struggle with..

A lot of people can build things. Some can even follow directions. But few can translate customer desire to true expectations...as the classic PM tree swing meme conveys:

tree swing

2

ChronoFish t1_j13m55z wrote

Well conversely, this opens up a whole new world of rapid prototyping that small businesses never had access to before. Instead of being reliant on the business owners skill, or maybe a single employee, those individuals can now ramp up their productivity to the point where they can service customers that were traditionally out of reach.

2

ExternaJudgment t1_j13n60q wrote

Sure, I've heard this expressed that it will make 10x programmers now 100x programmers.

And as an AI using 10x programmer, I couldn't agree more.

ChatGPT gives me such a boost it is getting ridiculous.

1

dc2b18b t1_j140mhm wrote

“Make me a game. It has these aspects. Also when a user clicks item X and they have an axe in their inventory and greater than 50% health and there are less than 10 people currently playing and one of the other people playing is an elf who is greater than level 5 but less than level 8, the item should trigger Y. Oh but if the elf is at level 9, then scratch all that because the item should then do Z.”

Yeah good luck describing aspects at a high enough level that the AI can make something usable but in enough detail that it’s actually creating the game you want it to create, without causing unintended side effects.

−2

MeatisOmalley t1_j14q6au wrote

You just have no clue what you're talking about.

I can ask the ai, "write a gravity simulation" and it can code it in 1 second, and I could plug it into some aspect of my game. I could ask the ai, "write a performant function that constantly checks for and gives coordinates" and plug that into my minimap UI. It's literally faster than typing. I could also give parts of my own code to the ai, and say "I got (x) error. What's wrong with my code?" And the AI will describe in detail what I did wrong, and provide a solution.

Game development is coded in much smaller pieces than you seem to think.

4

SnooPuppers1978 t1_j164hcj wrote

Or you could just use an existing library that also takes "1 second".

I think so many examples people bring up are just something that you already can do with few clicks, existing livrary, existing framework, assets, etc.

In reality code is prompts. And you can always abstract code into functions which are like promots.

If prompts are that beneficial to some it sounds like their code has way too much boilerplate.

2

No_Introduction_3881 t1_j16k9f4 wrote

Silly question, I dont see it available in my open ai account? Where do I get it from?

1

thisdesignup t1_j186y3w wrote

Yep, I'm trying to create a personal assistant. So I'm teaching Davinci 003 how to respond to voice commands and return keywords to trigger python scripts.

I basically feel like I'm programming in plain English. It's easier only because I don't have to know what code to use but I still have to know how the process of teaching. Teaching is not easy.

2

BlitzBlotz t1_j189jo9 wrote

Making 3D models is super easy and super fast already, the anoying and hard part is making them in a way that doesnt cause glitches, errors or takes a huge toll on the graphic budget of your game.

Currently AI 3D models are pretty basic and have shitty geometry, will take some time til they fixed that. Not saying it doesnt happen but the current version of that AI is more a proof of concept than anything else.

1

sanman t1_j1fuowj wrote

Since a 3D object is a vector object, it would be nice if they could also come out with AI that generates 2D vector art as well, since that would also help to reinforce the 2D art in general. Then 2D vector art could be seamlessly integrated with the 2D raster art.

1