isn't one of the top players in the AI game today, but the company's new open source AI model for image editing shows what it's capable of contributing to the space. The model called MLLM-Guided Image Editing , which uses multimodal large language models to interpret text-based commands when manipulating images. In other words, the tool has the ability to edit photos based on the text the user types in.
The company developed MGIE with researchers from the University of California, Santa Barbara. MLLMs have the power to transform simple or ambiguous text prompts into more detailed and clear instructions the photo editor itself can follow. For instance, if a user wants to edit a photo of a pepperoni pizza to "make it more healthy," MLLMs can interpret it as "add vegetable toppings" and edit the photo as such.