Technologies new blogTechnologies new blog

For now, these features will only be available in the beta version of Photoshop. Adobe is also making some of these capabilities available to Firefly beta users on the web (Firefly users, by the way, have now created over 100 million images on the service). The neat thing here is that this integration allows Photoshop users to use natural language text prompts to describe the kind of image or object they want Firefly to create. As with all generative AI tools, the results can occasionally be somewhat unpredictable. By default, Adobe will provide users with three variations for every prompt, though unlike with the Firefly web app, there is currently no option to iterate on one of these to see similar variations on a given result. To do all of this, Photoshop sends parts of a given image to Firefly — not the entire image, though the company is also experimenting with that — and creates a new layer for the results. Maria Yap, the vice president of Digital Imaging at Adobe, gave me a demo of these new features ahead of today’s announcement. As with all things generative AI, it’s often hard to predict what the model will return, but some of the results were surprisingly good. For instance, when requested to generate a puddle beneath a running corgi, Firefly appeared to take the overall lighting of the image into account, even generating a realistic reflection. Not every result worked quite as well — a bright purple puddle was also an option — but the model does seem to do a pretty good job at adding object and especially at extending existing images beyond their frame. Given that Firefly was trained on the photos available in Adobe Stock (as well as other commercially safe images), it’s maybe no surprise that it does especially well with landscapes. Similar to most generative image generators, Firefly struggles with text. Adobe also ensured that the model returns safe results. This is partly due to the training set used, but Adobe has also implemented additional safeguards. “We married that with a series of prompt engineering things that we know,” explained Yap. “We exclude certain terms, certain words that we feel aren’t safe. And then we’re even looking into another hierarchy of ‘if Maria selects an area that has a lot of skin in it, maybe right now — and you’ll actually see warning messages at times — we won’t expand a prompt on that one, just because it’s unpredictable. We just don’t want to go into a place that doesn’t feel comfortable for us.” As with all Firefly images, Adobe will automatically apply its Content Credentials to any images utilizing these AI-based features. A lot of these features would also be quite useful in Lightroom. Yap agreed and while she wouldn’t commit to a timeline, she did confirm that the company is planning on bringing Firefly to its photo management tool as well. Adobe brings Firefly’s generative AI to Photoshop by Frederic Lardinois originally published on TechCrunch


Press ESC to close