Adobe MAX 2025: All the Top Announcements for Adobe’s Creative Suite
Summary
Adobe used MAX 2025 to push AI deeper into its creative tools. The biggest updates centre on Firefly — custom models for individuals, Image Model 5 with layered editing, and new generative media features (soundtrack and speech). Adobe also unveiled a browser-based multi-track Firefly video editor, AI assistants for Photoshop and Express, and previews of Project Moonlight (cross-app context) plus a planned ChatGPT integration. Several features are rolling out soon with waiting lists for early access.
Key Points
- Custom Firefly models: individuals can train character or tone models with as few as 6–12 images; based on Adobe’s proprietary, commercially safe Firefly model.
- Firefly Image Model 5: native 4‑megapixel outputs (2K) and improved layered image editing—move, resize and replace scene elements with minimal artifacts.
- Generate Soundtrack: analyses uploaded video and suggests soundtrack prompts; you can pick vibe, style and purpose to refine results.
- Generate Speech: text‑to‑speech in Firefly (15 languages at launch) with emotion tags and multi-model support (including ElevenLabs).
- Browser-based Firefly video editor: full multi-track editor combining generated and captured assets; public release timing TBD (waiting list available).
- AI assistants in Photoshop and Express: agentic yet guided helpers to suggest tools and complete tasks while keeping user control.
- Project Moonlight: carries context across Adobe apps and can ingest social/account context so outputs match your style and tone.
- ChatGPT integration teased: Adobe exploring embedding Adobe model capabilities directly into ChatGPT (early collaboration via Microsoft/OpenAI).
Context and Relevance
Adobe is turning generative AI from an add‑on into the backbone of its creative suite. Firefly’s move to allow custom, user-trained models broadens creative control and brand consistency for small teams and individuals. Layered editing and the 2K native output from Image Model 5 make generative edits more practical for real projects. Audio generation (soundtracks and speech) plus a browser multi-track editor indicate Adobe wants to collapse the gap between ideation and finished media inside the cloud. Project Moonlight and potential ChatGPT links point to a future where context and conversational interfaces speed workflows across apps.
Why should I read this?
Quick and dirty: if you make images, video or audio (or manage creatives), these updates change how fast you can prototype and ship stuff. Firefly now does more than pretty images — think automated soundtracks, TTS, better scene-aware edits and an actual multi-track editor in the browser. Bookmark the waiting lists or you’ll miss the early access perks.
Author take
Punchy and short: this is a big step towards Adobe making AI an everyday tool, not just a gimmick. Expect faster iterations, new workflow hooks, and a stronger platform lock‑in—plus fresh questions about provenance and creative ownership as people train custom models on their work.
Source
Source: https://www.wired.com/story/adobe-max-2025-firefly-photoshop-updates/