From: redpointai

Adobe leverages its millions of users to refine its AI models, ensuring they meet user expectations and maintain high standards of safety and quality [00:00:29]. This extensive user engagement provides crucial data for model refinement and improvement.

Data Collection and User Feedback Mechanisms

Adobe’s terms of use for platforms like Firefly.com allow the company to store user prompts and the resulting generated images [00:00:31]. This data is then used in training their models, which is essentially their way of doing reinforcement learning from human feedback (RHF) [00:34:05].

Key feedback mechanisms include:

  • Direct Feedback Beta customers on Firefly.com can explicitly provide feedback on what they like or dislike, and report issues like bias or harmful content [00:19:08]. This explicit feedback includes “like/dislike” and “report” signals [00:34:18].
  • Implicit Signals Actions such as downloading, saving, or sharing generated content are strong implicit signals indicating a user liked a particular generation [00:34:25].
  • Inspiration Stream Users can submit their own creations to an inspiration stream, which is then curated by Adobe’s design team to showcase the best examples for new users to build upon [00:19:16][00:20:46]. This allows new users to find something they like and start “riffing from it,” seeing the prompt and making their own variations [00:19:27].

Model Refinement through Feedback

The collected feedback, both explicit and implicit, is logged and integrated into the feedback loop for future model training [00:34:32]. This process teaches the generative models to produce content that users prefer and to avoid generating content that users dislike [00:34:40].

Addressing Bias and Harmful Content

A significant aspect of model refinement is addressing bias and harmful content. Adobe trains its Firefly models exclusively on data from Adobe Stock, which undergoes both automated and manual curation to ensure high-quality and compliant content [00:27:19][00:27:25]. This process helps reduce the risk of harmful or infringing content being generated [00:27:39].

Despite data curation, inherent biases can exist [00:28:07]. Adobe conducts extensive internal testing with tens of thousands of employees to stress-test models and identify areas for improvement regarding bias and harm [00:28:29].

Specific methods for debiasing include:

  • Person Detector A model detects if a prompt references a person, especially in relation to a job [00:29:21].
  • Distribution Adjustment To counter bias (e.g., a “lawyer” prompt generating only older Caucasian males), the model introduces a fair distribution of skin tones, genders, and age groups based on the country of origin of the request [00:29:52].
  • Toxicity Detectors and Filters Various toxicity detector models, deny lists, and block lists are employed to prevent the generation of “not safe for work” content [00:31:01]. NSFW filters are also applied at the end of the generation process [00:31:13].
  • Child-Referencing Systems Specialized systems detect prompts referencing children and implement negative prompting to avoid generating inappropriate content in relation to them (e.g., children with tobacco) [00:31:31].

This combination of machine learning models, algorithms, and rules helps continuously refine the output, although the process is acknowledged as not perfect [00:31:49]. User feedback continues to provide new training data points for these rules and models [00:32:11].

Differentiating Content

Adobe ensures that models are not trained on customer data stored in Creative Cloud, which might contain proprietary brand campaigns. Instead, they focus on Adobe Stock data and potentially licensed museum/art content [00:33:33].

Organizational Structure for AI Refinement

Adobe employs a hybrid organizational structure that supports continuous AI model refinement:

  • Horizontal AI Teams Adobe Research, part of the CTO organization, conducts foundational research across modalities (images, video, vector, 3D) and publishes papers [00:41:52]. The Sensei team provides a horizontal, distributed training platform and curated high-quality data sets for both research and applied research [00:42:19].
  • Applied Research Once models are ready for pre-production, they move to applied research groups. Here, models are refined, which can involve additional training or even architectural changes, preparing them for productization [00:42:50][00:39:51].
  • Gen Services Team This team creates services that run the refined models efficiently, building optimized inference platforms and APIs for product teams [00:43:09].
  • Vertical “Tech Transfer” Subgroups Within each product line (e.g., Photoshop, Adobe Express, Illustrator), there are “Tech Transfer” subgroups. These teams leverage the APIs and services from the horizontal teams to build tailored user experiences around the AI capabilities, acting as product experts [00:44:19]. This approach allows product teams to focus on core workflows that benefit most from AI [00:44:42].

This “AI Super Highway” model, which acts as a managed model zoo, accelerates innovation and the transfer of advanced AI capabilities to customers [00:45:01].