Loading
Salesforce now sends email only from verified domains. Read More
Agentforce and Einstein Generative AI
Table of Contents
Select Filters

          No results
          No results
          Here are some search tips

          Check the spelling of your keywords.
          Use more general search terms.
          Select fewer filters to broaden your search.

          Search all of Salesforce Help
          Einstein Generative AI & Trust

          Einstein Generative AI & Trust

          It’s important that your data stays safe while you innovate with new technology. Einstein generative AI keeps trust as its #1 value, and we strive to make sure your data is secure while also creating experiences that are accurate and safe.

          Einstein Generative AI and Your Data

          At Salesforce, trust is our #1 value. To keep your data secure, Salesforce has agreements in place with large language model (LLM) providers, such as OpenAI. These agreements allow organizations to use generative AI capabilities without their private data being retained by the LLM providers.

          Trusted Generative AI

          Salesforce’s Einstein generative AI solutions are designed, developed, and delivered based on our five principles for trusted generative AI.

          • Accuracy: We back up model responses with explanations and sources whenever possible. We recommend that a human checks model responses before sharing with end users for the majority of use cases.
          • Safety: We work to detect and mitigate bias, toxicity, and harmful responses from models used in our products through industry-leading detection and mitigation techniques.
          • Transparency: We ensure that our models and features respect data provenance and are grounded in your data whenever possible.
          • Empowerment: We believe our products should augment people’s capabilities and make them more efficient and purposeful in their work.
          • Sustainability: We strive towards building right-sized models that prioritize accuracy and reducing our carbon footprint.

          Reviewing Generative AI Responses

          Generative AI is a tool that helps you be more creative, productive, and make smarter business decisions. This technology isn’t a replacement for human judgment. You’re ultimately responsible for any LLM-generated response you share with your customers. Whether text is human- or LLM-generated, your customers associate it with your organization’s brand and use it to make decisions. So it’s important to make sure that LLM-generated responses intended for external audiences are accurate and helpful, and that they align with your company’s values, voice, and tone.

          When your end users review generated responses for external audiences, focus on the accuracy and safety of the content.

          • Accuracy: Generative AI can sometimes “hallucinate”—fabricate responses that aren’t grounded in fact or existing sources. Before you publish a response, check to make sure that key details are correct. For example, is the customer service suggestion based on an existing and up-to-date knowledge article?
          • Bias and Toxicity: Because AI is created by humans and trained on data created by humans, it can also contain bias against historically marginalized groups. Rarely, some responses can contain harmful language. Check your responses to make sure they’re appropriate for your customers.

          If the response doesn’t meet your standards or your company’s business needs, you don’t have to use it. Some solutions allow end users to edit the response directly, and if not, it’s best to start over and generate another response.

           
          Loading
          Salesforce Help | Article