Security and privacy for ​BMC HelixGPT​


BMC HelixGPT is a platform service inside the BMC Helix tenant. It is subject to the same security policies and role-based access as other Helix applications.

BMC HelixGPT connects to the generative AI service provider, such as Azure OpenAI, Azure ML, or Google Cloud Platform Vertex, and all transactions occur through the service. The prompts and contextual data are sent to the third-party Large Language Model (LLM) provider or the BMC Helix LLM running in customer's environment. Depending on the use cases, either the generic third-party model or the BMC Helix-trained model is used. BMC Helix does not train on customer data for any AI use cases. 

Warning

Important

  • We recommend using Azure AI Content Safety Prompt Shields to check your large language model (LLM) inputs for potential threats in user prompts and third-party data.
    To learn more about Prompt Shields and to use the API for securing your LLM, see Prompt Shields.

  • We also recommend setting up processes to review and sanitize any content before it is ingested into BMC HelixGPT, to safeguard the systems against malicious or manipulative attempts to disrupt data.

BMC HelixGPT prioritizes data security and privacy for every tenant through the following security measures:

Data security and privacy for BMC HelixGPT

StatelessDesignIcon.pngStateless design

For every prompt, BMC HelixGPT seamlessly incorporates contextual data, such as the text obtained from knowledge articles, into the input before submitting it to the LLM. Contextual data is always under the control of the HelixGPT or product administrators.

Warning

Important

  • The third-party hosted LLM discards both the prompt and the contextual data depending on the customer’s agreement with the third-party LLM provider.
  • The Helix LLM also discards both the prompt and contextual data regardless of the customer’s agreement with the vendor.
  • The contextual data (including potentially PII) is exclusively utilized for inference and not incorporated into the LLM.

The following workflow shows how data flows in the BMC HelixGPT architecture for BMC Helix ITSM without being persisted in HelixGPT Agent Studio, Vector DB or the third-party LLMs:

Security_and_privacy.jpg

The following table lists the steps in which data flows:

Step

Performed by

Action

Example

1

UserIcon.png

User

A user submits a query through one of the BMC Helix products.

“I cannot connect to the office Wi-Fi.”

2

GearIcon.png

HelixGPT Agent Studio

HelixGPT Agent Studio converts the user query into a partial BMC HelixGPT prompt.

Predefined prompt: “I cannot connect to the office Wi-Fi”

Placeholders are added for contextual data.

3

VectorDBIcon.png

Vector DB

The Vector DB performs a semantic search on knowledge articles based on the prompt.

Relevant knowledge article is found: Connecting to the office Wi-Fi

4

LLMIcon.png

LLM

The LLM provides the response to the user query in natural language.

Troubleshooting advice is provided in natural language.

5

GearIcon.png

HelixGPT Agent Studio

HelixGPT Agent Studio generates a comprehensive answer to the user query containing the natural language response from the LLM and relevant knowledge articles. 

An answer containing the LLM-generated response and links to the relevant knowledge article is returned to the user.

6

UserIcon.png

User

The user receives the response through the BMC Helix product from which the query was sent.

"Steps to connect to the Wi-Fi.

Sources of information."

 Additionally, the following workflow is an example of the root cause analysis (RCA) use case from BMC Helix AIOps. In this example, data is queried during runtime but is not persisted in BMC HelixGPT Agent Studio or in the BMC Helix LLM: 

AIOpsFlow.png

StepPerformed byAction
1UserAn operator or a site reliability engineer (SRE) requests RCA of a situation in BMC Helix AIOps.
2BMC Helix AIOpsBMC Helix AIOps queries BMC Helix ITSM to retrieve any changes related to the CIs in the system.
3BMC Helix AIOpsBMC Helix AIOps queries log tools, such as BMC Helix Log Analytics, to retrieve relevant log data.
4LLMBMC Helix AIOps queries the BMC Helix fine-tuned LLM, and the LLM responds with the root cause analysis.
5UserThe analysis is displayed to the operator or SRE. 

SecurePermissionModelIcon.pngSecure permission model

BMC HelixGPT follows a multilevel security model to secure the data:

  • At the infrastructure level: BMC Helix uses many data stores as part of the overall platform. Every data store utilizes logical segregation of data and encryption at rest. Specifically for BMC Helix ITSM use cases where vectorized data is used, the Vector DB is built on the OpenSearch cluster and follows the same data security standards as other OpenSearch clusters for BMC Helix ITSM that run in the SaaS infrastructure, which includes encryption at rest for disks.
  • At the application level: All data is segregated based on user RBAC within the system. If a user is not authorized for a set of data, then that data is not sent to the third-party LLM provider.

With multilevel security, data is accessible based on a user’s access permissions defined in the BMC Helix ITSM permissions model or in BMC Helix Portal. Any change to permissions is immediately synced with BMC HelixGPT.

BMC HelixGPT uses the end user’s existing permissions to enforce access control. The end user (typically a support agent) has access to the following resources:

  • Knowledge articles
  • Support tickets
  • Enterprise data sources
  • Observability data

With BMC HelixGPT, an end user obtains actionable summaries from the resources based on the permissions assigned to the end user and the resources. An access control list (ACL) check is also conducted before the contextual data is returned to BMC HelixGPT to ensure the user has access to the related contextual data.

Warning

Important

  • The HelixGPT administrator user maintains full access control over enterprise data sources.
  • The HelixGPT administrator user can choose whether internal knowledge articles should be used for composing contextual data or training the model.

Examples

  • User permission model:
    • An IT Support agent cannot access HR support tickets.
    • A support agent and an administrator receive different answers to the same question based on their user access.
  • Resource permission model: Internal KAs and documents are inaccessible externally or publicly.

DataEncryptionIcon.pngData encryption

BMC HelixGPT leverages data stored in several data stores, which are securely hosted within BMC data centers. These include VictoriaMetrics, BMC AgileGraphDB, and OpenSearch (standard and vector DB). Security is reinforced through full disk encryption (AES-256), ensuring protection for data at rest. BMC manages all encryption keys.

All external API calls crossing the system boundary on the Internet are authenticated and encrypted by using TLSv1.2 and later. BMC data centers use an industry-leading edge firewall and web application firewall for all traffic in and out of the BMC Helix network.

Warning

Important

Bring Your Own Key (BYOK) and Hold Your Own Key (HYOK) are planned for future releases.

DataIsolationIcon.pngData isolation

All data in the various data stores is logically separated by tenant and managed by BMC Helix. Data isolation is implemented in BMC HelixGPT in the following ways:

  • Tenant data is logically separate.
  • Tenant access to the data is also controlled based on the role-based access control (RBAC) and permissions defined in BMC Helix ITSM or BMC Helix Portal. 
  • Cross-tenant searches are not permitted.
  • Access to the prompt library and Vector DB is provided per tenant. You must have the HelixGPT Admin role to view and update the prompts in the prompt library.
  • Other observability data is stored in existing data stores with logical tenant separation.

BMC HelixGPT supports “Bring Your Own Model” from generative AI vendors such as Azure OpenAI and Google Cloud Platform Vector AI for certain use cases.

DataRetentionIcon.pngData retention

The prompts and contextual data are sent to the third-party Large Language Model (LLM) provider, and the customer is responsible for determining the data retention policy with the third-party or customer-hosted LLMs. The customer is responsible for determining the data retention policy with the third party (Microsoft, Google, and so on) through the customer's secure and private agreement with the third-party LLM provider.

If the target LLM is the BMC Helix model, a similar stateless design applies.  

OpenSearch (Vector DB) follows the data retention policy of the data sources. The OpenSearch (Vector DB) is synchronized with all its data sources for data retention. For example, if a knowledge article has been unpublished, it is also removed from the Vector DB.

The identities of the stored documents depend on their data sources, such as BMC Helix Knowledge Management by ComAround, SharePoint, or BMC Helix Business Workflows. The documents are stored and identified as a part of the metadata of chunks. Each chunk stores the document ID and other metadata associated with the chunk, chunk content, and chunk embeddings. 

Other observability data, which is only queried at inference time, is retained based on published standards

PIIDataIcon.pngHandling PII data

Depending on a customer's agreement with the third-party LLM provider, the contextual data, (including potential personal identifiable information (PII)), or any sensitive information provided by the end user in the user query is exclusively utilized for inference and not incorporated into the LLM. It is, however, sent to the third-party or customer-hosted LLM.

Warning

Important

  • The definition of sensitive and PII data may vary depending on the customer’s data security and privacy policy, and the use cases for which the customer uses BMC HelixGPT, such as an HR chatbot vs IT chatbot.
  • The customer is responsible for removing PII data from support tickets, knowledge articles, and other enterprise data sources. BMC Helix Agent Studio allows fields to be remapped. 

Secure deployment for custom LLMs

BMC supports custom LLMs hosted on Google’s Vertex AI and Microsoft Azure ML for certain use cases. The custom LLM must be securely deployed according to Google Vertex AI security best practices or Microsoft Azure best practices.  

  • Microsoft Azure/Google’s responsibility
    • Protect the infrastructure
    • Secure the platform
    • Maintain compliance
  • Customer’s responsibility
    • Use the latest versions of Vertex AI Containers and VM Images
    • Manage access controls
    • Secure applications through encryption and the service perimeter
    • Monitor for security incidents
    • Comply with applicable laws and regulations for your use case

Related topics

Adding data sources in BMC HelixGPT

Ingesting data into BMC HelixGPT

 

Tip: For faster searching, add an asterisk to the end of your partial query. Example: cert*

BMC HelixGPT 26.2