Cookie の設定を選択する

当社は、当社のサイトおよびサービスを提供するために必要な必須 Cookie および類似のツールを使用しています。当社は、パフォーマンス Cookie を使用して匿名の統計情報を収集することで、お客様が当社のサイトをどのように利用しているかを把握し、改善に役立てています。必須 Cookie は無効化できませんが、[カスタマイズ] または [拒否] をクリックしてパフォーマンス Cookie を拒否することはできます。

お客様が同意した場合、AWS および承認された第三者は、Cookie を使用して便利なサイト機能を提供したり、お客様の選択を記憶したり、関連する広告を含む関連コンテンツを表示したりします。すべての必須ではない Cookie を受け入れるか拒否するには、[受け入れる] または [拒否] をクリックしてください。より詳細な選択を行うには、[カスタマイズ] をクリックしてください。

Prerequisites for connecting Amazon Q Business to Web Crawler

フォーカスモード
Prerequisites for connecting Amazon Q Business to Web Crawler - Amazon Q Business
このページはお客様の言語に翻訳されていません。 翻訳のリクエスト

Before you begin, make sure that you have completed the following prerequisites.

Note

For more information on connecting Web Crawler to Amazon Q Business, see Index website contents using the Amazon Q Web Crawler connector for Amazon Q Business in the AWS Machine Learning Blog.

For Amazon Q Web Crawler, make sure you have:

  • Copied the seed or sitemap URLs of the websites that you want to index and stored them in a text file or an Amazon S3 bucket. Each URL must be included on a separate line.

  • For XML sitemaps: Copied the sitemap XML and saved it in an XML file in an Amazon S3 bucket. You can also combine multiple sitemap XML files into a .zip file.

  • For websites that require basic, NTLM, or Kerberos authentication:

    • Noted your website authentication credentials, which include a username and password.

      Note

      Amazon Q Web Crawler supports the NTLM authentication protocol that includes password hashing, and Kerberos authentication protocol that includes password encryption.

  • For websites that require SAML or login form authentication:

    • Noted your website authentication credentials, which include a username and password.

    • Copied the XPaths (XML Path Language) of the username field (and the username button if using SAML), password field and button, and copied the login page URL. You can find the XPaths of elements using your web browser’s developer tools. XPaths follow this format: //tagname[@Attribute='Value'].

      Note

      Amazon Q Web Crawler uses a headless Chrome browser and the information from the form to authenticate and authorize access with an OAuth 2.0 protected URL.

  • Optional: Copied the host name and the port number of the web proxy server if you want to use a web proxy to connect to internal websites that you want to crawl. The web proxy must be public facing. Amazon Q supports connecting to web proxy servers backed by basic authentication, or you can connect with no authentication.

  • Optional: Copied the virtual private cloud (VPC) subnet ID if you want to use a VPC to connect to internal websites you want to crawl. For more information, see Using Amazon VPC.

In your AWS account, make sure you have:

  • Created an IAM role for your data source and, if using the API, noted the ARN of the IAM role.

  • For websites that require authentication credentials to crawl: Stored your Web Crawler authentication credentials in an AWS Secrets Manager secret and, if using the API, noted the ARN of the secret.

    Note

    If you’re a console user, you can create the IAM role and Secrets Manager secret as part of configuring your Amazon Q application environment on the console.

プライバシーサイト規約Cookie の設定
© 2025, Amazon Web Services, Inc. or its affiliates.All rights reserved.