Skip to content

Commit

Permalink
Use AWSDocument
Browse files Browse the repository at this point in the history
  • Loading branch information
adam-fowler committed Feb 27, 2025
1 parent f0721d7 commit 29b82a4
Show file tree
Hide file tree
Showing 35 changed files with 210 additions and 210 deletions.
2 changes: 1 addition & 1 deletion Package.swift
Original file line number Diff line number Diff line change
Expand Up @@ -432,7 +432,7 @@ let package = Package(
.library(name: "SotoXRay", targets: ["SotoXRay"]),
],
dependencies: [
.package(url: "https://github.com/soto-project/soto-core.git", from: "7.3.0")
.package(url: "https://github.com/soto-project/soto-core.git", branch: "document")
],
targets: [
.target(
Expand Down
12 changes: 6 additions & 6 deletions Sources/Soto/Services/Bedrock/Bedrock_shapes.swift
Original file line number Diff line number Diff line change
Expand Up @@ -2165,7 +2165,7 @@ extension Bedrock {

public struct ExternalSourcesGenerationConfiguration: AWSEncodableShape & AWSDecodableShape {
/// Additional model parameters and their corresponding values not included in the text inference configuration for an external source. Takes in custom model parameters specific to the language model being used.
public let additionalModelRequestFields: [String: String]?
public let additionalModelRequestFields: [String: AWSDocument]?
/// Configuration details for the guardrail.
public let guardrailConfiguration: GuardrailConfiguration?
/// Configuration details for inference when using RetrieveAndGenerate to generate responses while using an external source.
Expand All @@ -2174,7 +2174,7 @@ extension Bedrock {
public let promptTemplate: PromptTemplate?

@inlinable
public init(additionalModelRequestFields: [String: String]? = nil, guardrailConfiguration: GuardrailConfiguration? = nil, kbInferenceConfig: KbInferenceConfig? = nil, promptTemplate: PromptTemplate? = nil) {
public init(additionalModelRequestFields: [String: AWSDocument]? = nil, guardrailConfiguration: GuardrailConfiguration? = nil, kbInferenceConfig: KbInferenceConfig? = nil, promptTemplate: PromptTemplate? = nil) {
self.additionalModelRequestFields = additionalModelRequestFields
self.guardrailConfiguration = guardrailConfiguration
self.kbInferenceConfig = kbInferenceConfig
Expand Down Expand Up @@ -2236,10 +2236,10 @@ extension Bedrock {
/// The name of metadata attribute/field, which must match the name in your data source/document metadata.
public let key: String
/// The value of the metadata attribute/field.
public let value: String
public let value: AWSDocument

@inlinable
public init(key: String, value: String) {
public init(key: String, value: AWSDocument) {
self.key = key
self.value = value
}
Expand Down Expand Up @@ -2371,7 +2371,7 @@ extension Bedrock {

public struct GenerationConfiguration: AWSEncodableShape & AWSDecodableShape {
/// Additional model parameters and corresponding values not included in the textInferenceConfig structure for a knowledge base. This allows you to provide custom model parameters specific to the language model being used.
public let additionalModelRequestFields: [String: String]?
public let additionalModelRequestFields: [String: AWSDocument]?
/// Contains configuration details for the guardrail.
public let guardrailConfiguration: GuardrailConfiguration?
/// Contains configuration details for inference for knowledge base retrieval and response generation.
Expand All @@ -2380,7 +2380,7 @@ extension Bedrock {
public let promptTemplate: PromptTemplate?

@inlinable
public init(additionalModelRequestFields: [String: String]? = nil, guardrailConfiguration: GuardrailConfiguration? = nil, kbInferenceConfig: KbInferenceConfig? = nil, promptTemplate: PromptTemplate? = nil) {
public init(additionalModelRequestFields: [String: AWSDocument]? = nil, guardrailConfiguration: GuardrailConfiguration? = nil, kbInferenceConfig: KbInferenceConfig? = nil, promptTemplate: PromptTemplate? = nil) {
self.additionalModelRequestFields = additionalModelRequestFields
self.guardrailConfiguration = guardrailConfiguration
self.kbInferenceConfig = kbInferenceConfig
Expand Down
16 changes: 8 additions & 8 deletions Sources/Soto/Services/BedrockAgent/BedrockAgent_shapes.swift
Original file line number Diff line number Diff line change
Expand Up @@ -7849,7 +7849,7 @@ extension BedrockAgent {

public struct PromptConfiguration: AWSEncodableShape & AWSDecodableShape {
/// If the Converse or ConverseStream operations support the model, additionalModelRequestFields contains additional inference parameters, beyond the base set of inference parameters in the inferenceConfiguration field. For more information, see Inference request parameters and response fields for foundation models in the Amazon Bedrock user guide.
public let additionalModelRequestFields: String?
public let additionalModelRequestFields: AWSDocument?
/// Defines the prompt template with which to replace the default prompt template. You can use placeholder variables in the base prompt template to customize the prompt. For more information, see Prompt template placeholder variables. For more information, see Configure the prompt templates.
public let basePromptTemplate: String?
/// The agent's foundation model.
Expand All @@ -7866,7 +7866,7 @@ extension BedrockAgent {
public let promptType: PromptType?

@inlinable
public init(additionalModelRequestFields: String? = nil, basePromptTemplate: String? = nil, foundationModel: String? = nil, inferenceConfiguration: InferenceConfiguration? = nil, parserMode: CreationMode? = nil, promptCreationMode: CreationMode? = nil, promptState: PromptState? = nil, promptType: PromptType? = nil) {
public init(additionalModelRequestFields: AWSDocument? = nil, basePromptTemplate: String? = nil, foundationModel: String? = nil, inferenceConfiguration: InferenceConfiguration? = nil, parserMode: CreationMode? = nil, promptCreationMode: CreationMode? = nil, promptState: PromptState? = nil, promptType: PromptType? = nil) {
self.additionalModelRequestFields = additionalModelRequestFields
self.basePromptTemplate = basePromptTemplate
self.foundationModel = foundationModel
Expand Down Expand Up @@ -7923,7 +7923,7 @@ extension BedrockAgent {

public struct PromptFlowNodeInlineConfiguration: AWSEncodableShape & AWSDecodableShape {
/// Additional fields to be included in the model request for the Prompt node.
public let additionalModelRequestFields: String?
public let additionalModelRequestFields: AWSDocument?
/// Contains inference configurations for the prompt.
public let inferenceConfiguration: PromptInferenceConfiguration?
/// The unique identifier of the model or inference profile to run inference with.
Expand All @@ -7934,7 +7934,7 @@ extension BedrockAgent {
public let templateType: PromptTemplateType

@inlinable
public init(additionalModelRequestFields: String? = nil, inferenceConfiguration: PromptInferenceConfiguration? = nil, modelId: String, templateConfiguration: PromptTemplateConfiguration, templateType: PromptTemplateType) {
public init(additionalModelRequestFields: AWSDocument? = nil, inferenceConfiguration: PromptInferenceConfiguration? = nil, modelId: String, templateConfiguration: PromptTemplateConfiguration, templateType: PromptTemplateType) {
self.additionalModelRequestFields = additionalModelRequestFields
self.inferenceConfiguration = inferenceConfiguration
self.modelId = modelId
Expand Down Expand Up @@ -8126,7 +8126,7 @@ extension BedrockAgent {

public struct PromptVariant: AWSEncodableShape & AWSDecodableShape {
/// Contains model-specific inference configurations that aren't in the inferenceConfiguration field. To see model-specific inference parameters, see Inference request parameters and response fields for foundation models.
public let additionalModelRequestFields: String?
public let additionalModelRequestFields: AWSDocument?
/// Specifies a generative AI resource with which to use the prompt.
public let genAiResource: PromptGenAiResource?
/// Contains inference configurations for the prompt variant.
Expand All @@ -8143,7 +8143,7 @@ extension BedrockAgent {
public let templateType: PromptTemplateType

@inlinable
public init(additionalModelRequestFields: String? = nil, genAiResource: PromptGenAiResource? = nil, inferenceConfiguration: PromptInferenceConfiguration? = nil, metadata: [PromptMetadataEntry]? = nil, modelId: String? = nil, name: String, templateConfiguration: PromptTemplateConfiguration, templateType: PromptTemplateType) {
public init(additionalModelRequestFields: AWSDocument? = nil, genAiResource: PromptGenAiResource? = nil, inferenceConfiguration: PromptInferenceConfiguration? = nil, metadata: [PromptMetadataEntry]? = nil, modelId: String? = nil, name: String, templateConfiguration: PromptTemplateConfiguration, templateType: PromptTemplateType) {
self.additionalModelRequestFields = additionalModelRequestFields
self.genAiResource = genAiResource
self.inferenceConfiguration = inferenceConfiguration
Expand Down Expand Up @@ -10920,10 +10920,10 @@ extension BedrockAgent {

public struct ToolInputSchema: AWSEncodableShape & AWSDecodableShape {
/// A JSON object defining the input schema for the tool.
public let json: String?
public let json: AWSDocument?

@inlinable
public init(json: String? = nil) {
public init(json: AWSDocument? = nil) {
self.json = json
}

Expand Down
Loading

0 comments on commit 29b82a4

Please sign in to comment.