Tag: bedrock

  • Product Name Detection with AWS Bedrock & Anthropic Claude

    Well, my AWS bill me a bit larger than normal this month due to testing this script. I thoroughly enjoy utilizing Generative AI to do work for me and I had some spare time to tackle this problem this week.

    A client sent me a bunch of product images that were not named properly. All of the files were named something like “IMG_123.jpeg”. There was 63 total files so I decided rather than going through them one by one I would see if I could get one of Anthropic’s models to handle it for me and low and behold it was very successful!

    I scripted out the workflow in Python and utilized AWS Bedrock’s platform to execute the interactions with the Claude 3 Haiku model. Take a look at the code below to see how this was executed.

    if __name__ == "__main__":
        print("Processing images")
        files = os.listdir("photos")
        print(len(files))
        for file in files:
            if file.endswith(".jpeg"):
                print(f"Sending {file} to Bedrock")
                with open(f"photos/{file}", "rb") as photo:
    
                    prompt = f"""
                        Looking at the image included, find and return the name of the product. 
                        Rules: 
                        1. Return only the product name that has been determined.
                        2. Do not include any other text in your response like "the product determined..."
                        """
                    model_response = bedrock_actions.converse(
                        prompt, 
                        image_format="jpeg",
                        encoded_image=photo.read(),
                        max_tokens="2000",
                        temperature=.01,
                        top_p=0.999
                        )
                    print(model_response['output'])
                    product_name = modify_product_name(model_response['output']['message']['content'][0]['text'])
                    
                    photo.close()
                    if os.system(f"cp photos/{file} renamed_photos/{product_name}.jpeg") != 0:
                        print("failed to move file")
                    else:
                        os.system(f"mv photos/{file} finished/{file}")
        sys.exit(0)

    The code will loop through all the files in a folder called “photos” passing each one to Bedrock and getting a response. There was a lot of characters that were returned that would either break the script or that are just not needed so I also wrote a function to handle those.

    Ultimately, the script will copy the photo to a file named after the product and then move the original file into a folder called “finished”.

    I’ve uploaded the code to GitHub and you can utilize it however you want!

  • Building a Generative AI Workflow with AWS Bedrock

    I’ve finally been tasked with a Generative AI project to work on. I’ve done this workflow manually with ChatGPT in the past and it works quite well but, for this project, the requirement was to use Amazon Web Services’ new product “AWS Bedrock”.

    The workflow takes in some code and writes a technical document to support a clear English understanding of what the code is going to accomplish. Using AWS Bedrock, the AI will write the document and output it to an S3 bucket.

    The architecture involves uploading the initial code to an S3 Bucket which will then send the request to an SQS queue and ultimately trigger a Lambda to prompt the AI and fulfill the output upload to a separate S3 bucket. Because this was a proof of concept, the Lambda function was a significant compute resource however going forward I am going to look at placing this code into a Docker container so that it can scale for larger code inputs.

    Here is the architecture diagram:

    Let’s take a look at some of the important code. First is the prompt management. I wrote a function that will take input of the code as well as a parameter of “prompt_type”. This will allow the function to be scalable to accommodate other prompts in the future.

    def return_prompt(code, prompt_type):
        if prompt_type == "testPrompt":
            prompt1 = f"Human: <your prompt>. Assistant:"
            return prompt1

    The important thing to look at here is the format of the message. You have to include the “Human:” and the “Assistant:”. Without this formatting, your API call will error.

    The next bit of code is what we use to prompt the Bedrock AI.

     prompt_to_send = prompts.return_prompt(report_file, "testPrompt")
            body = {
                "prompt": prompt_to_send,
                "max_tokens_to_sample": 300,
                "temperature": 0.1,
                "top_p": 0.9
            }
            accept = 'application/json'
            contentType = 'application/json'
    
    
            # Return Psuedo code
            bedrock_response = h.bedrock_actions.invoke_model(json.dumps(body, indent=2).encode('utf-8'), contentType, accept, modelId=modelid)
        def invoke_model(body, contentType, accept, modelId):
            print(f"Body being sent: {body}")
            try:
                response = bedrock_runtime.invoke_model(
                    body=body,
                    contentType=contentType,
                    accept=accept,
                    modelId=modelId
                )
                return response
            except ClientError as e:
                print("Failed to invoke Bedrock model")
                print(e)
                return False

    The body of our request is what configures Bedrock to run and create a response. These values can be tweaked as follows:

    max_tokens_to_sample: This specifies the number of tokens to sample in your request. Amazon recommends setting this to 4000
    TopP: Use a lower value to ignore less probable options.
    Top K: Specify the number of token choices the model uses to generate the next token.
    Temperature: Use a lower value to decrease randomness in the response.

    You can read more about the inputs here.

    If you want to see more of this code take a look at my GitHub repository below. Feel free to use it wherever you want. If you have any questions be sure to reach out to me!

    GitHub: https://github.com/avansledright/bedrock-poc-public