admin管理员组文章数量:1400211
Issue: I am attempting to create a fine-tuning job on Amazon Bedrock via the AWS Web console. The base model selected for the task is AWS Nova Micro. The training data - which resides in an S3 bucket in a .jsonl file - is saved in the required format as per the Amazon Bedrock User Guide and contains around 3000 records:-
{"prompt": "What is the capital of France?", "completion": "The capital of France is Paris."}
The IAM role used contains the correct level of permissions for reading & writing. The job saves and runs for up to about 14-16 minutes before failing. The error message flashed on the top of the page is
Unable to parse S3 file.
I have enabled Model Invocation logging for Bedrock but nothing related to this job is written to the CloudWatch. Additionally, the standard dashboards in CloudWatch don't seem to have any helpful information either.
In order to test the quality of my data, I converted the data into Open AI's required format:-
{"messages": [{"role": "user", "content": "Where do babies come from"}, {"role": "assistant", "content": "That is really something you should be asking your dad about"}]}
I then performed a fine-tuning job on Open AI's console and the job completed successfully and the custom model was created as required.
So, my assumption is that the data is relatively clean and that the error lies somewhere with Bedrock - but how can I get more specific information on what exactly is causing the failure?
I am looking for guidance on how to go about debugging this.
Issue: I am attempting to create a fine-tuning job on Amazon Bedrock via the AWS Web console. The base model selected for the task is AWS Nova Micro. The training data - which resides in an S3 bucket in a .jsonl file - is saved in the required format as per the Amazon Bedrock User Guide and contains around 3000 records:-
{"prompt": "What is the capital of France?", "completion": "The capital of France is Paris."}
The IAM role used contains the correct level of permissions for reading & writing. The job saves and runs for up to about 14-16 minutes before failing. The error message flashed on the top of the page is
Unable to parse S3 file.
I have enabled Model Invocation logging for Bedrock but nothing related to this job is written to the CloudWatch. Additionally, the standard dashboards in CloudWatch don't seem to have any helpful information either.
In order to test the quality of my data, I converted the data into Open AI's required format:-
{"messages": [{"role": "user", "content": "Where do babies come from"}, {"role": "assistant", "content": "That is really something you should be asking your dad about"}]}
I then performed a fine-tuning job on Open AI's console and the job completed successfully and the custom model was created as required.
So, my assumption is that the data is relatively clean and that the error lies somewhere with Bedrock - but how can I get more specific information on what exactly is causing the failure?
I am looking for guidance on how to go about debugging this.
Share Improve this question asked Mar 25 at 13:27 ReegzReegz 6041 gold badge7 silver badges15 bronze badges1 Answer
Reset to default 0Ok so it turns out that this was indeed a format error on my part.
The below format is the correct per-line format to use for each record in the JSONL file.
{
"schemaVersion":"bedrock-conversation-2024",
"system":[
{
"text":"You are a digital assistant with a friendly personality"
}
],
"messages":[
{
"role":"user",
"content":[
{
"text":"Question"
}
]
},
{
"role":"assistant",
"content":[
{
"text":"Answer"
}
]
}
]
}
本文标签: Amazon Bedrock Fine Tune Job Unable to Parse FileStack Overflow
版权声明:本文标题:Amazon Bedrock Fine Tune Job Unable to Parse File - Stack Overflow 内容由网友自发贡献,该文观点仅代表作者本人, 转载请联系作者并注明出处:http://www.betaflare.com/web/1744193175a2594615.html, 本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌抄袭侵权/违法违规的内容,一经查实,本站将立刻删除。
发表评论