Questions tagged with Machine Learning & AI

Content language: English

Sort by most recent

Browse through the questions and answers listed below or filter and sort to narrow down your results.

Which aws service should we start with to build a chatbot with our private data in AWS? Should we go to Kendra or any LLM solution?
1
answers
0
votes
5
views
JJone
asked 12 hours ago
regarding sagemaker experiments, we can add additonal arguments or parameters as part of tracking for sagemaker experiments . is there anything similar for sagemaker training jobs? ``` from smexperiments.tracker import tracker with Tracker.create(...) as tracker: tracker.log_parameters (.... ```
1
answers
0
votes
6
views
asked 21 hours ago
So the notebook is still running, name of the notebook is **Studio-Notebook:ml.t3.medium** ,please help me i am about to exceed the free tier limit . Please tell me how to fix this so that i wont be billed as i am not even using sagemaker rightnow.
3
answers
1
votes
16
views
asked 2 days ago
I have used **ml.g4dn.2xlarge** instance on SageMaker to test GPT-J6B model from HuggingFace using Transformer. I am using `revision=float16` and `low_cpu_mem_usage=True` so that the model is only of 12GB. It is downloaded but ***after*** that it suddenly crashes the kernel. Please share the workaround. The memory of that instance is 32 GB wit 4 vCPU. ```python !pip install transformers from transformers import AutoTokenizer, AutoModelForCasualLM model = AutoModelForCasualLM.from_pretrained("EleutherAI/gpt-j-6B", revision="float16", low_cpu_mem_usage=True) # It crashes here tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-j-6B") ``` It downloads 12GB model but after that, it crashes. I tried to follow this thread [here](https://repost.aws/questions/QUsO3sfUGpTKeHiU8W9k1Kwg/why-does-my-kernal-keep-dying-when-i-try-to-import-hugging-face-bert-models-to-amazon-sage-maker) but still there I can't update the sentencepiece. Please help. Thanks
0
answers
0
votes
22
views
EM_User
asked 3 days ago
i need to pull data from different sources and run processing then training jobs, I want to set it up such that , I spin up one or more ec2 instance , collect the data , copy it to some s3 , then use sagemaker api to create processing and training jobs. I want to do this via code, provisioning ec2 instances and the code to download data and running training and processing jobs, all of it. one question, i had was, once i download the data, can i copy the data directly to whatever data storage/volume comes attached with the sagemaker training/processing instance. I'm not sure what's it called . i know there are options to stream data? any thoughts. this would be my first training job , so apologies for basic question here
1
answers
0
votes
13
views
asked 6 days ago
How can I extend the default time out period for SageMaker pre-signed URL ?
2
answers
0
votes
11
views
asked 7 days ago
I want to specify a country/region for ip verification against ip geo location for SageMaker Studio access
1
answers
0
votes
8
views
asked 7 days ago
I want to read data from Databricks output and format the data for SageMaker training
1
answers
0
votes
8
views
asked 7 days ago
Use case : New documents are added through a web application on ongoing basis to S3. I am trying to build a document search for the documents stored in S3 that can display documents uploaded in near real time. Does Kendra sync data source with index based on an event trigger?
1
answers
0
votes
8
views
asked 8 days ago
Hi All, I'm having an issue running enhanced scanning in ECR for my Docker image. To replicate the issue, I have tested this on some sample base images that I'm using from Nvidia's container registry. When uploading the base Nvidia TensorRT image for Cuda 11.6, I am able to receive a vulnerability report. This is the tag: `nvcr.io/nvidia/tensorrt:21.07-py3` However, a newer CUDA version variant (which is still Ubuntu 20 based) is showing `UNSUPPORTED_IMAGE` in the vulnerability report: `nvcr.io/nvidia/tensorrt:22.12-py3` According to AWS docs, Ubuntu 20 images should still be supported. Is there any way to remediate this?
1
answers
0
votes
8
views
honson1
asked 8 days ago
Is it possible to deploy a model to an endpoint along with a large amount of data for it to use? (~50GB), this can be though of as an internal database for my model to use. I only require such functionality temporary, since I am trying to avoid using database services (e.g DynamoDB etc) for the moment. Thanks
3
answers
0
votes
21
views
Ori
asked 9 days ago
https://aws.amazon.com/cn/blogs/machine-learning/analyze-us-census-data-for-population-segmentation-using-amazon-sagemaker/#Comments,官方示例文档中的美国人口普查数据无法下载,我执行aws s3 ls s3://aws-ml-blog-sagemaker-census-segmentation。提示我“An error occurred (AccessDenied) when calling the ListObjectsV2 operation: Access Denied”拒绝访问,IAM账号已经给了AmazonS3FullAccess权限
1
answers
0
votes
13
views
asked 10 days ago