How can I connect to to inference machine in order to debug the custom image? #5087
              
                Unanswered
              
          
                  
                    
                      andrew-aladjev
                    
                  
                
                  asked this question in
                Help
              
            Replies: 0 comments
  
    Sign up for free
    to join this conversation on GitHub.
    Already have an account?
    Sign in to comment
  
        
    
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello, I've derived my custom docker image from your sagemaker image. Dockerfile is the following:
I built this image, uploaded it to ecr and used new custom image ARN in
image_uriofPyTorchModelas follows:It works: sagemaker deployed my model, but inference code is still not working: torchaudio can't see
ffmpegbackend. Now I want to debug my machine. In order to debug it I want to make any kind of terminal connection. But I am unable to do it, because I can't find any ability to do it.I've clicked everything in Amazon UI: Amazon SageMaker AI, Inference menu, Notebook menu, Jupiter menu. I have even opened fluffy toy "Studio" and I can't see anything. I see just endpoint without any ability to connect to the related machine(s). What should I do? For now I am thinking to move away from Amazon.
Beta Was this translation helpful? Give feedback.
All reactions