Replies: 1 comment
-
The algorithm supports JSON format for sparse data as described in https://docs.aws.amazon.com/sagemaker/latest/dg/cdf-inference.html#common-in-formats. Sparse data can also be provided in protobuf format. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
ISSUE
I trained Factorization machine model on sagemaker with MXNET. Now for training I followed this notebook . My train data is sparse(300 M rows, 4M columns) I used
smac.write_spmatrix_to_sparse_tensor
to write my sparse training data to s3 for training. I was able to train and deploy an endpoint for inference. But problem starts when I try to inference, my feature dimension is 4 Million. So when I try to pass a dense vector of this size topredict
for inference, I getRequest Entity Too Large
. On checking I found that sagemaker has 5mb size limit. only option I can think of is to somehow pass the sparse vector butpredict
is not accepting that. Can you guys help?I tried sending sparse tensor, with content-type as
protobuf
, didn't work.I am using
RealTimePredictor
class aspredictor_cls
, is there anything I can do with this class so that it acceptsparse vector
and maybe convert it on server side if needed. Any Suggestions?Beta Was this translation helpful? Give feedback.
All reactions