Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Help with Training Inputs #7

Open
avichapman opened this issue Feb 10, 2025 · 2 comments
Open

Help with Training Inputs #7

avichapman opened this issue Feb 10, 2025 · 2 comments

Comments

@avichapman
Copy link

Hi.

I am attempting to replicate the training of the SFPNet in order to investigate other training improvements.

Can you please give me some insight on how to do the training? I don't see any example training code in this repo.

Specifically, in the Semantic model's forward method, there are three inputs - input_data, xyz and batch.

I assume that input_data is a float tensor of shape BxNxF, where the first three in the F dimension are also x, y and z.

Is the xyz the voxel coordinates as a long tensor or the original lidar coordinates as a float tensor? Is the shape BxNx3?

Is the batch coordinate just torch.arange(batch_size)? Is the shape B?

Thank you for your help.

Avi

@Cavendish518
Copy link
Owner

Hi,

input is SparseConvTensor (check spconv lib). batch record the ID (due to spconv lib) xyz is the lidar coordinates (check the code in dataloader and the function related to xyz in backbone.py)

For training details, you can refer to issue #2.

@avichapman
Copy link
Author

Ah, ok.

So in training mode for a single entry in a minibatch, this would be torch.arange([number of samples])?
And in evaluation mode, this would be set in collation_fn_voxelmean_tta using the inds_recons value returned from the loader?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants